Uploaded image for project: 'Hadoop Distributed Data Store'
  1. Hadoop Distributed Data Store
  2. HDDS-352

Separate install and testing phases in acceptance tests.

    XMLWordPrintableJSON

    Details

    • Type: Improvement
    • Status: Resolved
    • Priority: Major
    • Resolution: Fixed
    • Affects Version/s: None
    • Fix Version/s: 0.2.1
    • Component/s: None
    • Labels:
    • Target Version/s:

      Description

      In the current acceptance tests (hadoop-ozone/acceptance-test) the robot files contain two kind of commands:

      1) starting and stopping clusters
      2) testing the basic behaviour with client calls

      It would be great to separate the two functionality and include only the testing part in the robot files.

      1. Ideally the tests could be executed in any environment. After a kubernetes install I would like to do a smoke test. It could be a different environment but I would like to execute most of the tests (check ozone cli, rest api, etc.)

      2. There could be multiple ozone environment (standlaone ozone cluster, hdfs + ozone cluster, etc.). We need to test all of them with all the tests.

      3. With this approach we can collect the docker-compose files just in one place (hadoop-dist project). After a docker-compose up there should be a way to execute the tests with an existing cluster. Something like this:

      docker run -it apache/hadoop-runner -v ./acceptance-test:/opt/acceptance-test -e SCM_URL=http://scm:9876 --network=composenetwork start-all-tests.sh
      

      4. It also means that we need to execute the tests from a separated container instance. We need a configuration parameter to define the cluster topology. Ideally it could be just one environment variables with the url of the scm and the scm could be used to discovery all of the required components + download the configuration files from there.

      5. Until now we used the log output of the docker-compose files to do some readiness probes. They should be converted to poll the jmx endpoints and check if the cluster is up and running. If we need the log files for additional testing we can create multiple implementations for different type of environments (docker-compose/kubernetes) and include the right set of functions based on an external parameters.

      6. Still we need a generic script under the ozone-acceptance test project to run all the tests (starting the docker-compose clusters, execute tests in a different container, stop the cluster)

        Attachments

        1. HDDS-352-ozone-0.2.006.patch
          55 kB
          Bharat Viswanadham
        2. TestRun.rtf
          61 kB
          Bharat Viswanadham
        3. HDDS-352.00.patch
          55 kB
          Bharat Viswanadham
        4. HDDS-352-ozone-0.2.005.patch
          51 kB
          Marton Elek
        5. HDDS-352-ozone-0.2.004.patch
          51 kB
          Marton Elek
        6. HDDS-352-ozone-0.2.003.patch
          51 kB
          Marton Elek
        7. HDDS-352-ozone-0.2.002.patch
          51 kB
          Marton Elek
        8. HDDS-352-ozone-0.2.001.patch
          51 kB
          Marton Elek

          Activity

            People

            • Assignee:
              elek Marton Elek
              Reporter:
              elek Marton Elek
            • Votes:
              0 Vote for this issue
              Watchers:
              6 Start watching this issue

              Dates

              • Created:
                Updated:
                Resolved: