X-Git-Url: http://git.lttng.org./?a=blobdiff_plain;f=scripts%2Flttng-baremetal-tests%2Flava-submit.py;h=391cda30fc8f83369b07c65df9d649bb71b93d62;hb=70e85c9830a708869b110e4083ff04fd3bf3cb57;hp=77abff970b13e0d1946ec47d7506821fc1959b44;hpb=64d4186531e1d38372a06e35100a70ea9bb7a7c1;p=lttng-ci.git diff --git a/scripts/lttng-baremetal-tests/lava-submit.py b/scripts/lttng-baremetal-tests/lava-submit.py index 77abff9..391cda3 100644 --- a/scripts/lttng-baremetal-tests/lava-submit.py +++ b/scripts/lttng-baremetal-tests/lava-submit.py @@ -70,7 +70,9 @@ def check_job_all_test_cases_state_count(server, job): # save them as CSV files localy def fetch_benchmark_results(server, job): content = get_job_bundle_content(server, job) - testcases = ['processed_results_close.csv', 'processed_results_open_enoent.csv', 'processed_results_open_efault.csv'] + testcases = ['processed_results_close.csv', + 'processed_results_open_efault.csv', + 'processed_results_dup_close.csv'] # The result bundle is a large JSON containing the results of every testcase # of the LAVA job as well as the files that were attached during the run. @@ -195,7 +197,7 @@ def get_baremetal_benchmarks_cmd(): { 'git-repo': 'https://github.com/lttng/lttng-ci.git', 'revision': 'master', - 'testdef': 'lava/baremetal-tests/failing-open-enoent.yml' + 'testdef': 'lava/baremetal-tests/success-dup-close.yml' } ], 'timeout': 18000