X-Git-Url: http://git.lttng.org./?a=blobdiff_plain;f=scripts%2Flttng-baremetal-tests%2Flava-submit.py;h=34ad7118563298c098527b7a54a75f1f165d90f6;hb=dff1609bad06d0b99beb61189ad7d4b3fe55b716;hp=391cda30fc8f83369b07c65df9d649bb71b93d62;hpb=65d04f1d172ffa4ee18b951ac9ab65fb3abd5f63;p=lttng-ci.git diff --git a/scripts/lttng-baremetal-tests/lava-submit.py b/scripts/lttng-baremetal-tests/lava-submit.py index 391cda3..34ad711 100644 --- a/scripts/lttng-baremetal-tests/lava-submit.py +++ b/scripts/lttng-baremetal-tests/lava-submit.py @@ -72,7 +72,8 @@ def fetch_benchmark_results(server, job): content = get_job_bundle_content(server, job) testcases = ['processed_results_close.csv', 'processed_results_open_efault.csv', - 'processed_results_dup_close.csv'] + 'processed_results_dup_close.csv', + 'processed_results_lttng_test_filter.csv'] # The result bundle is a large JSON containing the results of every testcase # of the LAVA job as well as the files that were attached during the run. @@ -80,7 +81,7 @@ def fetch_benchmark_results(server, job): # benchmark results produced during the run. for run in content['test_runs']: # We only care of the benchmark testcases - if 'benchmark-syscall-' in run['test_id']: + if 'benchmark-' in run['test_id']: if 'test_results' in run: for res in run['test_results']: if 'attachments' in res: @@ -198,6 +199,11 @@ def get_baremetal_benchmarks_cmd(): 'git-repo': 'https://github.com/lttng/lttng-ci.git', 'revision': 'master', 'testdef': 'lava/baremetal-tests/success-dup-close.yml' + }, + { + 'git-repo': 'https://github.com/lttng/lttng-ci.git', + 'revision': 'master', + 'testdef': 'lava/baremetal-tests/lttng-test-filter.yml' } ], 'timeout': 18000