You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
174 lines
7.0 KiB
174 lines
7.0 KiB
#!/usr/bin/python3
|
|
#
|
|
# Copyright (c) 2018 Google Inc.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
# Author: William Henning <whenning@google.com>
|
|
#
|
|
# This script parses the validation layers test continuous integration ouput
|
|
# and reports the number of tests that passed, failured, ouput unexpected errors,
|
|
# or were skipped. As such, the script is only designed to parse the ouput
|
|
# generated by the existing CI implementation.
|
|
#
|
|
# usage:
|
|
# for profile in tests/device_profiles/*.json; do echo Testing with
|
|
# profile $profile; VK_LAYER_PATH=DEVSIM_AND_VALIDATION_PATHS
|
|
# VK_DEVSIM_FILE=$profile VK_ICD_FILENAMES=MOCK_ICD_PATH
|
|
# ./build/tests/vk_layer_validation_tests --devsim; done
|
|
# | python3 parse_test_results.py [--fail_on_skip] [--fail_on_unexpected]
|
|
#
|
|
# --fail_on_skip causes the script to exit with a non-zero exit code if a test
|
|
# didn't run on any device profile
|
|
#
|
|
# --fail_on_unexpected causes the script to exit with a non-zero exit code if
|
|
# a test printed unexpected errors
|
|
#
|
|
|
|
import argparse
|
|
import re
|
|
import sys
|
|
from collections import defaultdict
|
|
|
|
class OutputStats(object):
|
|
def __init__(self):
|
|
self.current_profile = ""
|
|
self.current_test = ""
|
|
self.current_test_output = ""
|
|
self.test_results = defaultdict(defaultdict)
|
|
self.unexpected_errors = defaultdict(defaultdict)
|
|
|
|
def match(self, line):
|
|
self.new_profile_match(line)
|
|
self.test_suite_end_match(line)
|
|
self.start_test_match(line)
|
|
if self.current_test != "":
|
|
self.current_test_output += line
|
|
self.skip_test_match(line)
|
|
self.pass_test_match(line)
|
|
self.fail_test_match(line)
|
|
self.unexpected_error_match(line)
|
|
|
|
def print_summary(self, skip_is_failure, unexpected_is_failure):
|
|
if self.current_test != "":
|
|
self.test_died()
|
|
|
|
passed_tests = 0
|
|
skipped_tests = 0
|
|
failed_tests = 0
|
|
unexpected_error_tests = 0
|
|
did_fail = False
|
|
|
|
for test_name, results in self.test_results.items():
|
|
skipped_profiles = 0
|
|
passed_profiles = 0
|
|
failed_profiles = 0
|
|
aborted_profiles = 0
|
|
unexpected_error_profiles = 0
|
|
for profile, result in results.items():
|
|
if result == "pass":
|
|
passed_profiles += 1
|
|
if result == "fail":
|
|
failed_profiles += 1
|
|
if result == "skip":
|
|
skipped_profiles += 1
|
|
if self.unexpected_errors.get(test_name, {}).get(profile, "") == "true":
|
|
unexpected_error_profiles += 1
|
|
if failed_profiles != 0:
|
|
print("TEST FAILED:", test_name)
|
|
failed_tests += 1
|
|
elif skipped_profiles == len(results):
|
|
print("TEST SKIPPED ALL DEVICES:", test_name)
|
|
skipped_tests += 1
|
|
else:
|
|
passed_tests += 1
|
|
if unexpected_error_profiles != 0:
|
|
print("UNEXPECTED ERRORS:", test_name)
|
|
unexpected_error_tests += 1
|
|
num_tests = len(self.test_results)
|
|
print("PASSED: ", passed_tests, "/", num_tests, " tests")
|
|
if skipped_tests != 0:
|
|
did_fail |= skip_is_failure
|
|
print("NEVER RAN: ", skipped_tests, "/", num_tests, " tests")
|
|
if failed_tests != 0:
|
|
did_fail = True
|
|
print("FAILED: ", failed_tests, "/", num_tests, "tests")
|
|
if unexpected_error_tests != 0:
|
|
did_fail |= unexpected_is_failure
|
|
print("UNEXPECTED OUPUT: ", unexpected_error_tests, "/", num_tests, "tests")
|
|
return did_fail
|
|
|
|
def new_profile_match(self, line):
|
|
if re.search(r'Testing with profile .*/(.*)', line) is not None:
|
|
self.current_profile = re.search(r'Testing with profile .*/(.*)', line).group(1)
|
|
|
|
def test_suite_end_match(self, line):
|
|
if re.search(r'\[-*\]', line) is not None:
|
|
if self.current_test != "":
|
|
# Here we see a message that starts [----------] before another test
|
|
# finished running. This should mean that that other test died.
|
|
self.test_died()
|
|
|
|
def start_test_match(self, line):
|
|
if re.search(r'\[ RUN\s*\]', line) is not None:
|
|
# This parser doesn't handle the case where one test's start comes between another
|
|
# test's start and result.
|
|
assert self.current_test == ""
|
|
self.current_test = re.search(r'] (.*)', line).group(1)
|
|
self.current_test_output = ""
|
|
|
|
def skip_test_match(self, line):
|
|
if re.search(r'TEST SKIPPED', line) is not None:
|
|
self.test_results[self.current_test][self.current_profile] = "skip"
|
|
|
|
def pass_test_match(self, line):
|
|
if re.search(r'\[\s*OK \]', line) is not None:
|
|
# If gtest says the test passed, check if it was skipped before marking it passed
|
|
if self.test_results.get(self.current_test, {}).get(self.current_profile, "") != "skip":
|
|
self.test_results[self.current_test][self.current_profile] = "pass"
|
|
self.current_test = ""
|
|
|
|
def fail_test_match(self, line):
|
|
if re.search(r'\[\s*FAILED\s*\]', line) is not None and self.current_test != "":
|
|
self.test_results[self.current_test][self.current_profile] = "fail"
|
|
self.current_test = ""
|
|
|
|
def unexpected_error_match(self, line):
|
|
if re.search(r'^Unexpected: ', line) is not None:
|
|
self.unexpected_errors[self.current_test][self.current_profile] = "true"
|
|
|
|
def test_died(self):
|
|
print("A test likely crashed. Testing is being aborted.")
|
|
print("Final test output: ")
|
|
print(self.current_test_output)
|
|
sys.exit(1)
|
|
|
|
def main():
|
|
parser = argparse.ArgumentParser(description='Parse the output from validation layer tests.')
|
|
parser.add_argument('--fail_on_skip', action='store_true', help="Makes the script exit with a "
|
|
"non-zero exit code if a test didn't run on any device profile.")
|
|
parser.add_argument('--fail_on_unexpected', action='store_true', help="Makes the script exit "
|
|
"with a non-zero exit code if a test causes unexpected errors.")
|
|
args = parser.parse_args()
|
|
|
|
stats = OutputStats()
|
|
for line in sys.stdin:
|
|
stats.match(line)
|
|
failed = stats.print_summary(args.fail_on_skip, args.fail_on_unexpected)
|
|
if failed == True:
|
|
print("\nFAILED CI")
|
|
sys.exit(1)
|
|
|
|
if __name__ == '__main__':
|
|
main()
|