googletest/googletest/test/googletest-json-output-unittest.py
Tom Hughes d1ad27e0a4 Fix formatting in subset of Python files
These files were formatted with automated tools. The remaining Python
files require some manual fix ups, so they will be fixed separately.

PiperOrigin-RevId: 504579820
Change-Id: I3923bd414bffe3ded6163ec496cd09ace3951928
2023-01-25 09:14:26 -08:00

821 lines
27 KiB
Python

#!/usr/bin/env python
# Copyright 2018, Google Inc.
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are
# met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
# * Redistributions in binary form must reproduce the above
# copyright notice, this list of conditions and the following disclaimer
# in the documentation and/or other materials provided with the
# distribution.
# * Neither the name of Google Inc. nor the names of its
# contributors may be used to endorse or promote products derived from
# this software without specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
"""Unit test for the gtest_json_output module."""
import datetime
import errno
import json
import os
import re
import sys
from googletest.test import gtest_json_test_utils
from googletest.test import gtest_test_utils
GTEST_FILTER_FLAG = '--gtest_filter'
GTEST_LIST_TESTS_FLAG = '--gtest_list_tests'
GTEST_OUTPUT_FLAG = '--gtest_output'
GTEST_DEFAULT_OUTPUT_FILE = 'test_detail.json'
GTEST_PROGRAM_NAME = 'gtest_xml_output_unittest_'
# The flag indicating stacktraces are not supported
NO_STACKTRACE_SUPPORT_FLAG = '--no_stacktrace_support'
SUPPORTS_STACK_TRACES = NO_STACKTRACE_SUPPORT_FLAG not in sys.argv
if SUPPORTS_STACK_TRACES:
STACK_TRACE_TEMPLATE = '\nStack trace:\n*'
else:
STACK_TRACE_TEMPLATE = ''
EXPECTED_NON_EMPTY = {
'tests': 26,
'failures': 5,
'disabled': 2,
'errors': 0,
'timestamp': '*',
'time': '*',
'ad_hoc_property': '42',
'name': 'AllTests',
'testsuites': [
{
'name': 'SuccessfulTest',
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'Succeeds',
'file': 'gtest_xml_output_unittest_.cc',
'line': 51,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'SuccessfulTest',
}],
},
{
'name': 'FailedTest',
'tests': 1,
'failures': 1,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'Fails',
'file': 'gtest_xml_output_unittest_.cc',
'line': 59,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'FailedTest',
'failures': [{
'failure': (
'gtest_xml_output_unittest_.cc:*\n'
'Expected equality of these values:\n'
' 1\n 2'
+ STACK_TRACE_TEMPLATE
),
'type': '',
}],
}],
},
{
'name': 'DisabledTest',
'tests': 1,
'failures': 0,
'disabled': 1,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'DISABLED_test_not_run',
'file': 'gtest_xml_output_unittest_.cc',
'line': 66,
'status': 'NOTRUN',
'result': 'SUPPRESSED',
'time': '*',
'timestamp': '*',
'classname': 'DisabledTest',
}],
},
{
'name': 'SkippedTest',
'tests': 3,
'failures': 1,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [
{
'name': 'Skipped',
'file': 'gtest_xml_output_unittest_.cc',
'line': 73,
'status': 'RUN',
'result': 'SKIPPED',
'time': '*',
'timestamp': '*',
'classname': 'SkippedTest',
},
{
'name': 'SkippedWithMessage',
'file': 'gtest_xml_output_unittest_.cc',
'line': 77,
'status': 'RUN',
'result': 'SKIPPED',
'time': '*',
'timestamp': '*',
'classname': 'SkippedTest',
},
{
'name': 'SkippedAfterFailure',
'file': 'gtest_xml_output_unittest_.cc',
'line': 81,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'SkippedTest',
'failures': [{
'failure': (
'gtest_xml_output_unittest_.cc:*\n'
'Expected equality of these values:\n'
' 1\n 2'
+ STACK_TRACE_TEMPLATE
),
'type': '',
}],
},
],
},
{
'name': 'MixedResultTest',
'tests': 3,
'failures': 1,
'disabled': 1,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [
{
'name': 'Succeeds',
'file': 'gtest_xml_output_unittest_.cc',
'line': 86,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'MixedResultTest',
},
{
'name': 'Fails',
'file': 'gtest_xml_output_unittest_.cc',
'line': 91,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'MixedResultTest',
'failures': [
{
'failure': (
'gtest_xml_output_unittest_.cc:*\n'
'Expected equality of these values:\n'
' 1\n 2'
+ STACK_TRACE_TEMPLATE
),
'type': '',
},
{
'failure': (
'gtest_xml_output_unittest_.cc:*\n'
'Expected equality of these values:\n'
' 2\n 3'
+ STACK_TRACE_TEMPLATE
),
'type': '',
},
],
},
{
'name': 'DISABLED_test',
'file': 'gtest_xml_output_unittest_.cc',
'line': 96,
'status': 'NOTRUN',
'result': 'SUPPRESSED',
'time': '*',
'timestamp': '*',
'classname': 'MixedResultTest',
},
],
},
{
'name': 'XmlQuotingTest',
'tests': 1,
'failures': 1,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'OutputsCData',
'file': 'gtest_xml_output_unittest_.cc',
'line': 100,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'XmlQuotingTest',
'failures': [{
'failure': (
'gtest_xml_output_unittest_.cc:*\n'
'Failed\nXML output: <?xml encoding="utf-8">'
'<top><![CDATA[cdata text]]></top>'
+ STACK_TRACE_TEMPLATE
),
'type': '',
}],
}],
},
{
'name': 'InvalidCharactersTest',
'tests': 1,
'failures': 1,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'InvalidCharactersInMessage',
'file': 'gtest_xml_output_unittest_.cc',
'line': 107,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'InvalidCharactersTest',
'failures': [{
'failure': (
'gtest_xml_output_unittest_.cc:*\n'
'Failed\nInvalid characters in brackets'
' [\x01\x02]'
+ STACK_TRACE_TEMPLATE
),
'type': '',
}],
}],
},
{
'name': 'PropertyRecordingTest',
'tests': 4,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'SetUpTestSuite': 'yes',
'TearDownTestSuite': 'aye',
'testsuite': [
{
'name': 'OneProperty',
'file': 'gtest_xml_output_unittest_.cc',
'line': 119,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'PropertyRecordingTest',
'key_1': '1',
},
{
'name': 'IntValuedProperty',
'file': 'gtest_xml_output_unittest_.cc',
'line': 123,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'PropertyRecordingTest',
'key_int': '1',
},
{
'name': 'ThreeProperties',
'file': 'gtest_xml_output_unittest_.cc',
'line': 127,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'PropertyRecordingTest',
'key_1': '1',
'key_2': '2',
'key_3': '3',
},
{
'name': 'TwoValuesForOneKeyUsesLastValue',
'file': 'gtest_xml_output_unittest_.cc',
'line': 133,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'PropertyRecordingTest',
'key_1': '2',
},
],
},
{
'name': 'NoFixtureTest',
'tests': 3,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [
{
'name': 'RecordProperty',
'file': 'gtest_xml_output_unittest_.cc',
'line': 138,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'NoFixtureTest',
'key': '1',
},
{
'name': 'ExternalUtilityThatCallsRecordIntValuedProperty',
'file': 'gtest_xml_output_unittest_.cc',
'line': 151,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'NoFixtureTest',
'key_for_utility_int': '1',
},
{
'name': (
'ExternalUtilityThatCallsRecordStringValuedProperty'
),
'file': 'gtest_xml_output_unittest_.cc',
'line': 155,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'NoFixtureTest',
'key_for_utility_string': '1',
},
],
},
{
'name': 'TypedTest/0',
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'HasTypeParamAttribute',
'type_param': 'int',
'file': 'gtest_xml_output_unittest_.cc',
'line': 171,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'TypedTest/0',
}],
},
{
'name': 'TypedTest/1',
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'HasTypeParamAttribute',
'type_param': 'long',
'file': 'gtest_xml_output_unittest_.cc',
'line': 171,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'TypedTest/1',
}],
},
{
'name': 'Single/TypeParameterizedTestSuite/0',
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'HasTypeParamAttribute',
'type_param': 'int',
'file': 'gtest_xml_output_unittest_.cc',
'line': 178,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'Single/TypeParameterizedTestSuite/0',
}],
},
{
'name': 'Single/TypeParameterizedTestSuite/1',
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'HasTypeParamAttribute',
'type_param': 'long',
'file': 'gtest_xml_output_unittest_.cc',
'line': 178,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'Single/TypeParameterizedTestSuite/1',
}],
},
{
'name': 'Single/ValueParamTest',
'tests': 4,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [
{
'name': 'HasValueParamAttribute/0',
'value_param': '33',
'file': 'gtest_xml_output_unittest_.cc',
'line': 162,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'Single/ValueParamTest',
},
{
'name': 'HasValueParamAttribute/1',
'value_param': '42',
'file': 'gtest_xml_output_unittest_.cc',
'line': 162,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'Single/ValueParamTest',
},
{
'name': 'AnotherTestThatHasValueParamAttribute/0',
'value_param': '33',
'file': 'gtest_xml_output_unittest_.cc',
'line': 163,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'Single/ValueParamTest',
},
{
'name': 'AnotherTestThatHasValueParamAttribute/1',
'value_param': '42',
'file': 'gtest_xml_output_unittest_.cc',
'line': 163,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'Single/ValueParamTest',
},
],
},
],
}
EXPECTED_FILTERED = {
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'name': 'AllTests',
'ad_hoc_property': '42',
'testsuites': [{
'name': 'SuccessfulTest',
'tests': 1,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': 'Succeeds',
'file': 'gtest_xml_output_unittest_.cc',
'line': 51,
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': 'SuccessfulTest',
}],
}],
}
EXPECTED_NO_TEST = {
'tests': 0,
'failures': 0,
'disabled': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'name': 'AllTests',
'testsuites': [{
'name': 'NonTestSuiteFailure',
'tests': 1,
'failures': 1,
'disabled': 0,
'skipped': 0,
'errors': 0,
'time': '*',
'timestamp': '*',
'testsuite': [{
'name': '',
'status': 'RUN',
'result': 'COMPLETED',
'time': '*',
'timestamp': '*',
'classname': '',
'failures': [{
'failure': (
'gtest_no_test_unittest.cc:*\n'
'Expected equality of these values:\n'
' 1\n 2'
+ STACK_TRACE_TEMPLATE
),
'type': '',
}],
}],
}],
}
GTEST_PROGRAM_PATH = gtest_test_utils.GetTestExecutablePath(GTEST_PROGRAM_NAME)
SUPPORTS_TYPED_TESTS = (
'TypedTest'
in gtest_test_utils.Subprocess(
[GTEST_PROGRAM_PATH, GTEST_LIST_TESTS_FLAG], capture_stderr=False
).output
)
class GTestJsonOutputUnitTest(gtest_test_utils.TestCase):
"""Unit test for Google Test's JSON output functionality."""
# This test currently breaks on platforms that do not support typed and
# type-parameterized tests, so we don't run it under them.
if SUPPORTS_TYPED_TESTS:
def testNonEmptyJsonOutput(self):
"""Verifies JSON output for a Google Test binary with non-empty output.
Runs a test program that generates a non-empty JSON output, and
tests that the JSON output is expected.
"""
self._TestJsonOutput(GTEST_PROGRAM_NAME, EXPECTED_NON_EMPTY, 1)
def testNoTestJsonOutput(self):
"""Verifies JSON output for a Google Test binary without actual tests.
Runs a test program that generates an JSON output for a binary with no
tests, and tests that the JSON output is expected.
"""
self._TestJsonOutput('gtest_no_test_unittest', EXPECTED_NO_TEST, 0)
def testTimestampValue(self):
"""Checks whether the timestamp attribute in the JSON output is valid.
Runs a test program that generates an empty JSON output, and checks if
the timestamp attribute in the testsuites tag is valid.
"""
actual = self._GetJsonOutput('gtest_no_test_unittest', [], 0)
date_time_str = actual['timestamp']
# datetime.strptime() is only available in Python 2.5+ so we have to
# parse the expected datetime manually.
match = re.match(r'(\d+)-(\d\d)-(\d\d)T(\d\d):(\d\d):(\d\d)', date_time_str)
self.assertTrue(
re.match,
'JSON datettime string %s has incorrect format' % date_time_str,
)
date_time_from_json = datetime.datetime(
year=int(match.group(1)),
month=int(match.group(2)),
day=int(match.group(3)),
hour=int(match.group(4)),
minute=int(match.group(5)),
second=int(match.group(6)),
)
time_delta = abs(datetime.datetime.now() - date_time_from_json)
# timestamp value should be near the current local time
self.assertTrue(
time_delta < datetime.timedelta(seconds=600),
'time_delta is %s' % time_delta,
)
def testDefaultOutputFile(self):
"""Verifies the default output file name.
Confirms that Google Test produces an JSON output file with the expected
default name if no name is explicitly specified.
"""
output_file = os.path.join(
gtest_test_utils.GetTempDir(), GTEST_DEFAULT_OUTPUT_FILE
)
gtest_prog_path = gtest_test_utils.GetTestExecutablePath(
'gtest_no_test_unittest'
)
try:
os.remove(output_file)
except OSError:
e = sys.exc_info()[1]
if e.errno != errno.ENOENT:
raise
p = gtest_test_utils.Subprocess(
[gtest_prog_path, '%s=json' % GTEST_OUTPUT_FLAG],
working_dir=gtest_test_utils.GetTempDir(),
)
self.assertTrue(p.exited)
self.assertEqual(0, p.exit_code)
self.assertTrue(os.path.isfile(output_file))
def testSuppressedJsonOutput(self):
"""Verifies that no JSON output is generated.
Tests that no JSON file is generated if the default JSON listener is
shut down before RUN_ALL_TESTS is invoked.
"""
json_path = os.path.join(
gtest_test_utils.GetTempDir(), GTEST_PROGRAM_NAME + 'out.json'
)
if os.path.isfile(json_path):
os.remove(json_path)
command = [
GTEST_PROGRAM_PATH,
'%s=json:%s' % (GTEST_OUTPUT_FLAG, json_path),
'--shut_down_xml',
]
p = gtest_test_utils.Subprocess(command)
if p.terminated_by_signal:
# p.signal is available only if p.terminated_by_signal is True.
self.assertFalse(
p.terminated_by_signal,
'%s was killed by signal %d' % (GTEST_PROGRAM_NAME, p.signal),
)
else:
self.assertTrue(p.exited)
self.assertEqual(
1,
p.exit_code,
"'%s' exited with code %s, which doesn't match "
'the expected exit code %s.' % (command, p.exit_code, 1),
)
self.assertTrue(not os.path.isfile(json_path))
def testFilteredTestJsonOutput(self):
"""Verifies JSON output when a filter is applied.
Runs a test program that executes only some tests and verifies that
non-selected tests do not show up in the JSON output.
"""
self._TestJsonOutput(
GTEST_PROGRAM_NAME,
EXPECTED_FILTERED,
0,
extra_args=['%s=SuccessfulTest.*' % GTEST_FILTER_FLAG],
)
def _GetJsonOutput(self, gtest_prog_name, extra_args, expected_exit_code):
"""Returns the JSON output generated by running the program gtest_prog_name.
Furthermore, the program's exit code must be expected_exit_code.
Args:
gtest_prog_name: Google Test binary name.
extra_args: extra arguments to binary invocation.
expected_exit_code: program's exit code.
"""
json_path = os.path.join(
gtest_test_utils.GetTempDir(), gtest_prog_name + 'out.json'
)
gtest_prog_path = gtest_test_utils.GetTestExecutablePath(gtest_prog_name)
command = [
gtest_prog_path,
'%s=json:%s' % (GTEST_OUTPUT_FLAG, json_path),
] + extra_args
p = gtest_test_utils.Subprocess(command)
if p.terminated_by_signal:
self.assertTrue(
False, '%s was killed by signal %d' % (gtest_prog_name, p.signal)
)
else:
self.assertTrue(p.exited)
self.assertEqual(
expected_exit_code,
p.exit_code,
"'%s' exited with code %s, which doesn't match "
'the expected exit code %s.'
% (command, p.exit_code, expected_exit_code),
)
with open(json_path) as f:
actual = json.load(f)
return actual
def _TestJsonOutput(
self, gtest_prog_name, expected, expected_exit_code, extra_args=None
):
"""Checks the JSON output generated by the Google Test binary.
Asserts that the JSON document generated by running the program
gtest_prog_name matches expected_json, a string containing another
JSON document. Furthermore, the program's exit code must be
expected_exit_code.
Args:
gtest_prog_name: Google Test binary name.
expected: expected output.
expected_exit_code: program's exit code.
extra_args: extra arguments to binary invocation.
"""
actual = self._GetJsonOutput(
gtest_prog_name, extra_args or [], expected_exit_code
)
self.assertEqual(expected, gtest_json_test_utils.normalize(actual))
if __name__ == '__main__':
if NO_STACKTRACE_SUPPORT_FLAG in sys.argv:
# unittest.main() can't handle unknown flags
sys.argv.remove(NO_STACKTRACE_SUPPORT_FLAG)
os.environ['GTEST_STACK_TRACE_DEPTH'] = '1'
gtest_test_utils.Main()