1
0
mirror of https://github.com/samba-team/samba.git synced 2025-03-30 06:50:24 +03:00

selftest: Display unexpected successes and expected failures.

This commit is contained in:
Jelmer Vernooij 2011-12-04 01:55:23 +01:00
parent e62654578b
commit 0fee2c01b8
2 changed files with 52 additions and 15 deletions

View File

@ -1,7 +1,22 @@
#!/usr/bin/env python
# Filter a subunit stream
# Copyright (C) Jelmer Vernooij <jelmer@samba.org>
# Published under the GNU GPL, v3 or later
# Copyright (C) 2009-2011 Jelmer Vernooij <jelmer@samba.org>
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 3 of the License, or
# (at your option) any later version.
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
# NOTE: This script is a hack, meant as a placeholder until we can migrate
# to upstream subunit's filtering tools.
import optparse
import os

View File

@ -22,8 +22,9 @@ import sys
import subunit
import subunit.iso8601
import testtools
from testtools import content, content_type
VALID_RESULTS = ['success', 'successful', 'failure', 'fail', 'skip', 'knownfail', 'error', 'xfail', 'skip-testsuite', 'testsuite-failure', 'testsuite-xfail', 'testsuite-success', 'testsuite-error']
VALID_RESULTS = ['success', 'successful', 'failure', 'fail', 'skip', 'knownfail', 'error', 'xfail', 'skip-testsuite', 'testsuite-failure', 'testsuite-xfail', 'testsuite-success', 'testsuite-error', 'uxsuccess']
class TestsuiteEnabledTestResult(testtools.testresult.TestResult):
@ -33,7 +34,6 @@ class TestsuiteEnabledTestResult(testtools.testresult.TestResult):
def parse_results(msg_ops, statistics, fh):
exitcode = 0
expected_fail = 0
open_tests = {}
while fh:
@ -111,7 +111,17 @@ def parse_results(msg_ops, statistics, fh):
else:
statistics['TESTS_EXPECTED_FAIL']+=1
msg_ops.addExpectedFailure(test, remote_error)
expected_fail+=1
elif result in ("uxsuccess", ):
try:
test = open_tests.pop(testname)
except KeyError:
statistics['TESTS_ERROR']+=1
exitcode = 1
msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"Test was never started"))
else:
statistics['TESTS_UNEXPECTED_OK']+=1
msg_ops.addUnexpectedSuccess(test, remote_error)
exitcode = 1
elif result in ("failure", "fail"):
try:
test = open_tests.pop(testname)
@ -285,6 +295,11 @@ class FilterOps(testtools.testresult.TestResult):
self._ops.addExpectedFailure(test, details)
self.output = None
def addUnexpectedSuccess(self, test, details=None):
test = self._add_prefix(test)
self._ops.addUnexpectedSuccess(test, details)
self.output = None
def addFailure(self, test, details=None):
test = self._add_prefix(test)
xfail_reason = find_in_list(self.expected_failures, test.id())
@ -314,10 +329,11 @@ class FilterOps(testtools.testresult.TestResult):
if xfail_reason is not None:
self.uxsuccess_added += 1
self.total_uxsuccess += 1
if details is not None:
details = subunit.RemoteError(unicode(details[1]) + xfail_reason.decode("utf-8"))
else:
details = subunit.RemoteError(xfail_reason.decode("utf-8"))
if details is None:
details = {}
details['reason'] = content.Content(
content_type.ContentType("text", "plain",
{"charset": "utf8"}), lambda: xfail_reason)
self._ops.addUnexpectedSuccess(test, details)
if self.output:
self._ops.output_msg(self.output)
@ -508,10 +524,13 @@ class PlainFormatter(TestsuiteEnabledTestResult):
def addSkip(self, test, details=None):
self.end_test(test.id(), "skip", False, details)
def addExpectedFail(self, test, details=None):
def addExpectedFailure(self, test, details=None):
self.end_test(test.id(), "xfail", False, details)
def end_test(self, testname, result, unexpected, reason=None):
def addUnexpectedSuccess(self, test, details=None):
self.end_test(test.id(), "uxsuccess", True, details)
def end_test(self, testname, result, unexpected, details=None):
if not unexpected:
self.test_output[self.name] = ""
if not self.immediate:
@ -526,17 +545,18 @@ class PlainFormatter(TestsuiteEnabledTestResult):
self.test_output[self.name] = ""
self.test_output[self.name] += "UNEXPECTED(%s): %s\n" % (result, testname)
if reason is not None:
self.test_output[self.name] += "REASON: %s\n" % (unicode(reason[1]).encode("utf-8").strip(),)
if details is not None:
self.test_output[self.name] += "REASON: %s\n" % (unicode(details[1]).encode("utf-8").strip(),)
if self.immediate and not self.verbose:
print self.test_output[self.name]
sys.stdout.write(self.test_output[self.name])
self.test_output[self.name] = ""
if not self.immediate:
sys.stdout.write({
'error': 'E',
'failure': 'F',
'uxsuccess': 'U',
'success': 'S'}.get(result, "?"))
def write_summary(self, path):
@ -570,14 +590,16 @@ class PlainFormatter(TestsuiteEnabledTestResult):
if (not self.suitesfailed and
not self.statistics['TESTS_UNEXPECTED_FAIL'] and
not self.statistics['TESTS_UNEXPECTED_OK'] and
not self.statistics['TESTS_ERROR']):
ok = (self.statistics['TESTS_EXPECTED_OK'] +
self.statistics['TESTS_EXPECTED_FAIL'])
print "\nALL OK (%d tests in %d testsuites)" % (ok, self.suites_ok)
else:
print "\nFAILED (%d failures and %d errors in %d testsuites)" % (
print "\nFAILED (%d failures, %d errors and %d unexpected successes in %d testsuites)" % (
self.statistics['TESTS_UNEXPECTED_FAIL'],
self.statistics['TESTS_ERROR'],
self.statistics['TESTS_UNEXPECTED_OK'],
len(self.suitesfailed))
def skip_testsuite(self, name, reason="UNKNOWN"):