# You should have received a copy of the GNU General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
+from __future__ import print_function
__all__ = ['parse_results']
+import datetime
import re
import sys
-import subunit
-import subunit.iso8601
-import testtools
+import os
+from samba import subunit
+from samba.subunit.run import TestProtocolClient
+from samba.subunit import iso8601
+import unittest
+from samba.compat import binary_type
-VALID_RESULTS = ['success', 'successful', 'failure', 'fail', 'skip', 'knownfail', 'error', 'xfail', 'skip-testsuite', 'testsuite-failure', 'testsuite-xfail', 'testsuite-success', 'testsuite-error']
-class TestsuiteEnabledTestResult(testtools.testresult.TestResult):
+VALID_RESULTS = set(['success', 'successful', 'failure', 'fail', 'skip',
+ 'knownfail', 'error', 'xfail', 'skip-testsuite',
+ 'testsuite-failure', 'testsuite-xfail',
+ 'testsuite-success', 'testsuite-error',
+ 'uxsuccess', 'testsuite-uxsuccess'])
+
+
+class TestsuiteEnabledTestResult(unittest.TestResult):
def start_testsuite(self, name):
raise NotImplementedError(self.start_testsuite)
def parse_results(msg_ops, statistics, fh):
- expected_fail = 0
- open_tests = []
+ exitcode = 0
+ open_tests = {}
while fh:
l = fh.readline()
arg = parts[1]
if command in ("test", "testing"):
msg_ops.control_msg(l)
- msg_ops.startTest(subunit.RemotedTestCase(arg.rstrip()))
- open_tests.append(arg.rstrip())
+ name = arg.rstrip()
+ test = subunit.RemotedTestCase(name)
+ if name in open_tests:
+ msg_ops.addError(open_tests.pop(name), subunit.RemoteError(u"Test already running"))
+ msg_ops.startTest(test)
+ open_tests[name] = test
elif command == "time":
msg_ops.control_msg(l)
try:
- dt = subunit.iso8601.parse_date(arg.rstrip("\n"))
- except TypeError, e:
- print "Unable to parse time line: %s" % arg.rstrip("\n")
+ dt = iso8601.parse_date(arg.rstrip("\n"))
+ except TypeError as e:
+ print("Unable to parse time line: %s" % arg.rstrip("\n"))
else:
msg_ops.time(dt)
elif command in VALID_RESULTS:
if l == "":
break
msg_ops.control_msg(l)
- if l == "]\n":
+ if l[-2:] == "]\n":
+ reason += l[:-2]
terminated = True
break
else:
reason += l
+ if isinstance(reason, binary_type):
+ remote_error = subunit.RemoteError(reason.decode("utf-8"))
+ else:
+ remote_error = subunit.RemoteError(reason)
+
if not terminated:
- statistics['TESTS_ERROR']+=1
- msg_ops.addError(subunit.RemotedTestCase(testname), "reason (%s) interrupted" % result)
+ statistics['TESTS_ERROR'] += 1
+ msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"reason (%s) interrupted" % result))
return 1
else:
reason = None
+ remote_error = subunit.RemoteError(u"No reason specified")
if result in ("success", "successful"):
try:
- open_tests.remove(testname)
- except ValueError:
- statistics['TESTS_ERROR']+=1
- msg_ops.addError(subunit.RemotedTestCase(testname), "Test was never started")
+ test = open_tests.pop(testname)
+ except KeyError:
+ statistics['TESTS_ERROR'] += 1
+ exitcode = 1
+ msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"Test was never started"))
else:
- statistics['TESTS_EXPECTED_OK']+=1
- msg_ops.end_test(testname, "success", False, reason)
+ statistics['TESTS_EXPECTED_OK'] += 1
+ msg_ops.addSuccess(test)
elif result in ("xfail", "knownfail"):
try:
- open_tests.remove(testname)
- except ValueError:
- statistics['TESTS_ERROR']+=1
- msg_ops.addError(subunit.RemotedTestCase(testname), "Test was never started")
+ test = open_tests.pop(testname)
+ except KeyError:
+ statistics['TESTS_ERROR'] += 1
+ exitcode = 1
+ msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"Test was never started"))
+ else:
+ statistics['TESTS_EXPECTED_FAIL'] += 1
+ msg_ops.addExpectedFailure(test, remote_error)
+ elif result in ("uxsuccess", ):
+ try:
+ test = open_tests.pop(testname)
+ except KeyError:
+ statistics['TESTS_ERROR'] += 1
+ exitcode = 1
+ msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"Test was never started"))
else:
- statistics['TESTS_EXPECTED_FAIL']+=1
- msg_ops.end_test(testname, "xfail", False, reason)
- expected_fail+=1
+ statistics['TESTS_UNEXPECTED_OK'] += 1
+ msg_ops.addUnexpectedSuccess(test)
+ exitcode = 1
elif result in ("failure", "fail"):
try:
- open_tests.remove(testname)
- except ValueError:
- statistics['TESTS_ERROR']+=1
- msg_ops.addError(subunit.RemotedTestCase(testname), "Test was never started")
+ test = open_tests.pop(testname)
+ except KeyError:
+ statistics['TESTS_ERROR'] += 1
+ exitcode = 1
+ msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"Test was never started"))
else:
- statistics['TESTS_UNEXPECTED_FAIL']+=1
- msg_ops.end_test(testname, "failure", True, reason)
+ statistics['TESTS_UNEXPECTED_FAIL'] += 1
+ exitcode = 1
+ msg_ops.addFailure(test, remote_error)
elif result == "skip":
- statistics['TESTS_SKIP']+=1
+ statistics['TESTS_SKIP'] += 1
# Allow tests to be skipped without prior announcement of test
- last = open_tests.pop()
- if last is not None and last != testname:
- open_tests.append(testname)
- msg_ops.end_test(testname, "skip", False, reason)
+ try:
+ test = open_tests.pop(testname)
+ except KeyError:
+ test = subunit.RemotedTestCase(testname)
+ msg_ops.addSkip(test, reason)
elif result == "error":
- statistics['TESTS_ERROR']+=1
+ statistics['TESTS_ERROR'] += 1
+ exitcode = 1
try:
- open_tests.remove(testname)
- except ValueError:
- pass
- msg_ops.addError(subunit.RemotedTestCase(testname), reason)
+ test = open_tests.pop(testname)
+ except KeyError:
+ test = subunit.RemotedTestCase(testname)
+ msg_ops.addError(test, remote_error)
elif result == "skip-testsuite":
msg_ops.skip_testsuite(testname)
elif result == "testsuite-success":
msg_ops.end_testsuite(testname, "success", reason)
elif result == "testsuite-failure":
msg_ops.end_testsuite(testname, "failure", reason)
+ exitcode = 1
elif result == "testsuite-xfail":
msg_ops.end_testsuite(testname, "xfail", reason)
+ elif result == "testsuite-uxsuccess":
+ msg_ops.end_testsuite(testname, "uxsuccess", reason)
+ exitcode = 1
elif result == "testsuite-error":
msg_ops.end_testsuite(testname, "error", reason)
+ exitcode = 1
else:
raise AssertionError("Recognized but unhandled result %r" %
- result)
+ result)
elif command == "testsuite":
msg_ops.start_testsuite(arg.strip())
elif command == "progress":
msg_ops.output_msg(l)
while open_tests:
- msg_ops.end_test(open_tests.pop(), "error", True,
- "was started but never finished!")
- statistics['TESTS_ERROR']+=1
-
- if statistics['TESTS_ERROR'] > 0:
- return 1
- if statistics['TESTS_UNEXPECTED_FAIL'] > 0:
- return 1
- return 0
-
+ test = subunit.RemotedTestCase(open_tests.popitem()[1])
+ msg_ops.addError(test, subunit.RemoteError(u"was started but never finished!"))
+ statistics['TESTS_ERROR'] += 1
+ exitcode = 1
-class SubunitOps(subunit.TestProtocolClient,TestsuiteEnabledTestResult):
+ return exitcode
- def addError(self, test, details=None):
- self.end_test(test.id(), "error", details)
- def end_test(self, name, result, reason=None):
- if reason:
- self._stream.write("%s: %s [\n%s\n]\n" % (result, name, reason))
- else:
- self._stream.write("%s: %s\n" % (result, name))
+class SubunitOps(TestProtocolClient, TestsuiteEnabledTestResult):
- def skip_test(self, name, reason=None):
- self.end_test(name, "skip", reason)
-
- def fail_test(self, name, reason=None):
- self.end_test(name, "fail", reason)
-
- def success_test(self, name, reason=None):
- self.end_test(name, "success", reason)
-
- def xfail_test(self, name, reason=None):
- self.end_test(name, "xfail", reason)
+ def progress(self, count, whence):
+ if whence == subunit.PROGRESS_POP:
+ self._stream.write("progress: pop\n")
+ elif whence == subunit.PROGRESS_PUSH:
+ self._stream.write("progress: push\n")
+ elif whence == subunit.PROGRESS_SET:
+ self._stream.write("progress: %d\n" % count)
+ elif whence == subunit.PROGRESS_CUR:
+ raise NotImplementedError
# The following are Samba extensions:
def start_testsuite(self, name):
else:
self._stream.write("testsuite-%s: %s\n" % (result, name))
+ def output_msg(self, msg):
+ self._stream.write(msg)
+
-def read_test_regexes(name):
+def read_test_regexes(*names):
ret = {}
- f = open(name, 'r')
- try:
- for l in f:
- l = l.strip()
- if l == "" or l[0] == "#":
- continue
- if "#" in l:
- (regex, reason) = l.split("#", 1)
- ret[regex.strip()] = reason.strip()
- else:
- ret[l] = None
- finally:
- f.close()
+ files = []
+ for name in names:
+ # if we are given a directory, we read all the files it contains
+ # (except the ones that end with "~").
+ if os.path.isdir(name):
+ files.extend([os.path.join(name, x)
+ for x in os.listdir(name)
+ if x[-1] != '~'])
+ else:
+ files.append(name)
+
+ for filename in files:
+ f = open(filename, 'r')
+ try:
+ for l in f:
+ l = l.strip()
+ if l == "" or l[0] == "#":
+ continue
+ if "#" in l:
+ (regex, reason) = l.split("#", 1)
+ ret[regex.strip()] = reason.strip()
+ else:
+ ret[l] = None
+ finally:
+ f.close()
return ret
def find_in_list(regexes, fullname):
- for regex, reason in regexes.iteritems():
+ for regex, reason in regexes.items():
if re.match(regex, fullname):
if reason is None:
return ""
return None
-class FilterOps(testtools.testresult.TestResult):
+class ImmediateFail(Exception):
+ """Raised to abort immediately."""
+
+ def __init__(self):
+ super(ImmediateFail, self).__init__("test failed and fail_immediately set")
+
+
+class FilterOps(unittest.TestResult):
def control_msg(self, msg):
- pass # We regenerate control messages, so ignore this
+ pass # We regenerate control messages, so ignore this
def time(self, time):
self._ops.time(time)
if self.output is None:
sys.stdout.write(msg)
else:
- self.output+=msg
+ self.output += msg
def startTest(self, test):
- if self.prefix is not None:
- test = subunit.RemotedTestCase(self.prefix + test.id())
-
+ self.seen_output = True
+ test = self._add_prefix(test)
if self.strip_ok_output:
- self.output = ""
+ self.output = ""
self._ops.startTest(test)
- def addError(self, test, details=None):
- self.end_test(test.id(), "error", details)
+ def _add_prefix(self, test):
+ return subunit.RemotedTestCase(self.prefix + test.id() + self.suffix)
- def end_test(self, testname, result, unexpected, reason):
- if self.prefix is not None:
- testname = self.prefix + testname
-
- if result in ("fail", "failure") and not unexpected:
- result = "xfail"
- self.xfail_added+=1
- self.total_xfail+=1
- xfail_reason = find_in_list(self.expected_failures, testname)
- if xfail_reason is not None and result in ("fail", "failure"):
- result = "xfail"
- self.xfail_added+=1
- self.total_xfail+=1
- reason += xfail_reason
+ def addError(self, test, err=None):
+ test = self._add_prefix(test)
+ self.error_added += 1
+ self.total_error += 1
+ self._ops.addError(test, err)
+ self.output = None
+ if self.fail_immediately:
+ raise ImmediateFail()
- if result in ("fail", "failure"):
- self.fail_added+=1
- self.total_fail+=1
+ def addSkip(self, test, reason=None):
+ self.seen_output = True
+ test = self._add_prefix(test)
+ self._ops.addSkip(test, reason)
+ self.output = None
- if result == "error":
- self.error_added+=1
- self.total_error+=1
+ def addExpectedFailure(self, test, err=None):
+ test = self._add_prefix(test)
+ self._ops.addExpectedFailure(test, err)
+ self.output = None
- if self.strip_ok_output:
- if result not in ("success", "xfail", "skip"):
- print self.output
+ def addUnexpectedSuccess(self, test):
+ test = self._add_prefix(test)
+ self.uxsuccess_added += 1
+ self.total_uxsuccess += 1
+ self._ops.addUnexpectedSuccess(test)
+ if self.output:
+ self._ops.output_msg(self.output)
+ self.output = None
+ if self.fail_immediately:
+ raise ImmediateFail()
+
+ def addFailure(self, test, err=None):
+ test = self._add_prefix(test)
+ xfail_reason = find_in_list(self.expected_failures, test.id())
+ if xfail_reason is None:
+ xfail_reason = find_in_list(self.flapping, test.id())
+ if xfail_reason is not None:
+ self.xfail_added += 1
+ self.total_xfail += 1
+ self._ops.addExpectedFailure(test, err)
+ else:
+ self.fail_added += 1
+ self.total_fail += 1
+ self._ops.addFailure(test, err)
+ if self.output:
+ self._ops.output_msg(self.output)
+ if self.fail_immediately:
+ raise ImmediateFail()
self.output = None
- self._ops.end_test(testname, result, reason)
+ def addSuccess(self, test):
+ test = self._add_prefix(test)
+ xfail_reason = find_in_list(self.expected_failures, test.id())
+ if xfail_reason is not None:
+ self.uxsuccess_added += 1
+ self.total_uxsuccess += 1
+ self._ops.addUnexpectedSuccess(test)
+ if self.output:
+ self._ops.output_msg(self.output)
+ if self.fail_immediately:
+ raise ImmediateFail()
+ else:
+ self._ops.addSuccess(test)
+ self.output = None
def skip_testsuite(self, name, reason=None):
self._ops.skip_testsuite(name, reason)
def start_testsuite(self, name):
self._ops.start_testsuite(name)
-
self.error_added = 0
self.fail_added = 0
self.xfail_added = 0
+ self.uxsuccess_added = 0
def end_testsuite(self, name, result, reason=None):
xfail = False
if self.xfail_added > 0:
xfail = True
- if self.fail_added > 0 or self.error_added > 0:
+ if self.fail_added > 0 or self.error_added > 0 or self.uxsuccess_added > 0:
xfail = False
if xfail and result in ("fail", "failure"):
result = "xfail"
+ if self.uxsuccess_added > 0 and result != "uxsuccess":
+ result = "uxsuccess"
+ if reason is None:
+ reason = "Subunit/Filter Reason"
+ reason += "\n uxsuccess[%d]" % self.uxsuccess_added
+
if self.fail_added > 0 and result != "failure":
result = "failure"
if reason is None:
reason += "\n errors[%d]" % self.error_added
self._ops.end_testsuite(name, result, reason)
+ if result not in ("success", "xfail"):
+ if self.output:
+ self._ops.output_msg(self.output)
+ if self.fail_immediately:
+ raise ImmediateFail()
+ self.output = None
- def __init__(self, out, prefix, expected_failures, strip_ok_output):
+ def __init__(self, out, prefix=None, suffix=None, expected_failures=None,
+ strip_ok_output=False, fail_immediately=False,
+ flapping=None):
self._ops = out
+ self.seen_output = False
self.output = None
self.prefix = prefix
- self.expected_failures = expected_failures
+ self.suffix = suffix
+ if expected_failures is not None:
+ self.expected_failures = expected_failures
+ else:
+ self.expected_failures = {}
+ if flapping is not None:
+ self.flapping = flapping
+ else:
+ self.flapping = {}
self.strip_ok_output = strip_ok_output
self.xfail_added = 0
+ self.fail_added = 0
+ self.uxsuccess_added = 0
self.total_xfail = 0
self.total_error = 0
self.total_fail = 0
+ self.total_uxsuccess = 0
+ self.error_added = 0
+ self.fail_immediately = fail_immediately
+
+
+class PerfFilterOps(unittest.TestResult):
+
+ def progress(self, delta, whence):
+ pass
+
+ def output_msg(self, msg):
+ pass
+
+ def control_msg(self, msg):
+ pass
+
+ def skip_testsuite(self, name, reason=None):
+ self._ops.skip_testsuite(name, reason)
+
+ def start_testsuite(self, name):
+ self.suite_has_time = False
+
+ def end_testsuite(self, name, result, reason=None):
+ pass
+
+ def _add_prefix(self, test):
+ return subunit.RemotedTestCase(self.prefix + test.id() + self.suffix)
+
+ def time(self, time):
+ self.latest_time = time
+ #self._ops.output_msg("found time %s\n" % time)
+ self.suite_has_time = True
+
+ def get_time(self):
+ if self.suite_has_time:
+ return self.latest_time
+ return datetime.datetime.utcnow()
+
+ def startTest(self, test):
+ self.seen_output = True
+ test = self._add_prefix(test)
+ self.starts[test.id()] = self.get_time()
+
+ def addSuccess(self, test):
+ test = self._add_prefix(test)
+ tid = test.id()
+ if tid not in self.starts:
+ self._ops.addError(test, "%s succeeded without ever starting!" % tid)
+ delta = self.get_time() - self.starts[tid]
+ self._ops.output_msg("elapsed-time: %s: %f\n" % (tid, delta.total_seconds()))
+
+ def addFailure(self, test, err=''):
+ tid = test.id()
+ delta = self.get_time() - self.starts[tid]
+ self._ops.output_msg("failure: %s failed after %f seconds (%s)\n" %
+ (tid, delta.total_seconds(), err))
+
+ def addError(self, test, err=''):
+ tid = test.id()
+ delta = self.get_time() - self.starts[tid]
+ self._ops.output_msg("error: %s failed after %f seconds (%s)\n" %
+ (tid, delta.total_seconds(), err))
+
+ def __init__(self, out, prefix='', suffix=''):
+ self._ops = out
+ self.prefix = prefix or ''
+ self.suffix = suffix or ''
+ self.starts = {}
+ self.seen_output = False
+ self.suite_has_time = False
+
+
+class PlainFormatter(TestsuiteEnabledTestResult):
+
+ def __init__(self, verbose, immediate, statistics,
+ totaltests=None):
+ super(PlainFormatter, self).__init__()
+ self.verbose = verbose
+ self.immediate = immediate
+ self.statistics = statistics
+ self.start_time = None
+ self.test_output = {}
+ self.suitesfailed = []
+ self.suites_ok = 0
+ self.skips = {}
+ self.index = 0
+ self.name = None
+ self._progress_level = 0
+ self.totalsuites = totaltests
+ self.last_time = None
+
+ @staticmethod
+ def _format_time(delta):
+ minutes, seconds = divmod(delta.seconds, 60)
+ hours, minutes = divmod(minutes, 60)
+ ret = ""
+ if hours:
+ ret += "%dh" % hours
+ if minutes:
+ ret += "%dm" % minutes
+ ret += "%ds" % seconds
+ return ret
+
+ def progress(self, offset, whence):
+ if whence == subunit.PROGRESS_POP:
+ self._progress_level -= 1
+ elif whence == subunit.PROGRESS_PUSH:
+ self._progress_level += 1
+ elif whence == subunit.PROGRESS_SET:
+ if self._progress_level == 0:
+ self.totalsuites = offset
+ elif whence == subunit.PROGRESS_CUR:
+ raise NotImplementedError
+
+ def time(self, dt):
+ if self.start_time is None:
+ self.start_time = dt
+ self.last_time = dt
+
+ def start_testsuite(self, name):
+ self.index += 1
+ self.name = name
+
+ if not self.verbose:
+ self.test_output[name] = ""
+
+ total_tests = (self.statistics['TESTS_EXPECTED_OK'] +
+ self.statistics['TESTS_EXPECTED_FAIL'] +
+ self.statistics['TESTS_ERROR'] +
+ self.statistics['TESTS_UNEXPECTED_FAIL'] +
+ self.statistics['TESTS_UNEXPECTED_OK'])
+
+ out = "[%d(%d)" % (self.index, total_tests)
+ if self.totalsuites is not None:
+ out += "/%d" % self.totalsuites
+ if self.start_time is not None:
+ out += " at " + self._format_time(self.last_time - self.start_time)
+ if self.suitesfailed:
+ out += ", %d errors" % (len(self.suitesfailed),)
+ out += "] %s" % name
+ if self.immediate:
+ sys.stdout.write(out + "\n")
+ else:
+ sys.stdout.write(out + ": ")
+
+ def output_msg(self, output):
+ if self.verbose:
+ sys.stdout.write(output)
+ elif self.name is not None:
+ self.test_output[self.name] += output
+ else:
+ sys.stdout.write(output)
+
+ def control_msg(self, output):
+ pass
+
+ def end_testsuite(self, name, result, reason):
+ out = ""
+ unexpected = False
+
+ if name not in self.test_output:
+ print("no output for name[%s]" % name)
+
+ if result in ("success", "xfail"):
+ self.suites_ok += 1
+ else:
+ self.output_msg("ERROR: Testsuite[%s]\n" % name)
+ if reason is not None:
+ self.output_msg("REASON: %s\n" % (reason,))
+ self.suitesfailed.append(name)
+ if self.immediate and not self.verbose and name in self.test_output:
+ out += self.test_output[name]
+ unexpected = True
+
+ if not self.immediate:
+ if not unexpected:
+ out += " ok\n"
+ else:
+ out += " " + result.upper() + "\n"
+
+ sys.stdout.write(out)
+
+ def startTest(self, test):
+ pass
+
+ def addSuccess(self, test):
+ self.end_test(test.id(), "success", False)
+
+ def addError(self, test, err=None):
+ self.end_test(test.id(), "error", True, err)
+
+ def addFailure(self, test, err=None):
+ self.end_test(test.id(), "failure", True, err)
+
+ def addSkip(self, test, reason=None):
+ self.end_test(test.id(), "skip", False, reason)
+
+ def addExpectedFailure(self, test, err=None):
+ self.end_test(test.id(), "xfail", False, err)
+
+ def addUnexpectedSuccess(self, test):
+ self.end_test(test.id(), "uxsuccess", True)
+
+ def end_test(self, testname, result, unexpected, err=None):
+ if not unexpected:
+ self.test_output[self.name] = ""
+ if not self.immediate:
+ sys.stdout.write({
+ 'failure': 'f',
+ 'xfail': 'X',
+ 'skip': 's',
+ 'success': '.'}.get(result, "?(%s)" % result))
+ return
+
+ if self.name not in self.test_output:
+ self.test_output[self.name] = ""
+
+ self.test_output[self.name] += "UNEXPECTED(%s): %s\n" % (result, testname)
+ if err is not None:
+ self.test_output[self.name] += "REASON: %s\n" % str(err[1]).strip()
+
+ if self.immediate and not self.verbose:
+ sys.stdout.write(self.test_output[self.name])
+ self.test_output[self.name] = ""
+
+ if not self.immediate:
+ sys.stdout.write({
+ 'error': 'E',
+ 'failure': 'F',
+ 'uxsuccess': 'U',
+ 'success': 'S'}.get(result, "?"))
+
+ def write_summary(self, path):
+ f = open(path, 'w+')
+
+ if self.suitesfailed:
+ f.write("= Failed tests =\n")
+
+ for suite in self.suitesfailed:
+ f.write("== %s ==\n" % suite)
+ if suite in self.test_output:
+ f.write(self.test_output[suite] + "\n\n")
+
+ f.write("\n")
+
+ if not self.immediate and not self.verbose:
+ for suite in self.suitesfailed:
+ print("=" * 78)
+ print("FAIL: %s" % suite)
+ if suite in self.test_output:
+ print(self.test_output[suite])
+ print("")
+
+ f.write("= Skipped tests =\n")
+ for reason in self.skips.keys():
+ f.write(reason + "\n")
+ for name in self.skips[reason]:
+ f.write("\t%s\n" % name)
+ f.write("\n")
+ f.close()
+
+ if (not self.suitesfailed and
+ not self.statistics['TESTS_UNEXPECTED_FAIL'] and
+ not self.statistics['TESTS_UNEXPECTED_OK'] and
+ not self.statistics['TESTS_ERROR']):
+ ok = (self.statistics['TESTS_EXPECTED_OK'] +
+ self.statistics['TESTS_EXPECTED_FAIL'])
+ print("\nALL OK (%d tests in %d testsuites)" % (ok, self.suites_ok))
+ else:
+ print("\nFAILED (%d failures, %d errors and %d unexpected successes in %d testsuites)" % (
+ self.statistics['TESTS_UNEXPECTED_FAIL'],
+ self.statistics['TESTS_ERROR'],
+ self.statistics['TESTS_UNEXPECTED_OK'],
+ len(self.suitesfailed)))
+
+ def skip_testsuite(self, name, reason="UNKNOWN"):
+ self.skips.setdefault(reason, []).append(name)
+ if self.totalsuites:
+ self.totalsuites -= 1