Add dedicated exception for immediate failure in filter-subunit, don't raise it on...
[sfrench/samba-autobuild/.git] / selftest / subunithelper.py
index 3cd0f013d03f6a2cea30881d6810c5a7f571c5b5..163f2ba7dbcb81765487ffa5cd48ebb815ea668f 100644 (file)
 __all__ = ['parse_results']
 
 import re
-import time
+import sys
+import subunit
+import subunit.iso8601
+import testtools
 
 VALID_RESULTS = ['success', 'successful', 'failure', 'fail', 'skip', 'knownfail', 'error', 'xfail', 'skip-testsuite', 'testsuite-failure', 'testsuite-xfail', 'testsuite-success', 'testsuite-error']
 
+class TestsuiteEnabledTestResult(testtools.testresult.TestResult):
+
+    def start_testsuite(self, name):
+        raise NotImplementedError(self.start_testsuite)
+
+
 def parse_results(msg_ops, statistics, fh):
     expected_fail = 0
-    open_tests = []
+    open_tests = {}
 
     while fh:
         l = fh.readline()
-        if l.startswith("test: "):
+        if l == "":
+            break
+        parts = l.split(None, 1)
+        if not len(parts) == 2 or not l.startswith(parts[0]):
+            msg_ops.output_msg(l)
+            continue
+        command = parts[0].rstrip(":")
+        arg = parts[1]
+        if command in ("test", "testing"):
+            msg_ops.control_msg(l)
+            name = arg.rstrip()
+            test = subunit.RemotedTestCase(name)
+            if name in open_tests:
+                msg_ops.addError(open_tests.pop(name), subunit.RemoteError(u"Test already running"))
+            msg_ops.startTest(test)
+            open_tests[name] = test
+        elif command == "time":
             msg_ops.control_msg(l)
-            name = l.split(":", 1)[1].strip()
-            msg_ops.start_test(name)
-            open_tests.append(name)
-        elif l.startswith("time: "):
-            (year, month, day, hour, minute, second) = re.match(
-                "^time: (\d+)-(\d+)-(\d+) (\d+):(\d+):(\d+)\n/", l)
-            msg_ops.report_time(time.mktime(second, minute, hour, day, month-1, year-1900))
-        elif re.match("^(" + "|".join(VALID_RESULTS) + "): (.*?)( \[)?([ \t]*)( multipart)?\n", l):
+            try:
+                dt = subunit.iso8601.parse_date(arg.rstrip("\n"))
+            except TypeError, e:
+                print "Unable to parse time line: %s" % arg.rstrip("\n")
+            else:
+                msg_ops.time(dt)
+        elif command in VALID_RESULTS:
             msg_ops.control_msg(l)
-            (result, testname, hasreason) = re.match("^(" + "|".join(VALID_RESULTS) + "): (.*?)( \[)?([ \t]*)( multipart)?\n", l)
+            result = command
+            grp = re.match("(.*?)( \[)?([ \t]*)( multipart)?\n", arg)
+            (testname, hasreason) = (grp.group(1), grp.group(2))
             if hasreason:
                 reason = ""
                 # reason may be specified in next lines
                 terminated = False
                 while fh:
                     l = fh.readline()
+                    if l == "":
+                        break
                     msg_ops.control_msg(l)
                     if l == "]\n":
                         terminated = True
                         break
                     else:
                         reason += l
-                
+
+                remote_error = subunit.RemoteError(reason.decode("utf-8"))
+
                 if not terminated:
                     statistics['TESTS_ERROR']+=1
-                    msg_ops.end_test(testname, "error", 1, 
-                                       "reason (%s) interrupted" % result)
+                    msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"reason (%s) interrupted" % result))
                     return 1
+            else:
+                reason = None
+                remote_error = subunit.RemoteError(u"No reason specified")
             if result in ("success", "successful"):
-                open_tests.pop() #FIXME: Check that popped value == $testname 
-                statistics['TESTS_EXPECTED_OK']+=1
-                msg_ops.end_test(testname, "success", 0, reason)
+                try:
+                    test = open_tests.pop(testname)
+                except KeyError:
+                    statistics['TESTS_ERROR']+=1
+                    msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"Test was never started"))
+                else:
+                    statistics['TESTS_EXPECTED_OK']+=1
+                    msg_ops.addSuccess(test)
             elif result in ("xfail", "knownfail"):
-                open_tests.pop() #FIXME: Check that popped value == $testname
-                statistics['TESTS_EXPECTED_FAIL']+=1
-                msg_ops.end_test(testname, "xfail", 0, reason)
-                expected_fail+=1
+                try:
+                    test = open_tests.pop(testname)
+                except KeyError:
+                    statistics['TESTS_ERROR']+=1
+                    msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"Test was never started"))
+                else:
+                    statistics['TESTS_EXPECTED_FAIL']+=1
+                    msg_ops.addExpectedFailure(test, remote_error)
+                    expected_fail+=1
             elif result in ("failure", "fail"):
-                open_tests.pop() #FIXME: Check that popped value == $testname
-                statistics['TESTS_UNEXPECTED_FAIL']+=1
-                msg_ops.end_test(testname, "failure", 1, reason)
+                try:
+                    test = open_tests.pop(testname)
+                except KeyError:
+                    statistics['TESTS_ERROR']+=1
+                    msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"Test was never started"))
+                else:
+                    statistics['TESTS_UNEXPECTED_FAIL']+=1
+                    msg_ops.addFailure(test, remote_error)
             elif result == "skip":
                 statistics['TESTS_SKIP']+=1
                 # Allow tests to be skipped without prior announcement of test
-                last = open_tests.pop()
-                if last is not None and last != testname:
-                    open_tests.append(testname)
-                msg_ops.end_test(testname, "skip", 0, reason)
+                try:
+                    test = open_tests.pop(testname)
+                except KeyError:
+                    test = subunit.RemotedTestCase(testname)
+                msg_ops.addSkip(test, reason)
             elif result == "error":
                 statistics['TESTS_ERROR']+=1
-                open_tests.pop() #FIXME: Check that popped value == $testname
-                msg_ops.end_test(testname, "error", 1, reason)
+                try:
+                    test = open_tests.pop(testname)
+                except KeyError:
+                    test = subunit.RemotedTestCase(testname)
+                msg_ops.addError(test, remote_error)
             elif result == "skip-testsuite":
                 msg_ops.skip_testsuite(testname)
             elif result == "testsuite-success":
@@ -92,79 +143,212 @@ def parse_results(msg_ops, statistics, fh):
                 msg_ops.end_testsuite(testname, "xfail", reason)
             elif result == "testsuite-error":
                 msg_ops.end_testsuite(testname, "error", reason)
-        elif l.startswith("testsuite: "):
-            msg_ops.start_testsuite(l.split(":", 1)[1].strip())
-        elif l.startswith("testsuite-count: "):
-            msg_ops.testsuite_count(int(l.split(":", 1)[1].strip()))
+            else:
+                raise AssertionError("Recognized but unhandled result %r" %
+                    result)
+        elif command == "testsuite":
+            msg_ops.start_testsuite(arg.strip())
+        elif command == "progress":
+            arg = arg.strip()
+            if arg == "pop":
+                msg_ops.progress(None, subunit.PROGRESS_POP)
+            elif arg == "push":
+                msg_ops.progress(None, subunit.PROGRESS_PUSH)
+            elif arg[0] in '+-':
+                msg_ops.progress(int(arg), subunit.PROGRESS_CUR)
+            else:
+                msg_ops.progress(int(arg), subunit.PROGRESS_SET)
         else:
             msg_ops.output_msg(l)
 
     while open_tests:
-        msg_ops.end_test(open_tests.pop(), "error", 1,
-                   "was started but never finished!")
+        test = subunit.RemotedTestCase(open_tests.popitem()[1])
+        msg_ops.addError(test, subunit.RemoteError(u"was started but never finished!"))
         statistics['TESTS_ERROR']+=1
 
-    # if the Filter module is in use, it will have the right counts
-    if 'total_error' in msg_ops:
-        statistics['TESTS_ERROR'] = msg_ops['total_error']
-        statistics['TESTS_UNEXPECTED_FAIL'] = msg_ops['total_fail']
-        statistics['TESTS_EXPECTED_FAIL'] = msg_ops['total_xfail']
-
     if statistics['TESTS_ERROR'] > 0:
         return 1
     if statistics['TESTS_UNEXPECTED_FAIL'] > 0:
-        return 1 
+        return 1
     return 0
 
 
-def start_test(testname):
-    print "test: %s" % testname
+class SubunitOps(subunit.TestProtocolClient,TestsuiteEnabledTestResult):
+
+    # The following are Samba extensions:
+    def start_testsuite(self, name):
+        self._stream.write("testsuite: %s\n" % name)
+
+    def skip_testsuite(self, name, reason=None):
+        if reason:
+            self._stream.write("skip-testsuite: %s [\n%s\n]\n" % (name, reason))
+        else:
+            self._stream.write("skip-testsuite: %s\n" % name)
+
+    def end_testsuite(self, name, result, reason=None):
+        if reason:
+            self._stream.write("testsuite-%s: %s [\n%s\n]\n" % (result, name, reason))
+        else:
+            self._stream.write("testsuite-%s: %s\n" % (result, name))
+
+    def output_msg(self, msg):
+        self._stream.write(msg)
+
+
+def read_test_regexes(name):
+    ret = {}
+    f = open(name, 'r')
+    try:
+        for l in f:
+            l = l.strip()
+            if l == "" or l[0] == "#":
+                continue
+            if "#" in l:
+                (regex, reason) = l.split("#", 1)
+                ret[regex.strip()] = reason.strip()
+            else:
+                ret[l] = None
+    finally:
+        f.close()
+    return ret
+
+
+def find_in_list(regexes, fullname):
+    for regex, reason in regexes.iteritems():
+        if re.match(regex, fullname):
+            if reason is None:
+                return ""
+            return reason
+    return None
+
 
-def end_test(name, result, reason=None):
-    if reason:
-        print "%s: %s [" % (result, name)
-        print "%s" % reason
-        print "]"
-    else:
-        print "%s: %s" % (result, name)
+class ImmediateFail(Exception):
+    """Raised to abort immediately."""
 
+    def __init__(self):
+        super(ImmediateFail, self).__init__("test failed and fail_immediately set")
 
-def skip_test(name, reason=None):
-    end_test(name, "skip", reason)
 
+class FilterOps(testtools.testresult.TestResult):
+
+    def control_msg(self, msg):
+        pass # We regenerate control messages, so ignore this
+
+    def time(self, time):
+        self._ops.time(time)
+
+    def progress(self, delta, whence):
+        self._ops.progress(delta, whence)
+
+    def output_msg(self, msg):
+        if self.output is None:
+            sys.stdout.write(msg)
+        else:
+            self.output+=msg
+
+    def startTest(self, test):
+        test = self._add_prefix(test)
+        if self.strip_ok_output:
+           self.output = ""
+
+        self._ops.startTest(test)
+
+    def _add_prefix(self, test):
+        if self.prefix is not None:
+            return subunit.RemotedTestCase(self.prefix + test.id())
+        else:
+            return test
+
+    def addError(self, test, details=None):
+        test = self._add_prefix(test)
+        self.error_added+=1
+        self.total_error+=1
+        self._ops.addError(test, details)
+        self.output = None
+        if self.fail_immediately:
+            raise ImmediateFail()
+
+    def addSkip(self, test, details=None):
+        test = self._add_prefix(test)
+        self._ops.addSkip(test, details)
+        self.output = None
+
+    def addExpectedFailure(self, test, details=None):
+        test = self._add_prefix(test)
+        self._ops.addExpectedFailure(test, details)
+        self.output = None
+
+    def addFailure(self, test, details=None):
+        test = self._add_prefix(test)
+        xfail_reason = find_in_list(self.expected_failures, test.id())
+        if xfail_reason is not None:
+            self.xfail_added+=1
+            self.total_xfail+=1
+            if details is not None:
+                details = subunit.RemoteError(details[1].message + xfail_reason.decode("utf-8"))
+            else:
+                details = subunit.RemoteError(xfail_reason.decode("utf-8"))
+            self._ops.addExpectedFailure(test, details)
+        else:
+            self.fail_added+=1
+            self.total_fail+=1
+            self._ops.addFailure(test, details)
+            if self.output:
+                self._ops.output_msg(self.output)
+            if self.fail_immediately:
+                raise ImmediateFail()
+        self.output = None
 
-def fail_test(name, reason=None):
-    end_test(name, "fail", reason)
+    def addSuccess(self, test, details=None):
+        test = self._add_prefix(test)
+        self._ops.addSuccess(test, details)
+        self.output = None
 
+    def skip_testsuite(self, name, reason=None):
+        self._ops.skip_testsuite(name, reason)
 
-def success_test(name, reason=None):
-    end_test(name, "success", reason)
+    def start_testsuite(self, name):
+        self._ops.start_testsuite(name)
+        self.error_added = 0
+        self.fail_added = 0
+        self.xfail_added = 0
 
-def xfail_test(name, reason=None):
-    end_test(name, "xfail", reason)
+    def end_testsuite(self, name, result, reason=None):
+        xfail = False
 
-def report_time(t):
-    (sec, min, hour, mday, mon, year, wday, yday, isdst) = time.localtimet(t)
-    print "time: %04d-%02d-%02d %02d:%02d:%02d" % (year+1900, mon+1, mday, hour, min, sec)
+        if self.xfail_added > 0:
+            xfail = True
+        if self.fail_added > 0 or self.error_added > 0:
+            xfail = False
 
+        if xfail and result in ("fail", "failure"):
+            result = "xfail"
 
-# The following are Samba extensions:
-def start_testsuite(name):
-    print "testsuite: %s" % name
+        if self.fail_added > 0 and result != "failure":
+            result = "failure"
+            if reason is None:
+                reason = "Subunit/Filter Reason"
+            reason += "\n failures[%d]" % self.fail_added
 
-def skip_testsuite(name, reason=None):
-    if reason:
-        print "skip-testsuite: %s [\n%s\n]" % (name, reason)
-    else:
-        print "skip-testsuite: %s" % name
+        if self.error_added > 0 and result != "error":
+            result = "error"
+            if reason is None:
+                reason = "Subunit/Filter Reason"
+            reason += "\n errors[%d]" % self.error_added
 
-def end_testsuite(name, result, reason=None):
-    if reason:
-        print "testsuite-$result: %s [" % name
-        print "%s" % reason
-        print "]"
-    else:
-        print "testsuite-$result: %s" % name
+        self._ops.end_testsuite(name, result, reason)
 
-def testsuite_count(count):
-    print "testsuite-count: %d" % count
+    def __init__(self, out, prefix, expected_failures, strip_ok_output, fail_immediately=False):
+        self._ops = out
+        self.output = None
+        self.prefix = prefix
+        self.expected_failures = expected_failures
+        self.strip_ok_output = strip_ok_output
+        self.xfail_added = 0
+        self.fail_added = 0
+        self.total_xfail = 0
+        self.total_error = 0
+        self.total_fail = 0
+        self.error_added = 0
+        self.fail_immediately = fail_immediately
+