diff options
author | Tim Peters <tim.peters@gmail.com> | 2004-09-04 17:21:02 (GMT) |
---|---|---|
committer | Tim Peters <tim.peters@gmail.com> | 2004-09-04 17:21:02 (GMT) |
commit | 1fbf9c5ec10d38d58837e20a681604440aa7b3da (patch) | |
tree | ee3e64b8066a13df7de071bdae9aef0bb4d01f12 /Lib | |
parent | ba6019691eddff901bbb7eb5d8a2cc66e5ebaaca (diff) | |
download | cpython-1fbf9c5ec10d38d58837e20a681604440aa7b3da.zip cpython-1fbf9c5ec10d38d58837e20a681604440aa7b3da.tar.gz cpython-1fbf9c5ec10d38d58837e20a681604440aa7b3da.tar.bz2 |
Added IGNORE_EXCEPTION_DETAIL comparison option. The need is explained
in the new docs.
DocTestRunner.__run: Separate the determination of the example outcome
from reporting that outcome, to squash brittle code duplication and
excessive nesting.
Diffstat (limited to 'Lib')
-rw-r--r-- | Lib/doctest.py | 79 | ||||
-rw-r--r-- | Lib/test/test_doctest.py | 37 |
2 files changed, 84 insertions, 32 deletions
diff --git a/Lib/doctest.py b/Lib/doctest.py index 0c2787f..d77fe15 100644 --- a/Lib/doctest.py +++ b/Lib/doctest.py @@ -176,6 +176,7 @@ __all__ = [ 'DONT_ACCEPT_BLANKLINE', 'NORMALIZE_WHITESPACE', 'ELLIPSIS', + 'IGNORE_EXCEPTION_DETAIL', 'COMPARISON_FLAGS', 'REPORT_UDIFF', 'REPORT_CDIFF', @@ -261,11 +262,13 @@ DONT_ACCEPT_TRUE_FOR_1 = register_optionflag('DONT_ACCEPT_TRUE_FOR_1') DONT_ACCEPT_BLANKLINE = register_optionflag('DONT_ACCEPT_BLANKLINE') NORMALIZE_WHITESPACE = register_optionflag('NORMALIZE_WHITESPACE') ELLIPSIS = register_optionflag('ELLIPSIS') +IGNORE_EXCEPTION_DETAIL = register_optionflag('IGNORE_EXCEPTION_DETAIL') COMPARISON_FLAGS = (DONT_ACCEPT_TRUE_FOR_1 | DONT_ACCEPT_BLANKLINE | NORMALIZE_WHITESPACE | - ELLIPSIS) + ELLIPSIS | + IGNORE_EXCEPTION_DETAIL) REPORT_UDIFF = register_optionflag('REPORT_UDIFF') REPORT_CDIFF = register_optionflag('REPORT_CDIFF') @@ -1293,6 +1296,10 @@ class DocTestRunner: # to modify them). original_optionflags = self.optionflags + SUCCESS, FAILURE, BOOM = range(3) # `outcome` state + + check = self._checker.check_output + # Process each example. for examplenum, example in enumerate(test.examples): @@ -1337,45 +1344,53 @@ class DocTestRunner: got = self._fakeout.getvalue() # the actual output self._fakeout.truncate(0) + outcome = FAILURE # guilty until proved innocent or insane # If the example executed without raising any exceptions, - # then verify its output and report its outcome. + # verify its output. if exception is None: - if self._checker.check_output(example.want, got, - self.optionflags): - if not quiet: - self.report_success(out, test, example, got) - else: - if not quiet: - self.report_failure(out, test, example, got) - failures += 1 - - # If the example raised an exception, then check if it was - # expected. + if check(example.want, got, self.optionflags): + outcome = SUCCESS + + # The example raised an exception: check if it was expected. else: exc_info = sys.exc_info() exc_msg = traceback.format_exception_only(*exc_info[:2])[-1] + if not quiet: + got += _exception_traceback(exc_info) - # If `example.exc_msg` is None, then we weren't - # expecting an exception. + # If `example.exc_msg` is None, then we weren't expecting + # an exception. if example.exc_msg is None: - if not quiet: - self.report_unexpected_exception(out, test, example, - exc_info) - failures += 1 - # If `example.exc_msg` matches the actual exception - # message (`exc_msg`), then the example succeeds. - elif (self._checker.check_output(example.exc_msg, exc_msg, - self.optionflags)): - if not quiet: - got += _exception_traceback(exc_info) - self.report_success(out, test, example, got) - # Otherwise, the example fails. - else: - if not quiet: - got += _exception_traceback(exc_info) - self.report_failure(out, test, example, got) - failures += 1 + outcome = BOOM + + # We expected an exception: see whether it matches. + elif check(example.exc_msg, exc_msg, self.optionflags): + outcome = SUCCESS + + # Another chance if they didn't care about the detail. + elif self.optionflags & IGNORE_EXCEPTION_DETAIL: + m1 = re.match(r'[^:]*:', example.exc_msg) + m2 = re.match(r'[^:]*:', exc_msg) + if m1 and m2 and check(m1.group(0), m2.group(0), + self.optionflags): + outcome = SUCCESS + + # Report the outcome. + if outcome is SUCCESS: + if not quiet: + self.report_success(out, test, example, got) + elif outcome is FAILURE: + if not quiet: + self.report_failure(out, test, example, got) + failures += 1 + elif outcome is BOOM: + if not quiet: + self.report_unexpected_exception(out, test, example, + exc_info) + failures += 1 + else: + assert False, ("unknown outcome", outcome) # Restore the option flags (in case they were modified) self.optionflags = original_optionflags diff --git a/Lib/test/test_doctest.py b/Lib/test/test_doctest.py index 2b39e33..7f51ab5 100644 --- a/Lib/test/test_doctest.py +++ b/Lib/test/test_doctest.py @@ -837,6 +837,43 @@ message is raised, then it is reported as a failure: ValueError: message (1, 1) +However, IGNORE_EXCEPTION_DETAIL can be used to allow a mismatch in the +detail: + + >>> def f(x): + ... r''' + ... >>> raise ValueError, 'message' #doctest: +IGNORE_EXCEPTION_DETAIL + ... Traceback (most recent call last): + ... ValueError: wrong message + ... ''' + >>> test = doctest.DocTestFinder().find(f)[0] + >>> doctest.DocTestRunner(verbose=False).run(test) + (0, 1) + +But IGNORE_EXCEPTION_DETAIL does not allow a mismatch in the exception type: + + >>> def f(x): + ... r''' + ... >>> raise ValueError, 'message' #doctest: +IGNORE_EXCEPTION_DETAIL + ... Traceback (most recent call last): + ... TypeError: wrong type + ... ''' + >>> test = doctest.DocTestFinder().find(f)[0] + >>> doctest.DocTestRunner(verbose=False).run(test) + ... # doctest: +ELLIPSIS + ********************************************************************** + Line 2, in f + Failed example: + raise ValueError, 'message' #doctest: +IGNORE_EXCEPTION_DETAIL + Expected: + Traceback (most recent call last): + TypeError: wrong type + Got: + Traceback (most recent call last): + ... + ValueError: message + (1, 1) + If an exception is raised but not expected, then it is reported as an unexpected exception: |