summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--Lib/test/test_baseexception.py (renamed from Lib/test/test_pep352.py)0
-rw-r--r--Lib/test/test_dict_version.py (renamed from Lib/test/test_pep509.py)0
-rw-r--r--Lib/test/test_exception_hierarchy.py (renamed from Lib/test/test_pep3151.py)0
-rw-r--r--Lib/test/test_generator_stop.py (renamed from Lib/test/test_pep479.py)0
-rw-r--r--Lib/test/test_tokenize.py11
-rw-r--r--Lib/test/test_unicode_file_functions.py (renamed from Lib/test/test_pep277.py)0
-rw-r--r--Lib/test/test_unicode_identifiers.py (renamed from Lib/test/test_pep3131.py)0
-rw-r--r--Lib/test/test_utf8source.py (renamed from Lib/test/test_pep3120.py)0
-rw-r--r--Lib/test/test_yield_from.py (renamed from Lib/test/test_pep380.py)0
9 files changed, 6 insertions, 5 deletions
diff --git a/Lib/test/test_pep352.py b/Lib/test/test_baseexception.py
index 27d514f..27d514f 100644
--- a/Lib/test/test_pep352.py
+++ b/Lib/test/test_baseexception.py
diff --git a/Lib/test/test_pep509.py b/Lib/test/test_dict_version.py
index 5671f9f..5671f9f 100644
--- a/Lib/test/test_pep509.py
+++ b/Lib/test/test_dict_version.py
diff --git a/Lib/test/test_pep3151.py b/Lib/test/test_exception_hierarchy.py
index 8649596..8649596 100644
--- a/Lib/test/test_pep3151.py
+++ b/Lib/test/test_exception_hierarchy.py
diff --git a/Lib/test/test_pep479.py b/Lib/test/test_generator_stop.py
index bc235ce..bc235ce 100644
--- a/Lib/test/test_pep479.py
+++ b/Lib/test/test_generator_stop.py
diff --git a/Lib/test/test_tokenize.py b/Lib/test/test_tokenize.py
index 77c0423..4c469a8 100644
--- a/Lib/test/test_tokenize.py
+++ b/Lib/test/test_tokenize.py
@@ -1529,12 +1529,13 @@ class TestRoundtrip(TestCase):
tempdir = os.path.dirname(fn) or os.curdir
testfiles = glob.glob(os.path.join(tempdir, "test*.py"))
- # Tokenize is broken on test_pep3131.py because regular expressions are
- # broken on the obscure unicode identifiers in it. *sigh*
- # With roundtrip extended to test the 5-tuple mode of untokenize,
- # 7 more testfiles fail. Remove them also until the failure is diagnosed.
+ # Tokenize is broken on test_unicode_identifiers.py because regular
+ # expressions are broken on the obscure unicode identifiers in it.
+ # *sigh* With roundtrip extended to test the 5-tuple mode of
+ # untokenize, 7 more testfiles fail. Remove them also until the
+ # failure is diagnosed.
- testfiles.remove(os.path.join(tempdir, "test_pep3131.py"))
+ testfiles.remove(os.path.join(tempdir, "test_unicode_identifiers.py"))
for f in ('buffer', 'builtin', 'fileio', 'inspect', 'os', 'platform', 'sys'):
testfiles.remove(os.path.join(tempdir, "test_%s.py") % f)
diff --git a/Lib/test/test_pep277.py b/Lib/test/test_unicode_file_functions.py
index 98c716b..98c716b 100644
--- a/Lib/test/test_pep277.py
+++ b/Lib/test/test_unicode_file_functions.py
diff --git a/Lib/test/test_pep3131.py b/Lib/test/test_unicode_identifiers.py
index 0679845..0679845 100644
--- a/Lib/test/test_pep3131.py
+++ b/Lib/test/test_unicode_identifiers.py
diff --git a/Lib/test/test_pep3120.py b/Lib/test/test_utf8source.py
index 97dced8..97dced8 100644
--- a/Lib/test/test_pep3120.py
+++ b/Lib/test/test_utf8source.py
diff --git a/Lib/test/test_pep380.py b/Lib/test/test_yield_from.py
index 23ffbed..23ffbed 100644
--- a/Lib/test/test_pep380.py
+++ b/Lib/test/test_yield_from.py