summaryrefslogtreecommitdiffstats
path: root/Lib
diff options
context:
space:
mode:
authorBenjamin Peterson <benjamin@python.org>2010-08-30 14:44:53 (GMT)
committerBenjamin Peterson <benjamin@python.org>2010-08-30 14:44:53 (GMT)
commit66428b2e5d456e0af609de8cdbcd9bae2a76d02d (patch)
tree51de6d4aaae68845249503d8b4835df0976f6fcf /Lib
parentf8a08d9d3618a8a3a53bdf1370daf07032313795 (diff)
downloadcpython-66428b2e5d456e0af609de8cdbcd9bae2a76d02d.zip
cpython-66428b2e5d456e0af609de8cdbcd9bae2a76d02d.tar.gz
cpython-66428b2e5d456e0af609de8cdbcd9bae2a76d02d.tar.bz2
Merged revisions 84364 via svnmerge from
svn+ssh://pythondev@svn.python.org/python/branches/py3k ........ r84364 | benjamin.peterson | 2010-08-30 09:41:20 -0500 (Mon, 30 Aug 2010) | 1 line handle names starting with non-ascii characters correctly #9712 ........
Diffstat (limited to 'Lib')
-rw-r--r--Lib/test/test_tokenize.py13
-rw-r--r--Lib/tokenize.py15
2 files changed, 23 insertions, 5 deletions
diff --git a/Lib/test/test_tokenize.py b/Lib/test/test_tokenize.py
index 7b91ab2..eeefce1 100644
--- a/Lib/test/test_tokenize.py
+++ b/Lib/test/test_tokenize.py
@@ -533,6 +533,7 @@ pass the '-ucompiler' option to process the full directory.
True
Evil tabs
+
>>> dump_tokens("def f():\\n\\tif x\\n \\tpass")
ENCODING 'utf-8' (0, 0) (0, 0)
NAME 'def' (1, 0) (1, 3)
@@ -549,6 +550,18 @@ Evil tabs
NAME 'pass' (3, 9) (3, 13)
DEDENT '' (4, 0) (4, 0)
DEDENT '' (4, 0) (4, 0)
+
+Non-ascii identifiers
+
+ >>> dump_tokens("Örter = 'places'\\ngrün = 'green'")
+ ENCODING 'utf-8' (0, 0) (0, 0)
+ NAME 'Örter' (1, 0) (1, 5)
+ OP '=' (1, 6) (1, 7)
+ STRING "'places'" (1, 8) (1, 16)
+ NEWLINE '\\n' (1, 16) (1, 17)
+ NAME 'grün' (2, 0) (2, 4)
+ OP '=' (2, 5) (2, 6)
+ STRING "'green'" (2, 7) (2, 14)
"""
from test import support
diff --git a/Lib/tokenize.py b/Lib/tokenize.py
index 9d2a6bb..e711a21 100644
--- a/Lib/tokenize.py
+++ b/Lib/tokenize.py
@@ -92,7 +92,7 @@ def maybe(*choices): return group(*choices) + '?'
Whitespace = r'[ \f\t]*'
Comment = r'#[^\r\n]*'
Ignore = Whitespace + any(r'\\\r?\n' + Whitespace) + maybe(Comment)
-Name = r'[a-zA-Z_]\w*'
+Name = r'\w+'
Hexnumber = r'0[xX][0-9a-fA-F]+'
Binnumber = r'0[bB][01]+'
@@ -142,9 +142,12 @@ ContStr = group(r"[bB]?[rR]?'[^\n'\\]*(?:\\.[^\n'\\]*)*" +
PseudoExtras = group(r'\\\r?\n', Comment, Triple)
PseudoToken = Whitespace + group(PseudoExtras, Number, Funny, ContStr, Name)
+def _compile(expr):
+ return re.compile(expr, re.UNICODE)
+
tokenprog, pseudoprog, single3prog, double3prog = map(
- re.compile, (Token, PseudoToken, Single3, Double3))
-endprogs = {"'": re.compile(Single), '"': re.compile(Double),
+ _compile, (Token, PseudoToken, Single3, Double3))
+endprogs = {"'": _compile(Single), '"': _compile(Double),
"'''": single3prog, '"""': double3prog,
"r'''": single3prog, 'r"""': double3prog,
"b'''": single3prog, 'b"""': double3prog,
@@ -171,6 +174,8 @@ for t in ("'", '"',
"bR'", 'bR"', "BR'", 'BR"' ):
single_quoted[t] = t
+del _compile
+
tabsize = 8
class TokenError(Exception): pass
@@ -392,7 +397,7 @@ def tokenize(readline):
def _tokenize(readline, encoding):
lnum = parenlev = continued = 0
- namechars, numchars = string.ascii_letters + '_', '0123456789'
+ numchars = '0123456789'
contstr, needcont = '', 0
contline = None
indents = [0]
@@ -516,7 +521,7 @@ def _tokenize(readline, encoding):
break
else: # ordinary string
yield TokenInfo(STRING, token, spos, epos, line)
- elif initial in namechars: # ordinary name
+ elif initial.isidentifier(): # ordinary name
yield TokenInfo(NAME, token, spos, epos, line)
elif initial == '\\': # continued stmt
continued = 1