From 78a7aeeb1a93b0a3b850355bc7f71dab00fa755a Mon Sep 17 00:00:00 2001 From: Raymond Hettinger Date: Tue, 5 Nov 2002 06:06:02 +0000 Subject: SF 633560: tokenize.__all__ needs "generate_tokens" --- Lib/tokenize.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/Lib/tokenize.py b/Lib/tokenize.py index 76ea7a2..37ce049 100644 --- a/Lib/tokenize.py +++ b/Lib/tokenize.py @@ -30,7 +30,8 @@ import string, re from token import * import token -__all__ = [x for x in dir(token) if x[0] != '_'] + ["COMMENT", "tokenize", "NL"] +__all__ = [x for x in dir(token) if x[0] != '_'] + ["COMMENT", "tokenize", + "generate_tokens", "NL"] del x del token -- cgit v0.12