summaryrefslogtreecommitdiffstats
path: root/Doc/lib/libtokenize.tex
diff options
context:
space:
mode:
authorFred Drake <fdrake@acm.org>1999-02-18 21:06:50 (GMT)
committerFred Drake <fdrake@acm.org>1999-02-18 21:06:50 (GMT)
commit6b103f1e128a4ae46bd38ae7b5d98bd49a65f5bd (patch)
treeab0ed6a453a02b1f617eed7771d94dfa8c79a89b /Doc/lib/libtokenize.tex
parent3d199af40d1c7d0068339ab1b03e8243105a4447 (diff)
downloadcpython-6b103f1e128a4ae46bd38ae7b5d98bd49a65f5bd.zip
cpython-6b103f1e128a4ae46bd38ae7b5d98bd49a65f5bd.tar.gz
cpython-6b103f1e128a4ae46bd38ae7b5d98bd49a65f5bd.tar.bz2
Documentation for pyclbr and tokenize modules.
Diffstat (limited to 'Doc/lib/libtokenize.tex')
-rw-r--r--Doc/lib/libtokenize.tex44
1 files changed, 44 insertions, 0 deletions
diff --git a/Doc/lib/libtokenize.tex b/Doc/lib/libtokenize.tex
new file mode 100644
index 0000000..2176173
--- /dev/null
+++ b/Doc/lib/libtokenize.tex
@@ -0,0 +1,44 @@
+\section{\module{tokenize} ---
+ Tokenizer for Python source}
+
+\declaremodule{standard}{tokenize}
+\modulesynopsis{Lexical scanner for Python source code.}
+\moduleauthor{Ka Ping Yee}{}
+\sectionauthor{Fred L. Drake, Jr.}{fdrake@acm.org}
+
+
+The \module{tokenize} module provides a lexical scanner for Python
+source code, implemented in Python. The scanner in this module
+returns comments as tokens as well, making it useful for implementing
+``pretty-printers,'' including colorizers for on-screen displays.
+
+The scanner is exposed via single function:
+
+
+\begin{funcdesc}{tokenize}{readline\optional{, tokeneater}}
+ The \function{tokenize()} function accepts two parameters: one
+ representing the input stream, and one providing an output mechanism
+ for \function{tokenize()}.
+
+ The first parameter, \var{readline}, must be a callable object which
+ provides the same interface as \method{readline()} method of
+ built-in file objects (see section~\ref{bltin-file-objects}). Each
+ call to the function should return one line of input as a string.
+
+ The second parameter, \var{tokeneater}, must also be a callable
+ object. It is called with five parameters: the token type, the
+ token string, a tuple \code{(\var{srow}, \var{scol})} specifying the
+ row and column where the token begins in the source, a tuple
+ \code{(\var{erow}, \var{ecol})} giving the ending position of the
+ token, and the line on which the token was found. The line passed
+ is the \emph{logical} line; continuation lines are included.
+\end{funcdesc}
+
+
+All constants from the \refmodule{token} module are also exported from
+\module{tokenize}, as is one additional token type value that might be
+passed to the \var{tokeneater} function by \function{tokenize()}:
+
+\begin{datadesc}{COMMENT}
+ Token value used to indicate a comment.
+\end{datadesc}