summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorFred Drake <fdrake@acm.org>1999-01-08 15:40:16 (GMT)
committerFred Drake <fdrake@acm.org>1999-01-08 15:40:16 (GMT)
commitdd21c3216925f855ff182799b6fbf1137a414c7a (patch)
treeda381c284729a8801a3b33156a6a3491a09f4180
parentf2b5374ba1a2b3f3ad941822b35342e275cdc818 (diff)
downloadcpython-dd21c3216925f855ff182799b6fbf1137a414c7a.zip
cpython-dd21c3216925f855ff182799b6fbf1137a414c7a.tar.gz
cpython-dd21c3216925f855ff182799b6fbf1137a414c7a.tar.bz2
It's finally gone!
-rw-r--r--Doc/tools/partparse.py2406
1 files changed, 0 insertions, 2406 deletions
diff --git a/Doc/tools/partparse.py b/Doc/tools/partparse.py
deleted file mode 100644
index 57ba6ef..0000000
--- a/Doc/tools/partparse.py
+++ /dev/null
@@ -1,2406 +0,0 @@
-#
-# partparse.py: parse a by-Guido-written-and-by-Jan-Hein-edited LaTeX file,
-# and generate texinfo source.
-#
-# This is *not* a good example of good programming practices. In fact, this
-# file could use a complete rewrite, in order to become faster, more
-# easily extensible and maintainable.
-#
-# However, I added some comments on a few places for the pityful person who
-# would ever need to take a look into this file.
-#
-# Have I been clear enough??
-#
-# -jh
-#
-# Yup. I made some performance improvements and hope this lasts a while;
-# I don't want to be the schmuck who ends up re-writting it!
-#
-# -fld
-#
-# (sometime later...)
-#
-# Ok, I've re-worked substantial chunks of this. It's only getting worse.
-# It just might be gone before the next source release. (Yeah!)
-#
-# -fld
-
-import sys, string, regex, getopt, os
-
-from types import IntType, ListType, StringType, TupleType
-
-release_version = string.split(sys.version)[0]
-
-# Different parse modes for phase 1
-MODE_REGULAR = 0
-MODE_VERBATIM = 1
-MODE_CS_SCAN = 2
-MODE_COMMENT = 3
-MODE_MATH = 4
-MODE_DMATH = 5
-MODE_GOBBLEWHITE = 6
-
-the_modes = (MODE_REGULAR, MODE_VERBATIM, MODE_CS_SCAN, MODE_COMMENT,
- MODE_MATH, MODE_DMATH, MODE_GOBBLEWHITE)
-
-# Show the neighbourhood of the scanned buffer
-def epsilon(buf, where):
- wmt, wpt = where - 10, where + 10
- if wmt < 0:
- wmt = 0
- if wpt > len(buf):
- wpt = len(buf)
- return ' Context ' + `buf[wmt:where]` + '.' + `buf[where:wpt]` + '.'
-
-# Should return the line number. never worked
-def lin():
- global lineno
- return ' Line ' + `lineno` + '.'
-
-# Displays the recursion level.
-def lv(lvl):
- return ' Level ' + `lvl` + '.'
-
-# Combine the three previous functions. Used often.
-def lle(lvl, buf, where):
- return lv(lvl) + lin() + epsilon(buf, where)
-
-
-# This class is only needed for _symbolic_ representation of the parse mode.
-class Mode:
- def __init__(self, arg):
- if arg not in the_modes:
- raise ValueError, 'mode not in the_modes'
- self.mode = arg
-
- def __cmp__(self, other):
- if type(self) != type(other):
- other = mode[other]
- return cmp(self.mode, other.mode)
-
- def __repr__(self):
- if self.mode == MODE_REGULAR:
- return 'MODE_REGULAR'
- elif self.mode == MODE_VERBATIM:
- return 'MODE_VERBATIM'
- elif self.mode == MODE_CS_SCAN:
- return 'MODE_CS_SCAN'
- elif self.mode == MODE_COMMENT:
- return 'MODE_COMMENT'
- elif self.mode == MODE_MATH:
- return 'MODE_MATH'
- elif self.mode == MODE_DMATH:
- return 'MODE_DMATH'
- elif self.mode == MODE_GOBBLEWHITE:
- return 'MODE_GOBBLEWHITE'
- else:
- raise ValueError, 'mode not in the_modes'
-
-# just a wrapper around a class initialisation
-mode = {}
-for t in the_modes:
- mode[t] = Mode(t)
-
-
-# After phase 1, the text consists of chunks, with a certain type
-# this type will be assigned to the chtype member of the chunk
-# the where-field contains the file position where this is found
-# and the data field contains (1): a tuple describing start- end end
-# positions of the substring (can be used as slice for the buf-variable),
-# (2) just a string, mostly generated by the changeit routine,
-# or (3) a list, describing a (recursive) subgroup of chunks
-PLAIN = 0 # ASSUME PLAINTEXT, data = the text
-GROUP = 1 # GROUP ({}), data = [chunk, chunk,..]
-CSNAME = 2 # CONTROL SEQ TOKEN, data = the command
-COMMENT = 3 # data is the actual comment
-DMATH = 4 # DISPLAYMATH, data = [chunk, chunk,..]
-MATH = 5 # MATH, see DISPLAYMATH
-OTHER = 6 # CHAR WITH CATCODE OTHER, data = char
-ACTIVE = 7 # ACTIVE CHAR
-GOBBLEDWHITE = 8 # Gobbled LWSP, after CSNAME
-ENDLINE = 9 # END-OF-LINE, data = '\n'
-DENDLINE = 10 # DOUBLE EOL, data='\n', indicates \par
-ENV = 11 # LaTeX-environment
- # data =(envname,[ch,ch,ch,.])
-CSLINE = 12 # for texi: next chunk will be one group
- # of args. Will be set all on 1 line
-IGNORE = 13 # IGNORE this data
-ENDENV = 14 # TEMP END OF GROUP INDICATOR
-IF = 15 # IF-directive
- # data = (flag,negate,[ch, ch, ch,...])
-
-the_types = (PLAIN, GROUP, CSNAME, COMMENT, DMATH, MATH, OTHER, ACTIVE,
- GOBBLEDWHITE, ENDLINE, DENDLINE, ENV, CSLINE, IGNORE, ENDENV, IF)
-
-# class, just to display symbolic name
-class ChunkType:
- def __init__(self, chunk_type):
- if chunk_type not in the_types:
- raise ValueError, 'chunk_type not in the_types'
- self.chunk_type = chunk_type
-
- def __cmp__(self, other):
- if type(self) != type(other):
- other = chunk_type[other]
- return cmp(self.chunk_type, other.chunk_type)
-
- def __repr__(self):
- if self.chunk_type == PLAIN:
- return 'PLAIN'
- elif self.chunk_type == GROUP:
- return 'GROUP'
- elif self.chunk_type == CSNAME:
- return 'CSNAME'
- elif self.chunk_type == COMMENT:
- return 'COMMENT'
- elif self.chunk_type == DMATH:
- return 'DMATH'
- elif self.chunk_type == MATH:
- return 'MATH'
- elif self.chunk_type == OTHER:
- return 'OTHER'
- elif self.chunk_type == ACTIVE:
- return 'ACTIVE'
- elif self.chunk_type == GOBBLEDWHITE:
- return 'GOBBLEDWHITE'
- elif self.chunk_type == DENDLINE:
- return 'DENDLINE'
- elif self.chunk_type == ENDLINE:
- return 'ENDLINE'
- elif self.chunk_type == ENV:
- return 'ENV'
- elif self.chunk_type == CSLINE:
- return 'CSLINE'
- elif self.chunk_type == IGNORE:
- return 'IGNORE'
- elif self.chunk_type == ENDENV:
- return 'ENDENV'
- elif self.chunk_type == IF:
- return 'IF'
- else:
- raise ValueError, 'chunk_type not in the_types'
-
-# ...and the wrapper
-chunk_type = {}
-for t in the_types:
- chunk_type[t] = ChunkType(t)
-
-# store a type object of the ChunkType-class-instance...
-chunk_type_type = type(chunk_type[PLAIN])
-
-# this class contains a part of the parsed buffer
-class Chunk:
- def __init__(self, chtype, where, data):
- if type(chtype) != chunk_type_type:
- chtype = chunk_type[chtype]
- self.chtype = chtype
- self.where = where
- self.data = data
-
- __datatypes = [chunk_type[CSNAME], chunk_type[PLAIN], chunk_type[CSLINE]]
-
- def __repr__(self):
- if self.chtype in self.__datatypes:
- data = s(self.buf, self.data)
- else:
- data = self.data
- return 'chunk' + `self.chtype, self.where, data`
-
-# and the wrapper
-chunk = Chunk
-
-
-error = 'partparse.error'
-
-#
-# TeX's catcodes...
-#
-CC_ESCAPE = 0
-CC_LBRACE = 1
-CC_RBRACE = 2
-CC_MATHSHIFT = 3
-CC_ALIGNMENT = 4
-CC_ENDLINE = 5
-CC_PARAMETER = 6
-CC_SUPERSCRIPT = 7
-CC_SUBSCRIPT = 8
-CC_IGNORE = 9
-CC_WHITE = 10
-CC_LETTER = 11
-CC_OTHER = 12
-CC_ACTIVE = 13
-CC_COMMENT = 14
-CC_INVALID = 15
-
-# and the names
-cc_names = [
- 'CC_ESCAPE',
- 'CC_LBRACE',
- 'CC_RBRACE',
- 'CC_MATHSHIFT',
- 'CC_ALIGNMENT',
- 'CC_ENDLINE',
- 'CC_PARAMETER',
- 'CC_SUPERSCRIPT',
- 'CC_SUBSCRIPT',
- 'CC_IGNORE',
- 'CC_WHITE',
- 'CC_LETTER',
- 'CC_OTHER',
- 'CC_ACTIVE',
- 'CC_COMMENT',
- 'CC_INVALID',
- ]
-
-# Show a list of catcode-name-symbols
-def pcl(codelist):
- result = ''
- for i in codelist:
- result = result + cc_names[i] + ', '
- return '[' + result[:-2] + ']'
-
-# the name of the catcode (ACTIVE, OTHER, etc.)
-def pc(code):
- return cc_names[code]
-
-
-# Which catcodes make the parser stop parsing regular plaintext
-regular_stopcodes = [CC_ESCAPE, CC_LBRACE, CC_RBRACE, CC_MATHSHIFT,
- CC_ALIGNMENT, CC_PARAMETER, CC_SUPERSCRIPT, CC_SUBSCRIPT,
- CC_IGNORE, CC_ACTIVE, CC_COMMENT, CC_INVALID, CC_ENDLINE]
-
-# same for scanning a control sequence name
-csname_scancodes = [CC_LETTER]
-
-# same for gobbling LWSP
-white_scancodes = [CC_WHITE]
-##white_scancodes = [CC_WHITE, CC_ENDLINE]
-
-# make a list of all catcode id's, except for catcode ``other''
-all_but_other_codes = range(16)
-del all_but_other_codes[CC_OTHER]
-##print all_but_other_codes
-
-# when does a comment end
-comment_stopcodes = [CC_ENDLINE]
-
-# gather all characters together, specified by a list of catcodes
-def code2string(cc, codelist):
- ##print 'code2string: codelist = ' + pcl(codelist),
- result = ''
- for category in codelist:
- if cc[category]:
- result = result + cc[category]
- ##print 'result = ' + `result`
- return result
-
-# automatically generate all characters of catcode other, being the
-# complement set in the ASCII range (128 characters)
-def make_other_codes(cc):
- otherchars = range(256) # could be made 256, no problem
- for category in all_but_other_codes:
- if cc[category]:
- for c in cc[category]:
- otherchars[ord(c)] = None
- result = ''
- for i in otherchars:
- if i != None:
- result = result + chr(i)
- return result
-
-# catcode dump (which characters have which catcodes).
-def dump_cc(name, cc):
- ##print '\t' + name
- ##print '=' * (8+len(name))
- if len(cc) != 16:
- raise TypeError, 'cc not good cat class'
-## for i in range(16):
-## print pc(i) + '\t' + `cc[i]`
-
-
-# In the beginning,....
-epoch_cc = [None] * 16
-##dump_cc('epoch_cc', epoch_cc)
-
-
-# INITEX
-initex_cc = epoch_cc[:]
-initex_cc[CC_ESCAPE] = '\\'
-initex_cc[CC_ENDLINE], initex_cc[CC_IGNORE], initex_cc[CC_WHITE] = \
- '\n', '\0', ' '
-initex_cc[CC_LETTER] = string.uppercase + string.lowercase
-initex_cc[CC_COMMENT], initex_cc[CC_INVALID] = '%', '\x7F'
-#initex_cc[CC_OTHER] = make_other_codes(initex_cc) I don't need them, anyway
-##dump_cc('initex_cc', initex_cc)
-
-
-# LPLAIN: LaTeX catcode setting (see lplain.tex)
-lplain_cc = initex_cc[:]
-lplain_cc[CC_LBRACE], lplain_cc[CC_RBRACE] = '{', '}'
-lplain_cc[CC_MATHSHIFT] = '$'
-lplain_cc[CC_ALIGNMENT] = '&'
-lplain_cc[CC_PARAMETER] = '#'
-lplain_cc[CC_SUPERSCRIPT] = '^\x0B' # '^' and C-k
-lplain_cc[CC_SUBSCRIPT] = '_\x01' # '_' and C-a
-lplain_cc[CC_WHITE] = lplain_cc[CC_WHITE] + '\t'
-lplain_cc[CC_ACTIVE] = '~\x0C' # '~' and C-l
-lplain_cc[CC_OTHER] = make_other_codes(lplain_cc)
-##dump_cc('lplain_cc', lplain_cc)
-
-
-# Guido's LaTeX environment catcoded '_' as ``other''
-# my own purpose catlist
-my_cc = lplain_cc[:]
-my_cc[CC_SUBSCRIPT] = my_cc[CC_SUBSCRIPT][1:] # remove '_' here
-my_cc[CC_OTHER] = my_cc[CC_OTHER] + '_' # add it to OTHER list
-dump_cc('my_cc', my_cc)
-
-
-
-# needed for un_re, my equivalent for regexp-quote in Emacs
-re_meaning = '\\[]^$'
-
-def un_re(str):
- result = ''
- for i in str:
- if i in re_meaning:
- result = result + '\\'
- result = result + i
- return result
-
-# NOTE the negate ('^') operator in *some* of the regexps below
-def make_rc_regular(cc):
- # problems here if '[]' are included!!
- return regex.compile('[' + code2string(cc, regular_stopcodes) + ']')
-
-def make_rc_cs_scan(cc):
- return regex.compile('[^' + code2string(cc, csname_scancodes) + ']')
-
-def make_rc_comment(cc):
- return regex.compile('[' + code2string(cc, comment_stopcodes) + ']')
-
-def make_rc_endwhite(cc):
- return regex.compile('[^' + code2string(cc, white_scancodes) + ']')
-
-
-
-# regular: normal mode:
-rc_regular = make_rc_regular(my_cc)
-
-# scan: scan a command sequence e.g. `newlength' or `mbox' or `;', `,' or `$'
-rc_cs_scan = make_rc_cs_scan(my_cc)
-rc_comment = make_rc_comment(my_cc)
-rc_endwhite = make_rc_endwhite(my_cc)
-
-
-# parseit (BUF, PARSEMODE=mode[MODE_REGULAR], START=0, RECURSION-LEVEL=0)
-# RECURSION-LEVEL will is incremented on entry.
-# result contains the list of chunks returned
-# together with this list, the buffer position is returned
-
-# RECURSION-LEVEL will be set to zero *again*, when recursively a
-# {,D}MATH-mode scan has been enetered.
-# This has been done in order to better check for environment-mismatches
-
-def parseit(buf, parsemode=mode[MODE_REGULAR], start=0, lvl=0):
- global lineno
-
- result = []
- end = len(buf)
- if lvl == 0 and parsemode == mode[MODE_REGULAR]:
- lineno = 1
- lvl = lvl + 1
-
- ##print 'parseit(' + epsilon(buf, start) + ', ' + `parsemode` + ', ' + `start` + ', ' + `lvl` + ')'
-
- #
- # some of the more regular modes...
- #
-
- if parsemode in (mode[MODE_REGULAR], mode[MODE_DMATH], mode[MODE_MATH]):
- cstate = []
- newpos = start
- curpmode = parsemode
- while 1:
- where = newpos
- #print '\tnew round: ' + epsilon(buf, where)
- if where == end:
- if lvl > 1 or curpmode != mode[MODE_REGULAR]:
- # not the way we started...
- raise EOFError, 'premature end of file.' + lle(lvl, buf, where)
- # the real ending of lvl-1 parse
- return end, result
-
- pos = rc_regular.search(buf, where)
-
- if pos < 0:
- pos = end
-
- if pos != where:
- newpos, c = pos, chunk(PLAIN, where, (where, pos))
- result.append(c)
- continue
-
-
- #
- # ok, pos == where and pos != end
- #
- foundchar = buf[where]
- if foundchar in my_cc[CC_LBRACE]:
- # recursive subgroup parse...
- newpos, data = parseit(buf, curpmode, where+1, lvl)
- result.append(chunk(GROUP, where, data))
-
- elif foundchar in my_cc[CC_RBRACE]:
- if lvl <= 1:
- raise error, 'ENDGROUP while in base level.' + lle(lvl, buf, where)
- if lvl == 1 and mode != mode[MODE_REGULAR]:
- raise error, 'endgroup while in math mode. +lin() + epsilon(buf, where)'
- return where + 1, result
-
- elif foundchar in my_cc[CC_ESCAPE]:
- #
- # call the routine that actually deals with
- # this problem. If do_ret is None, than
- # return the value of do_ret
- #
- # Note that handle_cs might call this routine
- # recursively again...
- #
- do_ret, newpos = handlecs(buf, where,
- curpmode, lvl, result, end)
- if do_ret != None:
- return do_ret
-
- elif foundchar in my_cc[CC_COMMENT]:
- newpos, data = parseit(buf,
- mode[MODE_COMMENT], where+1, lvl)
- result.append(chunk(COMMENT, where, data))
-
- elif foundchar in my_cc[CC_MATHSHIFT]:
- # note that recursive calls to math-mode
- # scanning are called with recursion-level 0
- # again, in order to check for bad mathend
- #
- if where + 1 != end and buf[where + 1] in my_cc[CC_MATHSHIFT]:
- #
- # double mathshift, e.g. '$$'
- #
- if curpmode == mode[MODE_REGULAR]:
- newpos, data = parseit(buf, mode[MODE_DMATH],
- where + 2, 0)
- result.append(chunk(DMATH, where, data))
- elif curpmode == mode[MODE_MATH]:
- raise error, 'wrong math delimiiter' + lin() + epsilon(buf, where)
- elif lvl != 1:
- raise error, 'bad mathend.' + lle(lvl, buf, where)
- else:
- return where + 2, result
- else:
- #
- # single math shift, e.g. '$'
- #
- if curpmode == mode[MODE_REGULAR]:
- newpos, data = parseit(buf, mode[MODE_MATH],
- where + 1, 0)
- result.append(chunk(MATH, where, data))
- elif curpmode == mode[MODE_DMATH]:
- raise error, 'wrong math delimiiter' + lin() + epsilon(buf, where)
- elif lvl != 1:
- raise error, 'bad mathend.' + lv(lvl, buf, where)
- else:
- return where + 1, result
-
- elif foundchar in my_cc[CC_IGNORE]:
- print 'warning: ignored char', `foundchar`
- newpos = where + 1
-
- elif foundchar in my_cc[CC_ACTIVE]:
- result.append(chunk(ACTIVE, where, foundchar))
- newpos = where + 1
-
- elif foundchar in my_cc[CC_INVALID]:
- raise error, 'invalid char ' + `foundchar`
- newpos = where + 1
-
- elif foundchar in my_cc[CC_ENDLINE]:
- #
- # after an end of line, eat the rest of
- # whitespace on the beginning of the next line
- # this is what LaTeX more or less does
- #
- # also, try to indicate double newlines (\par)
- #
- lineno = lineno + 1
- savedwhere = where
- newpos, dummy = parseit(buf, mode[MODE_GOBBLEWHITE], where + 1, lvl)
- if newpos != end and buf[newpos] in my_cc[CC_ENDLINE]:
- result.append(chunk(DENDLINE, savedwhere, foundchar))
- else:
- result.append(chunk(ENDLINE, savedwhere, foundchar))
- else:
- result.append(chunk(OTHER, where, foundchar))
- newpos = where + 1
-
- elif parsemode == mode[MODE_CS_SCAN]:
- #
- # scan for a control sequence token. `\ape', `\nut' or `\%'
- #
- if start == end:
- raise EOFError, 'can\'t find end of csname'
- pos = rc_cs_scan.search(buf, start)
- if pos < 0:
- pos = end
- if pos == start:
- # first non-letter right where we started the search
- # ---> the control sequence name consists of one single
- # character. Also: don't eat white space...
- if buf[pos] in my_cc[CC_ENDLINE]:
- lineno = lineno + 1
- pos = pos + 1
- return pos, (start, pos)
- else:
- spos = pos
- if buf[pos] == '\n':
- lineno = lineno + 1
- spos = pos + 1
- pos2, dummy = parseit(buf, mode[MODE_GOBBLEWHITE], spos, lvl)
- return pos2, (start, pos)
-
- elif parsemode == mode[MODE_GOBBLEWHITE]:
- if start == end:
- return start, ''
- pos = rc_endwhite.search(buf, start)
- if pos < 0:
- pos = start
- return pos, (start, pos)
-
- elif parsemode == mode[MODE_COMMENT]:
- pos = rc_comment.search(buf, start)
- lineno = lineno + 1
- if pos < 0:
- print 'no newline perhaps?'
- raise EOFError, 'can\'t find end of comment'
- pos = pos + 1
- pos2, dummy = parseit(buf, mode[MODE_GOBBLEWHITE], pos, lvl)
- return pos2, (start, pos)
-
- else:
- raise error, 'Unknown mode (' + `parsemode` + ')'
-
-
-#moreresult = cswitch(buf[x1:x2], buf, newpos, parsemode, lvl)
-
-#boxcommands = 'mbox', 'fbox'
-#defcommands = 'def', 'newcommand'
-
-endverbstr = '\\end{verbatim}'
-
-re_endverb = regex.compile(un_re(endverbstr))
-
-#
-# handlecs: helper function for parseit, for the special thing we might
-# wanna do after certain command control sequences
-# returns: None or return_data, newpos
-#
-# in the latter case, the calling function is instructed to immediately
-# return with the data in return_data
-#
-def handlecs(buf, where, curpmode, lvl, result, end):
- global lineno
-
- # get the control sequence name...
- newpos, data = parseit(buf, mode[MODE_CS_SCAN], where+1, lvl)
- saveddata = data
- s_buf_data = s(buf, data)
-
- if s_buf_data in ('begin', 'end'):
- # skip the expected '{' and get the LaTeX-envname '}'
- newpos, data = parseit(buf, mode[MODE_REGULAR], newpos+1, lvl)
- if len(data) != 1:
- raise error, 'expected 1 chunk of data.' + lle(lvl, buf, where)
-
- # yucky, we've got an environment
- envname = s(buf, data[0].data)
- s_buf_saveddata = s(buf, saveddata)
- ##print 'FOUND ' + s(buf, saveddata) + '. Name ' + `envname` + '.' + lv(lvl)
- if s_buf_saveddata == 'begin' and envname == 'verbatim':
- # verbatim deserves special treatment
- pos = re_endverb.search(buf, newpos)
- if pos < 0:
- raise error, "%s not found.%s" \
- % (`endverbstr`, lle(lvl, buf, where))
- result.append(chunk(ENV, where, (envname, [chunk(PLAIN, newpos, (newpos, pos))])))
- newpos = pos + len(endverbstr)
-
- elif s_buf_saveddata == 'begin':
- # start parsing recursively... If that parse returns
- # from an '\end{...}', then should the last item of
- # the returned data be a string containing the ended
- # environment
- newpos, data = parseit(buf, curpmode, newpos, lvl)
- if not data or type(data[-1]) is not StringType:
- raise error, "missing 'end'" + lle(lvl, buf, where) \
- + epsilon(buf, newpos)
- retenv = data[-1]
- del data[-1]
- if retenv != envname:
- #[`retenv`, `envname`]
- raise error, 'environments do not match.%s%s' \
- % (lle(lvl, buf, where), epsilon(buf, newpos))
- result.append(chunk(ENV, where, (retenv, data)))
- else:
- # 'end'... append the environment name, as just
- # pointed out, and order parsit to return...
- result.append(envname)
- ##print 'POINT of return: ' + epsilon(buf, newpos)
- # the tuple will be returned by parseit
- return (newpos, result), newpos
-
- # end of \begin ... \end handling
-
- elif s_buf_data[0:2] == 'if':
- # another scary monster: the 'if' directive
- flag = s_buf_data[2:]
-
- # recursively call parseit, just like environment above..
- # the last item of data should contain the if-termination
- # e.g., 'else' of 'fi'
- newpos, data = parseit(buf, curpmode, newpos, lvl)
- if not data or data[-1] not in ('else', 'fi'):
- raise error, 'wrong if... termination' + \
- lle(lvl, buf, where) + epsilon(buf, newpos)
-
- ifterm = data[-1]
- del data[-1]
- # 0 means dont_negate flag
- result.append(chunk(IF, where, (flag, 0, data)))
- if ifterm == 'else':
- # do the whole thing again, there is only one way
- # to end this one, by 'fi'
- newpos, data = parseit(buf, curpmode, newpos, lvl)
- if not data or data[-1] not in ('fi', ):
- raise error, 'wrong if...else... termination' \
- + lle(lvl, buf, where) \
- + epsilon(buf, newpos)
-
- ifterm = data[-1]
- del data[-1]
- result.append(chunk(IF, where, (flag, 1, data)))
- #done implicitely: return None, newpos
-
- elif s_buf_data in ('else', 'fi'):
- result.append(s(buf, data))
- # order calling party to return tuple
- return (newpos, result), newpos
-
- # end of \if, \else, ... \fi handling
-
- elif s(buf, saveddata) == 'verb':
- x2 = saveddata[1]
- result.append(chunk(CSNAME, where, data))
- if x2 == end:
- raise error, 'premature end of command.' + lle(lvl, buf, where)
- delimchar = buf[x2]
- ##print 'VERB: delimchar ' + `delimchar`
- pos = regex.compile(un_re(delimchar)).search(buf, x2 + 1)
- if pos < 0:
- raise error, 'end of \'verb\' argument (' + \
- `delimchar` + ') not found.' + \
- lle(lvl, buf, where)
- result.append(chunk(GROUP, x2, [chunk(PLAIN, x2+1, (x2+1, pos))]))
- newpos = pos + 1
- else:
- result.append(chunk(CSNAME, where, data))
- return None, newpos
-
-# this is just a function to get the string value if the possible data-tuple
-def s(buf, data):
- if type(data) is StringType:
- return data
- if len(data) != 2 or not (type(data[0]) is type(data[1]) is IntType):
- raise TypeError, 'expected tuple of 2 integers'
- x1, x2 = data
- return buf[x1:x2]
-
-
-##length, data1, i = getnextarg(length, buf, pp, i + 1)
-
-# make a deep-copy of some chunks
-def crcopy(r):
- return map(chunkcopy, r)
-
-
-# copy a chunk, would better be a method of class Chunk...
-def chunkcopy(ch):
- if ch.chtype == chunk_type[GROUP]:
- return chunk(GROUP, ch.where, map(chunkcopy, ch.data))
- else:
- return chunk(ch.chtype, ch.where, ch.data)
-
-
-# get next argument for TeX-macro, flatten a group (insert between)
-# or return Command Sequence token, or give back one character
-def getnextarg(length, buf, pp, item):
-
- ##wobj = Wobj()
- ##dumpit(buf, wobj.write, pp[item:min(length, item + 5)])
- ##print 'GETNEXTARG, (len, item) =', `length, item` + ' ---> ' + wobj.data + ' <---'
-
- while item < length and pp[item].chtype == chunk_type[ENDLINE]:
- del pp[item]
- length = length - 1
- if item >= length:
- raise error, 'no next arg.' + epsilon(buf, pp[-1].where)
- if pp[item].chtype == chunk_type[GROUP]:
- newpp = pp[item].data
- del pp[item]
- length = length - 1
- changeit(buf, newpp)
- length = length + len(newpp)
- pp[item:item] = newpp
- item = item + len(newpp)
- if len(newpp) < 10:
- wobj = Wobj()
- dumpit(buf, wobj.write, newpp)
- ##print 'GETNEXTARG: inserted ' + `wobj.data`
- return length, item
- elif pp[item].chtype == chunk_type[PLAIN]:
- #grab one char
- print 'WARNING: grabbing one char'
- if len(s(buf, pp[item].data)) > 1:
- pp.insert(item, chunk(PLAIN, pp[item].where, s(buf, pp[item].data)[:1]))
- item, length = item+1, length+1
- pp[item].data = s(buf, pp[item].data)[1:]
- else:
- item = item+1
- return length, item
- else:
- ch = pp[item]
- try:
- str = `s(buf, ch.data)`
- except TypeError:
- str = `ch.data`
- if len(str) > 400:
- str = str[:400] + '...'
- print 'GETNEXTARG:', ch.chtype, 'not handled, data ' + str
- return length, item
-
-
-# this one is needed to find the end of LaTeX's optional argument, like
-# item[...]
-re_endopt = regex.compile(']')
-
-# get a LaTeX-optional argument, you know, the square braces '[' and ']'
-def getoptarg(length, buf, pp, item):
-
- wobj = Wobj()
- dumpit(buf, wobj.write, pp[item:min(length, item + 5)])
- ##print 'GETOPTARG, (len, item) =', `length, item` + ' ---> ' + wobj.data + ' <---'
-
- if item >= length or \
- pp[item].chtype != chunk_type[PLAIN] or \
- s(buf, pp[item].data)[0] != '[':
- return length, item
-
- pp[item].data = s(buf, pp[item].data)[1:]
- if len(pp[item].data) == 0:
- del pp[item]
- length = length-1
-
- while 1:
- if item == length:
- raise error, 'No end of optional arg found'
- if pp[item].chtype == chunk_type[PLAIN]:
- text = s(buf, pp[item].data)
- pos = re_endopt.search(text)
- if pos >= 0:
- pp[item].data = text[:pos]
- if pos == 0:
- del pp[item]
- length = length-1
- else:
- item=item+1
- text = text[pos+1:]
-
- while text and text[0] in ' \t':
- text = text[1:]
-
- if text:
- pp.insert(item, chunk(PLAIN, 0, text))
- length = length + 1
- return length, item
-
- item = item+1
-
-
-# Wobj just add write-requests to the ``data'' attribute
-class Wobj:
- data = ''
-
- def write(self, data):
- self.data = self.data + data
-
-# ignore these commands
-ignoredcommands = ('hline', 'small', '/', 'tableofcontents', 'Large')
-# map commands like these to themselves as plaintext
-wordsselves = ('UNIX', 'ABC', 'C', 'ASCII', 'EOF', 'LaTeX', 'POSIX', 'TeX',
- 'SliTeX')
-# \{ --> {, \} --> }, etc
-themselves = ('{', '}', ',', '.', '@', ' ', '\n') + wordsselves
-# these ones also themselves (see argargs macro in myformat.sty)
-inargsselves = (',', '[', ']', '(', ')')
-# this is how *I* would show the difference between emph and strong
-# code 1 means: fold to uppercase
-markcmds = {'code': ('', ''), 'var': 1, 'emph': ('_', '_'),
- 'strong': ('*', '*')}
-
-# recognise patter {\FONTCHANGE-CMD TEXT} to \MAPPED-FC-CMD{TEXT}
-fontchanges = {'rm': 'r', 'it': 'i', 'em': 'emph', 'bf': 'b', 'tt': 't'}
-
-
-# try to remove macros and return flat text
-def flattext(buf, pp):
- pp = crcopy(pp)
- ##print '---> FLATTEXT ' + `pp`
- wobj = Wobj()
-
- i, length = 0, len(pp)
- while 1:
- if len(pp) != length:
- raise 'FATAL', 'inconsistent length'
- if i >= length:
- break
- ch = pp[i]
- i = i+1
- if ch.chtype == chunk_type[PLAIN]:
- pass
- elif ch.chtype == chunk_type[CSNAME]:
- s_buf_data = s(buf, ch.data)
- if convertible_csname(s_buf_data):
- ch.chtype, ch.data, nix = conversion(s_buf_data)
- if hist.inargs and s_buf_data in inargsselves:
- ch.chtype = chunk_type[PLAIN]
- elif len(s_buf_data) == 1 \
- and s_buf_data in onlylatexspecial:
- ch.chtype = chunk_type[PLAIN]
- # if it is followed by an empty group,
- # remove that group, it was needed for
- # a true space
- if i < length \
- and pp[i].chtype==chunk_type[GROUP] \
- and len(pp[i].data) == 0:
- del pp[i]
- length = length-1
-
- elif s_buf_data in markcmds.keys():
- length, newi = getnextarg(length, buf, pp, i)
- str = flattext(buf, pp[i:newi])
- del pp[i:newi]
- length = length - (newi - i)
- ch.chtype = chunk_type[PLAIN]
- markcmd = s_buf_data
- x = markcmds[markcmd]
- if type(x) == TupleType:
- pre, after = x
- str = pre+str+after
- elif x == 1:
- str = string.upper(str)
- else:
- raise 'FATAL', 'corrupt markcmds'
- ch.data = str
- else:
- if s_buf_data not in ignoredcommands:
- print 'WARNING: deleting command ' + s_buf_data
- print 'PP' + `pp[i-1]`
- del pp[i-1]
- i, length = i-1, length-1
- elif ch.chtype == chunk_type[GROUP]:
- length, newi = getnextarg(length, buf, pp, i-1)
- i = i-1
-## str = flattext(buf, crcopy(pp[i-1:newi]))
-## del pp[i:newi]
-## length = length - (newi - i)
-## ch.chtype = chunk_type[PLAIN]
-## ch.data = str
- else:
- pass
-
- dumpit(buf, wobj.write, pp)
- ##print 'FLATTEXT: RETURNING ' + `wobj.data`
- return wobj.data
-
-# try to generate node names (a bit shorter than the chapter title)
-# note that the \nodename command (see elsewhere) overules these efforts
-def invent_node_names(text):
- words = string.split(text)
-
- ##print 'WORDS ' + `words`
-
- if len(words) == 2 \
- and string.lower(words[0]) == 'built-in' \
- and string.lower(words[1]) not in ('modules', 'functions'):
- return words[1]
- if len(words) == 3 and string.lower(words[1]) == 'module':
- return words[2]
- if len(words) == 3 and string.lower(words[1]) == 'object':
- return string.join(words[0:2])
- if len(words) > 4 \
- and (string.lower(string.join(words[-4:])) \
- == 'methods and data attributes'):
- return string.join(words[:2])
- return text
-
-re_commas_etc = regex.compile('[,`\'@{}]')
-
-re_whitespace = regex.compile('[ \t]*')
-
-
-##nodenamecmd = next_command_p(length, buf, pp, newi, 'nodename')
-
-# look if the next non-white stuff is also a command, resulting in skipping
-# double endlines (DENDLINE) too, and thus omitting \par's
-# Sometimes this is too much, maybe consider DENDLINE's as stop
-def next_command_p(length, buf, pp, i, cmdname):
-
- while 1:
- if i >= len(pp):
- break
- ch = pp[i]
- i = i+1
- if ch.chtype == chunk_type[ENDLINE]:
- continue
- if ch.chtype == chunk_type[DENDLINE]:
- continue
- if ch.chtype == chunk_type[PLAIN]:
- if re_whitespace.search(s(buf, ch.data)) == 0 and \
- re_whitespace.match(s(buf, ch.data)) == len(s(buf, ch.data)):
- continue
- return -1
- if ch.chtype == chunk_type[CSNAME]:
- if s(buf, ch.data) == cmdname:
- return i # _after_ the command
- return -1
- return -1
-
-
-# things that are special to LaTeX, but not to texi..
-onlylatexspecial = '_~^$#&%'
-
-class Struct: pass
-
-hist = Struct()
-out = Struct()
-
-def startchange():
- global hist, out
-
- hist.chaptertype = "chapter"
- hist.inenv = []
- hist.nodenames = []
- hist.cindex = []
- hist.inargs = 0
- hist.enumeratenesting, hist.itemizenesting = 0, 0
- hist.this_module = None
-
- out.doublenodes = []
- out.doublecindeces = []
-
-
-spacech = [chunk(PLAIN, 0, ' ')]
-commach = [chunk(PLAIN, 0, ', ')]
-cindexch = [chunk(CSLINE, 0, 'cindex')]
-
-# the standard variation in symbols for itemize
-itemizesymbols = ['bullet', 'minus', 'dots']
-
-# same for enumerate
-enumeratesymbols = ['1', 'A', 'a']
-
-# Map of things that convert one-to-one. Each entry is a 3-tuple:
-#
-# new_chtype, new_data, nix_trailing_empty_group
-#
-d = {}
-# add stuff that converts from one name to another:
-for name in ('url', 'module', 'function', 'cfunction',
- 'keyword', 'method', 'exception', 'constant',
- 'email', 'class', 'member', 'cdata', 'ctype',
- 'member', 'sectcode', 'verb',
- 'cfunction', 'cdata', 'ctype',
- ):
- d[name] = chunk_type[CSNAME], 'code', 0
-for name in ('emph', 'var', 'strong', 'code', 'kbd', 'key',
- 'dfn', 'samp', 'file', 'r', 'i', 't'):
- d[name] = chunk_type[CSNAME], name, 0
-d['character'] = chunk_type[CSNAME], 'samp', 0
-d['url'] = chunk_type[CSNAME], 'code', 0
-d['email'] = chunk_type[CSNAME], 'code', 0
-d['mimetype'] = chunk_type[CSNAME], 'code', 0
-d['newsgroup'] = chunk_type[CSNAME], 'code', 0
-d['program'] = chunk_type[CSNAME], 'strong', 0
-d['\\'] = chunk_type[CSNAME], '*', 0
-# add stuff that converts to text:
-for name in themselves:
- d[name] = chunk_type[PLAIN], name, 0
-for name in wordsselves:
- d[name] = chunk_type[PLAIN], name, 1
-for name in ',[]()':
- d[name] = chunk_type[PLAIN], name, 0
-# a lot of these are LaTeX2e additions
-for name, value in [('quotedblbase', ',,'), ('quotesinglbase', ','),
- ('textquotedbl', '"'), ('LaTeXe', 'LaTeX2e'),
- ('e', '\\'), ('textquotedblleft', "``"),
- ('textquotedblright', "''"), ('textquoteleft', "`"),
- ('textquoteright', "'"), ('textbackslash', '\\'),
- ('textbar', '|'), ('textless', '<'),
- ('textgreater', '>'), ('textasciicircum', '^'),
- ('Cpp', 'C++'), ('copyright', '')]:
- d[name] = chunk_type[PLAIN], value, 1
-convertible_csname = d.has_key
-conversion = d.get
-del d, name, value
-
-##
-## \begin{ {func,data,exc}desc }{name}...
-## the resulting texi-code is dependent on the contents of indexsubitem
-##
-
-# indexsubitem: `['XXX', 'function']
-# funcdesc:
-# deffn {`idxsi`} NAME (FUNCARGS)
-
-# indexsubitem: `['XXX', 'method']`
-# funcdesc:
-# defmethod {`idxsi[0]`} NAME (FUNCARGS)
-
-# indexsubitem: `['in', 'module', 'MODNAME']'
-# datadesc:
-# defcv data {`idxsi[1:]`} NAME
-# excdesc:
-# defcv exception {`idxsi[1:]`} NAME
-# funcdesc:
-# deffn {function of `idxsi[1:]`} NAME (FUNCARGS)
-
-# indexsubitem: `['OBJECT', 'attribute']'
-# datadesc
-# defcv attribute {`OBJECT`} NAME
-
-
-## this routine will be called on \begin{funcdesc}{NAME}{ARGS}
-## or \funcline{NAME}{ARGS}
-##
-def do_funcdesc(length, buf, pp, i, index=1):
- startpoint = i-1
- ch = pp[startpoint]
- wh = ch.where
- length, newi = getnextarg(length, buf, pp, i)
- funcname = chunk(GROUP, wh, pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
- save = hist.inargs
- hist.inargs = 1
- length, newi = getnextarg(length, buf, pp, i)
- hist.inargs = save
- del save
- the_args = [chunk(PLAIN, wh, '()'[0])] + pp[i:newi] + \
- [chunk(PLAIN, wh, '()'[1])]
- del pp[i:newi]
- length = length - (newi-i)
-
- idxsi = hist.indexsubitem # words
- command = 'deffn'
- if hist.this_module:
- cat_class = 'function of ' + hist.this_module
- else:
- cat_class = 'built-in function'
- ch.chtype = chunk_type[CSLINE]
- ch.data = command
-
- cslinearg = [chunk(GROUP, wh, [chunk(PLAIN, wh, cat_class)])]
- cslinearg.append(chunk(PLAIN, wh, ' '))
- cslinearg.append(funcname)
- cslinearg.append(chunk(PLAIN, wh, ' '))
- l = len(cslinearg)
- cslinearg[l:l] = the_args
-
- pp.insert(i, chunk(GROUP, wh, cslinearg))
- i, length = i+1, length+1
- hist.command = command
- return length, i
-
-
-## this routine will be called on \begin{excdesc}{NAME}
-## or \excline{NAME}
-##
-def do_excdesc(length, buf, pp, i):
- startpoint = i-1
- ch = pp[startpoint]
- wh = ch.where
- length, newi = getnextarg(length, buf, pp, i)
- excname = chunk(GROUP, wh, pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
-
- idxsi = hist.indexsubitem # words
- command = ''
- cat_class = ''
- class_class = ''
- if idxsi == ['built-in', 'exception', 'base', 'class']:
- command = 'defvr'
- cat_class = 'exception base class'
- else:
- command = 'defcv'
- cat_class = 'exception'
-
- ch.chtype = chunk_type[CSLINE]
- ch.data = command
-
- cslinearg = [chunk(GROUP, wh, [chunk(PLAIN, wh, cat_class)])]
- cslinearg.append(chunk(PLAIN, wh, ' '))
- if class_class:
- cslinearg.append(chunk(GROUP, wh, [chunk(PLAIN, wh, class_class)]))
- cslinearg.append(chunk(PLAIN, wh, ' '))
- cslinearg.append(excname)
-
- pp.insert(i, chunk(GROUP, wh, cslinearg))
- i, length = i+1, length+1
- hist.command = command
- return length, i
-
-## same for datadesc or dataline...
-def do_datadesc(length, buf, pp, i, index=1):
- startpoint = i-1
- ch = pp[startpoint]
- wh = ch.where
- length, newi = getnextarg(length, buf, pp, i)
- dataname = chunk(GROUP, wh, pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
-
- idxsi = hist.indexsubitem # words
- command = 'defcv'
- cat_class = 'data'
- class_class = ''
- if idxsi[-1] in ('attribute', 'option'):
- cat_class = idxsi[-1]
- class_class = string.join(idxsi[:-1])
- elif len(idxsi) == 3 and idxsi[:2] == ['in', 'module']:
- class_class = string.join(idxsi[1:])
- elif len(idxsi) == 4 and idxsi[:3] == ['data', 'in', 'module']:
- class_class = string.join(idxsi[2:])
- else:
- class_class = string.join(idxsi)
-
- ch.chtype = chunk_type[CSLINE]
- ch.data = command
-
- cslinearg = [chunk(GROUP, wh, [chunk(PLAIN, wh, cat_class)])]
- cslinearg.append(chunk(PLAIN, wh, ' '))
- if class_class:
- cslinearg.append(chunk(GROUP, wh, [chunk(PLAIN, wh, class_class)]))
- cslinearg.append(chunk(PLAIN, wh, ' '))
- cslinearg.append(dataname)
-
- pp.insert(i, chunk(GROUP, wh, cslinearg))
- i, length = i+1, length+1
- hist.command = command
- return length, i
-
-
-def do_opcodedesc(length, buf, pp, i):
- startpoint = i-1
- ch = pp[startpoint]
- wh = ch.where
- length, newi = getnextarg(length, buf, pp, i)
- dataname = chunk(GROUP, wh, pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
-
- ch.chtype = CSLINE
- ch.data = "deffn"
-
- cslinearg = [chunk(PLAIN, wh, 'byte\ code\ instruction'),
- chunk(GROUP, wh, [chunk(PLAIN, wh, "byte code instruction")]),
- chunk(PLAIN, wh, ' '),
- dataname,
- chunk(PLAIN, wh, ' '),
- pp[i],
- ]
-
- pp[i] = chunk(GROUP, wh, cslinearg)
- hist.command = ch.data
- return length, i
-
-
-def add_module_index(pp, length, i, buf, ch, extra, ref=1):
- ch.chtype = chunk_type[CSLINE]
- ch.data = 'pindex'
- length, newi = getnextarg(length, buf, pp, i)
- ingroupch = pp[i:newi]
- del pp[i:newi]
- length = length - (newi-i)
- if not ref:
- if len(ingroupch) == 1:
- hist.this_module = s(buf, ch.data)
- else:
- hist.this_module = None
- print 'add_module_index() error ==>', ingroupch
-
- if extra:
- ingroupch.append(chunk(PLAIN, ch.where, ' '))
- ingroupch.append(chunk(CSNAME, ch.where, 'r'))
- ingroupch.append(chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, extra)]))
-
- pp.insert(i, chunk(GROUP, ch.where, ingroupch))
- return length+1, i+1
-
-
-def yank_indexsubitem(pp, length, i, buf, ch, cmdname):
- stuff = pp[i].data
- if len(stuff) != 1:
- raise error, "first parameter to \\%s too long" % cmdname
- if pp[i].chtype != chunk_type[GROUP]:
- raise error, "bad chunk type following \\%s" \
- "\nexpected GROUP, got %s" + (cmdname, str(ch.chtype))
- text = s(buf, stuff[0].data)
- if text[:1] != '(' or text[-1:] != ')':
- raise error, \
- 'expected indexsubitem enclosed in parenteses'
- hist.indexsubitem = string.split(text[1:-1])
- del pp[i-1:i+1]
- return length - 2, i - 1
-
-
-# regular indices: those that are not set in tt font by default....
-regindices = ('cindex', )
-
-# remove illegal characters from node names
-def rm_commas_etc(text):
- result = ''
- changed = 0
- while 1:
- pos = re_commas_etc.search(text)
- if pos >= 0:
- changed = 1
- result = result + text[:pos]
- text = text[pos+1:]
- else:
- result = result + text
- break
- if changed:
- print 'Warning: nodename changed to ' + `result`
-
- return result
-
-# boolean flags
-flags = {'texi': 1}
-
-
-# map of \label{} to node names
-label_nodes = {}
-
-
-##
-## changeit: the actual routine, that changes the contents of the parsed
-## chunks
-##
-
-def changeit(buf, pp):
- global onlylatexspecial, hist, out
-
- i, length = 0, len(pp)
- while 1:
- # sanity check: length should always equal len(pp)
- if len(pp) != length:
- print i, pp[i]
- raise 'FATAL', 'inconsistent length. thought ' + `length` + ', but should really be ' + `len(pp)`
- if i >= length:
- break
- ch = pp[i]
- i = i + 1
-
- if type(ch) is StringType:
- #normally, only chunks are present in pp,
- # but in some cases, some extra info
- # has been inserted, e.g., the \end{...} clauses
- raise 'FATAL', 'got string, probably too many ' + `end`
-
- if ch.chtype == chunk_type[GROUP]:
- # check for {\em ...} constructs
- data = ch.data
- if data and \
- data[0].chtype == chunk_type[CSNAME] and \
- fontchanges.has_key(s(buf, data[0].data)):
- k = s(buf, data[0].data)
- del data[0]
- pp.insert(i-1, chunk(CSNAME, ch.where, fontchanges[k]))
- length, i = length+1, i+1
-
- elif data:
- if len(data) \
- and data[0].chtype == chunk_type[GROUP] \
- and len(data[0].data) \
- and data[0].data[0].chtype == chunk_type[CSNAME] \
- and s(buf, data[0].data[0].data) == 'e':
- data[0] = data[0].data[0]
- print "invoking \\e magic group transform..."
- else:
-## print "GROUP -- ch.data[0].data =", ch.data[0].data
- k = s(buf, data[0].data)
- if k == "fulllineitems":
- del data[0]
- pp[i-1:i] = data
- i = i - 1
- length = length + len(data) - 1
- continue
-
- # recursively parse the contents of the group
- changeit(buf, data)
-
- elif ch.chtype == chunk_type[IF]:
- # \if...
- flag, negate, data = ch.data
- ##print 'IF: flag, negate = ' + `flag, negate`
- if flag not in flags.keys():
- raise error, 'unknown flag ' + `flag`
-
- value = flags[flag]
- if negate:
- value = (not value)
- del pp[i-1]
- length, i = length-1, i-1
- if value:
- pp[i:i] = data
- length = length + len(data)
-
-
- elif ch.chtype == chunk_type[ENV]:
- # \begin{...} ....
- envname, data = ch.data
-
- #push this environment name on stack
- hist.inenv.insert(0, envname)
-
- #append an endenv chunk after grouped data
- data.append(chunk(ENDENV, ch.where, envname))
- ##[`data`]
-
- #delete this object
- del pp[i-1]
- i, length = i-1, length-1
-
- #insert found data
- pp[i:i] = data
- length = length + len(data)
-
- if envname == 'verbatim':
- pp[i:i] = [chunk(CSLINE, ch.where, 'example'),
- chunk(GROUP, ch.where, [])]
- length, i = length+2, i+2
-
- elif envname in ('itemize', 'list', 'fulllineitems'):
- if hist.itemizenesting > len(itemizesymbols):
- raise error, 'too deep itemize nesting'
- if envname == 'list':
- del pp[i:i+2]
- length = length - 2
- ingroupch = [chunk(CSNAME, ch.where,
- itemizesymbols[hist.itemizenesting])]
- hist.itemizenesting = hist.itemizenesting + 1
- pp[i:i] = [chunk(CSLINE, ch.where, 'itemize'),
- chunk(GROUP, ch.where, ingroupch)]
- length, i = length+2, i+2
-
- elif envname == 'enumerate':
- if hist.enumeratenesting > len(enumeratesymbols):
- raise error, 'too deep enumerate nesting'
- ingroupch = [chunk(PLAIN, ch.where,
- enumeratesymbols[hist.enumeratenesting])]
- hist.enumeratenesting = hist.enumeratenesting + 1
- pp[i:i] = [chunk(CSLINE, ch.where, 'enumerate'),
- chunk(GROUP, ch.where, ingroupch)]
- length, i = length+2, i+2
-
- elif envname == 'description':
- ingroupch = [chunk(CSNAME, ch.where, 'b')]
- pp[i:i] = [chunk(CSLINE, ch.where, 'table'),
- chunk(GROUP, ch.where, ingroupch)]
- length, i = length+2, i+2
-
- elif (envname == 'tableiii') or (envname == 'tableii'):
- if (envname == 'tableii'):
- ltable = 2
- else:
- ltable = 3
- wh = ch.where
- newcode = []
-
- #delete tabular format description
- # e.g., {|l|c|l|}
- length, newi = getnextarg(length, buf, pp, i)
- del pp[i:newi]
- length = length - (newi-i)
-
- newcode.append(chunk(CSLINE, wh, 'table'))
- ingroupch = [chunk(CSNAME, wh, 'asis')]
- newcode.append(chunk(GROUP, wh, ingroupch))
- newcode.append(chunk(CSLINE, wh, 'item'))
-
- #get the name of macro for @item
- # e.g., {code}
- length, newi = getnextarg(length, buf, pp, i)
-
- if newi-i != 1:
- raise error, 'Sorry, expected 1 chunk argument'
- if pp[i].chtype != chunk_type[PLAIN]:
- raise error, 'Sorry, expected plain text argument'
- hist.itemargmacro = s(buf, pp[i].data)
- if convertible_csname(hist.itemargmacro):
- hist.itemargmacro = conversion(hist.itemargmacro)[1]
- del pp[i:newi]
- length = length - (newi-i)
-
- itembody = []
- for count in range(ltable):
- length, newi = getnextarg(length, buf, pp, i)
- emphgroup = [
- chunk(CSNAME, wh, 'emph'),
- chunk(GROUP, 0, pp[i:newi])]
- del pp[i:newi]
- length = length - (newi-i)
- if count == 0:
- itemarg = emphgroup
- elif count == ltable-1:
- itembody = itembody + \
- [chunk(PLAIN, wh, ' --- ')] + emphgroup
- else:
- itembody = emphgroup
- newcode.append(chunk(GROUP, wh, itemarg))
- newcode = newcode + itembody + [chunk(DENDLINE, wh, '\n')]
- pp[i:i] = newcode
- l = len(newcode)
- length, i = length+l, i+l
- del newcode, l
-
- if length != len(pp):
- raise 'STILL, SOMETHING wrong', `i`
-
- elif envname in ('methoddesc', 'methoddescni'):
- length, newi = getoptarg(length, buf, pp, i)
- ingroupch = pp[i:newi]
- del pp[i:newi]
- length = length - (newi-i)
- #
- pp.insert(i, chunk(PLAIN, ch.where, ''))
- i, length = i+1, length+1
- length, i = do_funcdesc(length, buf, pp, i,
- envname[-2:] != "ni")
-
- elif envname in ('memberdesc', 'memberdescni'):
- length, newi = getoptarg(length, buf, pp, i)
- ingroupch = pp[i:newi]
- del pp[i:newi]
- length = length - (newi-i)
- #
- pp.insert(i, chunk(PLAIN, ch.where, ''))
- i, length = i+1, length+1
- length, i = do_datadesc(length, buf, pp, i,
- envname[-2:] != "ni")
-
- elif envname in ('funcdesc', 'funcdescni', 'classdesc'):
- pp.insert(i, chunk(PLAIN, ch.where, ''))
- i, length = i+1, length+1
- length, i = do_funcdesc(length, buf, pp, i,
- envname[-2:] != "ni")
-
- elif envname == 'excdesc':
- pp.insert(i, chunk(PLAIN, ch.where, ''))
- i, length = i+1, length+1
- length, i = do_excdesc(length, buf, pp, i)
-
- elif envname in ('datadesc', 'datadescni'):
- pp.insert(i, chunk(PLAIN, ch.where, ''))
- i, length = i+1, length+1
- length, i = do_datadesc(length, buf, pp, i,
- envname[-2:] != "ni")
-
- elif envname == 'opcodedesc':
- pp.insert(i, chunk(PLAIN, ch.where, ''))
- i, length = i+1, length+1
- length, i = do_opcodedesc(length, buf, pp, i)
-
- elif envname == 'seealso':
- chunks = [chunk(ENDLINE, ch.where, "\n"),
- chunk(CSNAME, ch.where, "b"),
- chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, "See also: ")]),
- chunk(ENDLINE, ch.where, "\n"),
- chunk(ENDLINE, ch.where, "\n")]
- pp[i-1:i] = chunks
- length = length + len(chunks) - 1
- i = i + len(chunks) - 1
-
- elif envname in ('sloppypar', 'flushleft', 'document'):
- pass
-
- else:
- print 'WARNING: don\'t know what to do with env ' + `envname`
-
- elif ch.chtype == chunk_type[ENDENV]:
- envname = ch.data
- if envname != hist.inenv[0]:
- raise error, '\'end\' does not match. Name ' + `envname` + ', expected ' + `hist.inenv[0]`
- del hist.inenv[0]
- del pp[i-1]
- i, length = i-1, length-1
-
- if envname == 'verbatim':
- pp[i:i] = [chunk(CSLINE, ch.where, 'end'),
- chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, 'example')])]
- i, length = i+2, length+2
- elif envname in ('itemize', 'list', 'fulllineitems'):
- hist.itemizenesting = hist.itemizenesting - 1
- pp[i:i] = [chunk(CSLINE, ch.where, 'end'),
- chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, 'itemize')])]
- i, length = i+2, length+2
- elif envname == 'enumerate':
- hist.enumeratenesting = hist.enumeratenesting-1
- pp[i:i] = [chunk(CSLINE, ch.where, 'end'),
- chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, 'enumerate')])]
- i, length = i+2, length+2
- elif envname == 'description':
- pp[i:i] = [chunk(CSLINE, ch.where, 'end'),
- chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, 'table')])]
- i, length = i+2, length+2
- elif (envname == 'tableiii') or (envname == 'tableii'):
- pp[i:i] = [chunk(CSLINE, ch.where, 'end'),
- chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, 'table')])]
- i, length = i+2, length + 2
- pp.insert(i, chunk(DENDLINE, ch.where, '\n'))
- i, length = i+1, length+1
-
- elif envname in ('funcdesc', 'excdesc', 'datadesc', 'classdesc',
- 'funcdescni', 'datadescni',
- 'methoddesc', 'memberdesc',
- 'methoddescni', 'memberdescni',
- ):
- pp[i:i] = [chunk(CSLINE, ch.where, 'end'),
- chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, hist.command)])]
- i, length = i+2, length+2
-
- elif envname == 'opcodedesc':
- pp[i:i] = [chunk(CSLINE, ch.where, 'end'),
- chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, "deffn")])]
- i, length = i+2, length+2
-
- elif envname in ('seealso', 'sloppypar', 'flushleft', 'document'):
- pass
-
- else:
- print 'WARNING: ending env %s has no actions' % `envname`
-
- elif ch.chtype == chunk_type[CSNAME]:
- # control name transformations
- s_buf_data = s(buf, ch.data)
- if s_buf_data == 'optional':
- pp[i-1].chtype = chunk_type[PLAIN]
- pp[i-1].data = '['
- if (i < length) and \
- (pp[i].chtype == chunk_type[GROUP]):
- cp=pp[i].data
- pp[i:i+1]=cp + [
- chunk(PLAIN, ch.where, ']')]
- length = length+len(cp)
-
- elif s_buf_data in ignoredcommands:
- del pp[i-1]
- i, length = i-1, length-1
-
- elif s_buf_data == '@' and \
- i != length and \
- pp[i].chtype == chunk_type[PLAIN] and \
- s(buf, pp[i].data)[0] == '.':
- # \@. --> \. --> @.
- ch.data = '.'
- del pp[i]
- length = length - 1
-
- elif convertible_csname(s_buf_data):
- ch.chtype, ch.data, nix = conversion(s_buf_data)
- try:
- if nix and pp[i].chtype == chunk_type[GROUP] \
- and len(pp[i].data) == 0:
- del pp[i]
- length = length - 1
- except IndexError:
- pass
-
- elif s_buf_data == '\\':
- # \\ --> \* --> @*
- ch.data = '*'
-
- elif len(s_buf_data) == 1 and \
- s_buf_data in onlylatexspecial:
- ch.chtype = chunk_type[PLAIN]
- # check if such a command is followed by
- # an empty group: e.g., `\%{}'. If so, remove
- # this empty group too
- if i < length and \
- pp[i].chtype == chunk_type[GROUP] \
- and len(pp[i].data) == 0:
- del pp[i]
- length = length-1
-
- elif s_buf_data == "appendix":
- hist.chaptertype = "appendix"
- del pp[i-1]
- i, length = i-1, length-1
-
- elif hist.inargs and s_buf_data in inargsselves:
- # This is the special processing of the
- # arguments of the \begin{funcdesc}... or
- # \funcline... arguments
- # \, --> , \[ --> [, \] --> ]
- ch.chtype = chunk_type[PLAIN]
-
- elif s_buf_data == 'setindexsubitem':
- length, i = yank_indexsubitem(pp, length, i, buf, ch,
- 'setindexsubitem')
-
- elif s_buf_data == 'withsubitem':
- oldsubitem = hist.indexsubitem
- try:
- length, i = yank_indexsubitem(pp, length, i, buf, ch,
- 'withsubitem')
- stuff = pp[i].data
- del pp[i]
- length = length - 1
- changeit(buf, stuff)
- stuff = None
- finally:
- hist.indexsubitem = oldsubitem
-
- elif s_buf_data in ('textrm', 'pytype'):
- stuff = pp[i].data
- pp[i-1:i+1] = stuff
- length = length - 2 + len(stuff)
- stuff = None
- i = i - 1
-
- elif s_buf_data == 'newcommand':
- print "ignoring definition of \\" + s(buf, pp[i].data[0].data)
- del pp[i-1:i+2]
- i = i - 1
- length = length - 3
-
- elif s_buf_data == 'renewcommand':
- print "ignoring redefinition of \\" \
- + s(buf, pp[i].data[0].data)
- del pp[i-1:i+2]
- i = i - 1
- length = length - 3
-
- elif s_buf_data == 'mbox':
- stuff = pp[i].data
- pp[i-1:i+1] = stuff
- i = i - 1
- length = length + len(stuff) - 2
- stuff = None
-
- elif s_buf_data == 'version':
- ch.chtype = chunk_type[PLAIN]
- ch.data = release_version
-
- elif s_buf_data == 'item':
- ch.chtype = chunk_type[CSLINE]
- length, newi = getoptarg(length, buf, pp, i)
- ingroupch = pp[i:newi]
- del pp[i:newi]
- length = length - (newi-i)
- changeit(buf, ingroupch) # catch stuff inside the optional arg
- pp.insert(i, chunk(GROUP, ch.where, ingroupch))
- i, length = i+1, length+1
-
- elif s_buf_data == 'ttindex':
- idxsi = hist.indexsubitem
-
- cat_class = ''
- if len(idxsi) >= 2 and idxsi[1] in \
- ('method', 'function', 'protocol'):
- command = 'findex'
- elif len(idxsi) >= 2 and idxsi[1] in \
- ('exception', 'object'):
- command = 'vindex'
- elif len(idxsi) == 3 and idxsi[:2] == ['in', 'module']:
- command = 'cindex'
- elif len(idxsi) == 3 and idxsi[:2] == ['class', 'in']:
- command = 'findex'
- else:
- print 'WARNING: can\'t categorize ' + `idxsi` \
- + ' for \'ttindex\' command'
- command = 'cindex'
-
- if not cat_class:
- cat_class = '(%s)' % string.join(idxsi)
-
- ch.chtype = chunk_type[CSLINE]
- ch.data = command
-
- length, newi = getnextarg(length, buf, pp, i)
- arg = pp[i:newi]
- del pp[i:newi]
- length = length - (newi-i)
-
- cat_arg = [chunk(PLAIN, ch.where, cat_class)]
-
- # determine what should be set in roman, and
- # what in tt-font
- if command in regindices:
-
- arg = [chunk(CSNAME, ch.where, 't'),
- chunk(GROUP, ch.where, arg)]
- else:
- cat_arg = [chunk(CSNAME, ch.where, 'r'),
- chunk(GROUP, ch.where, cat_arg)]
-
- ingroupch = arg + \
- [chunk(PLAIN, ch.where, ' ')] + \
- cat_arg
-
- pp.insert(i, chunk(GROUP, ch.where, ingroupch))
- length, i = length+1, i+1
-
- elif s_buf_data == 'ldots':
- # \ldots --> \dots{} --> @dots{}
- ch.data = 'dots'
- if i == length \
- or pp[i].chtype != chunk_type[GROUP] \
- or pp[i].data != []:
- pp.insert(i, chunk(GROUP, ch.where, []))
- i, length = i+1, length+1
-
- elif s_buf_data in themselves:
- # \UNIX --> &UNIX;
- ch.chtype = chunk_type[PLAIN]
- if i != length \
- and pp[i].chtype == chunk_type[GROUP] \
- and pp[i].data == []:
- del pp[i]
- length = length-1
-
- elif s_buf_data == 'manpage':
- ch.data = 'emph'
- sect = s(buf, pp[i+1].data[0].data)
- pp[i+1].data = "(%s)" % sect
- pp[i+1].chtype = chunk_type[PLAIN]
-
- elif s_buf_data == 'envvar':
- # this should do stuff in the index, too...
- ch.data = "$"
- ch.chtype = chunk_type[PLAIN]
- pp[i] = pp[i].data[0]
-
- elif s_buf_data == 'regexp':
- ch.data = 'code'
- pp.insert(i+1, chunk(PLAIN, ch.where, '"'))
- pp.insert(i-1, chunk(PLAIN, ch.where, '"'))
- length = length + 2
- i = i + 1
-
- elif s_buf_data in ('lineiii', 'lineii'):
- # This is the most tricky one
- # \lineiii{a1}{a2}[{a3}] -->
- # @item @<cts. of itemargmacro>{a1}
- # a2 [ -- a3]
- #
- if not hist.inenv:
- raise error, 'no environment for lineiii'
- if (hist.inenv[0] != 'tableiii') and \
- (hist.inenv[0] != 'tableii'):
- raise error, \
- 'wrong command (%s) in wrong environment (%s)' \
- % (s_buf_data, `hist.inenv[0]`)
- ch.chtype = chunk_type[CSLINE]
- ch.data = 'item'
- length, newi = getnextarg(length, buf, pp, i)
- ingroupch = [chunk(CSNAME, 0, hist.itemargmacro),
- chunk(GROUP, 0, pp[i:newi])]
- del pp[i:newi]
- length = length - (newi-i)
- pp.insert(i, chunk(GROUP, ch.where, ingroupch))
- grouppos = i
- i, length = i+1, length+1
- length, i = getnextarg(length, buf, pp, i)
- length, newi = getnextarg(length, buf, pp, i)
- if newi > i:
- # we have a 3rd arg
- pp.insert(i, chunk(PLAIN, ch.where, ' --- '))
- i = newi + 1
- length = length + 1
- if length != len(pp):
- raise 'IN LINEIII IS THE ERR', `i`
-
- elif s_buf_data in ('chapter', 'section',
- 'subsection', 'subsubsection'):
- #\xxxsection{A} ---->
- # @node A, , ,
- # @xxxsection A
- ## also: remove commas and quotes
- hist.this_module = None
- if s_buf_data == "chapter":
- ch.data = hist.chaptertype
- ch.chtype = chunk_type[CSLINE]
- length, newi = getnextarg(length, buf, pp, i)
- afternodenamecmd = next_command_p(length, buf,
- pp, newi, 'nodename')
- if afternodenamecmd < 0:
- cp1 = crcopy(pp[i:newi])
- pp[i:newi] = [chunk(GROUP, ch.where, pp[i:newi])]
- length, newi = length - (newi-i) + 1, i+1
- text = flattext(buf, cp1)
- text = invent_node_names(text)
- else:
- length, endarg = getnextarg(length, buf,
- pp, afternodenamecmd)
- cp1 = crcopy(pp[afternodenamecmd:endarg])
- del pp[newi:endarg]
- length = length - (endarg-newi)
-
- pp[i:newi] = [chunk(GROUP, ch.where, pp[i:newi])]
- length, newi = length - (newi-i) + 1, i + 1
- text = flattext(buf, cp1)
- if text[-1] == '.':
- text = text[:-1]
- if text in hist.nodenames:
- print 'WARNING: node name ' + `text` + ' already used'
- out.doublenodes.append(text)
- else:
- hist.nodenames.append(text)
- text = rm_commas_etc(text)
- pp[i-1:i-1] = [chunk(CSLINE, ch.where, 'node'),
- chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, text+', , ,')
- ])]
- i, length = newi+2, length+2
-
- elif s_buf_data == 'funcline':
- # fold it to a very short environment
- pp[i-1:i-1] = [chunk(CSLINE, ch.where, 'end'),
- chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, hist.command)])]
- i, length = i+2, length+2
- length, i = do_funcdesc(length, buf, pp, i)
-
- elif s_buf_data == 'dataline':
- pp[i-1:i-1] = [chunk(CSLINE, ch.where, 'end'),
- chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, hist.command)])]
- i, length = i+2, length+2
- length, i = do_datadesc(length, buf, pp, i)
-
- elif s_buf_data == 'excline':
- pp[i-1:i-1] = [chunk(CSLINE, ch.where, 'end'),
- chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where, hist.command)])]
- i, length = i+2, length+2
- length, i = do_excdesc(length, buf, pp, i)
-
- elif s_buf_data == 'index':
- #\index{A} --->
- # @cindex A
- ch.chtype = chunk_type[CSLINE]
- ch.data = 'cindex'
- length, newi = getnextarg(length, buf, pp, i)
-
- ingroupch = pp[i:newi]
- del pp[i:newi]
- length = length - (newi-i)
- pp.insert(i, chunk(GROUP, ch.where, ingroupch))
- length, i = length+1, i+1
-
- elif s_buf_data == 'bifuncindex':
- ch.chtype = chunk_type[CSLINE]
- ch.data = 'findex'
- length, newi = getnextarg(length, buf, pp, i)
- ingroupch = pp[i:newi]
- del pp[i:newi]
- length = length - (newi-i)
-
- ingroupch.append(chunk(PLAIN, ch.where, ' '))
- ingroupch.append(chunk(CSNAME, ch.where, 'r'))
- ingroupch.append(chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where,
- '(built-in function)')]))
-
- pp.insert(i, chunk(GROUP, ch.where, ingroupch))
- length, i = length+1, i+1
-
- elif s_buf_data == 'obindex':
- ch.chtype = chunk_type[CSLINE]
- ch.data = 'findex'
- length, newi = getnextarg(length, buf, pp, i)
- ingroupch = pp[i:newi]
- del pp[i:newi]
- length = length - (newi-i)
-
- ingroupch.append(chunk(PLAIN, ch.where, ' '))
- ingroupch.append(chunk(CSNAME, ch.where, 'r'))
- ingroupch.append(chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where,
- '(object)')]))
-
- pp.insert(i, chunk(GROUP, ch.where, ingroupch))
- length, i = length+1, i+1
-
- elif s_buf_data == 'opindex':
- ch.chtype = chunk_type[CSLINE]
- ch.data = 'findex'
- length, newi = getnextarg(length, buf, pp, i)
- ingroupch = pp[i:newi]
- del pp[i:newi]
- length = length - (newi-i)
-
- ingroupch.append(chunk(PLAIN, ch.where, ' '))
- ingroupch.append(chunk(CSNAME, ch.where, 'r'))
- ingroupch.append(chunk(GROUP, ch.where, [
- chunk(PLAIN, ch.where,
- '(operator)')]))
-
- pp.insert(i, chunk(GROUP, ch.where, ingroupch))
- length, i = length+1, i+1
-
- elif s_buf_data in ('bimodindex', 'refbimodindex'):
- length, i = add_module_index(
- pp, length, i, buf, ch, '(built-in)',
- (s_buf_data[:3] == 'ref'))
-
- elif s_buf_data in ('modindex', 'refmodindex'):
- length, i = add_module_index(
- pp, length, i, buf, ch, '',
- (s_buf_data[:3] == 'ref'))
-
- elif s_buf_data in ('stmodindex', 'refstmodindex'):
- length, i = add_module_index(
- pp, length, i, buf, ch, '(standard)',
- (s_buf_data[:3] == 'ref'))
-
- elif s_buf_data in ('exmodindex', 'refexmodindex'):
- length, i = add_module_index(
- pp, length, i, buf, ch, '(extension)',
- (s_buf_data[:3] == 'ref'))
-
- elif s_buf_data == 'stindex':
- # XXX must actually go to newindex st
- what = (s_buf_data[:2] == "st") and "statement" or "keyword"
- wh = ch.where
- ch.chtype = chunk_type[CSLINE]
- ch.data = 'cindex'
- length, newi = getnextarg(length, buf, pp, i)
- ingroupch = [chunk(CSNAME, wh, 'code'),
- chunk(GROUP, wh, pp[i:newi])]
-
- del pp[i:newi]
- length = length - (newi-i)
-
- t = ingroupch[:]
- t.append(chunk(PLAIN, wh, ' ' + what))
-
- pp.insert(i, chunk(GROUP, wh, t))
- i, length = i+1, length+1
-
- pp.insert(i, chunk(CSLINE, wh, 'cindex'))
- i, length = i+1, length+1
-
- t = ingroupch[:]
- t.insert(0, chunk(PLAIN, wh, what + ', '))
-
- pp.insert(i, chunk(GROUP, wh, t))
- i, length = i+1, length+1
-
- elif s_buf_data == 'indexii':
- #\indexii{A}{B} --->
- # @cindex A B
- # @cindex B, A
- length, newi = getnextarg(length, buf, pp, i)
- cp11 = pp[i:newi]
- cp21 = crcopy(pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
- length, newi = getnextarg(length, buf, pp, i)
- cp12 = pp[i:newi]
- cp22 = crcopy(pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
-
- ch.chtype = chunk_type[CSLINE]
- ch.data = 'cindex'
- pp.insert(i, chunk(GROUP, ch.where, cp11 + [
- chunk(PLAIN, ch.where, ' ')] + cp12))
- i, length = i+1, length+1
- pp[i:i] = [chunk(CSLINE, ch.where, 'cindex'),
- chunk(GROUP, ch.where, cp22 + [
- chunk(PLAIN, ch.where, ', ')]+ cp21)]
- i, length = i+2, length+2
-
- elif s_buf_data == 'indexiii':
- length, newi = getnextarg(length, buf, pp, i)
- cp11 = pp[i:newi]
- cp21 = crcopy(pp[i:newi])
- cp31 = crcopy(pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
- length, newi = getnextarg(length, buf, pp, i)
- cp12 = pp[i:newi]
- cp22 = crcopy(pp[i:newi])
- cp32 = crcopy(pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
- length, newi = getnextarg(length, buf, pp, i)
- cp13 = pp[i:newi]
- cp23 = crcopy(pp[i:newi])
- cp33 = crcopy(pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
-
- ch.chtype = chunk_type[CSLINE]
- ch.data = 'cindex'
- pp.insert(i, chunk(GROUP, ch.where, cp11 + [
- chunk(PLAIN, ch.where, ' ')] + cp12
- + [chunk(PLAIN, ch.where, ' ')]
- + cp13))
- i, length = i+1, length+1
- pp[i:i] = [chunk(CSLINE, ch.where, 'cindex'),
- chunk(GROUP, ch.where, cp22 + [
- chunk(PLAIN, ch.where, ' ')]+ cp23
- + [chunk(PLAIN, ch.where, ', ')] +
- cp21)]
- i, length = i+2, length+2
- pp[i:i] = [chunk(CSLINE, ch.where, 'cindex'),
- chunk(GROUP, ch.where, cp33 + [
- chunk(PLAIN, ch.where, ', ')]+ cp31
- + [chunk(PLAIN, ch.where, ' ')] +
- cp32)]
- i, length = i+2, length+2
-
- elif s_buf_data == 'indexiv':
- length, newi = getnextarg(length, buf, pp, i)
- cp11 = pp[i:newi]
- cp21 = crcopy(pp[i:newi])
- cp31 = crcopy(pp[i:newi])
- cp41 = crcopy(pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
- length, newi = getnextarg(length, buf, pp, i)
- cp12 = pp[i:newi]
- cp22 = crcopy(pp[i:newi])
- cp32 = crcopy(pp[i:newi])
- cp42 = crcopy(pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
- length, newi = getnextarg(length, buf, pp, i)
- cp13 = pp[i:newi]
- cp23 = crcopy(pp[i:newi])
- cp33 = crcopy(pp[i:newi])
- cp43 = crcopy(pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
- length, newi = getnextarg(length, buf, pp, i)
- cp14 = pp[i:newi]
- cp24 = crcopy(pp[i:newi])
- cp34 = crcopy(pp[i:newi])
- cp44 = crcopy(pp[i:newi])
- del pp[i:newi]
- length = length - (newi-i)
-
- ch.chtype = chunk_type[CSLINE]
- ch.data = 'cindex'
- ingroupch = cp11 + \
- spacech + cp12 + \
- spacech + cp13 + \
- spacech + cp14
- pp.insert(i, chunk(GROUP, ch.where, ingroupch))
- i, length = i+1, length+1
- ingroupch = cp22 + \
- spacech + cp23 + \
- spacech + cp24 + \
- commach + cp21
- pp[i:i] = cindexch + [
- chunk(GROUP, ch.where, ingroupch)]
- i, length = i+2, length+2
- ingroupch = cp33 + \
- spacech + cp34 + \
- commach + cp31 + \
- spacech + cp32
- pp[i:i] = cindexch + [
- chunk(GROUP, ch.where, ingroupch)]
- i, length = i+2, length+2
- ingroupch = cp44 + \
- commach + cp41 + \
- spacech + cp42 + \
- spacech + cp43
- pp[i:i] = cindexch + [
- chunk(GROUP, ch.where, ingroupch)]
- i, length = i+2, length+2
-
- elif s_buf_data == 'seemodule':
- # discard optional arg first:
- length, newi = getoptarg(length, buf, pp, i)
- ingroupch = pp[i:newi]
- del pp[i:newi]
- length = length - (newi-i)
- #
- ch.data = "code"
- data = pp[i+1].data
- data.insert(0, chunk(PLAIN, ch.where, " ("))
- data.append(chunk(PLAIN, ch.where, ")"))
- pp[i+1:i+2] = data
- length = length + len(data) - 1
-
- elif s_buf_data == 'seetext':
- data = pp[i].data
- data.insert(0, chunk(ENDLINE, ch.where, "\n"))
- pp[i-1:i+1] = data
- i = i - 1
- length = length + len(data) - 2
-
- elif s_buf_data == 'deprecated':
- length, newi = getnextarg(length, buf, pp, i)
- version = pp[i:newi][0]
- length, newi2 = getnextarg(length, buf, pp, newi)
- action = pp[newi:newi2]
- del pp[i-1:newi2]
- length = length - (newi2 - i) - 1
- stuff = [chunk(PLAIN, ch.where, 'Deprecated since release '),
- version,
- chunk(PLAIN, ch.where, '.')]
- chunks = [chunk(CSNAME, ch.where, 'strong'),
- chunk(GROUP, ch.where, stuff),
- chunk(PLAIN, ch.where, ' ')] + action \
- + [chunk(DENDLINE, ch.where, '\n')]
- stuff = None
- i = i - 1
- pp[i:i] = chunks
- length = length + len(chunks)
-
- elif s_buf_data == "quad":
- ch.chtype = PLAIN
- ch.data = " "
-
- elif s_buf_data in ('usepackage', 'input'):
- del pp[i-1:i+1]
- i, length = i-1, length-2
-
- elif s_buf_data in ('noindent', 'indexsubitem', 'footnote'):
- pass
-
- elif s_buf_data == 'label':
- name = s(buf, pp[i].data[0].data)
- del pp[i-1:i+1]
- length = length - 2
- i = i - 1
- label_nodes[name] = hist.nodenames[-1]
-
- elif s_buf_data == 'rfc':
- ch.chtype = chunk_type[PLAIN]
- ch.data = "RFC " + s(buf, pp[i].data[0].data)
- del pp[i]
- length = length - 1
-
- elif s_buf_data == 'ref':
- name = s(buf, pp[i].data[0].data)
- if label_nodes.has_key(name):
- pp[i].data[0].data = label_nodes[name]
- else:
- pp[i-1:i+1] = [
- chunk(PLAIN, ch.where,
- "(unknown node reference: %s)" % name)]
- length = length - 1
- print "WARNING: unknown node label", `name`
-
- else:
- print "don't know what to do with keyword " + s_buf_data
-
-
-re_atsign = regex.compile('[@{}]')
-re_newline = regex.compile('\n')
-
-def dumpit(buf, wm, pp):
-
- global out
-
- i, length = 0, len(pp)
-
- addspace = 0
-
- while 1:
- if len(pp) != length:
- raise 'FATAL', 'inconsistent length'
- if i == length:
- break
- ch = pp[i]
- i = i + 1
-
- dospace = addspace
- addspace = 0
-
- if ch.chtype == chunk_type[CSNAME]:
- s_buf_data = s(buf, ch.data)
-## if s_buf_data == 'e':
-## wm('\\')
-## continue
-## if s_buf_data == '$':
-## wm('$')
-## continue
- wm('@' + s_buf_data)
- if s_buf_data == 'node' and \
- pp[i].chtype == chunk_type[PLAIN] and \
- s(buf, pp[i].data) in out.doublenodes:
- ##XXX doesnt work yet??
- wm(' ZZZ-' + zfill(`i`, 4))
- if s_buf_data[0] in string.letters:
- addspace = 1
- elif ch.chtype == chunk_type[PLAIN]:
- if dospace and s(buf, ch.data) not in (' ', '\t'):
- wm(' ')
- text = s(buf, ch.data)
- while 1:
- pos = re_atsign.search(text)
- if pos < 0:
- break
- wm(text[:pos] + '@' + text[pos])
- text = text[pos+1:]
- wm(text)
- elif ch.chtype == chunk_type[GROUP]:
- wm('{')
- dumpit(buf, wm, ch.data)
- wm('}')
- elif ch.chtype == chunk_type[DENDLINE]:
- wm('\n\n')
- while i != length and pp[i].chtype in \
- (chunk_type[DENDLINE], chunk_type[ENDLINE]):
- i = i + 1
- elif ch.chtype == chunk_type[OTHER]:
- wm(s(buf, ch.data))
- elif ch.chtype == chunk_type[ACTIVE]:
- wm(s(buf, ch.data))
- elif ch.chtype == chunk_type[ENDLINE]:
- wm('\n')
- elif ch.chtype == chunk_type[CSLINE]:
- if i >= 2 and pp[i-2].chtype not in \
- (chunk_type[ENDLINE], chunk_type[DENDLINE]) \
- and (pp[i-2].chtype != chunk_type[PLAIN]
- or s(buf, pp[i-2].data)[-1] != '\n'):
-
- wm('\n')
- wm('@' + s(buf, ch.data))
- if i == length:
- raise error, 'CSLINE expected another chunk'
- if pp[i].chtype != chunk_type[GROUP]:
- raise error, 'CSLINE expected GROUP'
- if type(pp[i].data) != ListType:
- raise error, 'GROUP chould contain []-data'
-
- wobj = Wobj()
- dumpit(buf, wobj.write, pp[i].data)
- i = i + 1
- text = wobj.data
- del wobj
- if text:
- wm(' ')
- while 1:
- pos = re_newline.search(text)
- if pos < 0:
- break
- # these seem to be completely harmless, so don't warn:
-## print 'WARNING: found newline in csline arg (%s)' \
-## % s(buf, ch.data)
- wm(text[:pos] + ' ')
- text = text[pos+1:]
- wm(text)
- if i >= length or \
- pp[i].chtype not in (chunk_type[CSLINE],
- chunk_type[ENDLINE], chunk_type[DENDLINE]) \
- and (pp[i].chtype != chunk_type[PLAIN]
- or s(buf, pp[i].data)[0] != '\n'):
- wm('\n')
-
- elif ch.chtype == chunk_type[COMMENT]:
- if s(buf, ch.data) and \
- regex.match('^[ \t]*$', s(buf, ch.data)) < 0:
- if i >= 2 \
- and pp[i-2].chtype not in (chunk_type[ENDLINE],
- chunk_type[DENDLINE]) \
- and not (pp[i-2].chtype == chunk_type[PLAIN]
- and regex.match('\\(.\\|\n\\)*[ \t]*\n$', s(buf, pp[i-2].data)) >= 0):
- wm('\n')
- wm('@c ' + s(buf, ch.data))
- elif ch.chtype == chunk_type[IGNORE]:
- pass
- else:
- try:
- str = `s(buf, ch.data)`
- except TypeError:
- str = `ch.data`
- if len(str) > 400:
- str = str[:400] + '...'
- print 'warning:', ch.chtype, 'not handled, data ' + str
-
-
-
-def main():
- global release_version
- outfile = None
- headerfile = 'texipre.dat'
- trailerfile = 'texipost.dat'
-
- try:
- opts, args = getopt.getopt(sys.argv[1:], 'o:h:t:v:')
- except getopt.error:
- args = []
-
- if not args:
- print 'usage: partparse [-o outfile] [-h headerfile]',
- print '[-t trailerfile] file ...'
- sys.exit(2)
-
- for opt, arg in opts:
- if opt == '-o': outfile = arg
- if opt == '-h': headerfile = arg
- if opt == '-t': trailerfile = arg
- if opt == '-v': release_version = arg
-
- if not outfile:
- root, ext = os.path.splitext(args[0])
- outfile = root + '.texi'
-
- if outfile in args:
- print 'will not overwrite input file', outfile
- sys.exit(2)
-
- outf = open(outfile, 'w')
- outf.write(open(headerfile, 'r').read())
-
- for file in args:
- if len(args) > 1: print '='*20, file, '='*20
- buf = open(file, 'r').read()
- chunk.buf = buf
- w, pp = parseit(buf)
- startchange()
- changeit(buf, pp)
- dumpit(buf, outf.write, pp)
-
- outf.write(open(trailerfile, 'r').read())
-
- outf.close()
-
-if __name__ == "__main__":
- main()