diff options
author | Fred Drake <fdrake@acm.org> | 1999-01-08 15:40:16 (GMT) |
---|---|---|
committer | Fred Drake <fdrake@acm.org> | 1999-01-08 15:40:16 (GMT) |
commit | dd21c3216925f855ff182799b6fbf1137a414c7a (patch) | |
tree | da381c284729a8801a3b33156a6a3491a09f4180 | |
parent | f2b5374ba1a2b3f3ad941822b35342e275cdc818 (diff) | |
download | cpython-dd21c3216925f855ff182799b6fbf1137a414c7a.zip cpython-dd21c3216925f855ff182799b6fbf1137a414c7a.tar.gz cpython-dd21c3216925f855ff182799b6fbf1137a414c7a.tar.bz2 |
It's finally gone!
-rw-r--r-- | Doc/tools/partparse.py | 2406 |
1 files changed, 0 insertions, 2406 deletions
diff --git a/Doc/tools/partparse.py b/Doc/tools/partparse.py deleted file mode 100644 index 57ba6ef..0000000 --- a/Doc/tools/partparse.py +++ /dev/null @@ -1,2406 +0,0 @@ -# -# partparse.py: parse a by-Guido-written-and-by-Jan-Hein-edited LaTeX file, -# and generate texinfo source. -# -# This is *not* a good example of good programming practices. In fact, this -# file could use a complete rewrite, in order to become faster, more -# easily extensible and maintainable. -# -# However, I added some comments on a few places for the pityful person who -# would ever need to take a look into this file. -# -# Have I been clear enough?? -# -# -jh -# -# Yup. I made some performance improvements and hope this lasts a while; -# I don't want to be the schmuck who ends up re-writting it! -# -# -fld -# -# (sometime later...) -# -# Ok, I've re-worked substantial chunks of this. It's only getting worse. -# It just might be gone before the next source release. (Yeah!) -# -# -fld - -import sys, string, regex, getopt, os - -from types import IntType, ListType, StringType, TupleType - -release_version = string.split(sys.version)[0] - -# Different parse modes for phase 1 -MODE_REGULAR = 0 -MODE_VERBATIM = 1 -MODE_CS_SCAN = 2 -MODE_COMMENT = 3 -MODE_MATH = 4 -MODE_DMATH = 5 -MODE_GOBBLEWHITE = 6 - -the_modes = (MODE_REGULAR, MODE_VERBATIM, MODE_CS_SCAN, MODE_COMMENT, - MODE_MATH, MODE_DMATH, MODE_GOBBLEWHITE) - -# Show the neighbourhood of the scanned buffer -def epsilon(buf, where): - wmt, wpt = where - 10, where + 10 - if wmt < 0: - wmt = 0 - if wpt > len(buf): - wpt = len(buf) - return ' Context ' + `buf[wmt:where]` + '.' + `buf[where:wpt]` + '.' - -# Should return the line number. never worked -def lin(): - global lineno - return ' Line ' + `lineno` + '.' - -# Displays the recursion level. -def lv(lvl): - return ' Level ' + `lvl` + '.' - -# Combine the three previous functions. Used often. -def lle(lvl, buf, where): - return lv(lvl) + lin() + epsilon(buf, where) - - -# This class is only needed for _symbolic_ representation of the parse mode. -class Mode: - def __init__(self, arg): - if arg not in the_modes: - raise ValueError, 'mode not in the_modes' - self.mode = arg - - def __cmp__(self, other): - if type(self) != type(other): - other = mode[other] - return cmp(self.mode, other.mode) - - def __repr__(self): - if self.mode == MODE_REGULAR: - return 'MODE_REGULAR' - elif self.mode == MODE_VERBATIM: - return 'MODE_VERBATIM' - elif self.mode == MODE_CS_SCAN: - return 'MODE_CS_SCAN' - elif self.mode == MODE_COMMENT: - return 'MODE_COMMENT' - elif self.mode == MODE_MATH: - return 'MODE_MATH' - elif self.mode == MODE_DMATH: - return 'MODE_DMATH' - elif self.mode == MODE_GOBBLEWHITE: - return 'MODE_GOBBLEWHITE' - else: - raise ValueError, 'mode not in the_modes' - -# just a wrapper around a class initialisation -mode = {} -for t in the_modes: - mode[t] = Mode(t) - - -# After phase 1, the text consists of chunks, with a certain type -# this type will be assigned to the chtype member of the chunk -# the where-field contains the file position where this is found -# and the data field contains (1): a tuple describing start- end end -# positions of the substring (can be used as slice for the buf-variable), -# (2) just a string, mostly generated by the changeit routine, -# or (3) a list, describing a (recursive) subgroup of chunks -PLAIN = 0 # ASSUME PLAINTEXT, data = the text -GROUP = 1 # GROUP ({}), data = [chunk, chunk,..] -CSNAME = 2 # CONTROL SEQ TOKEN, data = the command -COMMENT = 3 # data is the actual comment -DMATH = 4 # DISPLAYMATH, data = [chunk, chunk,..] -MATH = 5 # MATH, see DISPLAYMATH -OTHER = 6 # CHAR WITH CATCODE OTHER, data = char -ACTIVE = 7 # ACTIVE CHAR -GOBBLEDWHITE = 8 # Gobbled LWSP, after CSNAME -ENDLINE = 9 # END-OF-LINE, data = '\n' -DENDLINE = 10 # DOUBLE EOL, data='\n', indicates \par -ENV = 11 # LaTeX-environment - # data =(envname,[ch,ch,ch,.]) -CSLINE = 12 # for texi: next chunk will be one group - # of args. Will be set all on 1 line -IGNORE = 13 # IGNORE this data -ENDENV = 14 # TEMP END OF GROUP INDICATOR -IF = 15 # IF-directive - # data = (flag,negate,[ch, ch, ch,...]) - -the_types = (PLAIN, GROUP, CSNAME, COMMENT, DMATH, MATH, OTHER, ACTIVE, - GOBBLEDWHITE, ENDLINE, DENDLINE, ENV, CSLINE, IGNORE, ENDENV, IF) - -# class, just to display symbolic name -class ChunkType: - def __init__(self, chunk_type): - if chunk_type not in the_types: - raise ValueError, 'chunk_type not in the_types' - self.chunk_type = chunk_type - - def __cmp__(self, other): - if type(self) != type(other): - other = chunk_type[other] - return cmp(self.chunk_type, other.chunk_type) - - def __repr__(self): - if self.chunk_type == PLAIN: - return 'PLAIN' - elif self.chunk_type == GROUP: - return 'GROUP' - elif self.chunk_type == CSNAME: - return 'CSNAME' - elif self.chunk_type == COMMENT: - return 'COMMENT' - elif self.chunk_type == DMATH: - return 'DMATH' - elif self.chunk_type == MATH: - return 'MATH' - elif self.chunk_type == OTHER: - return 'OTHER' - elif self.chunk_type == ACTIVE: - return 'ACTIVE' - elif self.chunk_type == GOBBLEDWHITE: - return 'GOBBLEDWHITE' - elif self.chunk_type == DENDLINE: - return 'DENDLINE' - elif self.chunk_type == ENDLINE: - return 'ENDLINE' - elif self.chunk_type == ENV: - return 'ENV' - elif self.chunk_type == CSLINE: - return 'CSLINE' - elif self.chunk_type == IGNORE: - return 'IGNORE' - elif self.chunk_type == ENDENV: - return 'ENDENV' - elif self.chunk_type == IF: - return 'IF' - else: - raise ValueError, 'chunk_type not in the_types' - -# ...and the wrapper -chunk_type = {} -for t in the_types: - chunk_type[t] = ChunkType(t) - -# store a type object of the ChunkType-class-instance... -chunk_type_type = type(chunk_type[PLAIN]) - -# this class contains a part of the parsed buffer -class Chunk: - def __init__(self, chtype, where, data): - if type(chtype) != chunk_type_type: - chtype = chunk_type[chtype] - self.chtype = chtype - self.where = where - self.data = data - - __datatypes = [chunk_type[CSNAME], chunk_type[PLAIN], chunk_type[CSLINE]] - - def __repr__(self): - if self.chtype in self.__datatypes: - data = s(self.buf, self.data) - else: - data = self.data - return 'chunk' + `self.chtype, self.where, data` - -# and the wrapper -chunk = Chunk - - -error = 'partparse.error' - -# -# TeX's catcodes... -# -CC_ESCAPE = 0 -CC_LBRACE = 1 -CC_RBRACE = 2 -CC_MATHSHIFT = 3 -CC_ALIGNMENT = 4 -CC_ENDLINE = 5 -CC_PARAMETER = 6 -CC_SUPERSCRIPT = 7 -CC_SUBSCRIPT = 8 -CC_IGNORE = 9 -CC_WHITE = 10 -CC_LETTER = 11 -CC_OTHER = 12 -CC_ACTIVE = 13 -CC_COMMENT = 14 -CC_INVALID = 15 - -# and the names -cc_names = [ - 'CC_ESCAPE', - 'CC_LBRACE', - 'CC_RBRACE', - 'CC_MATHSHIFT', - 'CC_ALIGNMENT', - 'CC_ENDLINE', - 'CC_PARAMETER', - 'CC_SUPERSCRIPT', - 'CC_SUBSCRIPT', - 'CC_IGNORE', - 'CC_WHITE', - 'CC_LETTER', - 'CC_OTHER', - 'CC_ACTIVE', - 'CC_COMMENT', - 'CC_INVALID', - ] - -# Show a list of catcode-name-symbols -def pcl(codelist): - result = '' - for i in codelist: - result = result + cc_names[i] + ', ' - return '[' + result[:-2] + ']' - -# the name of the catcode (ACTIVE, OTHER, etc.) -def pc(code): - return cc_names[code] - - -# Which catcodes make the parser stop parsing regular plaintext -regular_stopcodes = [CC_ESCAPE, CC_LBRACE, CC_RBRACE, CC_MATHSHIFT, - CC_ALIGNMENT, CC_PARAMETER, CC_SUPERSCRIPT, CC_SUBSCRIPT, - CC_IGNORE, CC_ACTIVE, CC_COMMENT, CC_INVALID, CC_ENDLINE] - -# same for scanning a control sequence name -csname_scancodes = [CC_LETTER] - -# same for gobbling LWSP -white_scancodes = [CC_WHITE] -##white_scancodes = [CC_WHITE, CC_ENDLINE] - -# make a list of all catcode id's, except for catcode ``other'' -all_but_other_codes = range(16) -del all_but_other_codes[CC_OTHER] -##print all_but_other_codes - -# when does a comment end -comment_stopcodes = [CC_ENDLINE] - -# gather all characters together, specified by a list of catcodes -def code2string(cc, codelist): - ##print 'code2string: codelist = ' + pcl(codelist), - result = '' - for category in codelist: - if cc[category]: - result = result + cc[category] - ##print 'result = ' + `result` - return result - -# automatically generate all characters of catcode other, being the -# complement set in the ASCII range (128 characters) -def make_other_codes(cc): - otherchars = range(256) # could be made 256, no problem - for category in all_but_other_codes: - if cc[category]: - for c in cc[category]: - otherchars[ord(c)] = None - result = '' - for i in otherchars: - if i != None: - result = result + chr(i) - return result - -# catcode dump (which characters have which catcodes). -def dump_cc(name, cc): - ##print '\t' + name - ##print '=' * (8+len(name)) - if len(cc) != 16: - raise TypeError, 'cc not good cat class' -## for i in range(16): -## print pc(i) + '\t' + `cc[i]` - - -# In the beginning,.... -epoch_cc = [None] * 16 -##dump_cc('epoch_cc', epoch_cc) - - -# INITEX -initex_cc = epoch_cc[:] -initex_cc[CC_ESCAPE] = '\\' -initex_cc[CC_ENDLINE], initex_cc[CC_IGNORE], initex_cc[CC_WHITE] = \ - '\n', '\0', ' ' -initex_cc[CC_LETTER] = string.uppercase + string.lowercase -initex_cc[CC_COMMENT], initex_cc[CC_INVALID] = '%', '\x7F' -#initex_cc[CC_OTHER] = make_other_codes(initex_cc) I don't need them, anyway -##dump_cc('initex_cc', initex_cc) - - -# LPLAIN: LaTeX catcode setting (see lplain.tex) -lplain_cc = initex_cc[:] -lplain_cc[CC_LBRACE], lplain_cc[CC_RBRACE] = '{', '}' -lplain_cc[CC_MATHSHIFT] = '$' -lplain_cc[CC_ALIGNMENT] = '&' -lplain_cc[CC_PARAMETER] = '#' -lplain_cc[CC_SUPERSCRIPT] = '^\x0B' # '^' and C-k -lplain_cc[CC_SUBSCRIPT] = '_\x01' # '_' and C-a -lplain_cc[CC_WHITE] = lplain_cc[CC_WHITE] + '\t' -lplain_cc[CC_ACTIVE] = '~\x0C' # '~' and C-l -lplain_cc[CC_OTHER] = make_other_codes(lplain_cc) -##dump_cc('lplain_cc', lplain_cc) - - -# Guido's LaTeX environment catcoded '_' as ``other'' -# my own purpose catlist -my_cc = lplain_cc[:] -my_cc[CC_SUBSCRIPT] = my_cc[CC_SUBSCRIPT][1:] # remove '_' here -my_cc[CC_OTHER] = my_cc[CC_OTHER] + '_' # add it to OTHER list -dump_cc('my_cc', my_cc) - - - -# needed for un_re, my equivalent for regexp-quote in Emacs -re_meaning = '\\[]^$' - -def un_re(str): - result = '' - for i in str: - if i in re_meaning: - result = result + '\\' - result = result + i - return result - -# NOTE the negate ('^') operator in *some* of the regexps below -def make_rc_regular(cc): - # problems here if '[]' are included!! - return regex.compile('[' + code2string(cc, regular_stopcodes) + ']') - -def make_rc_cs_scan(cc): - return regex.compile('[^' + code2string(cc, csname_scancodes) + ']') - -def make_rc_comment(cc): - return regex.compile('[' + code2string(cc, comment_stopcodes) + ']') - -def make_rc_endwhite(cc): - return regex.compile('[^' + code2string(cc, white_scancodes) + ']') - - - -# regular: normal mode: -rc_regular = make_rc_regular(my_cc) - -# scan: scan a command sequence e.g. `newlength' or `mbox' or `;', `,' or `$' -rc_cs_scan = make_rc_cs_scan(my_cc) -rc_comment = make_rc_comment(my_cc) -rc_endwhite = make_rc_endwhite(my_cc) - - -# parseit (BUF, PARSEMODE=mode[MODE_REGULAR], START=0, RECURSION-LEVEL=0) -# RECURSION-LEVEL will is incremented on entry. -# result contains the list of chunks returned -# together with this list, the buffer position is returned - -# RECURSION-LEVEL will be set to zero *again*, when recursively a -# {,D}MATH-mode scan has been enetered. -# This has been done in order to better check for environment-mismatches - -def parseit(buf, parsemode=mode[MODE_REGULAR], start=0, lvl=0): - global lineno - - result = [] - end = len(buf) - if lvl == 0 and parsemode == mode[MODE_REGULAR]: - lineno = 1 - lvl = lvl + 1 - - ##print 'parseit(' + epsilon(buf, start) + ', ' + `parsemode` + ', ' + `start` + ', ' + `lvl` + ')' - - # - # some of the more regular modes... - # - - if parsemode in (mode[MODE_REGULAR], mode[MODE_DMATH], mode[MODE_MATH]): - cstate = [] - newpos = start - curpmode = parsemode - while 1: - where = newpos - #print '\tnew round: ' + epsilon(buf, where) - if where == end: - if lvl > 1 or curpmode != mode[MODE_REGULAR]: - # not the way we started... - raise EOFError, 'premature end of file.' + lle(lvl, buf, where) - # the real ending of lvl-1 parse - return end, result - - pos = rc_regular.search(buf, where) - - if pos < 0: - pos = end - - if pos != where: - newpos, c = pos, chunk(PLAIN, where, (where, pos)) - result.append(c) - continue - - - # - # ok, pos == where and pos != end - # - foundchar = buf[where] - if foundchar in my_cc[CC_LBRACE]: - # recursive subgroup parse... - newpos, data = parseit(buf, curpmode, where+1, lvl) - result.append(chunk(GROUP, where, data)) - - elif foundchar in my_cc[CC_RBRACE]: - if lvl <= 1: - raise error, 'ENDGROUP while in base level.' + lle(lvl, buf, where) - if lvl == 1 and mode != mode[MODE_REGULAR]: - raise error, 'endgroup while in math mode. +lin() + epsilon(buf, where)' - return where + 1, result - - elif foundchar in my_cc[CC_ESCAPE]: - # - # call the routine that actually deals with - # this problem. If do_ret is None, than - # return the value of do_ret - # - # Note that handle_cs might call this routine - # recursively again... - # - do_ret, newpos = handlecs(buf, where, - curpmode, lvl, result, end) - if do_ret != None: - return do_ret - - elif foundchar in my_cc[CC_COMMENT]: - newpos, data = parseit(buf, - mode[MODE_COMMENT], where+1, lvl) - result.append(chunk(COMMENT, where, data)) - - elif foundchar in my_cc[CC_MATHSHIFT]: - # note that recursive calls to math-mode - # scanning are called with recursion-level 0 - # again, in order to check for bad mathend - # - if where + 1 != end and buf[where + 1] in my_cc[CC_MATHSHIFT]: - # - # double mathshift, e.g. '$$' - # - if curpmode == mode[MODE_REGULAR]: - newpos, data = parseit(buf, mode[MODE_DMATH], - where + 2, 0) - result.append(chunk(DMATH, where, data)) - elif curpmode == mode[MODE_MATH]: - raise error, 'wrong math delimiiter' + lin() + epsilon(buf, where) - elif lvl != 1: - raise error, 'bad mathend.' + lle(lvl, buf, where) - else: - return where + 2, result - else: - # - # single math shift, e.g. '$' - # - if curpmode == mode[MODE_REGULAR]: - newpos, data = parseit(buf, mode[MODE_MATH], - where + 1, 0) - result.append(chunk(MATH, where, data)) - elif curpmode == mode[MODE_DMATH]: - raise error, 'wrong math delimiiter' + lin() + epsilon(buf, where) - elif lvl != 1: - raise error, 'bad mathend.' + lv(lvl, buf, where) - else: - return where + 1, result - - elif foundchar in my_cc[CC_IGNORE]: - print 'warning: ignored char', `foundchar` - newpos = where + 1 - - elif foundchar in my_cc[CC_ACTIVE]: - result.append(chunk(ACTIVE, where, foundchar)) - newpos = where + 1 - - elif foundchar in my_cc[CC_INVALID]: - raise error, 'invalid char ' + `foundchar` - newpos = where + 1 - - elif foundchar in my_cc[CC_ENDLINE]: - # - # after an end of line, eat the rest of - # whitespace on the beginning of the next line - # this is what LaTeX more or less does - # - # also, try to indicate double newlines (\par) - # - lineno = lineno + 1 - savedwhere = where - newpos, dummy = parseit(buf, mode[MODE_GOBBLEWHITE], where + 1, lvl) - if newpos != end and buf[newpos] in my_cc[CC_ENDLINE]: - result.append(chunk(DENDLINE, savedwhere, foundchar)) - else: - result.append(chunk(ENDLINE, savedwhere, foundchar)) - else: - result.append(chunk(OTHER, where, foundchar)) - newpos = where + 1 - - elif parsemode == mode[MODE_CS_SCAN]: - # - # scan for a control sequence token. `\ape', `\nut' or `\%' - # - if start == end: - raise EOFError, 'can\'t find end of csname' - pos = rc_cs_scan.search(buf, start) - if pos < 0: - pos = end - if pos == start: - # first non-letter right where we started the search - # ---> the control sequence name consists of one single - # character. Also: don't eat white space... - if buf[pos] in my_cc[CC_ENDLINE]: - lineno = lineno + 1 - pos = pos + 1 - return pos, (start, pos) - else: - spos = pos - if buf[pos] == '\n': - lineno = lineno + 1 - spos = pos + 1 - pos2, dummy = parseit(buf, mode[MODE_GOBBLEWHITE], spos, lvl) - return pos2, (start, pos) - - elif parsemode == mode[MODE_GOBBLEWHITE]: - if start == end: - return start, '' - pos = rc_endwhite.search(buf, start) - if pos < 0: - pos = start - return pos, (start, pos) - - elif parsemode == mode[MODE_COMMENT]: - pos = rc_comment.search(buf, start) - lineno = lineno + 1 - if pos < 0: - print 'no newline perhaps?' - raise EOFError, 'can\'t find end of comment' - pos = pos + 1 - pos2, dummy = parseit(buf, mode[MODE_GOBBLEWHITE], pos, lvl) - return pos2, (start, pos) - - else: - raise error, 'Unknown mode (' + `parsemode` + ')' - - -#moreresult = cswitch(buf[x1:x2], buf, newpos, parsemode, lvl) - -#boxcommands = 'mbox', 'fbox' -#defcommands = 'def', 'newcommand' - -endverbstr = '\\end{verbatim}' - -re_endverb = regex.compile(un_re(endverbstr)) - -# -# handlecs: helper function for parseit, for the special thing we might -# wanna do after certain command control sequences -# returns: None or return_data, newpos -# -# in the latter case, the calling function is instructed to immediately -# return with the data in return_data -# -def handlecs(buf, where, curpmode, lvl, result, end): - global lineno - - # get the control sequence name... - newpos, data = parseit(buf, mode[MODE_CS_SCAN], where+1, lvl) - saveddata = data - s_buf_data = s(buf, data) - - if s_buf_data in ('begin', 'end'): - # skip the expected '{' and get the LaTeX-envname '}' - newpos, data = parseit(buf, mode[MODE_REGULAR], newpos+1, lvl) - if len(data) != 1: - raise error, 'expected 1 chunk of data.' + lle(lvl, buf, where) - - # yucky, we've got an environment - envname = s(buf, data[0].data) - s_buf_saveddata = s(buf, saveddata) - ##print 'FOUND ' + s(buf, saveddata) + '. Name ' + `envname` + '.' + lv(lvl) - if s_buf_saveddata == 'begin' and envname == 'verbatim': - # verbatim deserves special treatment - pos = re_endverb.search(buf, newpos) - if pos < 0: - raise error, "%s not found.%s" \ - % (`endverbstr`, lle(lvl, buf, where)) - result.append(chunk(ENV, where, (envname, [chunk(PLAIN, newpos, (newpos, pos))]))) - newpos = pos + len(endverbstr) - - elif s_buf_saveddata == 'begin': - # start parsing recursively... If that parse returns - # from an '\end{...}', then should the last item of - # the returned data be a string containing the ended - # environment - newpos, data = parseit(buf, curpmode, newpos, lvl) - if not data or type(data[-1]) is not StringType: - raise error, "missing 'end'" + lle(lvl, buf, where) \ - + epsilon(buf, newpos) - retenv = data[-1] - del data[-1] - if retenv != envname: - #[`retenv`, `envname`] - raise error, 'environments do not match.%s%s' \ - % (lle(lvl, buf, where), epsilon(buf, newpos)) - result.append(chunk(ENV, where, (retenv, data))) - else: - # 'end'... append the environment name, as just - # pointed out, and order parsit to return... - result.append(envname) - ##print 'POINT of return: ' + epsilon(buf, newpos) - # the tuple will be returned by parseit - return (newpos, result), newpos - - # end of \begin ... \end handling - - elif s_buf_data[0:2] == 'if': - # another scary monster: the 'if' directive - flag = s_buf_data[2:] - - # recursively call parseit, just like environment above.. - # the last item of data should contain the if-termination - # e.g., 'else' of 'fi' - newpos, data = parseit(buf, curpmode, newpos, lvl) - if not data or data[-1] not in ('else', 'fi'): - raise error, 'wrong if... termination' + \ - lle(lvl, buf, where) + epsilon(buf, newpos) - - ifterm = data[-1] - del data[-1] - # 0 means dont_negate flag - result.append(chunk(IF, where, (flag, 0, data))) - if ifterm == 'else': - # do the whole thing again, there is only one way - # to end this one, by 'fi' - newpos, data = parseit(buf, curpmode, newpos, lvl) - if not data or data[-1] not in ('fi', ): - raise error, 'wrong if...else... termination' \ - + lle(lvl, buf, where) \ - + epsilon(buf, newpos) - - ifterm = data[-1] - del data[-1] - result.append(chunk(IF, where, (flag, 1, data))) - #done implicitely: return None, newpos - - elif s_buf_data in ('else', 'fi'): - result.append(s(buf, data)) - # order calling party to return tuple - return (newpos, result), newpos - - # end of \if, \else, ... \fi handling - - elif s(buf, saveddata) == 'verb': - x2 = saveddata[1] - result.append(chunk(CSNAME, where, data)) - if x2 == end: - raise error, 'premature end of command.' + lle(lvl, buf, where) - delimchar = buf[x2] - ##print 'VERB: delimchar ' + `delimchar` - pos = regex.compile(un_re(delimchar)).search(buf, x2 + 1) - if pos < 0: - raise error, 'end of \'verb\' argument (' + \ - `delimchar` + ') not found.' + \ - lle(lvl, buf, where) - result.append(chunk(GROUP, x2, [chunk(PLAIN, x2+1, (x2+1, pos))])) - newpos = pos + 1 - else: - result.append(chunk(CSNAME, where, data)) - return None, newpos - -# this is just a function to get the string value if the possible data-tuple -def s(buf, data): - if type(data) is StringType: - return data - if len(data) != 2 or not (type(data[0]) is type(data[1]) is IntType): - raise TypeError, 'expected tuple of 2 integers' - x1, x2 = data - return buf[x1:x2] - - -##length, data1, i = getnextarg(length, buf, pp, i + 1) - -# make a deep-copy of some chunks -def crcopy(r): - return map(chunkcopy, r) - - -# copy a chunk, would better be a method of class Chunk... -def chunkcopy(ch): - if ch.chtype == chunk_type[GROUP]: - return chunk(GROUP, ch.where, map(chunkcopy, ch.data)) - else: - return chunk(ch.chtype, ch.where, ch.data) - - -# get next argument for TeX-macro, flatten a group (insert between) -# or return Command Sequence token, or give back one character -def getnextarg(length, buf, pp, item): - - ##wobj = Wobj() - ##dumpit(buf, wobj.write, pp[item:min(length, item + 5)]) - ##print 'GETNEXTARG, (len, item) =', `length, item` + ' ---> ' + wobj.data + ' <---' - - while item < length and pp[item].chtype == chunk_type[ENDLINE]: - del pp[item] - length = length - 1 - if item >= length: - raise error, 'no next arg.' + epsilon(buf, pp[-1].where) - if pp[item].chtype == chunk_type[GROUP]: - newpp = pp[item].data - del pp[item] - length = length - 1 - changeit(buf, newpp) - length = length + len(newpp) - pp[item:item] = newpp - item = item + len(newpp) - if len(newpp) < 10: - wobj = Wobj() - dumpit(buf, wobj.write, newpp) - ##print 'GETNEXTARG: inserted ' + `wobj.data` - return length, item - elif pp[item].chtype == chunk_type[PLAIN]: - #grab one char - print 'WARNING: grabbing one char' - if len(s(buf, pp[item].data)) > 1: - pp.insert(item, chunk(PLAIN, pp[item].where, s(buf, pp[item].data)[:1])) - item, length = item+1, length+1 - pp[item].data = s(buf, pp[item].data)[1:] - else: - item = item+1 - return length, item - else: - ch = pp[item] - try: - str = `s(buf, ch.data)` - except TypeError: - str = `ch.data` - if len(str) > 400: - str = str[:400] + '...' - print 'GETNEXTARG:', ch.chtype, 'not handled, data ' + str - return length, item - - -# this one is needed to find the end of LaTeX's optional argument, like -# item[...] -re_endopt = regex.compile(']') - -# get a LaTeX-optional argument, you know, the square braces '[' and ']' -def getoptarg(length, buf, pp, item): - - wobj = Wobj() - dumpit(buf, wobj.write, pp[item:min(length, item + 5)]) - ##print 'GETOPTARG, (len, item) =', `length, item` + ' ---> ' + wobj.data + ' <---' - - if item >= length or \ - pp[item].chtype != chunk_type[PLAIN] or \ - s(buf, pp[item].data)[0] != '[': - return length, item - - pp[item].data = s(buf, pp[item].data)[1:] - if len(pp[item].data) == 0: - del pp[item] - length = length-1 - - while 1: - if item == length: - raise error, 'No end of optional arg found' - if pp[item].chtype == chunk_type[PLAIN]: - text = s(buf, pp[item].data) - pos = re_endopt.search(text) - if pos >= 0: - pp[item].data = text[:pos] - if pos == 0: - del pp[item] - length = length-1 - else: - item=item+1 - text = text[pos+1:] - - while text and text[0] in ' \t': - text = text[1:] - - if text: - pp.insert(item, chunk(PLAIN, 0, text)) - length = length + 1 - return length, item - - item = item+1 - - -# Wobj just add write-requests to the ``data'' attribute -class Wobj: - data = '' - - def write(self, data): - self.data = self.data + data - -# ignore these commands -ignoredcommands = ('hline', 'small', '/', 'tableofcontents', 'Large') -# map commands like these to themselves as plaintext -wordsselves = ('UNIX', 'ABC', 'C', 'ASCII', 'EOF', 'LaTeX', 'POSIX', 'TeX', - 'SliTeX') -# \{ --> {, \} --> }, etc -themselves = ('{', '}', ',', '.', '@', ' ', '\n') + wordsselves -# these ones also themselves (see argargs macro in myformat.sty) -inargsselves = (',', '[', ']', '(', ')') -# this is how *I* would show the difference between emph and strong -# code 1 means: fold to uppercase -markcmds = {'code': ('', ''), 'var': 1, 'emph': ('_', '_'), - 'strong': ('*', '*')} - -# recognise patter {\FONTCHANGE-CMD TEXT} to \MAPPED-FC-CMD{TEXT} -fontchanges = {'rm': 'r', 'it': 'i', 'em': 'emph', 'bf': 'b', 'tt': 't'} - - -# try to remove macros and return flat text -def flattext(buf, pp): - pp = crcopy(pp) - ##print '---> FLATTEXT ' + `pp` - wobj = Wobj() - - i, length = 0, len(pp) - while 1: - if len(pp) != length: - raise 'FATAL', 'inconsistent length' - if i >= length: - break - ch = pp[i] - i = i+1 - if ch.chtype == chunk_type[PLAIN]: - pass - elif ch.chtype == chunk_type[CSNAME]: - s_buf_data = s(buf, ch.data) - if convertible_csname(s_buf_data): - ch.chtype, ch.data, nix = conversion(s_buf_data) - if hist.inargs and s_buf_data in inargsselves: - ch.chtype = chunk_type[PLAIN] - elif len(s_buf_data) == 1 \ - and s_buf_data in onlylatexspecial: - ch.chtype = chunk_type[PLAIN] - # if it is followed by an empty group, - # remove that group, it was needed for - # a true space - if i < length \ - and pp[i].chtype==chunk_type[GROUP] \ - and len(pp[i].data) == 0: - del pp[i] - length = length-1 - - elif s_buf_data in markcmds.keys(): - length, newi = getnextarg(length, buf, pp, i) - str = flattext(buf, pp[i:newi]) - del pp[i:newi] - length = length - (newi - i) - ch.chtype = chunk_type[PLAIN] - markcmd = s_buf_data - x = markcmds[markcmd] - if type(x) == TupleType: - pre, after = x - str = pre+str+after - elif x == 1: - str = string.upper(str) - else: - raise 'FATAL', 'corrupt markcmds' - ch.data = str - else: - if s_buf_data not in ignoredcommands: - print 'WARNING: deleting command ' + s_buf_data - print 'PP' + `pp[i-1]` - del pp[i-1] - i, length = i-1, length-1 - elif ch.chtype == chunk_type[GROUP]: - length, newi = getnextarg(length, buf, pp, i-1) - i = i-1 -## str = flattext(buf, crcopy(pp[i-1:newi])) -## del pp[i:newi] -## length = length - (newi - i) -## ch.chtype = chunk_type[PLAIN] -## ch.data = str - else: - pass - - dumpit(buf, wobj.write, pp) - ##print 'FLATTEXT: RETURNING ' + `wobj.data` - return wobj.data - -# try to generate node names (a bit shorter than the chapter title) -# note that the \nodename command (see elsewhere) overules these efforts -def invent_node_names(text): - words = string.split(text) - - ##print 'WORDS ' + `words` - - if len(words) == 2 \ - and string.lower(words[0]) == 'built-in' \ - and string.lower(words[1]) not in ('modules', 'functions'): - return words[1] - if len(words) == 3 and string.lower(words[1]) == 'module': - return words[2] - if len(words) == 3 and string.lower(words[1]) == 'object': - return string.join(words[0:2]) - if len(words) > 4 \ - and (string.lower(string.join(words[-4:])) \ - == 'methods and data attributes'): - return string.join(words[:2]) - return text - -re_commas_etc = regex.compile('[,`\'@{}]') - -re_whitespace = regex.compile('[ \t]*') - - -##nodenamecmd = next_command_p(length, buf, pp, newi, 'nodename') - -# look if the next non-white stuff is also a command, resulting in skipping -# double endlines (DENDLINE) too, and thus omitting \par's -# Sometimes this is too much, maybe consider DENDLINE's as stop -def next_command_p(length, buf, pp, i, cmdname): - - while 1: - if i >= len(pp): - break - ch = pp[i] - i = i+1 - if ch.chtype == chunk_type[ENDLINE]: - continue - if ch.chtype == chunk_type[DENDLINE]: - continue - if ch.chtype == chunk_type[PLAIN]: - if re_whitespace.search(s(buf, ch.data)) == 0 and \ - re_whitespace.match(s(buf, ch.data)) == len(s(buf, ch.data)): - continue - return -1 - if ch.chtype == chunk_type[CSNAME]: - if s(buf, ch.data) == cmdname: - return i # _after_ the command - return -1 - return -1 - - -# things that are special to LaTeX, but not to texi.. -onlylatexspecial = '_~^$#&%' - -class Struct: pass - -hist = Struct() -out = Struct() - -def startchange(): - global hist, out - - hist.chaptertype = "chapter" - hist.inenv = [] - hist.nodenames = [] - hist.cindex = [] - hist.inargs = 0 - hist.enumeratenesting, hist.itemizenesting = 0, 0 - hist.this_module = None - - out.doublenodes = [] - out.doublecindeces = [] - - -spacech = [chunk(PLAIN, 0, ' ')] -commach = [chunk(PLAIN, 0, ', ')] -cindexch = [chunk(CSLINE, 0, 'cindex')] - -# the standard variation in symbols for itemize -itemizesymbols = ['bullet', 'minus', 'dots'] - -# same for enumerate -enumeratesymbols = ['1', 'A', 'a'] - -# Map of things that convert one-to-one. Each entry is a 3-tuple: -# -# new_chtype, new_data, nix_trailing_empty_group -# -d = {} -# add stuff that converts from one name to another: -for name in ('url', 'module', 'function', 'cfunction', - 'keyword', 'method', 'exception', 'constant', - 'email', 'class', 'member', 'cdata', 'ctype', - 'member', 'sectcode', 'verb', - 'cfunction', 'cdata', 'ctype', - ): - d[name] = chunk_type[CSNAME], 'code', 0 -for name in ('emph', 'var', 'strong', 'code', 'kbd', 'key', - 'dfn', 'samp', 'file', 'r', 'i', 't'): - d[name] = chunk_type[CSNAME], name, 0 -d['character'] = chunk_type[CSNAME], 'samp', 0 -d['url'] = chunk_type[CSNAME], 'code', 0 -d['email'] = chunk_type[CSNAME], 'code', 0 -d['mimetype'] = chunk_type[CSNAME], 'code', 0 -d['newsgroup'] = chunk_type[CSNAME], 'code', 0 -d['program'] = chunk_type[CSNAME], 'strong', 0 -d['\\'] = chunk_type[CSNAME], '*', 0 -# add stuff that converts to text: -for name in themselves: - d[name] = chunk_type[PLAIN], name, 0 -for name in wordsselves: - d[name] = chunk_type[PLAIN], name, 1 -for name in ',[]()': - d[name] = chunk_type[PLAIN], name, 0 -# a lot of these are LaTeX2e additions -for name, value in [('quotedblbase', ',,'), ('quotesinglbase', ','), - ('textquotedbl', '"'), ('LaTeXe', 'LaTeX2e'), - ('e', '\\'), ('textquotedblleft', "``"), - ('textquotedblright', "''"), ('textquoteleft', "`"), - ('textquoteright', "'"), ('textbackslash', '\\'), - ('textbar', '|'), ('textless', '<'), - ('textgreater', '>'), ('textasciicircum', '^'), - ('Cpp', 'C++'), ('copyright', '')]: - d[name] = chunk_type[PLAIN], value, 1 -convertible_csname = d.has_key -conversion = d.get -del d, name, value - -## -## \begin{ {func,data,exc}desc }{name}... -## the resulting texi-code is dependent on the contents of indexsubitem -## - -# indexsubitem: `['XXX', 'function'] -# funcdesc: -# deffn {`idxsi`} NAME (FUNCARGS) - -# indexsubitem: `['XXX', 'method']` -# funcdesc: -# defmethod {`idxsi[0]`} NAME (FUNCARGS) - -# indexsubitem: `['in', 'module', 'MODNAME']' -# datadesc: -# defcv data {`idxsi[1:]`} NAME -# excdesc: -# defcv exception {`idxsi[1:]`} NAME -# funcdesc: -# deffn {function of `idxsi[1:]`} NAME (FUNCARGS) - -# indexsubitem: `['OBJECT', 'attribute']' -# datadesc -# defcv attribute {`OBJECT`} NAME - - -## this routine will be called on \begin{funcdesc}{NAME}{ARGS} -## or \funcline{NAME}{ARGS} -## -def do_funcdesc(length, buf, pp, i, index=1): - startpoint = i-1 - ch = pp[startpoint] - wh = ch.where - length, newi = getnextarg(length, buf, pp, i) - funcname = chunk(GROUP, wh, pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - save = hist.inargs - hist.inargs = 1 - length, newi = getnextarg(length, buf, pp, i) - hist.inargs = save - del save - the_args = [chunk(PLAIN, wh, '()'[0])] + pp[i:newi] + \ - [chunk(PLAIN, wh, '()'[1])] - del pp[i:newi] - length = length - (newi-i) - - idxsi = hist.indexsubitem # words - command = 'deffn' - if hist.this_module: - cat_class = 'function of ' + hist.this_module - else: - cat_class = 'built-in function' - ch.chtype = chunk_type[CSLINE] - ch.data = command - - cslinearg = [chunk(GROUP, wh, [chunk(PLAIN, wh, cat_class)])] - cslinearg.append(chunk(PLAIN, wh, ' ')) - cslinearg.append(funcname) - cslinearg.append(chunk(PLAIN, wh, ' ')) - l = len(cslinearg) - cslinearg[l:l] = the_args - - pp.insert(i, chunk(GROUP, wh, cslinearg)) - i, length = i+1, length+1 - hist.command = command - return length, i - - -## this routine will be called on \begin{excdesc}{NAME} -## or \excline{NAME} -## -def do_excdesc(length, buf, pp, i): - startpoint = i-1 - ch = pp[startpoint] - wh = ch.where - length, newi = getnextarg(length, buf, pp, i) - excname = chunk(GROUP, wh, pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - - idxsi = hist.indexsubitem # words - command = '' - cat_class = '' - class_class = '' - if idxsi == ['built-in', 'exception', 'base', 'class']: - command = 'defvr' - cat_class = 'exception base class' - else: - command = 'defcv' - cat_class = 'exception' - - ch.chtype = chunk_type[CSLINE] - ch.data = command - - cslinearg = [chunk(GROUP, wh, [chunk(PLAIN, wh, cat_class)])] - cslinearg.append(chunk(PLAIN, wh, ' ')) - if class_class: - cslinearg.append(chunk(GROUP, wh, [chunk(PLAIN, wh, class_class)])) - cslinearg.append(chunk(PLAIN, wh, ' ')) - cslinearg.append(excname) - - pp.insert(i, chunk(GROUP, wh, cslinearg)) - i, length = i+1, length+1 - hist.command = command - return length, i - -## same for datadesc or dataline... -def do_datadesc(length, buf, pp, i, index=1): - startpoint = i-1 - ch = pp[startpoint] - wh = ch.where - length, newi = getnextarg(length, buf, pp, i) - dataname = chunk(GROUP, wh, pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - - idxsi = hist.indexsubitem # words - command = 'defcv' - cat_class = 'data' - class_class = '' - if idxsi[-1] in ('attribute', 'option'): - cat_class = idxsi[-1] - class_class = string.join(idxsi[:-1]) - elif len(idxsi) == 3 and idxsi[:2] == ['in', 'module']: - class_class = string.join(idxsi[1:]) - elif len(idxsi) == 4 and idxsi[:3] == ['data', 'in', 'module']: - class_class = string.join(idxsi[2:]) - else: - class_class = string.join(idxsi) - - ch.chtype = chunk_type[CSLINE] - ch.data = command - - cslinearg = [chunk(GROUP, wh, [chunk(PLAIN, wh, cat_class)])] - cslinearg.append(chunk(PLAIN, wh, ' ')) - if class_class: - cslinearg.append(chunk(GROUP, wh, [chunk(PLAIN, wh, class_class)])) - cslinearg.append(chunk(PLAIN, wh, ' ')) - cslinearg.append(dataname) - - pp.insert(i, chunk(GROUP, wh, cslinearg)) - i, length = i+1, length+1 - hist.command = command - return length, i - - -def do_opcodedesc(length, buf, pp, i): - startpoint = i-1 - ch = pp[startpoint] - wh = ch.where - length, newi = getnextarg(length, buf, pp, i) - dataname = chunk(GROUP, wh, pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - - ch.chtype = CSLINE - ch.data = "deffn" - - cslinearg = [chunk(PLAIN, wh, 'byte\ code\ instruction'), - chunk(GROUP, wh, [chunk(PLAIN, wh, "byte code instruction")]), - chunk(PLAIN, wh, ' '), - dataname, - chunk(PLAIN, wh, ' '), - pp[i], - ] - - pp[i] = chunk(GROUP, wh, cslinearg) - hist.command = ch.data - return length, i - - -def add_module_index(pp, length, i, buf, ch, extra, ref=1): - ch.chtype = chunk_type[CSLINE] - ch.data = 'pindex' - length, newi = getnextarg(length, buf, pp, i) - ingroupch = pp[i:newi] - del pp[i:newi] - length = length - (newi-i) - if not ref: - if len(ingroupch) == 1: - hist.this_module = s(buf, ch.data) - else: - hist.this_module = None - print 'add_module_index() error ==>', ingroupch - - if extra: - ingroupch.append(chunk(PLAIN, ch.where, ' ')) - ingroupch.append(chunk(CSNAME, ch.where, 'r')) - ingroupch.append(chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, extra)])) - - pp.insert(i, chunk(GROUP, ch.where, ingroupch)) - return length+1, i+1 - - -def yank_indexsubitem(pp, length, i, buf, ch, cmdname): - stuff = pp[i].data - if len(stuff) != 1: - raise error, "first parameter to \\%s too long" % cmdname - if pp[i].chtype != chunk_type[GROUP]: - raise error, "bad chunk type following \\%s" \ - "\nexpected GROUP, got %s" + (cmdname, str(ch.chtype)) - text = s(buf, stuff[0].data) - if text[:1] != '(' or text[-1:] != ')': - raise error, \ - 'expected indexsubitem enclosed in parenteses' - hist.indexsubitem = string.split(text[1:-1]) - del pp[i-1:i+1] - return length - 2, i - 1 - - -# regular indices: those that are not set in tt font by default.... -regindices = ('cindex', ) - -# remove illegal characters from node names -def rm_commas_etc(text): - result = '' - changed = 0 - while 1: - pos = re_commas_etc.search(text) - if pos >= 0: - changed = 1 - result = result + text[:pos] - text = text[pos+1:] - else: - result = result + text - break - if changed: - print 'Warning: nodename changed to ' + `result` - - return result - -# boolean flags -flags = {'texi': 1} - - -# map of \label{} to node names -label_nodes = {} - - -## -## changeit: the actual routine, that changes the contents of the parsed -## chunks -## - -def changeit(buf, pp): - global onlylatexspecial, hist, out - - i, length = 0, len(pp) - while 1: - # sanity check: length should always equal len(pp) - if len(pp) != length: - print i, pp[i] - raise 'FATAL', 'inconsistent length. thought ' + `length` + ', but should really be ' + `len(pp)` - if i >= length: - break - ch = pp[i] - i = i + 1 - - if type(ch) is StringType: - #normally, only chunks are present in pp, - # but in some cases, some extra info - # has been inserted, e.g., the \end{...} clauses - raise 'FATAL', 'got string, probably too many ' + `end` - - if ch.chtype == chunk_type[GROUP]: - # check for {\em ...} constructs - data = ch.data - if data and \ - data[0].chtype == chunk_type[CSNAME] and \ - fontchanges.has_key(s(buf, data[0].data)): - k = s(buf, data[0].data) - del data[0] - pp.insert(i-1, chunk(CSNAME, ch.where, fontchanges[k])) - length, i = length+1, i+1 - - elif data: - if len(data) \ - and data[0].chtype == chunk_type[GROUP] \ - and len(data[0].data) \ - and data[0].data[0].chtype == chunk_type[CSNAME] \ - and s(buf, data[0].data[0].data) == 'e': - data[0] = data[0].data[0] - print "invoking \\e magic group transform..." - else: -## print "GROUP -- ch.data[0].data =", ch.data[0].data - k = s(buf, data[0].data) - if k == "fulllineitems": - del data[0] - pp[i-1:i] = data - i = i - 1 - length = length + len(data) - 1 - continue - - # recursively parse the contents of the group - changeit(buf, data) - - elif ch.chtype == chunk_type[IF]: - # \if... - flag, negate, data = ch.data - ##print 'IF: flag, negate = ' + `flag, negate` - if flag not in flags.keys(): - raise error, 'unknown flag ' + `flag` - - value = flags[flag] - if negate: - value = (not value) - del pp[i-1] - length, i = length-1, i-1 - if value: - pp[i:i] = data - length = length + len(data) - - - elif ch.chtype == chunk_type[ENV]: - # \begin{...} .... - envname, data = ch.data - - #push this environment name on stack - hist.inenv.insert(0, envname) - - #append an endenv chunk after grouped data - data.append(chunk(ENDENV, ch.where, envname)) - ##[`data`] - - #delete this object - del pp[i-1] - i, length = i-1, length-1 - - #insert found data - pp[i:i] = data - length = length + len(data) - - if envname == 'verbatim': - pp[i:i] = [chunk(CSLINE, ch.where, 'example'), - chunk(GROUP, ch.where, [])] - length, i = length+2, i+2 - - elif envname in ('itemize', 'list', 'fulllineitems'): - if hist.itemizenesting > len(itemizesymbols): - raise error, 'too deep itemize nesting' - if envname == 'list': - del pp[i:i+2] - length = length - 2 - ingroupch = [chunk(CSNAME, ch.where, - itemizesymbols[hist.itemizenesting])] - hist.itemizenesting = hist.itemizenesting + 1 - pp[i:i] = [chunk(CSLINE, ch.where, 'itemize'), - chunk(GROUP, ch.where, ingroupch)] - length, i = length+2, i+2 - - elif envname == 'enumerate': - if hist.enumeratenesting > len(enumeratesymbols): - raise error, 'too deep enumerate nesting' - ingroupch = [chunk(PLAIN, ch.where, - enumeratesymbols[hist.enumeratenesting])] - hist.enumeratenesting = hist.enumeratenesting + 1 - pp[i:i] = [chunk(CSLINE, ch.where, 'enumerate'), - chunk(GROUP, ch.where, ingroupch)] - length, i = length+2, i+2 - - elif envname == 'description': - ingroupch = [chunk(CSNAME, ch.where, 'b')] - pp[i:i] = [chunk(CSLINE, ch.where, 'table'), - chunk(GROUP, ch.where, ingroupch)] - length, i = length+2, i+2 - - elif (envname == 'tableiii') or (envname == 'tableii'): - if (envname == 'tableii'): - ltable = 2 - else: - ltable = 3 - wh = ch.where - newcode = [] - - #delete tabular format description - # e.g., {|l|c|l|} - length, newi = getnextarg(length, buf, pp, i) - del pp[i:newi] - length = length - (newi-i) - - newcode.append(chunk(CSLINE, wh, 'table')) - ingroupch = [chunk(CSNAME, wh, 'asis')] - newcode.append(chunk(GROUP, wh, ingroupch)) - newcode.append(chunk(CSLINE, wh, 'item')) - - #get the name of macro for @item - # e.g., {code} - length, newi = getnextarg(length, buf, pp, i) - - if newi-i != 1: - raise error, 'Sorry, expected 1 chunk argument' - if pp[i].chtype != chunk_type[PLAIN]: - raise error, 'Sorry, expected plain text argument' - hist.itemargmacro = s(buf, pp[i].data) - if convertible_csname(hist.itemargmacro): - hist.itemargmacro = conversion(hist.itemargmacro)[1] - del pp[i:newi] - length = length - (newi-i) - - itembody = [] - for count in range(ltable): - length, newi = getnextarg(length, buf, pp, i) - emphgroup = [ - chunk(CSNAME, wh, 'emph'), - chunk(GROUP, 0, pp[i:newi])] - del pp[i:newi] - length = length - (newi-i) - if count == 0: - itemarg = emphgroup - elif count == ltable-1: - itembody = itembody + \ - [chunk(PLAIN, wh, ' --- ')] + emphgroup - else: - itembody = emphgroup - newcode.append(chunk(GROUP, wh, itemarg)) - newcode = newcode + itembody + [chunk(DENDLINE, wh, '\n')] - pp[i:i] = newcode - l = len(newcode) - length, i = length+l, i+l - del newcode, l - - if length != len(pp): - raise 'STILL, SOMETHING wrong', `i` - - elif envname in ('methoddesc', 'methoddescni'): - length, newi = getoptarg(length, buf, pp, i) - ingroupch = pp[i:newi] - del pp[i:newi] - length = length - (newi-i) - # - pp.insert(i, chunk(PLAIN, ch.where, '')) - i, length = i+1, length+1 - length, i = do_funcdesc(length, buf, pp, i, - envname[-2:] != "ni") - - elif envname in ('memberdesc', 'memberdescni'): - length, newi = getoptarg(length, buf, pp, i) - ingroupch = pp[i:newi] - del pp[i:newi] - length = length - (newi-i) - # - pp.insert(i, chunk(PLAIN, ch.where, '')) - i, length = i+1, length+1 - length, i = do_datadesc(length, buf, pp, i, - envname[-2:] != "ni") - - elif envname in ('funcdesc', 'funcdescni', 'classdesc'): - pp.insert(i, chunk(PLAIN, ch.where, '')) - i, length = i+1, length+1 - length, i = do_funcdesc(length, buf, pp, i, - envname[-2:] != "ni") - - elif envname == 'excdesc': - pp.insert(i, chunk(PLAIN, ch.where, '')) - i, length = i+1, length+1 - length, i = do_excdesc(length, buf, pp, i) - - elif envname in ('datadesc', 'datadescni'): - pp.insert(i, chunk(PLAIN, ch.where, '')) - i, length = i+1, length+1 - length, i = do_datadesc(length, buf, pp, i, - envname[-2:] != "ni") - - elif envname == 'opcodedesc': - pp.insert(i, chunk(PLAIN, ch.where, '')) - i, length = i+1, length+1 - length, i = do_opcodedesc(length, buf, pp, i) - - elif envname == 'seealso': - chunks = [chunk(ENDLINE, ch.where, "\n"), - chunk(CSNAME, ch.where, "b"), - chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, "See also: ")]), - chunk(ENDLINE, ch.where, "\n"), - chunk(ENDLINE, ch.where, "\n")] - pp[i-1:i] = chunks - length = length + len(chunks) - 1 - i = i + len(chunks) - 1 - - elif envname in ('sloppypar', 'flushleft', 'document'): - pass - - else: - print 'WARNING: don\'t know what to do with env ' + `envname` - - elif ch.chtype == chunk_type[ENDENV]: - envname = ch.data - if envname != hist.inenv[0]: - raise error, '\'end\' does not match. Name ' + `envname` + ', expected ' + `hist.inenv[0]` - del hist.inenv[0] - del pp[i-1] - i, length = i-1, length-1 - - if envname == 'verbatim': - pp[i:i] = [chunk(CSLINE, ch.where, 'end'), - chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, 'example')])] - i, length = i+2, length+2 - elif envname in ('itemize', 'list', 'fulllineitems'): - hist.itemizenesting = hist.itemizenesting - 1 - pp[i:i] = [chunk(CSLINE, ch.where, 'end'), - chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, 'itemize')])] - i, length = i+2, length+2 - elif envname == 'enumerate': - hist.enumeratenesting = hist.enumeratenesting-1 - pp[i:i] = [chunk(CSLINE, ch.where, 'end'), - chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, 'enumerate')])] - i, length = i+2, length+2 - elif envname == 'description': - pp[i:i] = [chunk(CSLINE, ch.where, 'end'), - chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, 'table')])] - i, length = i+2, length+2 - elif (envname == 'tableiii') or (envname == 'tableii'): - pp[i:i] = [chunk(CSLINE, ch.where, 'end'), - chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, 'table')])] - i, length = i+2, length + 2 - pp.insert(i, chunk(DENDLINE, ch.where, '\n')) - i, length = i+1, length+1 - - elif envname in ('funcdesc', 'excdesc', 'datadesc', 'classdesc', - 'funcdescni', 'datadescni', - 'methoddesc', 'memberdesc', - 'methoddescni', 'memberdescni', - ): - pp[i:i] = [chunk(CSLINE, ch.where, 'end'), - chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, hist.command)])] - i, length = i+2, length+2 - - elif envname == 'opcodedesc': - pp[i:i] = [chunk(CSLINE, ch.where, 'end'), - chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, "deffn")])] - i, length = i+2, length+2 - - elif envname in ('seealso', 'sloppypar', 'flushleft', 'document'): - pass - - else: - print 'WARNING: ending env %s has no actions' % `envname` - - elif ch.chtype == chunk_type[CSNAME]: - # control name transformations - s_buf_data = s(buf, ch.data) - if s_buf_data == 'optional': - pp[i-1].chtype = chunk_type[PLAIN] - pp[i-1].data = '[' - if (i < length) and \ - (pp[i].chtype == chunk_type[GROUP]): - cp=pp[i].data - pp[i:i+1]=cp + [ - chunk(PLAIN, ch.where, ']')] - length = length+len(cp) - - elif s_buf_data in ignoredcommands: - del pp[i-1] - i, length = i-1, length-1 - - elif s_buf_data == '@' and \ - i != length and \ - pp[i].chtype == chunk_type[PLAIN] and \ - s(buf, pp[i].data)[0] == '.': - # \@. --> \. --> @. - ch.data = '.' - del pp[i] - length = length - 1 - - elif convertible_csname(s_buf_data): - ch.chtype, ch.data, nix = conversion(s_buf_data) - try: - if nix and pp[i].chtype == chunk_type[GROUP] \ - and len(pp[i].data) == 0: - del pp[i] - length = length - 1 - except IndexError: - pass - - elif s_buf_data == '\\': - # \\ --> \* --> @* - ch.data = '*' - - elif len(s_buf_data) == 1 and \ - s_buf_data in onlylatexspecial: - ch.chtype = chunk_type[PLAIN] - # check if such a command is followed by - # an empty group: e.g., `\%{}'. If so, remove - # this empty group too - if i < length and \ - pp[i].chtype == chunk_type[GROUP] \ - and len(pp[i].data) == 0: - del pp[i] - length = length-1 - - elif s_buf_data == "appendix": - hist.chaptertype = "appendix" - del pp[i-1] - i, length = i-1, length-1 - - elif hist.inargs and s_buf_data in inargsselves: - # This is the special processing of the - # arguments of the \begin{funcdesc}... or - # \funcline... arguments - # \, --> , \[ --> [, \] --> ] - ch.chtype = chunk_type[PLAIN] - - elif s_buf_data == 'setindexsubitem': - length, i = yank_indexsubitem(pp, length, i, buf, ch, - 'setindexsubitem') - - elif s_buf_data == 'withsubitem': - oldsubitem = hist.indexsubitem - try: - length, i = yank_indexsubitem(pp, length, i, buf, ch, - 'withsubitem') - stuff = pp[i].data - del pp[i] - length = length - 1 - changeit(buf, stuff) - stuff = None - finally: - hist.indexsubitem = oldsubitem - - elif s_buf_data in ('textrm', 'pytype'): - stuff = pp[i].data - pp[i-1:i+1] = stuff - length = length - 2 + len(stuff) - stuff = None - i = i - 1 - - elif s_buf_data == 'newcommand': - print "ignoring definition of \\" + s(buf, pp[i].data[0].data) - del pp[i-1:i+2] - i = i - 1 - length = length - 3 - - elif s_buf_data == 'renewcommand': - print "ignoring redefinition of \\" \ - + s(buf, pp[i].data[0].data) - del pp[i-1:i+2] - i = i - 1 - length = length - 3 - - elif s_buf_data == 'mbox': - stuff = pp[i].data - pp[i-1:i+1] = stuff - i = i - 1 - length = length + len(stuff) - 2 - stuff = None - - elif s_buf_data == 'version': - ch.chtype = chunk_type[PLAIN] - ch.data = release_version - - elif s_buf_data == 'item': - ch.chtype = chunk_type[CSLINE] - length, newi = getoptarg(length, buf, pp, i) - ingroupch = pp[i:newi] - del pp[i:newi] - length = length - (newi-i) - changeit(buf, ingroupch) # catch stuff inside the optional arg - pp.insert(i, chunk(GROUP, ch.where, ingroupch)) - i, length = i+1, length+1 - - elif s_buf_data == 'ttindex': - idxsi = hist.indexsubitem - - cat_class = '' - if len(idxsi) >= 2 and idxsi[1] in \ - ('method', 'function', 'protocol'): - command = 'findex' - elif len(idxsi) >= 2 and idxsi[1] in \ - ('exception', 'object'): - command = 'vindex' - elif len(idxsi) == 3 and idxsi[:2] == ['in', 'module']: - command = 'cindex' - elif len(idxsi) == 3 and idxsi[:2] == ['class', 'in']: - command = 'findex' - else: - print 'WARNING: can\'t categorize ' + `idxsi` \ - + ' for \'ttindex\' command' - command = 'cindex' - - if not cat_class: - cat_class = '(%s)' % string.join(idxsi) - - ch.chtype = chunk_type[CSLINE] - ch.data = command - - length, newi = getnextarg(length, buf, pp, i) - arg = pp[i:newi] - del pp[i:newi] - length = length - (newi-i) - - cat_arg = [chunk(PLAIN, ch.where, cat_class)] - - # determine what should be set in roman, and - # what in tt-font - if command in regindices: - - arg = [chunk(CSNAME, ch.where, 't'), - chunk(GROUP, ch.where, arg)] - else: - cat_arg = [chunk(CSNAME, ch.where, 'r'), - chunk(GROUP, ch.where, cat_arg)] - - ingroupch = arg + \ - [chunk(PLAIN, ch.where, ' ')] + \ - cat_arg - - pp.insert(i, chunk(GROUP, ch.where, ingroupch)) - length, i = length+1, i+1 - - elif s_buf_data == 'ldots': - # \ldots --> \dots{} --> @dots{} - ch.data = 'dots' - if i == length \ - or pp[i].chtype != chunk_type[GROUP] \ - or pp[i].data != []: - pp.insert(i, chunk(GROUP, ch.where, [])) - i, length = i+1, length+1 - - elif s_buf_data in themselves: - # \UNIX --> &UNIX; - ch.chtype = chunk_type[PLAIN] - if i != length \ - and pp[i].chtype == chunk_type[GROUP] \ - and pp[i].data == []: - del pp[i] - length = length-1 - - elif s_buf_data == 'manpage': - ch.data = 'emph' - sect = s(buf, pp[i+1].data[0].data) - pp[i+1].data = "(%s)" % sect - pp[i+1].chtype = chunk_type[PLAIN] - - elif s_buf_data == 'envvar': - # this should do stuff in the index, too... - ch.data = "$" - ch.chtype = chunk_type[PLAIN] - pp[i] = pp[i].data[0] - - elif s_buf_data == 'regexp': - ch.data = 'code' - pp.insert(i+1, chunk(PLAIN, ch.where, '"')) - pp.insert(i-1, chunk(PLAIN, ch.where, '"')) - length = length + 2 - i = i + 1 - - elif s_buf_data in ('lineiii', 'lineii'): - # This is the most tricky one - # \lineiii{a1}{a2}[{a3}] --> - # @item @<cts. of itemargmacro>{a1} - # a2 [ -- a3] - # - if not hist.inenv: - raise error, 'no environment for lineiii' - if (hist.inenv[0] != 'tableiii') and \ - (hist.inenv[0] != 'tableii'): - raise error, \ - 'wrong command (%s) in wrong environment (%s)' \ - % (s_buf_data, `hist.inenv[0]`) - ch.chtype = chunk_type[CSLINE] - ch.data = 'item' - length, newi = getnextarg(length, buf, pp, i) - ingroupch = [chunk(CSNAME, 0, hist.itemargmacro), - chunk(GROUP, 0, pp[i:newi])] - del pp[i:newi] - length = length - (newi-i) - pp.insert(i, chunk(GROUP, ch.where, ingroupch)) - grouppos = i - i, length = i+1, length+1 - length, i = getnextarg(length, buf, pp, i) - length, newi = getnextarg(length, buf, pp, i) - if newi > i: - # we have a 3rd arg - pp.insert(i, chunk(PLAIN, ch.where, ' --- ')) - i = newi + 1 - length = length + 1 - if length != len(pp): - raise 'IN LINEIII IS THE ERR', `i` - - elif s_buf_data in ('chapter', 'section', - 'subsection', 'subsubsection'): - #\xxxsection{A} ----> - # @node A, , , - # @xxxsection A - ## also: remove commas and quotes - hist.this_module = None - if s_buf_data == "chapter": - ch.data = hist.chaptertype - ch.chtype = chunk_type[CSLINE] - length, newi = getnextarg(length, buf, pp, i) - afternodenamecmd = next_command_p(length, buf, - pp, newi, 'nodename') - if afternodenamecmd < 0: - cp1 = crcopy(pp[i:newi]) - pp[i:newi] = [chunk(GROUP, ch.where, pp[i:newi])] - length, newi = length - (newi-i) + 1, i+1 - text = flattext(buf, cp1) - text = invent_node_names(text) - else: - length, endarg = getnextarg(length, buf, - pp, afternodenamecmd) - cp1 = crcopy(pp[afternodenamecmd:endarg]) - del pp[newi:endarg] - length = length - (endarg-newi) - - pp[i:newi] = [chunk(GROUP, ch.where, pp[i:newi])] - length, newi = length - (newi-i) + 1, i + 1 - text = flattext(buf, cp1) - if text[-1] == '.': - text = text[:-1] - if text in hist.nodenames: - print 'WARNING: node name ' + `text` + ' already used' - out.doublenodes.append(text) - else: - hist.nodenames.append(text) - text = rm_commas_etc(text) - pp[i-1:i-1] = [chunk(CSLINE, ch.where, 'node'), - chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, text+', , ,') - ])] - i, length = newi+2, length+2 - - elif s_buf_data == 'funcline': - # fold it to a very short environment - pp[i-1:i-1] = [chunk(CSLINE, ch.where, 'end'), - chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, hist.command)])] - i, length = i+2, length+2 - length, i = do_funcdesc(length, buf, pp, i) - - elif s_buf_data == 'dataline': - pp[i-1:i-1] = [chunk(CSLINE, ch.where, 'end'), - chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, hist.command)])] - i, length = i+2, length+2 - length, i = do_datadesc(length, buf, pp, i) - - elif s_buf_data == 'excline': - pp[i-1:i-1] = [chunk(CSLINE, ch.where, 'end'), - chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, hist.command)])] - i, length = i+2, length+2 - length, i = do_excdesc(length, buf, pp, i) - - elif s_buf_data == 'index': - #\index{A} ---> - # @cindex A - ch.chtype = chunk_type[CSLINE] - ch.data = 'cindex' - length, newi = getnextarg(length, buf, pp, i) - - ingroupch = pp[i:newi] - del pp[i:newi] - length = length - (newi-i) - pp.insert(i, chunk(GROUP, ch.where, ingroupch)) - length, i = length+1, i+1 - - elif s_buf_data == 'bifuncindex': - ch.chtype = chunk_type[CSLINE] - ch.data = 'findex' - length, newi = getnextarg(length, buf, pp, i) - ingroupch = pp[i:newi] - del pp[i:newi] - length = length - (newi-i) - - ingroupch.append(chunk(PLAIN, ch.where, ' ')) - ingroupch.append(chunk(CSNAME, ch.where, 'r')) - ingroupch.append(chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, - '(built-in function)')])) - - pp.insert(i, chunk(GROUP, ch.where, ingroupch)) - length, i = length+1, i+1 - - elif s_buf_data == 'obindex': - ch.chtype = chunk_type[CSLINE] - ch.data = 'findex' - length, newi = getnextarg(length, buf, pp, i) - ingroupch = pp[i:newi] - del pp[i:newi] - length = length - (newi-i) - - ingroupch.append(chunk(PLAIN, ch.where, ' ')) - ingroupch.append(chunk(CSNAME, ch.where, 'r')) - ingroupch.append(chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, - '(object)')])) - - pp.insert(i, chunk(GROUP, ch.where, ingroupch)) - length, i = length+1, i+1 - - elif s_buf_data == 'opindex': - ch.chtype = chunk_type[CSLINE] - ch.data = 'findex' - length, newi = getnextarg(length, buf, pp, i) - ingroupch = pp[i:newi] - del pp[i:newi] - length = length - (newi-i) - - ingroupch.append(chunk(PLAIN, ch.where, ' ')) - ingroupch.append(chunk(CSNAME, ch.where, 'r')) - ingroupch.append(chunk(GROUP, ch.where, [ - chunk(PLAIN, ch.where, - '(operator)')])) - - pp.insert(i, chunk(GROUP, ch.where, ingroupch)) - length, i = length+1, i+1 - - elif s_buf_data in ('bimodindex', 'refbimodindex'): - length, i = add_module_index( - pp, length, i, buf, ch, '(built-in)', - (s_buf_data[:3] == 'ref')) - - elif s_buf_data in ('modindex', 'refmodindex'): - length, i = add_module_index( - pp, length, i, buf, ch, '', - (s_buf_data[:3] == 'ref')) - - elif s_buf_data in ('stmodindex', 'refstmodindex'): - length, i = add_module_index( - pp, length, i, buf, ch, '(standard)', - (s_buf_data[:3] == 'ref')) - - elif s_buf_data in ('exmodindex', 'refexmodindex'): - length, i = add_module_index( - pp, length, i, buf, ch, '(extension)', - (s_buf_data[:3] == 'ref')) - - elif s_buf_data == 'stindex': - # XXX must actually go to newindex st - what = (s_buf_data[:2] == "st") and "statement" or "keyword" - wh = ch.where - ch.chtype = chunk_type[CSLINE] - ch.data = 'cindex' - length, newi = getnextarg(length, buf, pp, i) - ingroupch = [chunk(CSNAME, wh, 'code'), - chunk(GROUP, wh, pp[i:newi])] - - del pp[i:newi] - length = length - (newi-i) - - t = ingroupch[:] - t.append(chunk(PLAIN, wh, ' ' + what)) - - pp.insert(i, chunk(GROUP, wh, t)) - i, length = i+1, length+1 - - pp.insert(i, chunk(CSLINE, wh, 'cindex')) - i, length = i+1, length+1 - - t = ingroupch[:] - t.insert(0, chunk(PLAIN, wh, what + ', ')) - - pp.insert(i, chunk(GROUP, wh, t)) - i, length = i+1, length+1 - - elif s_buf_data == 'indexii': - #\indexii{A}{B} ---> - # @cindex A B - # @cindex B, A - length, newi = getnextarg(length, buf, pp, i) - cp11 = pp[i:newi] - cp21 = crcopy(pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - length, newi = getnextarg(length, buf, pp, i) - cp12 = pp[i:newi] - cp22 = crcopy(pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - - ch.chtype = chunk_type[CSLINE] - ch.data = 'cindex' - pp.insert(i, chunk(GROUP, ch.where, cp11 + [ - chunk(PLAIN, ch.where, ' ')] + cp12)) - i, length = i+1, length+1 - pp[i:i] = [chunk(CSLINE, ch.where, 'cindex'), - chunk(GROUP, ch.where, cp22 + [ - chunk(PLAIN, ch.where, ', ')]+ cp21)] - i, length = i+2, length+2 - - elif s_buf_data == 'indexiii': - length, newi = getnextarg(length, buf, pp, i) - cp11 = pp[i:newi] - cp21 = crcopy(pp[i:newi]) - cp31 = crcopy(pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - length, newi = getnextarg(length, buf, pp, i) - cp12 = pp[i:newi] - cp22 = crcopy(pp[i:newi]) - cp32 = crcopy(pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - length, newi = getnextarg(length, buf, pp, i) - cp13 = pp[i:newi] - cp23 = crcopy(pp[i:newi]) - cp33 = crcopy(pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - - ch.chtype = chunk_type[CSLINE] - ch.data = 'cindex' - pp.insert(i, chunk(GROUP, ch.where, cp11 + [ - chunk(PLAIN, ch.where, ' ')] + cp12 - + [chunk(PLAIN, ch.where, ' ')] - + cp13)) - i, length = i+1, length+1 - pp[i:i] = [chunk(CSLINE, ch.where, 'cindex'), - chunk(GROUP, ch.where, cp22 + [ - chunk(PLAIN, ch.where, ' ')]+ cp23 - + [chunk(PLAIN, ch.where, ', ')] + - cp21)] - i, length = i+2, length+2 - pp[i:i] = [chunk(CSLINE, ch.where, 'cindex'), - chunk(GROUP, ch.where, cp33 + [ - chunk(PLAIN, ch.where, ', ')]+ cp31 - + [chunk(PLAIN, ch.where, ' ')] + - cp32)] - i, length = i+2, length+2 - - elif s_buf_data == 'indexiv': - length, newi = getnextarg(length, buf, pp, i) - cp11 = pp[i:newi] - cp21 = crcopy(pp[i:newi]) - cp31 = crcopy(pp[i:newi]) - cp41 = crcopy(pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - length, newi = getnextarg(length, buf, pp, i) - cp12 = pp[i:newi] - cp22 = crcopy(pp[i:newi]) - cp32 = crcopy(pp[i:newi]) - cp42 = crcopy(pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - length, newi = getnextarg(length, buf, pp, i) - cp13 = pp[i:newi] - cp23 = crcopy(pp[i:newi]) - cp33 = crcopy(pp[i:newi]) - cp43 = crcopy(pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - length, newi = getnextarg(length, buf, pp, i) - cp14 = pp[i:newi] - cp24 = crcopy(pp[i:newi]) - cp34 = crcopy(pp[i:newi]) - cp44 = crcopy(pp[i:newi]) - del pp[i:newi] - length = length - (newi-i) - - ch.chtype = chunk_type[CSLINE] - ch.data = 'cindex' - ingroupch = cp11 + \ - spacech + cp12 + \ - spacech + cp13 + \ - spacech + cp14 - pp.insert(i, chunk(GROUP, ch.where, ingroupch)) - i, length = i+1, length+1 - ingroupch = cp22 + \ - spacech + cp23 + \ - spacech + cp24 + \ - commach + cp21 - pp[i:i] = cindexch + [ - chunk(GROUP, ch.where, ingroupch)] - i, length = i+2, length+2 - ingroupch = cp33 + \ - spacech + cp34 + \ - commach + cp31 + \ - spacech + cp32 - pp[i:i] = cindexch + [ - chunk(GROUP, ch.where, ingroupch)] - i, length = i+2, length+2 - ingroupch = cp44 + \ - commach + cp41 + \ - spacech + cp42 + \ - spacech + cp43 - pp[i:i] = cindexch + [ - chunk(GROUP, ch.where, ingroupch)] - i, length = i+2, length+2 - - elif s_buf_data == 'seemodule': - # discard optional arg first: - length, newi = getoptarg(length, buf, pp, i) - ingroupch = pp[i:newi] - del pp[i:newi] - length = length - (newi-i) - # - ch.data = "code" - data = pp[i+1].data - data.insert(0, chunk(PLAIN, ch.where, " (")) - data.append(chunk(PLAIN, ch.where, ")")) - pp[i+1:i+2] = data - length = length + len(data) - 1 - - elif s_buf_data == 'seetext': - data = pp[i].data - data.insert(0, chunk(ENDLINE, ch.where, "\n")) - pp[i-1:i+1] = data - i = i - 1 - length = length + len(data) - 2 - - elif s_buf_data == 'deprecated': - length, newi = getnextarg(length, buf, pp, i) - version = pp[i:newi][0] - length, newi2 = getnextarg(length, buf, pp, newi) - action = pp[newi:newi2] - del pp[i-1:newi2] - length = length - (newi2 - i) - 1 - stuff = [chunk(PLAIN, ch.where, 'Deprecated since release '), - version, - chunk(PLAIN, ch.where, '.')] - chunks = [chunk(CSNAME, ch.where, 'strong'), - chunk(GROUP, ch.where, stuff), - chunk(PLAIN, ch.where, ' ')] + action \ - + [chunk(DENDLINE, ch.where, '\n')] - stuff = None - i = i - 1 - pp[i:i] = chunks - length = length + len(chunks) - - elif s_buf_data == "quad": - ch.chtype = PLAIN - ch.data = " " - - elif s_buf_data in ('usepackage', 'input'): - del pp[i-1:i+1] - i, length = i-1, length-2 - - elif s_buf_data in ('noindent', 'indexsubitem', 'footnote'): - pass - - elif s_buf_data == 'label': - name = s(buf, pp[i].data[0].data) - del pp[i-1:i+1] - length = length - 2 - i = i - 1 - label_nodes[name] = hist.nodenames[-1] - - elif s_buf_data == 'rfc': - ch.chtype = chunk_type[PLAIN] - ch.data = "RFC " + s(buf, pp[i].data[0].data) - del pp[i] - length = length - 1 - - elif s_buf_data == 'ref': - name = s(buf, pp[i].data[0].data) - if label_nodes.has_key(name): - pp[i].data[0].data = label_nodes[name] - else: - pp[i-1:i+1] = [ - chunk(PLAIN, ch.where, - "(unknown node reference: %s)" % name)] - length = length - 1 - print "WARNING: unknown node label", `name` - - else: - print "don't know what to do with keyword " + s_buf_data - - -re_atsign = regex.compile('[@{}]') -re_newline = regex.compile('\n') - -def dumpit(buf, wm, pp): - - global out - - i, length = 0, len(pp) - - addspace = 0 - - while 1: - if len(pp) != length: - raise 'FATAL', 'inconsistent length' - if i == length: - break - ch = pp[i] - i = i + 1 - - dospace = addspace - addspace = 0 - - if ch.chtype == chunk_type[CSNAME]: - s_buf_data = s(buf, ch.data) -## if s_buf_data == 'e': -## wm('\\') -## continue -## if s_buf_data == '$': -## wm('$') -## continue - wm('@' + s_buf_data) - if s_buf_data == 'node' and \ - pp[i].chtype == chunk_type[PLAIN] and \ - s(buf, pp[i].data) in out.doublenodes: - ##XXX doesnt work yet?? - wm(' ZZZ-' + zfill(`i`, 4)) - if s_buf_data[0] in string.letters: - addspace = 1 - elif ch.chtype == chunk_type[PLAIN]: - if dospace and s(buf, ch.data) not in (' ', '\t'): - wm(' ') - text = s(buf, ch.data) - while 1: - pos = re_atsign.search(text) - if pos < 0: - break - wm(text[:pos] + '@' + text[pos]) - text = text[pos+1:] - wm(text) - elif ch.chtype == chunk_type[GROUP]: - wm('{') - dumpit(buf, wm, ch.data) - wm('}') - elif ch.chtype == chunk_type[DENDLINE]: - wm('\n\n') - while i != length and pp[i].chtype in \ - (chunk_type[DENDLINE], chunk_type[ENDLINE]): - i = i + 1 - elif ch.chtype == chunk_type[OTHER]: - wm(s(buf, ch.data)) - elif ch.chtype == chunk_type[ACTIVE]: - wm(s(buf, ch.data)) - elif ch.chtype == chunk_type[ENDLINE]: - wm('\n') - elif ch.chtype == chunk_type[CSLINE]: - if i >= 2 and pp[i-2].chtype not in \ - (chunk_type[ENDLINE], chunk_type[DENDLINE]) \ - and (pp[i-2].chtype != chunk_type[PLAIN] - or s(buf, pp[i-2].data)[-1] != '\n'): - - wm('\n') - wm('@' + s(buf, ch.data)) - if i == length: - raise error, 'CSLINE expected another chunk' - if pp[i].chtype != chunk_type[GROUP]: - raise error, 'CSLINE expected GROUP' - if type(pp[i].data) != ListType: - raise error, 'GROUP chould contain []-data' - - wobj = Wobj() - dumpit(buf, wobj.write, pp[i].data) - i = i + 1 - text = wobj.data - del wobj - if text: - wm(' ') - while 1: - pos = re_newline.search(text) - if pos < 0: - break - # these seem to be completely harmless, so don't warn: -## print 'WARNING: found newline in csline arg (%s)' \ -## % s(buf, ch.data) - wm(text[:pos] + ' ') - text = text[pos+1:] - wm(text) - if i >= length or \ - pp[i].chtype not in (chunk_type[CSLINE], - chunk_type[ENDLINE], chunk_type[DENDLINE]) \ - and (pp[i].chtype != chunk_type[PLAIN] - or s(buf, pp[i].data)[0] != '\n'): - wm('\n') - - elif ch.chtype == chunk_type[COMMENT]: - if s(buf, ch.data) and \ - regex.match('^[ \t]*$', s(buf, ch.data)) < 0: - if i >= 2 \ - and pp[i-2].chtype not in (chunk_type[ENDLINE], - chunk_type[DENDLINE]) \ - and not (pp[i-2].chtype == chunk_type[PLAIN] - and regex.match('\\(.\\|\n\\)*[ \t]*\n$', s(buf, pp[i-2].data)) >= 0): - wm('\n') - wm('@c ' + s(buf, ch.data)) - elif ch.chtype == chunk_type[IGNORE]: - pass - else: - try: - str = `s(buf, ch.data)` - except TypeError: - str = `ch.data` - if len(str) > 400: - str = str[:400] + '...' - print 'warning:', ch.chtype, 'not handled, data ' + str - - - -def main(): - global release_version - outfile = None - headerfile = 'texipre.dat' - trailerfile = 'texipost.dat' - - try: - opts, args = getopt.getopt(sys.argv[1:], 'o:h:t:v:') - except getopt.error: - args = [] - - if not args: - print 'usage: partparse [-o outfile] [-h headerfile]', - print '[-t trailerfile] file ...' - sys.exit(2) - - for opt, arg in opts: - if opt == '-o': outfile = arg - if opt == '-h': headerfile = arg - if opt == '-t': trailerfile = arg - if opt == '-v': release_version = arg - - if not outfile: - root, ext = os.path.splitext(args[0]) - outfile = root + '.texi' - - if outfile in args: - print 'will not overwrite input file', outfile - sys.exit(2) - - outf = open(outfile, 'w') - outf.write(open(headerfile, 'r').read()) - - for file in args: - if len(args) > 1: print '='*20, file, '='*20 - buf = open(file, 'r').read() - chunk.buf = buf - w, pp = parseit(buf) - startchange() - changeit(buf, pp) - dumpit(buf, outf.write, pp) - - outf.write(open(trailerfile, 'r').read()) - - outf.close() - -if __name__ == "__main__": - main() |