summaryrefslogtreecommitdiffstats
path: root/doc/translator.py
diff options
context:
space:
mode:
authorDimitri van Heesch <dimitri@stack.nl>2013-03-24 11:51:51 (GMT)
committerDimitri van Heesch <dimitri@stack.nl>2013-03-24 11:51:51 (GMT)
commitdeaa34e0c1d990f37fe00e465ac7a22f705904f0 (patch)
tree8d5749ba8d587d7b753162c524b18f7e7cdf6a0b /doc/translator.py
parentaacd1557af129e2bff3514169e04168376a2431b (diff)
downloadDoxygen-deaa34e0c1d990f37fe00e465ac7a22f705904f0.zip
Doxygen-deaa34e0c1d990f37fe00e465ac7a22f705904f0.tar.gz
Doxygen-deaa34e0c1d990f37fe00e465ac7a22f705904f0.tar.bz2
Release-1.8.3.1-20130324
Diffstat (limited to 'doc/translator.py')
-rw-r--r--doc/translator.py747
1 files changed, 379 insertions, 368 deletions
diff --git a/doc/translator.py b/doc/translator.py
index 5a78ee2..816482b 100644
--- a/doc/translator.py
+++ b/doc/translator.py
@@ -4,25 +4,25 @@
related to internationalization (the translator classes). It uses the
information to generate documentation (language.doc,
translator_report.txt) from templates (language.tpl, maintainers.txt).
-
+
Simply run the script without parameters to get the reports and
documentation for all supported languages. If you want to generate the
translator report only for some languages, pass their codes as arguments
to the script. In that case, the language.doc will not be generated.
Example:
-
+
python translator.py en nl cz
-
+
Originally, the script was written in Perl and was known as translator.pl.
The last Perl version was dated 2002/05/21 (plus some later corrections)
$Id$
-
+
Petr Prikryl (prikrylp@skil.cz)
-
+
History:
--------
- 2002/05/21 - This was the last Perl version.
+ 2002/05/21 - This was the last Perl version.
2003/05/16 - List of language marks can be passed as arguments.
2004/01/24 - Total reimplementation started: classes TrManager, and Transl.
2004/02/05 - First version that produces translator report. No language.doc yet.
@@ -46,7 +46,7 @@
2004/10/04 - Reporting of not called translator methods added.
2004/10/05 - Modified to check only doxygen/src sources for the previous report.
2005/02/28 - Slight modification to generate "mailto.txt" auxiliary file.
- 2005/08/15 - Doxygen's root directory determined primarily from DOXYGEN
+ 2005/08/15 - Doxygen's root directory determined primarily from DOXYGEN
environment variable. When not found, then relatively to the script.
2007/03/20 - The "translate me!" searched in comments and reported if found.
2008/06/09 - Warning when the MAX_DOT_GRAPH_HEIGHT is still part of trLegendDocs().
@@ -62,7 +62,8 @@
2010/08/30 - Highlighting in what will be the table in langhowto.html modified.
2010/09/27 - The underscore in \latexonly part of the generated language.doc
was prefixed by backslash (was LaTeX related error).
- """
+ 2013/02/19 - Better diagnostics when translator_xx.h is too crippled.
+ """
from __future__ import generators
import codecs
@@ -73,16 +74,16 @@ import sys
def fill(s):
"""Returns string formated to the wrapped paragraph multiline string.
-
+
Replaces whitespaces by one space and then uses he textwrap.fill()."""
-
+
# Replace all whitespace by spaces, remove whitespaces that are not
- # necessary, strip the left and right whitespaces, and break the string
+ # necessary, strip the left and right whitespaces, and break the string
# to list of words.
rexWS = re.compile(r'\s+')
lst = rexWS.sub(' ', s).strip().split()
-
- # If the list is not empty, put the words together and form the lines
+
+ # If the list is not empty, put the words together and form the lines
# of maximum 70 characters. Build the list of lines.
lines = []
if lst:
@@ -96,17 +97,17 @@ def fill(s):
lines.append(line) # the last line
return '\n'.join(lines)
-
+
# The following function dedent() is the verbatim copy from the textwrap.py
# module. The textwrap.py was introduced in Python 2.3. To make this script
-# working also in older Python versions, I have decided to copy it.
+# working also in older Python versions, I have decided to copy it.
# Notice that the textwrap.py is copyrighted:
#
# Copyright (C) 1999-2001 Gregory P. Ward.
# Copyright (C) 2002, 2003 Python Software Foundation.
# Written by Greg Ward <gward@python.net>
#
-# The explicit permission to use the code here was sent by Guido van Rossum
+# The explicit permission to use the code here was sent by Guido van Rossum
# (4th June, 2004).
#
def dedent(text):
@@ -147,29 +148,29 @@ def dedent(text):
lines[i] = lines[i][margin:]
return '\n'.join(lines)
-
+
class Transl:
"""One instance is build for each translator.
-
+
The abbreviation of the source file--part after 'translator_'--is used as
the identification of the object. The empty string is used for the
abstract Translator class from translator.h. The other information is
extracted from inside the source file."""
-
+
def __init__(self, fname, manager):
"""Bind to the manager and initialize."""
-
+
# Store the filename and the reference to the manager object.
self.fname = fname
self.manager = manager
-
+
# The instance is responsible for loading the source file, so it checks
# for its existence and quits if something goes wrong.
if not os.path.isfile(fname):
sys.stderr.write("\a\nFile '%s' not found!\n" % fname)
sys.exit(1)
-
+
# Initialize the other collected information.
self.classId = None
self.baseClassId = None
@@ -186,11 +187,11 @@ class Transl:
self.missingMethods = None # list of prototypes to be implemented
self.implementedMethods = None # list of implemented required methods
self.adaptMinClass = None # The newest adapter class that can be used
- self.isDecodedTranslator = None # Flag related to internal usage of UTF-8
-
+ self.isDecodedTranslator = None # Flag related to internal usage of UTF-8
+
def __tokenGenerator(self):
"""Generator that reads the file and yields tokens as 4-tuples.
-
+
The tokens have the form (tokenId, tokenString, lineNo). The
last returned token has the form ('eof', None, None). When trying
to access next token afer that, the exception would be raised."""
@@ -205,14 +206,14 @@ class Transl:
'private': 'private',
'static': 'static',
'virtual': 'virtual',
- ':': 'colon',
+ ':': 'colon',
';': 'semic',
',': 'comma',
- '[': 'lsqbra',
+ '[': 'lsqbra',
']': 'rsqbra',
- '(': 'lpar',
+ '(': 'lpar',
')': 'rpar',
- '{': 'lcurly',
+ '{': 'lcurly',
'}': 'rcurly',
'=': 'assign',
'*': 'star',
@@ -230,7 +231,7 @@ class Transl:
'~': 'tilde',
'^': 'caret',
}
-
+
# Regular expression for recognizing identifiers.
rexId = re.compile(r'^[a-zA-Z]\w*$')
@@ -241,14 +242,14 @@ class Transl:
line = '' # init -- see the pos initialization below
linelen = 0 # init
pos = 100 # init -- pos after the end of line
- status = 0
-
+ status = 0
+
tokenId = None # init
tokenStr = '' # init -- the characters will be appended.
tokenLineNo = 0
while status != 777:
-
+
# Get the next character. Read next line first, if necessary.
if pos < linelen:
c = line[pos]
@@ -263,18 +264,18 @@ class Transl:
status = 777
else:
c = line[pos]
-
+
# Consume the character based on the status
if status == 0: # basic status
-
+
# This is the initial status. If tokenId is set, yield the
# token here and only here (except when eof is found).
# Initialize the token variables after the yield.
if tokenId:
# If it is an unknown item, it can still be recognized
# here. Keywords and separators are the example.
- if tokenId == 'unknown':
+ if tokenId == 'unknown':
if tokenDic.has_key(tokenStr):
tokenId = tokenDic[tokenStr]
elif tokenStr.isdigit():
@@ -283,28 +284,28 @@ class Transl:
tokenId = 'id'
else:
msg = '\aWarning: unknown token "' + tokenStr + '"'
- msg += '\tfound on line %d' % tokenLineNo
+ msg += '\tfound on line %d' % tokenLineNo
msg += ' in "' + self.fname + '".\n'
sys.stderr.write(msg)
-
+
yield (tokenId, tokenStr, tokenLineNo)
- # If it is a comment that contains the self.translateMeText
+ # If it is a comment that contains the self.translateMeText
# string, set the flag -- the situation will be reported.
if tokenId == 'comment' and tokenStr.find(self.translateMeText) >= 0:
self.translateMeFlag = True
-
+
tokenId = None
tokenStr = ''
- tokenLineNo = 0
-
+ tokenLineNo = 0
+
# Now process the character. When we just skip it (spaces),
# stay in this status. All characters that will be part of
# some token cause moving to the specific status. And only
- # when moving to the status == 0 (or the final state 777),
+ # when moving to the status == 0 (or the final state 777),
# the token is yielded. With respect to that the automaton
- # behaves as Moore's one (output bound to status). When
- # collecting tokens, the automaton is the Mealy's one
+ # behaves as Moore's one (output bound to status). When
+ # collecting tokens, the automaton is the Mealy's one
# (actions bound to transitions).
if c.isspace():
pass # just skip whitespace characters
@@ -328,7 +329,7 @@ class Transl:
tokenStr = c
tokenLineNo = lineNo
status = 8
- elif tokenDic.has_key(c): # known one-char token
+ elif tokenDic.has_key(c): # known one-char token
tokenId = tokenDic[c]
tokenStr = c
tokenLineNo = lineNo
@@ -338,30 +339,30 @@ class Transl:
tokenStr = c
tokenLineNo = lineNo
status = 333
-
+
pos += 1 # move position in any case
-
+
elif status == 1: # possibly a comment
if c == '/': # ... definitely the C++ comment
tokenId = 'comment'
tokenStr += c
- pos += 1
+ pos += 1
status = 2
elif c == '*': # ... definitely the C comment
tokenId = 'comment'
tokenStr += c
- pos += 1
+ pos += 1
status = 3
else:
status = 0 # unrecognized, don't move pos
-
+
elif status == 2: # inside the C++ comment
if c == '\n': # the end of C++ comment
status = 0 # yield the token
else:
tokenStr += c # collect the C++ comment
pos += 1
-
+
elif status == 3: # inside the C comment
if c == '*': # possibly the end of the C comment
tokenStr += c
@@ -387,7 +388,7 @@ class Transl:
else:
tokenStr += c # collect the preproc
pos += 1
-
+
elif status == 6: # inside the string
if c == '\\': # escaped char inside the string
tokenStr += c
@@ -398,17 +399,17 @@ class Transl:
else:
tokenStr += c # collect the chars of the string
pos += 1
-
+
elif status == 7: # escaped char inside the string
tokenStr += c # collect the char of the string
status = 6
pos += 1
-
+
elif status == 8: # inside the char literal
tokenStr += c # collect the char of the literal
status = 9
pos += 1
-
+
elif status == 9: # end of char literal expected
if c == "'": # ... and found
tokenStr += c
@@ -428,32 +429,32 @@ class Transl:
else:
tokenStr += c # collect
pos += 1
-
- # We should have finished in the final status. If some token
+
+ # We should have finished in the final status. If some token
# have been extracted, yield it first.
assert(status == 777)
if tokenId:
yield (tokenId, tokenStr, tokenLineNo)
tokenId = None
tokenStr = ''
- tokenLineNo = 0
+ tokenLineNo = 0
- # The file content is processed. Close the file. Then always yield
+ # The file content is processed. Close the file. Then always yield
# the eof token.
f.close()
yield ('eof', None, None)
-
-
+
+
def __collectClassInfo(self, tokenIterator):
"""Collect the information about the class and base class.
-
+
The tokens including the opening left curly brace of the class are
consumed."""
-
+
status = 0 # initial state
-
+
while status != 777: # final state
-
+
# Always assume that the previous tokens were processed. Get
# the next one.
tokenId, tokenStr, tokenLineNo = tokenIterator.next()
@@ -462,14 +463,14 @@ class Transl:
if status == 0: # waiting for the 'class' keyword.
if tokenId == 'class':
status = 1
-
+
elif status == 1: # expecting the class identification
if tokenId == 'id':
self.classId = tokenStr
status = 2
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 2: # expecting the curly brace or base class info
if tokenId == 'lcurly':
status = 777 # correctly finished
@@ -477,20 +478,20 @@ class Transl:
status = 3
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 3: # expecting the 'public' in front of base class id
if tokenId == 'public':
status = 4
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 4: # expecting the base class id
if tokenId == 'id':
self.baseClassId = tokenStr
status = 5
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 5: # expecting the curly brace and quitting
if tokenId == 'lcurly':
status = 777 # correctly finished
@@ -498,7 +499,7 @@ class Transl:
pass
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
# Extract the status of the TranslatorXxxx class. The readable form
# will be used in reports the status form is a string that can be
# compared lexically (unified length, padding with zeros, etc.).
@@ -523,13 +524,13 @@ class Transl:
else:
self.readableStatus = 'obsolete'
self.status = '0.0.00'
-
+
# Check whether status was set, or set 'strange'.
- if self.status == None:
+ if self.status == None:
self.status = 'strange'
- if not self.readableStatus:
+ if not self.readableStatus:
self.readableStatus = 'strange'
-
+
# Extract the name of the language and the readable form.
self.lang = self.classId[10:] # without 'Translator'
if self.lang == 'Brazilian':
@@ -538,11 +539,11 @@ class Transl:
self.langReadable = 'Chinese Traditional'
else:
self.langReadable = self.lang
-
+
def __unexpectedToken(self, status, tokenId, tokenLineNo):
"""Reports unexpected token and quits with exit code 1."""
-
+
import inspect
calledFrom = inspect.stack()[1][3]
msg = "\a\nUnexpected token '%s' on the line %d in '%s'.\n"
@@ -550,25 +551,25 @@ class Transl:
msg += 'status = %d in %s()\n' % (status, calledFrom)
sys.stderr.write(msg)
sys.exit(1)
-
-
+
+
def collectPureVirtualPrototypes(self):
"""Returns dictionary 'unified prototype' -> 'full prototype'.
The method is expected to be called only for the translator.h. It
extracts only the pure virtual method and build the dictionary where
key is the unified prototype without argument identifiers."""
-
+
# Prepare empty dictionary that will be returned.
resultDic = {}
-
+
# Start the token generator which parses the class source file.
tokenIterator = self.__tokenGenerator()
# Collect the class and the base class identifiers.
self.__collectClassInfo(tokenIterator)
assert(self.classId == 'Translator')
-
+
# Let's collect readable form of the public virtual pure method
# prototypes in the readable form -- as defined in translator.h.
# Let's collect also unified form of the same prototype that omits
@@ -576,7 +577,7 @@ class Transl:
# identifiers.
prototype = '' # readable prototype (with everything)
uniPrototype = '' # unified prototype (without arg. identifiers)
-
+
# Collect the pure virtual method prototypes. Stop on the closing
# curly brace followed by the semicolon (end of class).
status = 0
@@ -584,21 +585,21 @@ class Transl:
# Loop until the final state 777 is reached. The errors are processed
# immediately. In this implementation, it always quits the application.
- while status != 777:
-
+ while status != 777:
+
# Get the next token.
tokenId, tokenStr, tokenLineNo = tokenIterator.next()
-
+
if status == 0: # waiting for 'public:'
if tokenId == 'public':
status = 1
-
+
elif status == 1: # colon after the 'public'
if tokenId == 'colon':
status = 2
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 2: # waiting for 'virtual'
if tokenId == 'virtual':
prototype = tokenStr # but not to unified prototype
@@ -609,7 +610,7 @@ class Transl:
status = 11 # expected end of class
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 3: # return type of the method expected
if tokenId == 'id':
prototype += ' ' + tokenStr
@@ -619,7 +620,7 @@ class Transl:
status = 4
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 4: # method identifier expected
if tokenId == 'id':
prototype += ' ' + tokenStr
@@ -627,7 +628,7 @@ class Transl:
status = 5
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 5: # left bracket of the argument list expected
if tokenId == 'lpar':
prototype += tokenStr
@@ -635,7 +636,7 @@ class Transl:
status = 6
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 6: # collecting arguments of the method
if tokenId == 'rpar':
prototype += tokenStr
@@ -651,22 +652,22 @@ class Transl:
status = 13
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 7: # assignment expected or left curly brace
if tokenId == 'assign':
status = 8
elif tokenId == 'lcurly':
- curlyCnt = 1 # method body entered
+ curlyCnt = 1 # method body entered
status = 10
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 8: # zero expected
if tokenId == 'num' and tokenStr == '0':
status = 9
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 9: # after semicolon, produce the dic item
if tokenId == 'semic':
assert(not resultDic.has_key(uniPrototype))
@@ -674,7 +675,7 @@ class Transl:
status = 2
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 10: # consuming the body of the method
if tokenId == 'rcurly':
curlyCnt -= 1
@@ -682,13 +683,13 @@ class Transl:
status = 2 # body consumed
elif tokenId == 'lcurly':
curlyCnt += 1
-
+
elif status == 11: # probably the end of class
if tokenId == 'semic':
status = 777
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 12: # type id for argument expected
if tokenId == 'id':
prototype += ' ' + tokenStr
@@ -696,7 +697,7 @@ class Transl:
status = 13
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 13: # namespace qualification or * or & expected
if tokenId == 'colon': # was namespace id
prototype += tokenStr
@@ -712,33 +713,33 @@ class Transl:
status = 17
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 14: # second colon for namespace:: expected
- if tokenId == 'colon':
+ if tokenId == 'colon':
prototype += tokenStr
uniPrototype += tokenStr
status = 15
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 15: # type after namespace:: expected
- if tokenId == 'id':
+ if tokenId == 'id':
prototype += tokenStr
uniPrototype += tokenStr
status = 13
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 16: # argument identifier expected
- if tokenId == 'id':
+ if tokenId == 'id':
prototype += ' ' + tokenStr
# don't put this into unified prototype
status = 17
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 17: # comma or ')' after argument identifier expected
- if tokenId == 'comma':
+ if tokenId == 'comma':
prototype += ', '
uniPrototype += ', '
status = 6
@@ -748,26 +749,26 @@ class Transl:
status = 7
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
# Eat the rest of the source to cause closing the file.
while tokenId != 'eof':
tokenId, tokenStr, tokenLineNo = tokenIterator.next()
-
+
# Return the resulting dictionary with 'uniPrototype -> prototype'.
return resultDic
-
-
- def __collectPublicMethodPrototypes(self, tokenIterator):
+
+
+ def __collectPublicMethodPrototypes(self, tokenIterator):
"""Collects prototypes of public methods and fills self.prototypeDic.
-
- The dictionary is filled by items: uniPrototype -> prototype.
+
+ The dictionary is filled by items: uniPrototype -> prototype.
The method is expected to be called only for TranslatorXxxx classes,
i.e. for the classes that implement translation to some language.
It assumes that the openning curly brace of the class was already
- consumed. The source is consumed until the end of the class.
- The caller should consume the source until the eof to cause closing
+ consumed. The source is consumed until the end of the class.
+ The caller should consume the source until the eof to cause closing
the source file."""
-
+
assert(self.classId != 'Translator')
assert(self.baseClassId != None)
@@ -779,7 +780,7 @@ class Transl:
# TranslatorAdapterBase (states 8 and 9). Argument identifier inside
# method argument lists can be omitted or commented.
#
- # Let's collect readable form of all public method prototypes in
+ # Let's collect readable form of all public method prototypes in
# the readable form -- as defined in the source file.
# Let's collect also unified form of the same prototype that omits
# everything that can be omitted, namely 'virtual' and argument
@@ -788,7 +789,7 @@ class Transl:
uniPrototype = '' # unified prototype (without arg. identifiers)
warning = '' # warning message -- if something special detected
methodId = None # processed method id
-
+
# Collect the method prototypes. Stop on the closing
# curly brace followed by the semicolon (end of class).
status = 0
@@ -796,23 +797,23 @@ class Transl:
# Loop until the final state 777 is reached. The errors are processed
# immediately. In this implementation, it always quits the application.
- while status != 777:
-
+ while status != 777:
+
# Get the next token.
tokenId, tokenStr, tokenLineNo = tokenIterator.next()
-
+
if status == 0: # waiting for 'public:'
if tokenId == 'public':
status = 1
elif tokenId == 'eof': # non-public things until the eof
status = 777
-
+
elif status == 1: # colon after the 'public'
if tokenId == 'colon':
status = 2
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 2: # waiting for 'virtual' (can be omitted)
if tokenId == 'virtual':
prototype = tokenStr # but not to unified prototype
@@ -829,7 +830,7 @@ class Transl:
status = 11 # expected end of class
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 3: # return type of the method expected
if tokenId == 'id':
prototype += ' ' + tokenStr
@@ -837,7 +838,7 @@ class Transl:
status = 4
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 4: # method identifier expected
if tokenId == 'id':
prototype += ' ' + tokenStr
@@ -846,7 +847,7 @@ class Transl:
status = 5
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 5: # left bracket of the argument list expected
if tokenId == 'lpar':
prototype += tokenStr
@@ -854,7 +855,7 @@ class Transl:
status = 6
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 6: # collecting arguments of the method
if tokenId == 'rpar':
prototype += tokenStr
@@ -870,10 +871,10 @@ class Transl:
status = 13
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
- elif status == 7: # left curly brace expected
+
+ elif status == 7: # left curly brace expected
if tokenId == 'lcurly':
- curlyCnt = 1 # method body entered
+ curlyCnt = 1 # method body entered
status = 10
elif tokenId == 'comment':
pass
@@ -882,14 +883,14 @@ class Transl:
status = 8
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 8: # zero expected (TranslatorAdapterBase)
assert(self.classId == 'TranslatorAdapterBase')
if tokenId == 'num' and tokenStr == '0':
status = 9
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 9: # after semicolon (TranslatorAdapterBase)
assert(self.classId == 'TranslatorAdapterBase')
if tokenId == 'semic':
@@ -901,26 +902,36 @@ class Transl:
if tokenId == 'rcurly':
curlyCnt -= 1
if curlyCnt == 0:
- # Insert new dictionary item.
+ # Check for possible copy/paste error when name
+ # of the method was not corrected (i.e. the same
+ # name already exists).
+ if uniPrototype in self.prototypeDic:
+ msg = "'%s' prototype found again (duplicity)\n"
+ msg += "in '%s'.\n" % self.fname
+ msg = msg % uniPrototype
+ sys.stderr.write(msg)
+ assert False
+
assert(not self.prototypeDic.has_key(uniPrototype))
+ # Insert new dictionary item.
self.prototypeDic[uniPrototype] = prototype
status = 2 # body consumed
methodId = None # outside of any method
elif tokenId == 'lcurly':
curlyCnt += 1
-
- # Warn in special case.
+
+ # Warn in special case.
elif methodId == 'trLegendDocs' and tokenId == 'string' \
and tokenStr.find('MAX_DOT_GRAPH_HEIGHT') >= 0:
- self.txtMAX_DOT_GRAPH_HEIGHT_flag = True
-
-
+ self.txtMAX_DOT_GRAPH_HEIGHT_flag = True
+
+
elif status == 11: # probably the end of class
if tokenId == 'semic':
status = 777
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 12: # type id for argument expected
if tokenId == 'id':
prototype += ' ' + tokenStr
@@ -928,7 +939,7 @@ class Transl:
status = 13
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 13: # :: or * or & or id or ) expected
if tokenId == 'colon': # was namespace id
prototype += tokenStr
@@ -948,31 +959,31 @@ class Transl:
prototype += tokenStr
uniPrototype += tokenStr
status = 7
- elif tokenId == 'comma':
+ elif tokenId == 'comma':
prototype += ', '
uniPrototype += ', '
status = 6
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 14: # second colon for namespace:: expected
- if tokenId == 'colon':
+ if tokenId == 'colon':
prototype += tokenStr
uniPrototype += tokenStr
status = 15
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 15: # type after namespace:: expected
- if tokenId == 'id':
+ if tokenId == 'id':
prototype += tokenStr
uniPrototype += tokenStr
status = 13
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 16: # argument identifier or ) expected
- if tokenId == 'id':
+ if tokenId == 'id':
prototype += ' ' + tokenStr
# don't put this into unified prototype
status = 17
@@ -984,9 +995,9 @@ class Transl:
prototype += tokenStr
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
+
elif status == 17: # comma or ')' after argument identifier expected
- if tokenId == 'comma':
+ if tokenId == 'comma':
prototype += ', '
uniPrototype += ', '
status = 6
@@ -996,17 +1007,17 @@ class Transl:
status = 7
else:
self.__unexpectedToken(status, tokenId, tokenLineNo)
-
-
+
+
def collectAdapterPrototypes(self):
"""Returns the dictionary of prototypes implemented by adapters.
-
- It is created to process the translator_adapter.h. The returned
+
+ It is created to process the translator_adapter.h. The returned
dictionary has the form: unifiedPrototype -> (version, classId)
- thus by looking for the prototype, we get the information what is
- the newest (least adapting) adapter that is sufficient for
- implementing the method."""
+ thus by looking for the prototype, we get the information what is
+ the newest (least adapting) adapter that is sufficient for
+ implementing the method."""
# Start the token generator which parses the class source file.
assert(os.path.split(self.fname)[1] == 'translator_adapter.h')
@@ -1017,8 +1028,8 @@ class Transl:
# Create the empty dictionary that will be returned.
adaptDic = {}
-
-
+
+
# Loop through the source of the adapter file until no other adapter
# class is found.
while True:
@@ -1038,16 +1049,16 @@ class Transl:
version += '.' + ('%02d' % int(lst[3]))
else:
version += '.00'
-
+
# Collect the prototypes of implemented public methods.
self.__collectPublicMethodPrototypes(tokenIterator)
-
- # For the required methods, update the dictionary of methods
+
+ # For the required methods, update the dictionary of methods
# implemented by the adapter.
for protoUni in self.prototypeDic:
if reqDic.has_key(protoUni):
# This required method will be marked as implemented
- # by this adapter class. This implementation assumes
+ # by this adapter class. This implementation assumes
# that newer adapters do not reimplement any required
# methods already implemented by older adapters.
assert(not adaptDic.has_key(protoUni))
@@ -1058,16 +1069,16 @@ class Transl:
self.prototypeDic.clear()
self.classId = None
self.baseClassId = None
-
+
except StopIteration:
break
# Return the result dictionary.
- return adaptDic
+ return adaptDic
+
-
def processing(self):
- """Processing of the source file -- only for TranslatorXxxx classes."""
+ """Processing of the source file -- only for TranslatorXxxx classes."""
# Start the token generator which parses the class source file.
tokenIterator = self.__tokenGenerator()
@@ -1076,10 +1087,10 @@ class Transl:
self.__collectClassInfo(tokenIterator)
assert(self.classId != 'Translator')
assert(self.classId[:17] != 'TranslatorAdapter')
-
+
# Collect the prototypes of implemented public methods.
self.__collectPublicMethodPrototypes(tokenIterator)
-
+
# Eat the rest of the source to cause closing the file.
while True:
try:
@@ -1091,13 +1102,13 @@ class Transl:
reqDic = self.manager.requiredMethodsDic
adaptDic = self.manager.adaptMethodsDic
myDic = self.prototypeDic
-
+
# Build the list of obsolete methods.
self.obsoleteMethods = []
for p in myDic:
if not reqDic.has_key(p):
self.obsoleteMethods.append(p)
-
+
# Build the list of missing methods and the list of implemented
# required methods.
self.missingMethods = []
@@ -1114,17 +1125,17 @@ class Transl:
self.isDecodedTranslator = self.classId in self.manager.decodedTranslators
if self.isDecodedTranslator:
self.note = 'Reimplementation using UTF-8 suggested.'
-
+
# Check whether adapter must be used or suggest the newest one.
# Change the status and set the note accordingly.
if self.baseClassId != 'Translator':
- if not self.missingMethods:
+ if not self.missingMethods:
self.note = 'Change the base class to Translator.'
self.status = ''
self.readableStatus = 'up-to-date'
elif self.baseClassId != 'TranslatorEnglish':
# The translator uses some of the adapters.
- # Look at the missing methods and check what adapter
+ # Look at the missing methods and check what adapter
# implements them. Remember the one with the lowest version.
adaptMinVersion = '9.9.99'
adaptMinClass = 'TranslatorAdapter_9_9_99'
@@ -1134,12 +1145,12 @@ class Transl:
if version < adaptMinVersion:
adaptMinVersion = version
adaptMinClass = cls
-
+
# Test against the current status -- preserve the self.status.
- # Possibly, the translator implements enough methods to
+ # Possibly, the translator implements enough methods to
# use some newer adapter.
status = self.status
-
+
# If the version of the used adapter is smaller than
# the required, set the note and update the status as if
# the newer adapter was used.
@@ -1148,40 +1159,40 @@ class Transl:
self.status = adaptMinVersion
self.adaptMinClass = adaptMinClass
self.readableStatus = adaptMinVersion # simplified
-
- # If everything seems OK, some explicit warning flags still could
- # be set.
+
+ # If everything seems OK, some explicit warning flags still could
+ # be set.
if not self.note and self.status == '' and \
(self.translateMeFlag or self.txtMAX_DOT_GRAPH_HEIGHT_flag):
- self.note = ''
+ self.note = ''
if self.translateMeFlag:
self.note += 'The "%s" found in a comment.' % self.translateMeText
if self.note != '':
- self.note += '\n\t\t'
+ self.note += '\n\t\t'
if self.txtMAX_DOT_GRAPH_HEIGHT_flag:
self.note += 'The MAX_DOT_GRAPH_HEIGHT found in trLegendDocs()'
-
- # If everything seems OK, but there are obsolete methods, set
+
+ # If everything seems OK, but there are obsolete methods, set
# the note to clean-up source. This note will be used only when
# the previous code did not set another note (priority).
if not self.note and self.status == '' and self.obsoleteMethods:
self.note = 'Remove the obsolete methods (never used).'
-
-
+
+
def report(self, fout):
"""Returns the report part for the source as a multiline string.
-
+
No output for up-to-date translators without problem."""
-
+
# If there is nothing to report, return immediately.
if self.status == '' and not self.note:
return
-
+
# Report the number of not implemented methods.
fout.write('\n\n\n')
fout.write(self.classId + ' (' + self.baseClassId + ')')
percentImplemented = 100 # init
- allNum = len(self.manager.requiredMethodsDic)
+ allNum = len(self.manager.requiredMethodsDic)
if self.missingMethods:
num = len(self.missingMethods)
percentImplemented = 100 * (allNum - num) / allNum
@@ -1191,11 +1202,11 @@ class Transl:
fout.write('s')
fout.write(' to implement (%d %%)' % (100 * num / allNum))
fout.write('\n' + '-' * len(self.classId))
-
+
# Write the info about the implemented required methods.
fout.write('\n\n Implements %d' % len(self.implementedMethods))
fout.write(' of the required methods (%d %%).' % percentImplemented)
-
+
# Report the missing method, but only when it is not English-based
# translator.
if self.missingMethods and self.status != 'En':
@@ -1219,67 +1230,67 @@ class Transl:
for p in self.implementedMethods:
fout.write('\n ' + reqDic[p])
-
+
def getmtime(self):
"""Returns the last modification time of the source file."""
assert(os.path.isfile(self.fname))
return os.path.getmtime(self.fname)
-
-
+
+
class TrManager:
"""Collects basic info and builds subordinate Transl objects."""
-
+
def __init__(self):
"""Determines paths, creates and initializes structures.
-
- The arguments of the script may explicitly say what languages should
+
+ The arguments of the script may explicitly say what languages should
be processed. Write the two letter identifications that are used
for composing the source filenames, so...
-
+
python translator.py cz
-
+
this will process only translator_cz.h source.
"""
-
+
# Determine the path to the script and its name.
- self.script = os.path.abspath(sys.argv[0])
- self.script_path, self.script_name = os.path.split(self.script)
+ self.script = os.path.abspath(sys.argv[0])
+ self.script_path, self.script_name = os.path.split(self.script)
self.script_path = os.path.abspath(self.script_path)
-
+
# Determine the absolute path to the Doxygen's root subdirectory.
# If DOXYGEN environment variable is not found, the directory is
# determined from the path of the script.
- doxy_default = os.path.join(self.script_path, '..')
+ doxy_default = os.path.join(self.script_path, '..')
self.doxy_path = os.path.abspath(os.getenv('DOXYGEN', doxy_default))
-
+
# Get the explicit arguments of the script.
self.script_argLst = sys.argv[1:]
-
+
# Build the path names based on the Doxygen's root knowledge.
self.doc_path = os.path.join(self.doxy_path, 'doc')
self.src_path = os.path.join(self.doxy_path, 'src')
-
+
# Create the empty dictionary for Transl object identitied by the
- # class identifier of the translator.
+ # class identifier of the translator.
self.__translDic = {}
# Create the None dictionary of required methods. The key is the
# unified prototype, the value is the full prototype. Set inside
# the self.__build().
self.requiredMethodsDic = None
-
+
# Create the empty dictionary that says what method is implemented
# by what adapter.
self.adaptMethodsDic = {}
-
+
# The last modification time will capture the modification of this
- # script, of the translator.h, of the translator_adapter.h (see the
+ # script, of the translator.h, of the translator_adapter.h (see the
# self.__build() for the last two) of all the translator_xx.h files
# and of the template for generating the documentation. So, this
- # time can be compared with modification time of the generated
+ # time can be compared with modification time of the generated
# documentation to decide, whether the doc should be re-generated.
self.lastModificationTime = os.path.getmtime(self.script)
-
+
# Set the names of the translator report text file, of the template
# for generating "Internationalization" document, for the generated
# file itself, and for the maintainers list.
@@ -1287,36 +1298,36 @@ class TrManager:
self.maintainersFileName = 'maintainers.txt'
self.languageTplFileName = 'language.tpl'
self.languageDocFileName = 'language.doc'
-
- # The information about the maintainers will be stored
+
+ # The information about the maintainers will be stored
# in the dictionary with the following name.
self.__maintainersDic = None
-
+
# Define the other used structures and variables for information.
self.langLst = None # including English based
self.supportedLangReadableStr = None # coupled En-based as a note
self.numLang = None # excluding coupled En-based
self.doxVersion = None # Doxygen version
-
+
# Capture the knowledge about translators that are not implemented
# to use UTF-8 internally.
self.decodedTranslators = self.getDecodedTranslators()
-
+
# Build objects where each one is responsible for one translator.
self.__build()
-
-
+
+
def getDecodedTranslators(self):
"""Parses language.cpp to find what translators do not use UTF-8 yet"""
decodedTranslators = []
-
+
# Regular expression to detect the lines like
# theTranslator=new TranslatorDecoder(new TranslatorSwedish);
rex = re.compile(r'^\s*theTranslator\s*=\s*new\s+.*$')
-
+
# Regular expression to get the (optional) TranslatorDecoder and TranslatorXXX
rex2 = re.compile(r'\bTranslator\w+')
-
+
# Parse the lines in the specific source code.
f = open(os.path.join(self.src_path, 'language.cpp'), 'rU')
for line in f:
@@ -1325,20 +1336,20 @@ class TrManager:
if lst[0] == 'TranslatorDecoder':
decodedTranslators.append(lst[1])
f.close()
-
- # Display warning when all translator implementations were converted
+
+ # Display warning when all translator implementations were converted
# to UTF-8.
if len(decodedTranslators) == 0:
print 'This script should be updated. All translators do use UTF-8'
print 'internally. The TranslatorDecoder adapter should be removed'
print 'from the code and its usage should not be checked any more.'
-
+
return decodedTranslators
-
-
+
+
def __build(self):
"""Find the translator files and build the objects for translators."""
-
+
# The translator.h must exist (the Transl object will check it),
# create the object for it and let it build the dictionary of
# required methods.
@@ -1347,7 +1358,7 @@ class TrManager:
tim = tr.getmtime()
if tim > self.lastModificationTime:
self.lastModificationTime = tim
-
+
# The translator_adapter.h must exist (the Transl object will check it),
# create the object for it and store the reference in the dictionary.
tr = Transl(os.path.join(self.src_path, 'translator_adapter.h'), self)
@@ -1355,7 +1366,7 @@ class TrManager:
tim = tr.getmtime()
if tim > self.lastModificationTime:
self.lastModificationTime = tim
-
+
# Create the list of the filenames with language translator sources.
# If the explicit arguments of the script were typed, process only
# those files.
@@ -1365,13 +1376,13 @@ class TrManager:
if not os.path.isfile(os.path.join(self.src_path, fname)):
sys.stderr.write("\a\nFile '%s' not found!\n" % fname)
sys.exit(1)
- else:
+ else:
lst = os.listdir(self.src_path)
lst = filter(lambda x: x[:11] == 'translator_'
and x[-2:] == '.h'
and x != 'translator_adapter.h', lst)
-
- # Build the object for the translator_xx.h files, and process the
+
+ # Build the object for the translator_xx.h files, and process the
# content of the file. Then insert the object to the dictionary
# accessed via classId.
for fname in lst:
@@ -1381,12 +1392,12 @@ class TrManager:
assert(tr.classId != 'Translator')
self.__translDic[tr.classId] = tr
- # Extract the global information of the processed info.
+ # Extract the global information of the processed info.
self.__extractProcessedInfo()
-
-
+
+
def __extractProcessedInfo(self):
- """Build lists and strings of the processed info."""
+ """Build lists and strings of the processed info."""
# Build the auxiliary list with strings compound of the status,
# readable form of the language, and classId.
@@ -1395,7 +1406,7 @@ class TrManager:
assert(obj.classId != 'Translator')
s = obj.status + '|' + obj.langReadable + '|' + obj.classId
statLst.append(s)
-
+
# Sort the list and extract the object identifiers (classId's) for
# the up-to-date translators and English-based translators.
statLst.sort()
@@ -1406,13 +1417,13 @@ class TrManager:
statLst.reverse()
self.adaptIdLst = [x.split('|')[2] for x in statLst if x[0].isdigit()]
- # Build the list of tuples that contain (langReadable, obj).
+ # Build the list of tuples that contain (langReadable, obj).
# Sort it by readable name.
self.langLst = []
for obj in self.__translDic.values():
self.langLst.append((obj.langReadable, obj))
self.langLst.sort(lambda a, b: cmp(a[0], b[0]))
-
+
# Create the list with readable language names. If the language has
# also the English-based version, modify the item by appending
# the note. Number of the supported languages is equal to the length
@@ -1420,28 +1431,28 @@ class TrManager:
langReadableLst = []
for name, obj in self.langLst:
if obj.status == 'En': continue
-
+
# Append the 'En' to the classId to possibly obtain the classId
# of the English-based object. If the object exists, modify the
# name for the readable list of supported languages.
classIdEn = obj.classId + 'En'
if self.__translDic.has_key(classIdEn):
name += ' (+En)'
-
+
# Append the result name of the language, possibly with note.
langReadableLst.append(name)
- # Create the multiline string of readable language names,
+ # Create the multiline string of readable language names,
# with punctuation, wrapped to paragraph.
if len(langReadableLst) == 1:
s = langReadableLst[0]
elif len(langReadableLst) == 2:
- s = ' and '.join(langReadableLst)
- else:
- s = ', '.join(langReadableLst[:-1]) + ', and '
+ s = ' and '.join(langReadableLst)
+ else:
+ s = ', '.join(langReadableLst[:-1]) + ', and '
s += langReadableLst[-1]
-
- self.supportedLangReadableStr = fill(s + '.')
+
+ self.supportedLangReadableStr = fill(s + '.')
# Find the number of the supported languages. The English based
# languages are not counted if the non-English based also exists.
@@ -1451,32 +1462,32 @@ class TrManager:
classId = obj.classId[:-2]
if self.__translDic.has_key(classId):
self.numLang -= 1 # the couple will be counted as one
-
+
# Extract the version of Doxygen.
f = open(os.path.join(self.doxy_path, 'VERSION'))
self.doxVersion = f.readline().strip()
f.close()
-
+
# Update the last modification time.
for tr in self.__translDic.values():
tim = tr.getmtime()
if tim > self.lastModificationTime:
self.lastModificationTime = tim
-
-
+
+
def __getNoTrSourceFilesLst(self):
"""Returns the list of sources to be checked.
-
+
All .cpp files and also .h files that do not declare or define
- the translator methods are included in the list. The file names
+ the translator methods are included in the list. The file names
are searched in doxygen/src directory.
- """
+ """
files = []
for item in os.listdir(self.src_path):
# Split the bare name to get the extension.
name, ext = os.path.splitext(item)
ext = ext.lower()
-
+
# Include only .cpp and .h files (case independent) and exclude
# the files where the checked identifiers are defined.
if ext == '.cpp' or (ext == '.h' and name.find('translator') == -1):
@@ -1484,37 +1495,37 @@ class TrManager:
assert os.path.isfile(fname) # assumes no directory with the ext
files.append(fname) # full name
return files
-
-
+
+
def __removeUsedInFiles(self, fname, dic):
"""Removes items for method identifiers that are found in fname.
-
+
The method reads the content of the file as one string and searches
for all identifiers from dic. The identifiers that were found in
the file are removed from the dictionary.
-
+
Note: If more files is to be checked, the files where most items are
- probably used should be checked first and the resulting reduced
+ probably used should be checked first and the resulting reduced
dictionary should be used for checking the next files (speed up).
"""
lst_in = dic.keys() # identifiers to be searched for
-
+
# Read content of the file as one string.
assert os.path.isfile(fname)
f = open(fname)
cont = f.read()
f.close()
-
+
# Remove the items for identifiers that were found in the file.
while lst_in:
item = lst_in.pop(0)
if cont.find(item) != -1:
del dic[item]
-
+
def __checkForNotUsedTrMethods(self):
"""Returns the dictionary of not used translator methods.
-
+
The method can be called only after self.requiredMethodsDic has been
built. The stripped prototypes are the values, the method identifiers
are the keys.
@@ -1526,45 +1537,45 @@ class TrManager:
ri = prototype.split('(')[0]
identifier = ri.split()[1].strip()
trdic[identifier] = prototype
-
+
# Build the list of source files where translator method identifiers
# can be used.
files = self.__getNoTrSourceFilesLst()
-
+
# Loop through the files and reduce the dictionary of id -> proto.
for fname in files:
self.__removeUsedInFiles(fname, trdic)
-
+
# Return the dictionary of not used translator methods.
return trdic
-
-
+
+
def __emails(self, classId):
"""Returns the list of maintainer emails.
-
+
The method returns the list of e-mail adresses for the translator
class, but only the addresses that were not marked as [xxx]."""
lst = []
for m in self.__maintainersDic[classId]:
if not m[1].startswith('['):
email = m[1]
- email = email.replace(' at ', '@') # Unmangle the mangled e-mail
+ email = email.replace(' at ', '@') # Unmangle the mangled e-mail
email = email.replace(' dot ', '.')
lst.append(email)
return lst
-
-
+
+
def generateTranslatorReport(self):
"""Generates the translator report."""
output = os.path.join(self.doc_path, self.translatorReportFileName)
-
+
# Open the textual report file for the output.
f = open(output, 'w')
# Output the information about the version.
f.write('(' + self.doxVersion + ')\n\n')
-
+
# Output the information about the number of the supported languages
# and the list of the languages, or only the note about the explicitly
# given languages to process.
@@ -1577,30 +1588,30 @@ class TrManager:
f.write(str(self.numLang))
f.write(' languages (sorted alphabetically):\n\n')
f.write(self.supportedLangReadableStr + '\n\n')
-
- # Write the summary about the status of language translators (how
+
+ # Write the summary about the status of language translators (how
# many translators) are up-to-date, etc.
s = 'Of them, %d translators are up-to-date, ' % len(self.upToDateIdLst)
- s += '%d translators are based on some adapter class, ' % len(self.adaptIdLst)
+ s += '%d translators are based on some adapter class, ' % len(self.adaptIdLst)
s += 'and %d are English based.' % len(self.EnBasedIdLst)
f.write(fill(s) + '\n\n')
-
- # The e-mail addresses of the maintainers will be collected to
+
+ # The e-mail addresses of the maintainers will be collected to
# the auxiliary file in the order of translator classes listed
# in the translator report.
fmail = open('mailto.txt', 'w')
-
- # Write the list of up-to-date translator classes.
+
+ # Write the list of up-to-date translator classes.
if self.upToDateIdLst:
s = '''The following translator classes are up-to-date (sorted
alphabetically). This means that they derive from the
Translator class and they implement all %d of the required
- methods. Anyway, there still may be some details listed even
+ methods. Anyway, there still may be some details listed even
for them:'''
- s = s % len(self.requiredMethodsDic)
+ s = s % len(self.requiredMethodsDic)
f.write('-' * 70 + '\n')
f.write(fill(s) + '\n\n')
-
+
mailtoLst = []
for x in self.upToDateIdLst:
obj = self.__translDic[x]
@@ -1609,24 +1620,24 @@ class TrManager:
f.write(' -- ' + obj.note)
f.write('\n')
mailtoLst.extend(self.__emails(obj.classId))
-
+
fmail.write('up-to-date\n')
fmail.write('; '.join(mailtoLst))
# Write the list of the adapter based classes. The very obsolete
# translators that derive from TranslatorEnglish are included.
if self.adaptIdLst:
- s = '''The following translator classes need some maintenance
+ s = '''The following translator classes need some maintenance
(the most obsolete at the end). The other info shows the
estimation of Doxygen version when the class was last
updated and number of methods that must be implemented to
become up-to-date:'''
f.write('\n' + '-' * 70 + '\n')
f.write(fill(s) + '\n\n')
-
+
# Find also whether some adapter classes may be removed.
adaptMinVersion = '9.9.99'
-
+
mailtoLst = []
numRequired = len(self.requiredMethodsDic)
for x in self.adaptIdLst:
@@ -1643,48 +1654,48 @@ class TrManager:
f.write('\n\tNote: ' + obj.note + '\n')
f.write('\n')
mailtoLst.extend(self.__emails(obj.classId)) # to maintainer
-
+
# Check the level of required adapter classes.
if obj.status != '0.0.00' and obj.status < adaptMinVersion:
adaptMinVersion = obj.status
fmail.write('\n\ntranslator based\n')
fmail.write('; '.join(mailtoLst))
-
- # Set the note if some old translator adapters are not needed
+
+ # Set the note if some old translator adapters are not needed
# any more. Do it only when the script is called without arguments,
- # i.e. all languages were checked against the needed translator
- # adapters.
+ # i.e. all languages were checked against the needed translator
+ # adapters.
if not self.script_argLst:
to_remove = {}
for version, adaptClassId in self.adaptMethodsDic.values():
if version < adaptMinVersion:
to_remove[adaptClassId] = True
-
+
if to_remove:
lst = to_remove.keys()
lst.sort()
plural = len(lst) > 1
note = 'Note: The adapter class'
if plural: note += 'es'
- note += ' ' + ', '.join(lst)
- if not plural:
+ note += ' ' + ', '.join(lst)
+ if not plural:
note += ' is'
else:
note += ' are'
note += ' not used and can be removed.'
f.write('\n' + fill(note) + '\n')
-
+
# Write the list of the English-based classes.
if self.EnBasedIdLst:
- s = '''The following translator classes derive directly from the
- TranslatorEnglish. The class identifier has the suffix 'En'
- that says that this is intentional. Usually, there is also
- a non-English based version of the translator for
+ s = '''The following translator classes derive directly from the
+ TranslatorEnglish. The class identifier has the suffix 'En'
+ that says that this is intentional. Usually, there is also
+ a non-English based version of the translator for
the language:'''
f.write('\n' + '-' * 70 + '\n')
f.write(fill(s) + '\n\n')
-
+
for x in self.EnBasedIdLst:
obj = self.__translDic[x]
f.write(' ' + obj.classId)
@@ -1692,7 +1703,7 @@ class TrManager:
if obj.note:
f.write(' -- ' + obj.note)
f.write('\n')
-
+
# Check for not used translator methods and generate warning if found.
# The check is rather time consuming, so it is not done when report
# is restricted to explicitly given language identifiers.
@@ -1700,42 +1711,42 @@ class TrManager:
dic = self.__checkForNotUsedTrMethods()
if dic:
s = '''WARNING: The following translator methods are declared
- in the Translator class but their identifiers do not appear
- in source files. The situation should be checked. The .cpp
- files and .h files excluding the '*translator*' files
- in doxygen/src directory were simply searched for occurence
+ in the Translator class but their identifiers do not appear
+ in source files. The situation should be checked. The .cpp
+ files and .h files excluding the '*translator*' files
+ in doxygen/src directory were simply searched for occurence
of the method identifiers:'''
f.write('\n' + '=' * 70 + '\n')
f.write(fill(s) + '\n\n')
-
+
keys = dic.keys()
keys.sort()
for key in keys:
f.write(' ' + dic[key] + '\n')
f.write('\n')
-
+
# Write the details for the translators.
f.write('\n' + '=' * 70)
f.write('\nDetails for translators (classes sorted alphabetically):\n')
-
+
cls = self.__translDic.keys()
cls.sort()
-
+
for c in cls:
obj = self.__translDic[c]
- assert(obj.classId != 'Translator')
+ assert(obj.classId != 'Translator')
obj.report(f)
-
- # Close the report file and the auxiliary file with e-mails.
+
+ # Close the report file and the auxiliary file with e-mails.
f.close()
fmail.close()
-
-
+
+
def __loadMaintainers(self):
"""Load and process the file with the maintainers.
-
+
Fills the dictionary classId -> [(name, e-mail), ...]."""
-
+
fname = os.path.join(self.doc_path, self.maintainersFileName)
# Include the maintainers file to the group of files checked with
@@ -1743,7 +1754,7 @@ class TrManager:
tim = os.path.getmtime(fname)
if tim > self.lastModificationTime:
self.lastModificationTime = tim
-
+
# Process the content of the maintainers file.
f = codecs.open(fname, 'r', 'utf-8')
inside = False # inside the record for the language
@@ -1754,26 +1765,26 @@ class TrManager:
while lineReady:
line = f.readline() # next line
lineReady = line != '' # when eof, then line == ''
-
+
line = line.strip() # eof should also behave as separator
if line != u'' and line[0] == u'%': # skip the comment line
- continue
-
+ continue
+
if not inside: # if outside of the record
if line != u'': # should be language identifier
classId = line
maintainersLst = []
inside = True
# Otherwise skip empty line that do not act as separator.
-
- else: # if inside the record
+
+ else: # if inside the record
if line == u'': # separator found
inside = False
else:
# If it is the first maintainer, create the empty list.
if not self.__maintainersDic.has_key(classId):
self.__maintainersDic[classId] = []
-
+
# Split the information about the maintainer and append
# the tuple. The address may be prefixed '[unreachable]'
# or whatever '[xxx]'. This will be processed later.
@@ -1783,11 +1794,11 @@ class TrManager:
self.__maintainersDic[classId].append(t)
f.close()
-
+
def generateLanguageDoc(self):
"""Checks the modtime of files and generates language.doc."""
self.__loadMaintainers()
-
+
# Check the last modification time of the template file. It is the
# last file from the group that decide whether the documentation
# should or should not be generated.
@@ -1796,14 +1807,14 @@ class TrManager:
if tim > self.lastModificationTime:
self.lastModificationTime = tim
- # If the generated documentation exists and is newer than any of
- # the source files from the group, do not generate it and quit
+ # If the generated documentation exists and is newer than any of
+ # the source files from the group, do not generate it and quit
# quietly.
fDocName = os.path.join(self.doc_path, self.languageDocFileName)
- if os.path.isfile(fDocName):
+ if os.path.isfile(fDocName):
if os.path.getmtime(fDocName) > self.lastModificationTime:
return
-
+
# The document or does not exist or is older than some of the
# sources. It must be generated again.
#
@@ -1817,15 +1828,15 @@ class TrManager:
assert pos != -1
doctpl = doctpl[pos:]
- # Fill the tplDic by symbols that will be inserted into the
+ # Fill the tplDic by symbols that will be inserted into the
# document template.
tplDic = {}
-
+
s = u'Do not edit this file. It was generated by the %s script.' % self.script_name
tplDic['editnote'] = s
-
+
tplDic['doxVersion'] = self.doxVersion
- tplDic['supportedLangReadableStr'] = self.supportedLangReadableStr
+ tplDic['supportedLangReadableStr'] = self.supportedLangReadableStr
tplDic['translatorReportFileName'] = self.translatorReportFileName
ahref = u'<a href="../doc/' + self.translatorReportFileName
@@ -1833,7 +1844,7 @@ class TrManager:
ahref += u'</code></a>'
tplDic['translatorReportLink'] = ahref
tplDic['numLangStr'] = str(self.numLang)
-
+
# Define templates for HTML table parts of the documentation.
htmlTableTpl = u'''\
\\htmlonly
@@ -1857,7 +1868,7 @@ class TrManager:
</table>
\\endhtmlonly
'''
- htmlTableTpl = dedent(htmlTableTpl)
+ htmlTableTpl = dedent(htmlTableTpl)
htmlTrTpl = u'\n <tr bgcolor="#ffffff">%s\n </tr>'
htmlTdTpl = u'\n <td>%s</td>'
@@ -1868,10 +1879,10 @@ class TrManager:
# Fill the table data elements for one row. The first element
# contains the readable name of the language.
lst = [ htmlTdTpl % obj.langReadable ]
-
- # The next two elements contain the list of maintainers
+
+ # The next two elements contain the list of maintainers
# and the list of their mangled e-mails. For English-based
- # translators that are coupled with the non-English based,
+ # translators that are coupled with the non-English based,
# insert the 'see' note.
mm = None # init -- maintainer
ee = None # init -- e-mail address
@@ -1882,11 +1893,11 @@ class TrManager:
lang = self.__translDic[classId].langReadable
mm = u'see the %s language' % lang
ee = u'&nbsp;'
-
+
if not mm and obj.classId in self.__maintainersDic:
# Build a string of names separated by the HTML break element.
# Special notes used instead of names are highlighted.
- lm = []
+ lm = []
for maintainer in self.__maintainersDic[obj.classId]:
name = maintainer[0]
if name.startswith(u'--'):
@@ -1894,8 +1905,8 @@ class TrManager:
+ name + u'</span>'
lm.append(name)
mm = u'<br/>'.join(lm)
-
- # The marked adresses (they start with the mark '[unreachable]',
+
+ # The marked adresses (they start with the mark '[unreachable]',
# '[resigned]', whatever '[xxx]') will not be displayed at all.
# Only the mark will be used instead.
rexMark = re.compile(ur'(?P<mark>\[.*?\])')
@@ -1906,28 +1917,28 @@ class TrManager:
if m is not None:
address = u'<span style="color: brown">'\
+ m.group(u'mark') + u'</span>'
- le.append(address)
+ le.append(address)
ee = u'<br/>'.join(le)
-
+
# Append the maintainer and e-mail elements.
lst.append(htmlTdTpl % mm)
lst.append(htmlTdTpl % ee)
-
+
# The last element contains the readable form of the status.
lst.append(htmlTdTpl % obj.readableStatus)
-
- # Join the table data to one table row.
+
+ # Join the table data to one table row.
trlst.append(htmlTrTpl % (''.join(lst)))
-
+
# Join the table rows and insert into the template.
htmlTable = htmlTableTpl % (''.join(trlst))
-
+
# Define templates for LaTeX table parts of the documentation.
latexTableTpl = ur'''
\latexonly
\footnotesize
\begin{longtable}{|l|l|l|l|}
- \hline
+ \hline
{\bf Language} & {\bf Maintainer} & {\bf Contact address} & {\bf Status} \\
\hline
%s
@@ -1936,7 +1947,7 @@ class TrManager:
\normalsize
\endlatexonly
'''
- latexTableTpl = dedent(latexTableTpl)
+ latexTableTpl = dedent(latexTableTpl)
latexLineTpl = u'\n' + r' %s & %s & {\tt\tiny %s} & %s \\'
# Loop through transl objects in the order of sorted readable names
@@ -1961,9 +1972,9 @@ class TrManager:
langNE = self.__translDic[classId].langReadable
maintainer = u'see the %s language' % langNE
email = u'~'
-
+
if not maintainer and (obj.classId in self.__maintainersDic):
- lm = [ m[0] for m in self.__maintainersDic[obj.classId] ]
+ lm = [ m[0] for m in self.__maintainersDic[obj.classId] ]
maintainer = maintainers[0][0]
email = maintainers[0][1]
@@ -1976,7 +1987,7 @@ class TrManager:
s = latexLineTpl % (lang, maintainer, email, status)
s = s.replace(u'_', u'\\_')
trlst.append(s)
-
+
# List the other maintainers for the language. Do not set
# lang and status for them.
lang = u'~'
@@ -1987,26 +1998,26 @@ class TrManager:
s = latexLineTpl % (lang, maintainer, email, status)
s = s.replace(u'_', u'\\_')
trlst.append(s)
-
+
# Join the table lines and insert into the template.
latexTable = latexTableTpl % (u''.join(trlst))
-
+
# Put the HTML and LaTeX parts together and define the dic item.
- tplDic['informationTable'] = htmlTable + u'\n' + latexTable
+ tplDic['informationTable'] = htmlTable + u'\n' + latexTable
# Insert the symbols into the document template and write it down.
f = codecs.open(fDocName, 'w', 'utf-8')
f.write(doctpl % tplDic)
f.close()
-
+
if __name__ == '__main__':
-
+
# Create the manager, build the transl objects, and parse the related
# sources.
trMan = TrManager()
# Generate the language.doc.
trMan.generateLanguageDoc()
-
+
# Generate the translator report.
trMan.generateTranslatorReport()