summaryrefslogtreecommitdiffstats
path: root/doc
diff options
context:
space:
mode:
authordimitri <dimitri@afe2bf4a-e733-0410-8a33-86f594647bc7>2003-07-25 12:37:34 (GMT)
committerdimitri <dimitri@afe2bf4a-e733-0410-8a33-86f594647bc7>2003-07-25 12:37:34 (GMT)
commit5f27d7a0fcea0154d7d68cfb0812e097bb0deff1 (patch)
tree5648b4ab8fb4ca6051647d47c3e8cf8d621e8782 /doc
parent655c12fb6a950ed1b59bbcc29852ce20e45fdc6a (diff)
downloadDoxygen-5f27d7a0fcea0154d7d68cfb0812e097bb0deff1.zip
Doxygen-5f27d7a0fcea0154d7d68cfb0812e097bb0deff1.tar.gz
Doxygen-5f27d7a0fcea0154d7d68cfb0812e097bb0deff1.tar.bz2
Release-1.3.3
Diffstat (limited to 'doc')
-rw-r--r--doc/Doxyfile1
-rw-r--r--doc/commands.doc11
-rw-r--r--doc/install.doc16
-rw-r--r--doc/language.doc27
-rw-r--r--doc/translator.py1119
5 files changed, 1157 insertions, 17 deletions
diff --git a/doc/Doxyfile b/doc/Doxyfile
index c25878e..acfec34 100644
--- a/doc/Doxyfile
+++ b/doc/Doxyfile
@@ -26,6 +26,7 @@ GENERATE_LATEX = YES
GENERATE_HTML = YES
GENERATE_HTMLHELP = YES
GENERATE_RTF = NO
+GENERATE_XML = NO
ENABLED_SECTIONS = logo_on
ENABLE_PREPROCESSING = NO
CASE_SENSE_NAMES = NO
diff --git a/doc/commands.doc b/doc/commands.doc
index 950d4e5..073e3d9 100644
--- a/doc/commands.doc
+++ b/doc/commands.doc
@@ -391,8 +391,15 @@ doxygen. Unrecognized commands are treated as normal text.
\addindex \\internal
This command writes the message `For internal use only' to the output and
- all text \e after a \c \\internal command until the end of the
- comment block is ignored.
+ all text \e after an \c \\internal command until the end of the
+ comment block or the end of the section (whichever comes first) is
+ marked as "internal".
+
+ If the \\internal command is put inside a section
+ (see for example \ref cmdsection "\\section") all subsection after the
+ command are considered to be internal as well. Only a new section at the
+ same level will be visible again.
+
You can use \ref cfg_internal_docs "INTERNAL_DOCS" in the config file
to show or hide the internal documentation.
diff --git a/doc/install.doc b/doc/install.doc
index 2dbbf51..3dfc2e7 100644
--- a/doc/install.doc
+++ b/doc/install.doc
@@ -376,9 +376,21 @@ perl script in <code>wintools/make.pl</code> a bit.
Let me know what you had to change if you got Doxygen working with another
compiler.
+If you have Visual C++ 6.0, and the source distribution, you can easily
+build doxygen using the project files in the \c wintools directory. If
+you want to build the CVS sources, or want to build from the command line,
+or with another compiler, you have to follow the steps below.
+
+Thomas Baust reported that if you have Visual Studio.NET (2003) then
+you should be aware that there is a problem with the _popen() and _pclose()
+implementation, which currently leaks handles, so if you build doxygen with
+it and use the INPUT_FILTER, you will run to risk of crashing Windows!
+The problem is reported to and confirmed by Microsoft so maybe it will
+fixed in the next service pack.
+
Since Windows comes without all the nice tools that Unix users are
-used to, you need to install a number of these tools before you can compile
-doxygen for Windows.
+used to, you'll need to install a number of these tools before you can compile
+doxygen for Windows from the command-line.
Here is what is required:
<ul>
diff --git a/doc/language.doc b/doc/language.doc
index 9b0e4dd..d3c9db1 100644
--- a/doc/language.doc
+++ b/doc/language.doc
@@ -25,7 +25,7 @@ Doxygen has built-in support for multiple languages. This means
that the text fragments that doxygen generates can be produced in
languages other than English (the default) at configuration time.
-Currently (version 1.3.2), 28 languages
+Currently (version 1.3.2-20030717), 28 languages
are supported (sorted alphabetically):
Brazilian Portuguese, Catalan, Chinese, Chinese Traditional, Croatian,
Czech, Danish, Dutch, English, Finnish,
@@ -55,7 +55,7 @@ when the translator was updated.
<TD>Brazilian Portuguese</TD>
<TD>Fabio "FJTC" Jun Takada Chino</TD>
<TD>chino@NOSPAM.icmc.sc.usp.br</TD>
- <TD>1.3.1</TD>
+ <TD>up-to-date</TD>
</TR>
<TR BGCOLOR="#ffffff">
<TD>Catalan</TD>
@@ -71,9 +71,9 @@ when the translator was updated.
</TR>
<TR BGCOLOR="#ffffff">
<TD>Chinese Traditional</TD>
- <TD>Gary Lee</TD>
- <TD>garylee@NOSPAM.ecosine.com.tw</TD>
- <TD>1.2.16</TD>
+ <TD>Daniel YC Lin<br>Gary Lee</TD>
+ <TD>daniel@NOSPAM.twpda.com<br>garylee@NOSPAM.ecosine.com.tw</TD>
+ <TD>up-to-date</TD>
</TR>
<TR BGCOLOR="#ffffff">
<TD>Croatian</TD>
@@ -145,7 +145,7 @@ when the translator was updated.
<TD>Japanese</TD>
<TD>Ryunosuke Satoh<br>Kenji Nagamatsu</TD>
<TD>sun594@NOSPAM.hotmail.com<br>naga@NOSPAM.joyful.club.ne.jp</TD>
- <TD>strange</TD>
+ <TD>obsolete</TD>
</TR>
<TR BGCOLOR="#ffffff">
<TD>Korean</TD>
@@ -169,7 +169,7 @@ when the translator was updated.
<TD>Portuguese</TD>
<TD>Rui Godinho Lopes</TD>
<TD>ruiglopes@NOSPAM.yahoo.com</TD>
- <TD>1.3.1</TD>
+ <TD>up-to-date</TD>
</TR>
<TR BGCOLOR="#ffffff">
<TD>Romanian</TD>
@@ -187,7 +187,7 @@ when the translator was updated.
<TD>Serbian</TD>
<TD>Dejan Milosavljevic</TD>
<TD>dmilos@NOSPAM.email.com</TD>
- <TD>1.3.1</TD>
+ <TD>up-to-date</TD>
</TR>
<TR BGCOLOR="#ffffff">
<TD>Slovak</TD>
@@ -230,14 +230,15 @@ when the translator was updated.
{\bf Language} & {\bf Maintainer} & {\bf Contact address} & {\bf Status} \\
\hline
\hline
- Brazilian Portuguese & Fabio "FJTC" Jun Takada Chino & {\tt chino@icmc.sc.usp.br} & 1.3.1 \\
+ Brazilian Portuguese & Fabio "FJTC" Jun Takada Chino & {\tt chino@icmc.sc.usp.br} & up-to-date \\
\hline
Catalan & Albert Mora & {\tt amora@iua.upf.es} & 1.2.17 \\
\hline
Chinese & Wei Liu & {\tt liuwei@asiainfo.com} & 1.2.13 \\
& Wang Weihan & {\tt wangweihan@capinfo.com.cn} & \\
\hline
- Chinese Traditional & Gary Lee & {\tt garylee@ecosine.com.tw} & 1.2.16 \\
+ Chinese Traditional & Daniel YC Lin & {\tt daniel@twpda.com} & up-to-date \\
+ & Gary Lee & {\tt garylee@ecosine.com.tw} & \\
\hline
Croatian & Boris Bralo & {\tt boris.bralo@zg.tel.hr} & 1.3.1 \\
\hline
@@ -262,7 +263,7 @@ when the translator was updated.
Italian & Alessandro Falappa & {\tt alessandro@falappa.net} & up-to-date \\
& Ahmed Aldo Faisal & {\tt aaf23@cam.ac.uk} & \\
\hline
- Japanese & Ryunosuke Satoh & {\tt sun594@hotmail.com} & strange \\
+ Japanese & Ryunosuke Satoh & {\tt sun594@hotmail.com} & obsolete \\
& Kenji Nagamatsu & {\tt naga@joyful.club.ne.jp} & \\
\hline
Korean & Richard Kim & {\tt ryk@dspwiz.com} & strange \\
@@ -272,13 +273,13 @@ when the translator was updated.
Polish & Piotr Kaminski & {\tt Piotr.Kaminski@ctm.gdynia.pl} & strange \\
& Grzegorz Kowal & {\tt g\_kowal@poczta.onet.pl} & \\
\hline
- Portuguese & Rui Godinho Lopes & {\tt ruiglopes@yahoo.com} & 1.3.1 \\
+ Portuguese & Rui Godinho Lopes & {\tt ruiglopes@yahoo.com} & up-to-date \\
\hline
Romanian & Alexandru Iosup & {\tt aiosup@yahoo.com} & 1.2.16 \\
\hline
Russian & Alexandr Chelpanov & {\tt cav@cryptopro.ru} & strange \\
\hline
- Serbian & Dejan Milosavljevic & {\tt dmilos@email.com} & 1.3.1 \\
+ Serbian & Dejan Milosavljevic & {\tt dmilos@email.com} & up-to-date \\
\hline
Slovak & Stanislav Kudl\'{a}\v{c} & {\tt skudlac@pobox.sk} & 1.2.18 \\
\hline
diff --git a/doc/translator.py b/doc/translator.py
new file mode 100644
index 0000000..177db97
--- /dev/null
+++ b/doc/translator.py
@@ -0,0 +1,1119 @@
+"""Script to generate reports on translator classes from Doxygen sources.
+
+ The main purpose of the script is to extract the information from sources
+ related to internationalization (the translator classes). It uses the
+ information to generate documentation (language.doc,
+ translator_report.txt) from templates (language.tpl, maintainers.txt).
+
+ Simply run the script without parameters to get the reports and
+ documentation for all supported languages. If you want to generate the
+ translator report only for some languages, pass their codes as arguments
+ to the script. In that case, the language.doc will not be generated.
+ Example:
+
+ python translator.py en nl cz
+
+ Originally, the script was written in Perl and was known as translator.pl.
+ The last Perl version was dated 2002/05/21
+
+ Petr Prikryl (prikrylp@skil.cz)
+"""
+
+# History:
+# --------
+# 2002/05/21
+# - This was the last Perl version.
+# 2003/05/16
+# - If the script is given list of languages, only the translator report
+# is generated and only for those languages.
+#
+################################################################
+
+import os, re, sys
+
+# Global informations are stored in the global 'info' dictionary.
+# This dictionary should be accessed only via GetInfo() because the first
+# call initializes the empty dictionary. (I was too lazy to create
+# a singleton.)
+#
+info = {}
+
+def GetInfo():
+ """Returns reference to the info dictionary.
+
+ If the dictionary is empty, it will be filled with some initial values.
+ """
+
+ global info
+
+ # If the dictionary with globally shared information is empty, then
+ # fill the static values.
+ if not info:
+ # Get the name of the script without the path and the path without name.
+ scriptpath, scriptname = os.path.split(os.path.abspath(sys.argv[0]))
+ info['scriptname'] = scriptname
+
+ # Determine the Doxygen's doc directory. If the DOXYGEN_DOCDIR
+ # environment variable is defined, then it says where the directory
+ # is. If it is not, then it will be directory where this script is
+ # placed.
+ docdir = os.getenv('DOXYGEN_DOCDIR', '*')
+ if docdir == '*':
+ docdir = scriptpath
+
+ docdir = os.path.abspath(docdir)
+ info['docdir'] = docdir
+
+ # Doxygen's root directory is just one above the docdir.
+ doxygenrootdir = os.path.abspath(os.path.join(docdir, '..'))
+ info['doxygenrootdir'] = doxygenrootdir
+
+ # Doxygen's src directory is just below its root.
+ info['srcdir'] = os.path.join(doxygenrootdir, 'src')
+
+ # Doxygen's current version is read from the 'version' file in the root.
+ try:
+ fver = file(os.path.join(doxygenrootdir, 'version'))
+ doxversion = fver.readline().strip()
+ fver.close()
+ info['doxversion'] = doxversion
+ except IOError:
+ info['doxversion'] = 'unknown'
+
+
+ # Names of the template files and other intput files (template for
+ # language.doc and the database of local language maintainers).
+ info['flangtplname'] = 'language.tpl'
+ info['flangtpl'] = os.path.join(docdir, info['flangtplname'])
+ info['fmaintainersname'] = 'maintainers.txt'
+ info['fmaintainers'] = os.path.join(docdir, info['fmaintainersname'])
+
+ # Names of the output files.
+ info['ftranslatortxtname'] = 'translator_rep.txt'
+ info['ftranslatortxt'] = os.path.join(docdir, info['ftranslatortxtname'])
+ info['flangdocname'] = 'language.doc'
+ info['flangdoc'] = os.path.join(docdir, info['flangdocname'])
+
+ # If the script is given one or more arguments, they should be codes
+ # of languages (two letters). Convert them into lower case and
+ # build the list of them. Empty list will be interpreted as the request
+ # for processing all languages.
+ langlist = []
+ if len(sys.argv) > 1:
+ langlist = sys.argv[1:]
+ info['languages'] = langlist
+
+ # Create the dictionary of the required method. Keys will be the unified
+ # method prototypes, values will be True (for easy testing).
+ info['required_methods'] = {}
+
+ return info
+
+
+def CopyTemplateToLanguageDoc():
+ """'flangtpl' + no src --> 'flangdoc'
+
+ The function takes the 'flangtpl' template and generates 'flangdoc'
+ without using information from other sources. This function is called
+ when source files were not found. The marks inside the template are
+ replaced by warning-like explanations that something could not be done
+ because sources were not available. Writes directly to the file, returns
+ nothing.
+
+ If the script was called only for selected languages, the documentation
+ is not generated.
+ """
+
+ # Get the reference to the initialized info dictionary. If the script was
+ # called for selected languages, return immediately.
+ info = GetInfo()
+ if not info['langlist']:
+ return
+
+ # Read the content of the template file.
+ fin = file(info['flangtpl'])
+ cont = fin.read()
+ fin.close()
+
+ # Replace the template marks by some notices.
+ cont = re.sub(r'(?s)<notice>.+?</notice>',
+ """Warning: this file was generated from the %(flangtplname)s template
+ * by the %(scriptname)s script. As doxygen sources were not available
+ * in that time, some information could not be extracted
+ * and inserted into this file.
+ *
+ * Do not edit this file. Edit the above mentioned files!""", cont)
+
+ cont = re.sub(r'(?s)\$version', '%(doxversion)s', cont)
+ cont = re.sub(r'(?s)\$numlang',
+ '<b>[number of supported languages could not be extracted -- no sources]</b>',
+ cont)
+
+ cont = re.sub(r'(?s)\$languages',
+ '<b>[names of languages could not be extracted -- no sources]</b>', cont)
+
+ cont = re.sub(r'(?s)\$information_table',
+ '<b>[Information table could not be extracted -- no sources.]</b>', cont)
+
+ cont = re.sub(r'(?s)\$translator_report_file_name',
+ '%(ftranslatortxt)s <b>[translator report could not be generated -- no sources]</b>',
+ cont)
+
+ cont = re.sub(r'(?s)\$translator_report_link',
+ '<b>[no sources, no link]</b>', cont)
+
+ # Replace the generated marks by the info from the info dictionary.
+ #
+ cont = cont % info
+
+ # Write the template with replacements to the output doc.
+ fout = file(info['flangdoc'], 'w')
+ fout.write(cont)
+ fout.close()
+
+
+def GetPureVirtualFrom(filename):
+ """Returns the list of pure virtual method prototypes from the filename.
+
+ Each method prototype is returned as one string, one line, one list item).
+ The input argument is the full name of the source file."""
+
+ # Read the content of the file to one string and remove C comments,
+ # one line comments, leading text to the first 'virtual' keyword,
+ # text behind the class, and finally empty lines.
+ f = file(filename)
+ cont = f.read()
+ f.close()
+
+ cont = re.sub(r'(?s)/\*.+?\*/', '', cont) # C comments
+ cont = re.sub(r'(?m)//.*$', '', cont) # C++ comments
+ cont = 'virtual ' + re.sub(r'(?s)^.+?virtual\s', '', cont) # prefix
+ cont = re.sub(r'(?s)};.+$', '', cont) # suffix
+ cont = re.sub(r'(?s)\n\s*\n', r'\n', cont) # empty lines
+
+ # Remove the empty implementation of the updateNeededMessage() method
+ # which is to be implemented by adapters only, not by translators.
+ cont = re.sub(r'(?s)\s*virtual\s+QCString\s+updateNeededMessage.+?}.*?\n',
+ '', cont)
+
+ # Erase anything between "=0;" and "virtual". Only the pure virtual
+ # methods will remain. Remove also the text behind the last "= 0;"
+ cont = re.sub(r'(?s)(=\s*0\s*;).*?(?P<vir>virtual)', r'=0;\n\g<vir>', cont)
+ cont = re.sub(r'(?s)^(?P<x>.+=\s*0\s*;).*$', r'\g<x>', cont)
+
+ # Replace all consequent white spaces (including \n) by a single
+ # space. Strip also the leading and the trailing space.
+ cont = re.sub(r'(?s)\s+', ' ', cont)
+ cont = cont.strip()
+
+ # Split internally the string into lines by replacing the '= 0;' by '\n'.
+ # Keep the string stil as one multiline string.
+ cont = re.sub(r'(?s)\s*=\s*0\s*;\s*', r'\n', cont)
+
+ # Remove the keyword 'virtual' because the derived classes may not use it.
+ cont = re.sub(r'(?m)^virtual\s+', '', cont)
+
+ # Split the string to the list of striped lines. Do strip the string
+ # first so that no empty line list item is generated.
+ L = cont.strip().split('\n')
+
+ # Build the list of unified prototypes and return it.
+ return L
+
+def StripArgIdentifiers(prototype):
+ """Returns the method prototype without argument identifiers.
+
+ The goal of the function is to get only the necessary, single form
+ of the method prototype. This way the prototypes from derived classes
+ can be compared one by one with the methods in the base class."""
+
+ # Parse the prototype, and split the string of arguments it can be empty).
+ m = re.match(r'^(?P<prefix>.+?)\((?P<args>.*?)\)(?P<suffix>.*)$', prototype)
+ str_prefix = m.group('prefix')
+ str_args = m.group('args')
+ str_suffix = m.group('suffix')
+ args = str_args.split(',')
+
+ # Each argument will be stripped and put to the new list. Only the types
+ # are important. Using the spaces has to be normalized because people
+ # differ in opinion where to put spaces. Let's prepare regular
+ # expressions for the tasks.
+ rex_type = re.compile(r'''^(?P<type> # name of the type group
+ \s* # there can be spaces behind comma,
+ (const\s+)? # possibly const at the beginning
+ [A-Za-z0-9_:]+ # type identifier can be qualified
+ (\s*[*&])? # could be reference or pointer
+ ) # ... the above is important,
+ .*$''', # the rest contains the identifier
+ re.VERBOSE)
+
+ # People may differ in opinion whether a space should or should not
+ # be written between a type identifier and the '*' or '&' (when
+ # the argument is a pointer or a reference).
+ rex_norm = re.compile(r'\s*(?P<x>[*&])')
+
+ # Strip each of the arguments and put them to the 'stripped' list.
+ # Only the type of the identifier is important. Extract it, normalize
+ # the using of spaces, and append the result to the list of striped
+ # arguments. (Sequence of more than one space is solved later.)
+ stripped = []
+ for arg in args:
+ arg = rex_type.sub(r'\g<type>', arg)
+ arg = rex_norm.sub(r' \g<x>', arg)
+ stripped.append(arg)
+
+ # Join the stripped arguments into one line again, and build the striped
+ # form of the prototype. Remove the duplicit spaces.
+ result = re.sub(r'\s+', ' ',
+ str_prefix + '(' + ', '.join(stripped) + ')' + str_suffix)
+
+ return result
+
+def GetInfoFrom(input_filename):
+ """Returns list of info related to the parsed file.
+
+ GetInfoFrom returns the list of information related to the
+ parsed source file. The input argument is the name of the
+ translator_xx.h file including path.
+
+ The output list contains the following items:
+ - class identifier
+ - base class identifier
+ - method prototypes (each in a separate item)"""
+
+ # Let's open the file and read it into a single string.
+ f = file(input_filename)
+ cont = f.read()
+ f.close()
+
+ # Remove comments and empty lines.
+ cont = re.sub(r'(?m)//.*$', '', cont) # C++ comments
+ cont = re.sub(r'(?s)/\*.+?\*/', '', cont) # C comments
+ cont = re.sub(r'(?s)\n\s*\n', r'\n', cont) # empty lines
+
+ # Extract the class and base class identifiers.
+ rex = re.compile(r'''^.*class\s+
+ (?P<class>Translator\w+?)\s*:
+ \s*public\s+(?P<base>\w+)\b
+ ''', re.VERBOSE | re.DOTALL)
+ m = rex.match(cont)
+ assert(m)
+ assert(m.group('class'))
+ assert(m.group('base'))
+
+ # Put the class and the base class into the output list.
+ result = [m.group('class'), m.group('base')]
+
+ # Remove the opening curly brace. Remove also the first "public:"
+ cont = re.sub(r'(?s)^.*?{', '', cont)
+ cont = re.sub(r'(?s)(^.*\spublic:\s+)?', '', cont)
+
+ # Cut the things after the class.
+ cont = re.sub(r'(?s)}\s*;\s*#endif\s*$', '', cont)
+
+ # Remove the "virtual" keyword, because the derived class is not forced
+ # to use it.
+ cont = re.sub(r'\bvirtual\s+', '', cont)
+
+ # Remove all strings from lines.
+ cont = re.sub(r'(?s)".*?"', '', cont)
+
+ # Remove all bodies of methods
+ while cont.find('{') >= 0:
+ cont = re.sub(r'(?s){[^{}]+?}', '', cont)
+
+ # Remove all private methods, i.e. from "private:" to "public:"
+ # included. Later, remove also all from "private:" to the end.
+ cont = re.sub(r'(?s)\bprivate\s*:.*?\bpublic\s*:', '', cont)
+ cont = re.sub(r'(?s)\bprivate\s*:.*$', '', cont)
+
+ # Some of the translators use conditional compilation where
+ # the branches define the body of the method twice. Remove
+ # the ifdef/endif block content.
+ cont = re.sub(r'(?s)#ifdef.*?#endif', '', cont)
+
+ # Now the string should containt only method prototypes. Let's unify
+ # their format by removing all spaces that are not necessary.
+ cont = re.sub(r'\s+', ' ', cont)
+ cont = re.sub(r'^\s+', '', cont)
+ cont = re.sub(r'\s+$', '', cont)
+
+ # Then let's put all of them on separate lines (one protototype --
+ # one line; no empty lines).
+ cont = re.sub(r'\s+\(', '(', cont)
+ cont = re.sub(r'(?s)\)\s*', ')\n', cont)
+
+ # Split the string, add it to the output list, and return the result.
+ result.extend(cont.strip().split('\n'))
+ return result
+
+
+def GetAdapterClassesInfo(required):
+ """Returns the list of strings with information related to the adapter classes.
+
+ Each one-line string contains the identifier of the adapter class and
+ the number of required methods that are implemented by the adapter.
+
+ The function takes one agument -- the reference to the hash with
+ stripped prototypes of the required methods."""
+
+ # Let's open the file with the translator adapter classes.
+ info = GetInfo()
+ filename = os.path.join(info['srcdir'], 'translator_adapter.h')
+ f = file(filename)
+ cont = f.read()
+ f.close()
+
+ # Remove the preprocessor directives.
+ cont = re.sub(r'(?m)^\s*#\w+.+$', '', cont)
+
+ # Remove comments and empty lines.
+ cont = re.sub(r'(?m)//.*$', '', cont) # C++ comments
+ cont = re.sub(r'(?s)/\*.+?\*/', '', cont) # C comments
+ cont = re.sub(r'(?s)\n\s*\n', r'\n', cont) # empty lines
+
+ # Place delimiters to separate the classes, and remove
+ # the TranslatorAdapterBase class.
+ #
+ cont = re.sub(r'(?s)};\s*class\s+', '<class>', cont)
+ cont = re.sub(r'(?s)class\s+TranslatorAdapterBase\s+.+?<class>', '<class>', cont)
+ cont = re.sub(r'(?s)};', '', cont)
+
+ # Remove the base classes and the beginning of the the class definitions.
+ cont = re.sub(r'(?s)(TranslatorAdapter[_0-9]+)\s*:.+?{\s*(public\s*:)?',
+ '\g<1>', cont)
+
+ # Remove all bodies of methods
+ while cont.find('{') >= 0:
+ cont = re.sub(r'(?s){[^{}]+?}', '', cont)
+
+ # Remove the "virtual" keywords.
+ cont = re.sub(r'(?m)^\s*virtual\s+', '', cont)
+
+ # Remove the empty lines.
+ cont = re.sub(r'(?s)\n\s*\n', '\n', cont)
+
+ # Trim the spaces.
+ cont = re.sub(r'(?m)^\s+', '', cont)
+ cont = re.sub(r'(?m)\s+$', '', cont)
+
+ # Split the string into the lines again.
+ content = cont.split('\n')
+
+ # Now the list contains only two kinds of lines. The first
+ # kind of lines starts with the <class> tag and contains the
+ # identifier of the class. The following lines list the
+ # non-stripped prototypes of implemented methods without the
+ # "virtual" keyword.
+ #
+ # Now we will produce the result by looping through all the
+ # lines and counting the prototypes of the required methods
+ # that are implemented by the adapter class.
+ #
+ cinfo = ''
+ cnt = 0
+ methods = ''
+ result = []
+
+ rex_class = re.compile(r'^<class>(?P<class>\w+)\s*$')
+ for line in content:
+ m = rex_class.match(line)
+ if m:
+ # Another adapter class found.
+ adapter_class = m.group('class')
+
+ # If the cinfo is not empty then it contains partial
+ # information about the previously processed adapter.
+ if cinfo != '':
+ # Complete the cinfo and push it into the result.
+ s = ''
+ if cnt != 1:
+ s = 's'
+ cinfo += '\timplements %2d required method%s...\n' % (cnt, s)
+ result.append(cinfo + methods)
+
+ # Initialize the counter and store the adapter class identifier
+ # in the cinfo.
+ #
+ cinfo = adapter_class
+ cnt = 0;
+ methods = ''
+ else:
+ # The line contains the prototype of the implemented method.
+ # If it is the required method, count it, and add it to the
+ # string of methods.
+ stripped_prototype = StripArgIdentifiers(line)
+
+ if required.has_key(stripped_prototype):
+ cnt += 1
+ methods += ' %s\n' % line
+
+ # If the cinfo is not empty then it contains partial
+ # information about the last processed adapter.
+ if cinfo != '':
+ # Complete the cinfo and push it into the @result.
+ s = ''
+ if cnt != 1:
+ s = 's'
+ cinfo += '\timplements %2d required method%s...\n' % (cnt, s)
+ result.append(cinfo + methods)
+
+ # Return the result list.
+ return result
+
+
+def GetLanguagesInfo(seq_or_dic):
+ """Returns (numlang, langlst, formated langstr).
+
+ The numlang is the number of supported languages. The langlst is the
+ list of pairs like ('TranslatorBrazilian', 'Brazilian Portuguese')
+ """
+
+ languages = []
+ for k in seq_or_dic:
+ # Remove the 'Translator' prefix from the class name to obtain
+ # the brief name of the language.
+ assert(k[:len('Translator')] == 'Translator')
+ lang = k[len('Translator'):]
+
+ # Do correction of the language name for the selected languages.
+ if lang == 'Brazilian': lang = 'Brazilian Portuguese'
+ if lang == 'Chinesetraditional': lang = 'Chinese Traditional'
+
+ # Append the language to the list.
+ languages.append((k, lang))
+
+ # Sort the languages and construct the output string. Insert new line
+ # after each line. Add the 'and' before the last language
+ languages.sort()
+ count = 0
+ output = ''
+ for L in languages:
+ output += L[1] # readable form of the language
+ count += 1
+
+ if count < len(languages) - 1: # separate by comma
+ output += ', '
+ elif count == len(languages) - 1: # separate by comma and 'and'
+ output += ', and '
+
+ if count % 5 == 0:
+ output += '\n'
+
+ return (len(languages), languages, output)
+
+
+def GenerateLanguageDoc(cb):
+ """Generates the content as expected in the 'flangdoc' file.
+
+ GenerateLanguageDoc takes document templates and code sources
+ generates the content as expected in the 'flangdoc' file (the
+ part of the Doxygen documentation), and returns the result as a
+ string.
+
+ The input parameter is the reference to the class/base dictionary."""
+
+ # Define templates for HTML table parts of the documentation.
+ htmlTableHead = r'''\htmlonly
+<TABLE ALIGN=center CELLSPACING=0 CELLPADDING=0 BORDER=0>
+<TR BGCOLOR="#000000">
+<TD>
+ <TABLE CELLSPACING=1 CELLPADDING=2 BORDER=0>
+ <TR BGCOLOR="#4040c0">
+ <TD ><b><font size=+1 color="#ffffff"> Language </font></b></TD>
+ <TD ><b><font size=+1 color="#ffffff"> Maintainer </font></b></TD>
+ <TD ><b><font size=+1 color="#ffffff"> Contact address </font>
+ <font size=-2 color="#ffffff">(remove the NOSPAM.)</font></b></TD>
+ <TD ><b><font size=+1 color="#ffffff"> Status </font></b></TD>
+ </TR>
+'''
+
+ htmlTableRow = r''' <TR BGCOLOR="#ffffff">
+ <TD>$lang</TD>
+ <TD>$maintainer</TD>
+ <TD>$email</TD>
+ <TD>$status</TD>
+ </TR>
+'''
+
+ htmlTableFoot = r''' </TABLE>
+</TD>
+</TR>
+</TABLE>
+\endhtmlonly
+'''
+
+ # Define templates for LaTeX table parts of the documentation.
+ latexTableHead = r'''\latexonly
+\begin{tabular}{|l|l|l|l|}
+ \hline
+ {\bf Language} & {\bf Maintainer} & {\bf Contact address} & {\bf Status} \\
+ \hline
+'''
+
+ latexTableRow = r''' $lang & $maintainer & {\tt $email} & $status \\
+'''
+
+ latexTableFoot = r''' \hline
+\end{tabular}
+\endlatexonly
+'''
+
+ # Read the template of the documentation, and join the content
+ # to a single string.
+ info = GetInfo()
+ filename = os.path.join(info['docdir'], info['flangtpl'])
+ f = file(filename)
+ output = f.read()
+ f.close()
+
+ # Get the number of languages, list of their names and formated string
+ # with the list in human readable English form.
+ (numlang, langlst, langstr) = GetLanguagesInfo(cb)
+
+ # Substitute the marks inside the template.
+ output = re.sub(r'(?s)\$version', info['doxversion'], output)
+ output = re.sub(r'(?s)\$numlang', str(numlang), output)
+ output = re.sub(r'(?s)\$languages', langstr, output)
+
+ # Create the dictionary for info for each language.
+ langinfo = {}
+ for (trClass, langname) in langlst:
+ langinfo[langname] = cb[trClass] + '<msep/>unknown: unknown'
+
+ # Read the information related to maintainers into the
+ # string using suitable separators -- one line, one language. #{{{
+ filename = os.path.join(info['docdir'], info['fmaintainers'])
+ print filename
+ f = file(filename)
+ maintainers = f.read()
+ f.close()
+
+ # Trim the spaces on the lines. Strip the comment lines that
+ # start with % sign.
+ maintainers = re.sub(r'(?m)^[ \t]+', '', maintainers)
+ maintainers = re.sub(r'(?m)[ \t]+$', '', maintainers)
+ maintainers = re.sub(r'(?m)^%.*$', '', maintainers)
+
+ # Join the information for one language into one line,
+ # and remove empty lines.
+ maintainers = re.sub(r'(?s)\b\n\b', '<sep/>', maintainers)
+ maintainers = re.sub(r'(?s)\n{2,}', '\n', maintainers)
+ maintainers = re.sub(r'(?s)^\n+', '', maintainers)
+ maintainers = re.sub(r'(?s)\n+$', '', maintainers)
+
+ # Split the string back to the list, and update the information
+ # in the hash with information for languages.
+ lst = maintainers.split('\n')
+ lst.sort()
+ for line in lst:
+ # Split the line for one language to separate lines for
+ # the language and one or more maintainers. Ensure that the language
+ # starts with uppercase and continues with lowercase. (It will be used
+ # for reconstructing the translator class identifier.)
+ linfo = line.split('<sep/>')
+ lang = linfo[0].capitalize()
+ del linfo[0]
+
+ # Add information to the langinfo dictionary. If the language
+ # was not defined in sources, add the question mark to the
+ # language identifier.
+ #
+ if langinfo.has_key(lang):
+ langinfo[lang] = cb['Translator' + lang] + '<msep/>' + \
+ '<sep/>'.join(linfo)
+ else:
+ lang += ' (?)'
+ langinfo[lang] = 'unknown<msep/>' + '<sep/>'.join(linfo)
+
+ # Now, the langinfo dictionary contains all the information needed for
+ # generating the tables (HTML and LaTeX). Define string variables
+ # for each of the tables, and initialize them.
+ #
+ tableHTML = htmlTableHead
+ tableLATEX = latexTableHead
+
+ # Loop through sorted keys for the languages, parse the
+ # information, and add it to the tables.
+ langs = langinfo.keys()
+ langs.sort()
+ for lang in langs:
+ # Transform the key for the language into more human readable
+ # form. Basically, only languages with two words are going to be
+ # corrected.
+ if lang == 'Brazilian':
+ lang_readable = 'Brazilian Portuguese'
+ elif lang == 'Chinesetraditional':
+ lang_readable = 'Chinese Traditional'
+ else:
+ lang_readable = lang
+
+ print lang, lang_readable
+ """
+
+ # Read the line with info for the language and separate
+ # the status. #{{{
+ #
+ my @list = split(/<msep\/>/, $language{$lang});
+ my $status = shift @list;
+
+ my $i = $status =~ s{^Translator$}{up-to-date};
+
+ if ($i == 0) {
+ $i = $status =~ s{^TranslatorAdapter_(\d)_(\d)_(\d)}
+ {$1.$2.$3}x;
+ }
+
+ if ($i == 0) {
+ $i = $status =~ s{^TranslatorEnglish$}
+ {obsolete}x;
+ }
+
+ if ($i == 0) { $status = 'strange'; }
+
+ ##}}}
+
+ # Split the rest of the list (should be a single item) into
+ # the list with one or more maintainers -- one line, one
+ # maintainer. #{{{
+ #
+ my $rest = shift @list;
+ @list = split(/<sep\/>/, $rest);
+ ##}}}
+
+ # In HTML table, maintainer names are placed in the same
+ # cell. Also their e-mails are placed in a single cell.
+ # Extract the string with concatenated names and the string
+ # with concatenated e-mails. Add the row to the HTML
+ # table. #{{{
+ #
+ my $name = '';
+ my $email = '';
+
+ foreach my $maintainer (@list) {
+
+ if ($name ne '') { $name .= '<br>'; }
+ if ($email ne '') { $email .= '<br>'; }
+
+ $maintainer =~ m{^\s*(.+?)\s*:\s*(.+?)\s*$};
+
+ $name .= $1;
+ $email .= $2;
+ }
+
+ # Prepare the HTML row template, modify it, and add the
+ # result to the HTML table.
+ #
+ my $item = $htmlTableRow;
+
+ $item =~ s{\$lang}{$lang_readable};
+ $item =~ s{\$maintainer}{$name};
+ $item =~ s{\$email}{$email};
+ $item =~ s{\$status}{$status};
+
+ $tableHTML .= $item;
+
+ ##}}}
+
+ # For LaTeX, more maintainers for the same language are
+ # placed on separate rows in the table. The line separator
+ # in the table is placed explicitly above the first
+ # maintainer. Add rows for all maintainers to the LaTeX
+ # table. #{{{
+ #
+ # Prepare the LATEX row template, modify it, and add the
+ # result to the LATEX table.
+ #
+ $item = $latexTableRow;
+
+ my $first = shift @list; # the first maintainer.
+ $first =~ m{^\s*(.+?)\s*:\s*(.+?)\s*$};
+
+ $name = $1;
+ $email = $2;
+
+ $item =~ s{\$lang}{$lang_readable};
+ $item =~ s{\$maintainer}{$name};
+ $item =~ s{\$email}{$email};
+ $item =~ s{\$status}{$status};
+
+ $tableLATEX .= " \\hline\n" . $item;
+
+ # List the other maintainers for the language. Do not set
+ # lang and status for them.
+ #
+ while (@list) {
+ my $next = shift @list;
+ $next =~ m{^\s*(.+?)\s*:\s*(.+?)\s*$};
+
+ my $name = $1;
+ my $email = $2;
+ my $item = $latexTableRow;
+
+ $item =~ s{\$lang}{};
+ $item =~ s{\$maintainer}{$name};
+ $item =~ s{\$email}{$email};
+ $item =~ s{\$status}{};
+
+ $tableLATEX .= $item;
+ }
+ ##}}}
+ }
+ ##}}}
+
+ # Finish the tables, and substitute the mark in the doc
+ # template by the concatenation of the tables. Add NOSPAM to
+ # email addresses in the HTML table. Replace the special
+ # character sequences. #{{{
+ #
+ $tableHTML .= $htmlTableFoot;
+ $tableLATEX .= $latexTableFoot;
+
+ $tableHTML =~ s{@}{\@NOSPAM.}sg;
+ $tableHTML =~ s{&ccaron;}{&#x010d;}sg;
+ $tableHTML =~ s{&rcaron;}{&#x0159;}sg;
+
+ $tableLATEX =~ s/&aacute;/\\'{a}/sg;
+ $tableLATEX =~ s/&auml;/\\"{a}/sg;
+ $tableLATEX =~ s/&ouml;/\\"{o}/sg;
+ $tableLATEX =~ s/&oslash;/\\o{}/sg;
+ $tableLATEX =~ s/&ccaron;/\\v{c}/sg;
+ $tableLATEX =~ s/&rcaron;/\\v{r}/sg;
+ $tableLATEX =~ s/_/\\_/sg;
+
+ $output =~ s{\$information_table}{$tableHTML$tableLATEX};
+
+ ##}}}
+
+ # Replace the other symbols in the template by the expected
+ # information. ##{{{
+ #
+ $output =~ s{\$version}{$doxversion};
+
+ $output =~ s{\$translator_report_file_name}
+ {<code>doxygen/doc/$ftranslatortxt</code>}x;
+
+ $output =~ s{\$translator_report_link}
+ {<a href=\"../doc/$ftranslatortxt\">
+ <code>doxygen/doc/$ftranslatortxt</code></a>}x;
+ ##}}}
+
+ # Replace the introduction notice in the output. #{{{
+ #
+ $output =~ s{<notice>.+?</notice>}
+{Warning: this file was generated from the $flangtpl template
+ * and the $fmaintainers files by the $0 script.
+ *
+ * Do not edit this file. Edit the above mentioned files!}sx;
+ ##}}}
+
+ # Return the content of the generated output file.
+ #
+ return $output;
+}
+"""
+ return output
+
+
+################################################################# Body
+if __name__ == '__main__':
+ # Get the reference to the initialized dictionary with the shared info.
+ info = GetInfo()
+
+ # File with the translator base class must be present. Exit otherwise,
+ # but be kind to those who already have the documentation
+ # generated by this script ready, but who do not have sources.
+ # If no 'flangdoc' is present, copy the template to it.
+ if not os.path.isfile(os.path.join(info['srcdir'], 'translator.h')):
+ sys.stderr.write(('\nThe %(scriptname)s warning:\n' +
+ '\tThe translator.h not found in %(srcdir)s.\n' +
+ '\tThe %(ftranslatortxtname)s will not be ' +
+ "generated (you don't need it).\n") % info)
+
+ # $flangdoc is not present, copy the template to it, and do the simplified
+ # replacement of the markers inside the template. Generate the warning
+ # about 'flangdoc' content.
+ if not os.path.isfile(info['flangdoc']):
+ CopyTemplateToLanguageDoc();
+ sys.stderr.write(('\nThe %(scriptname)s warning:\n' +
+ "\tThe %(flangdoc)s not found in the '%(docdir)s' directory.\n" +
+ '\tThe %(flangtpl)s template content copied into it.\n' +
+ '\tAs the sources are not available, some information\n' +
+ '\tcould not be extracted and inserted into %(flangdoc)s.\n') % info)
+
+ # Exit as if nothing happened.
+ sys.exit(0)
+
+ # Create the list of all translator_xxxx.h file names. If the script was
+ # given a list of languages (two letters for each language), then fill
+ # the list only by the translator files for the languages.
+ directory = info['srcdir']
+ langlist = info['languages']
+ if langlist:
+ files = [os.path.join(directory, 'translator_%s.h' % lang)
+ for lang in langlist
+ if os.path.isfile(os.path.join(directory,
+ 'translator_%s.h' % lang))]
+ else:
+ rex_tr = re.compile(r'^translator_\w\w\w?\.h$', re.IGNORECASE)
+ files = [os.path.join(directory, f)
+ for f in os.listdir(directory)
+ if os.path.isfile(os.path.join(directory, f))
+ and rex_tr.match(f)]
+
+ # Get only the pure virtual methods from the Translator class
+ # into a list for later testing present/not present.
+ expected_lst = GetPureVirtualFrom(os.path.join(info['srcdir'],
+ 'translator.h'))
+ # Fill the 'required_methods' dictionary for unified form
+ # of the prototypes.
+ required = info['required_methods']
+ for method in expected_lst:
+ prototype = StripArgIdentifiers(method)
+ required[prototype] = True
+
+ # The details for translators will be collected into the output
+ # string. If some details are listed for a translator, the flag
+ # will be set to produce possible warning to the list of
+ # up-to-date translators.
+ output = ''
+ details = {}
+
+ # Collect base classes of translators in the hash. CB stands
+ # for Class and Base.
+ cb = {}
+
+ # Loop through all translator files. Extract the implemented
+ # virtual methods and compare it with the requirements. Prepare
+ # the output.
+ rex_trAdapter = re.compile(r'^TranslatorAdapter_')
+ rex_trEN = re.compile(r'^TranslatorEnglish$')
+
+ for filename in files:
+ # Get the information from the sources. Remember the base
+ # class for each of the classes. Clear the flag for
+ # details for the class.
+ finfo = GetInfoFrom(filename)
+ (class_, base_) = finfo[0:2]
+ cb[class_] = base_
+ details[class_] = False
+
+ # Set the value of the required methods to 1 (true). Let
+ # this indicate that the method was not defined in the
+ # translator class.
+ for method in required:
+ required[method] = True
+
+ # Loop through all items and compare the prototypes. Mark
+ # the implemented method and collect the old ones.
+ old_methods = []
+ for method in finfo[2:]:
+ # Get only the necessary form of the prototype.
+ prototype = StripArgIdentifiers(method)
+
+ # Mark as recognized when the prototype is required.
+ # Otherwise, remember it as old method which is
+ # implemented, but not required.
+ if (required.has_key(prototype)):
+ required[prototype] = False # satisfaction
+ else:
+ old_methods.append(method)
+
+ # Loop through the list of expected methods and collect
+ # the missing (new) methods. Do this only when it derives
+ # from Translator or TranslatorAdapter classes (i.e. ignore
+ # any unusual kind of TranslatorXxxx implementation).
+ # Accept also deriving from TranslatorEnglish, that can
+ # be done by doxygen developers to solve problems with
+ # some really outdated translators.
+ missing_methods = []
+ if rex_trAdapter.match(base_) or rex_trEN.match(base_):
+
+ for method in expected_lst:
+ # Get the stripped version of the prototype.
+ prototype = StripArgIdentifiers(method)
+
+ # If the prototype is stored in the required
+ # table, and if it was not marked as implemented,
+ # then it should be. It is a missing method.
+ #try:
+ if required[prototype]:
+ missing_methods.append(method)
+
+ # The detailed output will be produced only when it is needed.
+ if old_methods or missing_methods or rex_trAdapter.match(base_):
+ output += '\n\n\n'
+ output += '%s (%s)\n%s\n' % (class_, base_, '-' * len(class_))
+
+ if rex_trEN.match(base_):
+ output += '''
+This translator is implemented via deriving from the English translator.
+This should be done only in the case when the language maintainer
+or the doxygen developers need to update some really old-dated translator.
+Otherwise, deriving from the translator adapter classes should be used
+for obsolete translators. If you still want some texts to be in English
+copy the sources of the English translator.
+
+The obsolete and missing method lists (below) reflect what have to be done
+to derive directly from the Translator class (i.e. to reach up-to-date status).
+'''
+
+ elif not rex_trAdapter.match(base_):
+ output += '''
+This is some unusual implementation of the translator class. It is derived
+from the %s base class. The usual translator class derives
+or from the Translator class or from some TranslatorAdapter_x_x_x classes.
+Because of that, nothing can be guessed about missing or obsolete methods.
+''' % base_
+
+ if missing_methods:
+ output += '\nMissing methods (should be implemented):\n\n'
+ for m in missing_methods:
+ output += ' ' + m + '\n'
+
+ if old_methods:
+ output += '\nObsolete methods (should be removed):\n\n'
+ old_methods.sort()
+ for m in old_methods:
+ output += ' ' + m + '\n'
+
+ # Some details were listed, set the details flag for the class.
+ details[class_] = 1;
+
+
+ # Generate the ASCII output file.
+ fout_name = info['ftranslatortxt']
+
+ # Open it first, and output the version information.
+ fout = file(fout_name, 'w')
+ fout.write('(version %s)\n\n' % info['doxversion'])
+
+ # List the supported languages.
+ (numlang, langlst, langstr) = GetLanguagesInfo(cb)
+ fout.write('Doxygen supports the following (' + str(numlang) +
+ ') languages (sorted alphabetically):\n\n')
+ fout.write(langstr + '.\n')
+
+ # If there are up-to-date translators, list them.
+ L = [k for k in cb if cb[k] == 'Translator']
+ L.sort()
+
+ if L:
+ fout.write('\n' + '-' * 70 + '\n')
+ fout.write('''The \
+following translator classes are up-to-date (sorted alphabetically).
+This means that they derive from the Translator class. Anyway, there still
+may be some details listed even for the up-to-date translators.
+Please, check the text below if the translator is marked so.
+
+''')
+ for tr in L:
+ # Print the class name. If some details were listed for
+ # the translator class, add a notice.
+ fout.write(' ' + tr)
+ if details[tr]:
+ fout.write('\t-- see details below in the report')
+ fout.write('\n')
+
+ # If there are obsolete translators, list them.
+ L = [k for k in cb if rex_trAdapter.match(cb[k])]
+ L.sort()
+
+ if L:
+ fout.write('\n' + '-' * 70 + '\n')
+ fout.write('''The \
+following translator classes are obsolete (sorted alphabetically).
+This means that they derive from some of the adapter classes.
+
+''')
+ for tr in L:
+ fout.write(' %s\t(%s)\n' % (tr, cb[tr]))
+
+ # If there are translators derived from TranslatorEnglish, list them
+ # and name them as really obsolete.
+ L = [k for k in cb if cb[k] == 'TranslatorEnglish']
+ L.sort()
+ if L:
+ fout.write('\n' + '-' * 70 + '\n')
+ fout.write('''The \
+following translator classes are implemented via deriving
+from the English translator. This should be done only in the case
+when the language maintainer or the doxygen developers need to update
+some really outdated translator. Otherwise, deriving from
+the translator adapter classes should be prefered for obsolete translators.
+See details below in the report.
+
+''')
+ for tr in L:
+ fout.write(' %s\t(%s)\n' % (tr, cb[tr]))
+
+ # If there are other translators, list them. #{{{
+ #
+ L = [k for k in cb
+ if not rex_trAdapter.match(cb[k])
+ and cb[k] != 'TranslatorEnglish'
+ and cb[k] != 'Translator'
+ ]
+ L.sort()
+
+ if L:
+ fout.write('\n' + '-' * 70 + '\n')
+ fout.write('''The \
+following translator classes are somehow different
+(sorted alphabetically). This means that they do not derive from
+the Translator class, nor from some of the adapter classes,
+nor from the TranslatorEnglish. Nothing can be guessed about the methods.
+
+''')
+ for tr in L:
+ fout.write(' %s\t(%s)\n' % (tr, cb[tr]))
+
+ # List all the translator adapter classes to show for which versions
+ # the adapters had to be created. Show, how many and what new methods
+ # are implemented by the adapters.
+ #
+ fout.write('\n' + '-' * 70 + '\n')
+ fout.write('''The \
+following translator adapter classes are implemented -- the older (with
+lower number) are always derived from the newer. They implement the
+listed required methods. Notice that some versions of doxygen did not
+introduce any changes related to the language translators. From here you may
+guess how much work should be done to update your translator:
+
+''')
+ adapter_info = GetAdapterClassesInfo(required)
+
+ for ad in adapter_info:
+ fout.write(' %s\n' % ad)
+
+ # List the methods that are expected to be implemented.
+ fout.write('\n' + '-' * 70 + '\n')
+ fout.write('''Localized \
+translators are expected to implement the following methods
+(prototypes sorted aplhabetically):
+
+''')
+
+ expected_lst.sort()
+ for m in expected_lst:
+ fout.write('%s\n' % m)
+
+ # If there are some details for the translators, show them.
+ if output != '':
+ fout.write('\n\n' + '=' * 70 + '\n')
+ fout.write('Details related to specific translator classes follow.\n')
+ fout.write(output + '\n')
+
+ # Close the ASCII output file
+ fout.close()
+
+ # Generate the same using the original perl script.
+ os.system('translator.pl')
+
+ # Generate the language.doc file.
+ filename = os.path.join(info['docdir'], info['flangdoc'])
+ f = file(filename, 'w')
+ f.write(GenerateLanguageDoc(cb))
+ f.close()
+
+ sys.exit(0)