summaryrefslogtreecommitdiffstats
path: root/Doc/liburllib.tex
diff options
context:
space:
mode:
authorFred Drake <fdrake@acm.org>1997-12-15 22:13:50 (GMT)
committerFred Drake <fdrake@acm.org>1997-12-15 22:13:50 (GMT)
commitf5eaa2efb8b0178c5750e69ca99bbdb208a25c29 (patch)
treeb9855e7ce5c591d0de6d731e20303d3b0a2d5aae /Doc/liburllib.tex
parenta51f5a48e599a7c35c0667f5d48b0f93ddb48e35 (diff)
downloadcpython-f5eaa2efb8b0178c5750e69ca99bbdb208a25c29.zip
cpython-f5eaa2efb8b0178c5750e69ca99bbdb208a25c29.tar.gz
cpython-f5eaa2efb8b0178c5750e69ca99bbdb208a25c29.tar.bz2
Fixed index references to modules.
Added new index entry for mimetools module.
Diffstat (limited to 'Doc/liburllib.tex')
-rw-r--r--Doc/liburllib.tex17
1 files changed, 9 insertions, 8 deletions
diff --git a/Doc/liburllib.tex b/Doc/liburllib.tex
index 51a700a..5f8297d 100644
--- a/Doc/liburllib.tex
+++ b/Doc/liburllib.tex
@@ -8,13 +8,13 @@
\renewcommand{\indexsubitem}{(in module urllib)}
This module provides a high-level interface for fetching data across
-the World-Wide Web. In particular, the \code{urlopen} function is
-similar to the built-in function \code{open}, but accepts URLs
+the World-Wide Web. In particular, the \code{urlopen()} function is
+similar to the built-in function \code{open()}, but accepts URLs
(Universal Resource Locators) instead of filenames. Some restrictions
apply --- it can only open URLs for reading, and no seek operations
are available.
-it defines the following public functions:
+It defines the following public functions:
\begin{funcdesc}{urlopen}{url}
Open a network object denoted by a URL for reading. If the URL does
@@ -36,6 +36,7 @@ The \code{info()} method returns an instance of the class
if the protocol uses such headers (currently the only supported
protocol that uses this is HTTP). See the description of the
\code{mimetools} module.
+\refstmodindex{mimetools}
\end{funcdesc}
\begin{funcdesc}{urlretrieve}{url}
@@ -118,15 +119,15 @@ looking at the \code{Content-type} header. For the Gopher protocol,
type information is encoded in the URL; there is currently no easy way
to extract it. If the returned data is HTML, you can use the module
\code{htmllib} to parse it.
-\index{HTML}
-\index{HTTP}
-\index{Gopher}
-\stmodindex{htmllib}
+\index{HTML}%
+\index{HTTP}%
+\index{Gopher}%
+\refstmodindex{htmllib}
\item
Although the \code{urllib} module contains (undocumented) routines to
parse and unparse URL strings, the recommended interface for URL
manipulation is in module \code{urlparse}.
-\stmodindex{urlparse}
+\refstmodindex{urlparse}
\end{itemize}