diff options
author | Éric Araujo <merwok@netwok.org> | 2011-09-21 14:28:03 (GMT) |
---|---|---|
committer | Éric Araujo <merwok@netwok.org> | 2011-09-21 14:28:03 (GMT) |
commit | dd2d55c8a145deffb521724e7cb159b453eeb6b8 (patch) | |
tree | 47c7689c78be49f88864c1077b7ce7e75d0fb154 | |
parent | 4c0b7070ce1461fd54059d1fdc435ab89bdb6555 (diff) | |
download | cpython-dd2d55c8a145deffb521724e7cb159b453eeb6b8.zip cpython-dd2d55c8a145deffb521724e7cb159b453eeb6b8.tar.gz cpython-dd2d55c8a145deffb521724e7cb159b453eeb6b8.tar.bz2 |
Remove two unneeded attributes in packaging
-rw-r--r-- | Doc/library/packaging.pypi.simple.rst | 2 | ||||
-rw-r--r-- | Lib/packaging/command/install_dist.py | 1 | ||||
-rw-r--r-- | Lib/packaging/pypi/simple.py | 8 |
3 files changed, 4 insertions, 7 deletions
diff --git a/Doc/library/packaging.pypi.simple.rst b/Doc/library/packaging.pypi.simple.rst index 92b3270..9153738 100644 --- a/Doc/library/packaging.pypi.simple.rst +++ b/Doc/library/packaging.pypi.simple.rst @@ -24,7 +24,7 @@ API prefer_final=False, prefer_source=True, \ hosts=('*',), follow_externals=False, \ mirrors_url=None, mirrors=None, timeout=15, \ - mirrors_max_tries=0, verbose=False) + mirrors_max_tries=0) *index_url* is the address of the index to use for requests. diff --git a/Lib/packaging/command/install_dist.py b/Lib/packaging/command/install_dist.py index 4fbca7e..ce7015e 100644 --- a/Lib/packaging/command/install_dist.py +++ b/Lib/packaging/command/install_dist.py @@ -176,7 +176,6 @@ class install_dist(Command): self.installer = None self.requested = None self.no_record = None - self.no_resources = None # -- Option finalizing methods ------------------------------------- # (This is rather more involved than for most commands, diff --git a/Lib/packaging/pypi/simple.py b/Lib/packaging/pypi/simple.py index 9faabaa..e01e033 100644 --- a/Lib/packaging/pypi/simple.py +++ b/Lib/packaging/pypi/simple.py @@ -119,10 +119,9 @@ class Crawler(BaseClient): def __init__(self, index_url=DEFAULT_SIMPLE_INDEX_URL, prefer_final=False, prefer_source=True, hosts=DEFAULT_HOSTS, follow_externals=False, mirrors_url=None, mirrors=None, - timeout=SOCKET_TIMEOUT, mirrors_max_tries=0, verbose=False): + timeout=SOCKET_TIMEOUT, mirrors_max_tries=0): super(Crawler, self).__init__(prefer_final, prefer_source) self.follow_externals = follow_externals - self.verbose = verbose # mirroring attributes. parsed = urllib.parse.urlparse(index_url) @@ -322,9 +321,8 @@ class Crawler(BaseClient): infos = get_infos_from_url(link, project_name, is_external=self.index_url not in url) except CantParseArchiveName as e: - if self.verbose: - logger.warning( - "version has not been parsed: %s", e) + logger.warning( + "version has not been parsed: %s", e) else: self._register_release(release_info=infos) else: |