Kaydet (Commit) dd2d55c8 authored tarafından Éric Araujo's avatar Éric Araujo

Remove two unneeded attributes in packaging

üst 4c0b7070
...@@ -24,7 +24,7 @@ API ...@@ -24,7 +24,7 @@ API
prefer_final=False, prefer_source=True, \ prefer_final=False, prefer_source=True, \
hosts=('*',), follow_externals=False, \ hosts=('*',), follow_externals=False, \
mirrors_url=None, mirrors=None, timeout=15, \ mirrors_url=None, mirrors=None, timeout=15, \
mirrors_max_tries=0, verbose=False) mirrors_max_tries=0)
*index_url* is the address of the index to use for requests. *index_url* is the address of the index to use for requests.
......
...@@ -176,7 +176,6 @@ class install_dist(Command): ...@@ -176,7 +176,6 @@ class install_dist(Command):
self.installer = None self.installer = None
self.requested = None self.requested = None
self.no_record = None self.no_record = None
self.no_resources = None
# -- Option finalizing methods ------------------------------------- # -- Option finalizing methods -------------------------------------
# (This is rather more involved than for most commands, # (This is rather more involved than for most commands,
......
...@@ -119,10 +119,9 @@ class Crawler(BaseClient): ...@@ -119,10 +119,9 @@ class Crawler(BaseClient):
def __init__(self, index_url=DEFAULT_SIMPLE_INDEX_URL, prefer_final=False, def __init__(self, index_url=DEFAULT_SIMPLE_INDEX_URL, prefer_final=False,
prefer_source=True, hosts=DEFAULT_HOSTS, prefer_source=True, hosts=DEFAULT_HOSTS,
follow_externals=False, mirrors_url=None, mirrors=None, follow_externals=False, mirrors_url=None, mirrors=None,
timeout=SOCKET_TIMEOUT, mirrors_max_tries=0, verbose=False): timeout=SOCKET_TIMEOUT, mirrors_max_tries=0):
super(Crawler, self).__init__(prefer_final, prefer_source) super(Crawler, self).__init__(prefer_final, prefer_source)
self.follow_externals = follow_externals self.follow_externals = follow_externals
self.verbose = verbose
# mirroring attributes. # mirroring attributes.
parsed = urllib.parse.urlparse(index_url) parsed = urllib.parse.urlparse(index_url)
...@@ -322,9 +321,8 @@ class Crawler(BaseClient): ...@@ -322,9 +321,8 @@ class Crawler(BaseClient):
infos = get_infos_from_url(link, project_name, infos = get_infos_from_url(link, project_name,
is_external=self.index_url not in url) is_external=self.index_url not in url)
except CantParseArchiveName as e: except CantParseArchiveName as e:
if self.verbose: logger.warning(
logger.warning( "version has not been parsed: %s", e)
"version has not been parsed: %s", e)
else: else:
self._register_release(release_info=infos) self._register_release(release_info=infos)
else: else:
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment