Commit 8b2c4909 authored by Jason R. Coombs's avatar Jason R. Coombs Committed by GitHub

Merge pull request #821 from stepshal/inline

Fix spacing after comment hash.
parents 7d984bbd df1bd4e1
...@@ -145,7 +145,7 @@ def get_module_constant(module, symbol, default=-1, paths=None): ...@@ -145,7 +145,7 @@ def get_module_constant(module, symbol, default=-1, paths=None):
try: try:
if kind == PY_COMPILED: if kind == PY_COMPILED:
f.read(8) # skip magic & date f.read(8) # skip magic & date
code = marshal.load(f) code = marshal.load(f)
elif kind == PY_FROZEN: elif kind == PY_FROZEN:
code = imp.get_frozen_object(module) code = imp.get_frozen_object(module)
......
...@@ -412,7 +412,7 @@ class Distribution(_Distribution): ...@@ -412,7 +412,7 @@ class Distribution(_Distribution):
) )
for key in list(opts): for key in list(opts):
if key not in keep: if key not in keep:
del opts[key] # don't use any other settings del opts[key] # don't use any other settings
if self.dependency_links: if self.dependency_links:
links = self.dependency_links[:] links = self.dependency_links[:]
if 'find_links' in opts: if 'find_links' in opts:
...@@ -650,7 +650,7 @@ class Distribution(_Distribution): ...@@ -650,7 +650,7 @@ class Distribution(_Distribution):
aliases = self.get_option_dict('aliases') aliases = self.get_option_dict('aliases')
while command in aliases: while command in aliases:
src, alias = aliases[command] src, alias = aliases[command]
del aliases[command] # ensure each alias can expand only once! del aliases[command] # ensure each alias can expand only once!
import shlex import shlex
args[:1] = shlex.split(alias, True) args[:1] = shlex.split(alias, True)
command = args[0] command = args[0]
......
...@@ -88,7 +88,7 @@ def egg_info_for_url(url): ...@@ -88,7 +88,7 @@ def egg_info_for_url(url):
parts = urllib.parse.urlparse(url) parts = urllib.parse.urlparse(url)
scheme, server, path, parameters, query, fragment = parts scheme, server, path, parameters, query, fragment = parts
base = urllib.parse.unquote(path.split('/')[-1]) base = urllib.parse.unquote(path.split('/')[-1])
if server == 'sourceforge.net' and base == 'download': # XXX Yuck if server == 'sourceforge.net' and base == 'download': # XXX Yuck
base = urllib.parse.unquote(path.split('/')[-2]) base = urllib.parse.unquote(path.split('/')[-2])
if '#' in base: if '#' in base:
base, fragment = base.split('#', 1) base, fragment = base.split('#', 1)
...@@ -112,7 +112,7 @@ def distros_for_url(url, metadata=None): ...@@ -112,7 +112,7 @@ def distros_for_url(url, metadata=None):
def distros_for_location(location, basename, metadata=None): def distros_for_location(location, basename, metadata=None):
"""Yield egg or source distribution objects based on basename""" """Yield egg or source distribution objects based on basename"""
if basename.endswith('.egg.zip'): if basename.endswith('.egg.zip'):
basename = basename[:-4] # strip the .zip basename = basename[:-4] # strip the .zip
if basename.endswith('.egg') and '-' in basename: if basename.endswith('.egg') and '-' in basename:
# only one, unambiguous interpretation # only one, unambiguous interpretation
return [Distribution.from_location(location, basename, metadata)] return [Distribution.from_location(location, basename, metadata)]
...@@ -334,17 +334,17 @@ class PackageIndex(Environment): ...@@ -334,17 +334,17 @@ class PackageIndex(Environment):
return return
self.info("Reading %s", url) self.info("Reading %s", url)
self.fetched_urls[url] = True # prevent multiple fetch attempts self.fetched_urls[url] = True # prevent multiple fetch attempts
tmpl = "Download error on %s: %%s -- Some packages may not be found!" tmpl = "Download error on %s: %%s -- Some packages may not be found!"
f = self.open_url(url, tmpl % url) f = self.open_url(url, tmpl % url)
if f is None: if f is None:
return return
self.fetched_urls[f.url] = True self.fetched_urls[f.url] = True
if 'html' not in f.headers.get('content-type', '').lower(): if 'html' not in f.headers.get('content-type', '').lower():
f.close() # not html, we can't process it f.close() # not html, we can't process it
return return
base = f.url # handle redirects base = f.url # handle redirects
page = f.read() page = f.read()
if not isinstance(page, str): # We are in Python 3 and got bytes. We want str. if not isinstance(page, str): # We are in Python 3 and got bytes. We want str.
if isinstance(f, urllib.error.HTTPError): if isinstance(f, urllib.error.HTTPError):
...@@ -438,7 +438,7 @@ class PackageIndex(Environment): ...@@ -438,7 +438,7 @@ class PackageIndex(Environment):
except ValueError: except ValueError:
pass pass
pkg, ver = scan(url) # ensure this page is in the page index pkg, ver = scan(url) # ensure this page is in the page index
if pkg: if pkg:
# process individual package page # process individual package page
for new_url in find_external_links(url, page): for new_url in find_external_links(url, page):
...@@ -455,7 +455,7 @@ class PackageIndex(Environment): ...@@ -455,7 +455,7 @@ class PackageIndex(Environment):
lambda m: '<a href="%s#md5=%s">%s</a>' % m.group(1, 3, 2), page lambda m: '<a href="%s#md5=%s">%s</a>' % m.group(1, 3, 2), page
) )
else: else:
return "" # no sense double-scanning non-package pages return "" # no sense double-scanning non-package pages
def need_version_info(self, url): def need_version_info(self, url):
self.scan_all( self.scan_all(
...@@ -530,12 +530,12 @@ class PackageIndex(Environment): ...@@ -530,12 +530,12 @@ class PackageIndex(Environment):
"""Scan urls scheduled for prescanning (e.g. --find-links)""" """Scan urls scheduled for prescanning (e.g. --find-links)"""
if self.to_scan: if self.to_scan:
list(map(self.scan_url, self.to_scan)) list(map(self.scan_url, self.to_scan))
self.to_scan = None # from now on, go ahead and process immediately self.to_scan = None # from now on, go ahead and process immediately
def not_found_in_index(self, requirement): def not_found_in_index(self, requirement):
if self[requirement.key]: # we've seen at least one distro if self[requirement.key]: # we've seen at least one distro
meth, msg = self.info, "Couldn't retrieve index page for %r" meth, msg = self.info, "Couldn't retrieve index page for %r"
else: # no distros seen for this name, might be misspelled else: # no distros seen for this name, might be misspelled
meth, msg = (self.warn, meth, msg = (self.warn,
"Couldn't find index page for %r (maybe misspelled?)") "Couldn't find index page for %r (maybe misspelled?)")
meth(msg, requirement.unsafe_name) meth(msg, requirement.unsafe_name)
...@@ -665,7 +665,7 @@ class PackageIndex(Environment): ...@@ -665,7 +665,7 @@ class PackageIndex(Environment):
interpret_distro_name(filename, match.group(1), None) if d.version interpret_distro_name(filename, match.group(1), None) if d.version
] or [] ] or []
if len(dists) == 1: # unambiguous ``#egg`` fragment if len(dists) == 1: # unambiguous ``#egg`` fragment
basename = os.path.basename(filename) basename = os.path.basename(filename)
# Make sure the file has been downloaded to the temp dir. # Make sure the file has been downloaded to the temp dir.
...@@ -738,7 +738,7 @@ class PackageIndex(Environment): ...@@ -738,7 +738,7 @@ class PackageIndex(Environment):
fp.close() fp.close()
def reporthook(self, url, filename, blocknum, blksize, size): def reporthook(self, url, filename, blocknum, blksize, size):
pass # no-op pass # no-op
def open_url(self, url, warning=None): def open_url(self, url, warning=None):
if url.startswith('file:'): if url.startswith('file:'):
...@@ -783,10 +783,10 @@ class PackageIndex(Environment): ...@@ -783,10 +783,10 @@ class PackageIndex(Environment):
while '..' in name: while '..' in name:
name = name.replace('..', '.').replace('\\', '_') name = name.replace('..', '.').replace('\\', '_')
else: else:
name = "__downloaded__" # default if URL has no path contents name = "__downloaded__" # default if URL has no path contents
if name.endswith('.egg.zip'): if name.endswith('.egg.zip'):
name = name[:-4] # strip the extra .zip before download name = name[:-4] # strip the extra .zip before download
filename = os.path.join(tmpdir, name) filename = os.path.join(tmpdir, name)
...@@ -801,7 +801,7 @@ class PackageIndex(Environment): ...@@ -801,7 +801,7 @@ class PackageIndex(Environment):
elif scheme == 'file': elif scheme == 'file':
return urllib.request.url2pathname(urllib.parse.urlparse(url)[2]) return urllib.request.url2pathname(urllib.parse.urlparse(url)[2])
else: else:
self.url_ok(url, True) # raises error if not allowed self.url_ok(url, True) # raises error if not allowed
return self._attempt_download(url, filename) return self._attempt_download(url, filename)
def scan_url(self, url): def scan_url(self, url):
...@@ -824,13 +824,13 @@ class PackageIndex(Environment): ...@@ -824,13 +824,13 @@ class PackageIndex(Environment):
file.close() file.close()
os.unlink(filename) os.unlink(filename)
return self._download_svn(url, filename) return self._download_svn(url, filename)
break # not an index page break # not an index page
file.close() file.close()
os.unlink(filename) os.unlink(filename)
raise DistutilsError("Unexpected HTML page found at " + url) raise DistutilsError("Unexpected HTML page found at " + url)
def _download_svn(self, url, filename): def _download_svn(self, url, filename):
url = url.split('#', 1)[0] # remove any fragment for svn's sake url = url.split('#', 1)[0] # remove any fragment for svn's sake
creds = '' creds = ''
if url.lower().startswith('svn:') and '@' in url: if url.lower().startswith('svn:') and '@' in url:
scheme, netloc, path, p, q, f = urllib.parse.urlparse(url) scheme, netloc, path, p, q, f = urllib.parse.urlparse(url)
...@@ -1082,7 +1082,7 @@ open_with_auth = socket_timeout(_SOCKET_TIMEOUT)(open_with_auth) ...@@ -1082,7 +1082,7 @@ open_with_auth = socket_timeout(_SOCKET_TIMEOUT)(open_with_auth)
def fix_sf_url(url): def fix_sf_url(url):
return url # backward compatibility return url # backward compatibility
def local_open(url): def local_open(url):
......
...@@ -40,13 +40,13 @@ def __boot(): ...@@ -40,13 +40,13 @@ def __boot():
known_paths = dict([(makepath(item)[1], 1) for item in sys.path]) # 2.2 comp known_paths = dict([(makepath(item)[1], 1) for item in sys.path]) # 2.2 comp
oldpos = getattr(sys, '__egginsert', 0) # save old insertion position oldpos = getattr(sys, '__egginsert', 0) # save old insertion position
sys.__egginsert = 0 # and reset the current one sys.__egginsert = 0 # and reset the current one
for item in PYTHONPATH: for item in PYTHONPATH:
addsitedir(item) addsitedir(item)
sys.__egginsert += oldpos # restore effective old position sys.__egginsert += oldpos # restore effective old position
d, nd = makepath(stdpath[0]) d, nd = makepath(stdpath[0])
insert_at = None insert_at = None
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment