mirror of
https://github.com/soimort/you-get.git
synced 2025-02-11 20:52:31 +03:00
Avoid using {}
as a default parameter.
This commit is contained in:
parent
546457dea2
commit
2cb209311a
@ -267,7 +267,7 @@ def get_location(url):
|
|||||||
# not to do that
|
# not to do that
|
||||||
return response.geturl()
|
return response.geturl()
|
||||||
|
|
||||||
def get_content(url, headers={}, decoded=True):
|
def get_content(url, headers=None, decoded=True):
|
||||||
"""Gets the content of a URL via sending a HTTP GET request.
|
"""Gets the content of a URL via sending a HTTP GET request.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
@ -281,7 +281,7 @@ def get_content(url, headers={}, decoded=True):
|
|||||||
|
|
||||||
logging.debug('get_content: %s' % url)
|
logging.debug('get_content: %s' % url)
|
||||||
|
|
||||||
req = request.Request(url, headers=headers)
|
req = request.Request(url, headers=headers or {})
|
||||||
if cookies:
|
if cookies:
|
||||||
cookies.add_cookie_header(req)
|
cookies.add_cookie_header(req)
|
||||||
req.headers.update(req.unredirected_hdrs)
|
req.headers.update(req.unredirected_hdrs)
|
||||||
@ -301,7 +301,7 @@ def get_content(url, headers={}, decoded=True):
|
|||||||
|
|
||||||
return data
|
return data
|
||||||
|
|
||||||
def url_size(url, faker = False, headers = {}):
|
def url_size(url, faker = False, headers = None):
|
||||||
headers = fake_headers if faker else headers
|
headers = fake_headers if faker else headers
|
||||||
if headers:
|
if headers:
|
||||||
response = request.urlopen(request.Request(url, headers=headers), None)
|
response = request.urlopen(request.Request(url, headers=headers), None)
|
||||||
@ -311,16 +311,16 @@ def url_size(url, faker = False, headers = {}):
|
|||||||
size = response.headers['content-length']
|
size = response.headers['content-length']
|
||||||
return int(size) if size!=None else float('inf')
|
return int(size) if size!=None else float('inf')
|
||||||
|
|
||||||
def urls_size(urls, faker = False, headers = {}):
|
def urls_size(urls, faker = False, headers = None):
|
||||||
return sum([url_size(url, faker=faker, headers=headers) for url in urls])
|
return sum([url_size(url, faker=faker, headers=headers) for url in urls])
|
||||||
|
|
||||||
def get_head(url, headers = {}):
|
def get_head(url, headers = None):
|
||||||
req = request.Request(url, headers = headers) if headers else request.Request(url)
|
req = request.Request(url, headers = headers) if headers else request.Request(url)
|
||||||
req.get_method = lambda : 'HEAD'
|
req.get_method = lambda : 'HEAD'
|
||||||
res = request.urlopen(req)
|
res = request.urlopen(req)
|
||||||
return dict(res.headers)
|
return dict(res.headers)
|
||||||
|
|
||||||
def url_info(url, faker = False, headers = {}):
|
def url_info(url, faker = False, headers = None):
|
||||||
headers = fake_headers if faker else headers
|
headers = fake_headers if faker else headers
|
||||||
if headers:
|
if headers:
|
||||||
response = request.urlopen(request.Request(url, headers = headers), None)
|
response = request.urlopen(request.Request(url, headers = headers), None)
|
||||||
@ -370,7 +370,7 @@ def url_info(url, faker = False, headers = {}):
|
|||||||
|
|
||||||
return type, ext, size
|
return type, ext, size
|
||||||
|
|
||||||
def url_locations(urls, faker = False, headers = {}):
|
def url_locations(urls, faker = False, headers = None):
|
||||||
headers = fake_headers if faker else headers
|
headers = fake_headers if faker else headers
|
||||||
locations = []
|
locations = []
|
||||||
for url in urls:
|
for url in urls:
|
||||||
@ -381,7 +381,8 @@ def url_locations(urls, faker = False, headers = {}):
|
|||||||
locations.append(response.url)
|
locations.append(response.url)
|
||||||
return locations
|
return locations
|
||||||
|
|
||||||
def url_save(url, filepath, bar, refer = None, is_part = False, faker = False, headers = {}):
|
def url_save(url, filepath, bar, refer = None, is_part = False, faker = False, headers = None):
|
||||||
|
headers = headers or {}
|
||||||
file_size = url_size(url, faker = faker, headers = headers)
|
file_size = url_size(url, faker = faker, headers = headers)
|
||||||
|
|
||||||
if os.path.exists(filepath):
|
if os.path.exists(filepath):
|
||||||
@ -415,7 +416,7 @@ def url_save(url, filepath, bar, refer = None, is_part = False, faker = False, h
|
|||||||
open_mode = 'wb'
|
open_mode = 'wb'
|
||||||
|
|
||||||
if received < file_size:
|
if received < file_size:
|
||||||
headers = fake_headers if faker else (headers or {})
|
headers = fake_headers if faker else headers
|
||||||
if received:
|
if received:
|
||||||
headers['Range'] = 'bytes=' + str(received) + '-'
|
headers['Range'] = 'bytes=' + str(received) + '-'
|
||||||
if refer:
|
if refer:
|
||||||
@ -456,7 +457,8 @@ def url_save(url, filepath, bar, refer = None, is_part = False, faker = False, h
|
|||||||
os.remove(filepath) # on Windows rename could fail if destination filepath exists
|
os.remove(filepath) # on Windows rename could fail if destination filepath exists
|
||||||
os.rename(temp_filepath, filepath)
|
os.rename(temp_filepath, filepath)
|
||||||
|
|
||||||
def url_save_chunked(url, filepath, bar, refer = None, is_part = False, faker = False, headers = {}):
|
def url_save_chunked(url, filepath, bar, refer = None, is_part = False, faker = False, headers=None):
|
||||||
|
headers = headers or {}
|
||||||
if os.path.exists(filepath):
|
if os.path.exists(filepath):
|
||||||
if not force:
|
if not force:
|
||||||
if not is_part:
|
if not is_part:
|
||||||
@ -487,7 +489,7 @@ def url_save_chunked(url, filepath, bar, refer = None, is_part = False, faker =
|
|||||||
else:
|
else:
|
||||||
open_mode = 'wb'
|
open_mode = 'wb'
|
||||||
|
|
||||||
headers = fake_headers if faker else (headers or {})
|
headers = fake_headers if faker else headers
|
||||||
if received:
|
if received:
|
||||||
headers['Range'] = 'bytes=' + str(received) + '-'
|
headers['Range'] = 'bytes=' + str(received) + '-'
|
||||||
if refer:
|
if refer:
|
||||||
@ -624,8 +626,9 @@ def get_output_filename(urls, title, ext, output_dir, merge):
|
|||||||
merged_ext = 'mkv' if has_ffmpeg_installed() else 'ts'
|
merged_ext = 'mkv' if has_ffmpeg_installed() else 'ts'
|
||||||
return '%s.%s' % (title, merged_ext)
|
return '%s.%s' % (title, merged_ext)
|
||||||
|
|
||||||
def download_urls(urls, title, ext, total_size, output_dir='.', refer=None, merge=True, faker=False, headers = {}, **kwargs):
|
def download_urls(urls, title, ext, total_size, output_dir='.', refer=None, merge=True, faker=False, headers = None, **kwargs):
|
||||||
assert urls
|
assert urls
|
||||||
|
headers = headers or {}
|
||||||
if json_output:
|
if json_output:
|
||||||
json_output_.download_urls(urls=urls, title=title, ext=ext, total_size=total_size, refer=refer)
|
json_output_.download_urls(urls=urls, title=title, ext=ext, total_size=total_size, refer=refer)
|
||||||
return
|
return
|
||||||
@ -742,8 +745,9 @@ def download_urls(urls, title, ext, total_size, output_dir='.', refer=None, merg
|
|||||||
|
|
||||||
print()
|
print()
|
||||||
|
|
||||||
def download_urls_chunked(urls, title, ext, total_size, output_dir='.', refer=None, merge=True, faker=False, headers = {}):
|
def download_urls_chunked(urls, title, ext, total_size, output_dir='.', refer=None, merge=True, faker=False, headers = None):
|
||||||
assert urls
|
assert urls
|
||||||
|
headers = headers or {}
|
||||||
if dry_run:
|
if dry_run:
|
||||||
print('Real URLs:\n%s\n' % urls)
|
print('Real URLs:\n%s\n' % urls)
|
||||||
return
|
return
|
||||||
|
Loading…
Reference in New Issue
Block a user