mirror of
https://github.com/soimort/you-get.git
synced 2025-02-11 20:52:31 +03:00
faker is true
This commit is contained in:
parent
0e8d9e54e5
commit
4ebb9611bb
@ -129,7 +129,7 @@ def undeflate(data):
|
|||||||
return decompressobj.decompress(data)+decompressobj.flush()
|
return decompressobj.decompress(data)+decompressobj.flush()
|
||||||
|
|
||||||
# DEPRECATED in favor of get_content()
|
# DEPRECATED in favor of get_content()
|
||||||
def get_response(url, faker = False):
|
def get_response(url, faker = True):
|
||||||
if faker:
|
if faker:
|
||||||
response = request.urlopen(request.Request(url, headers = fake_headers), None)
|
response = request.urlopen(request.Request(url, headers = fake_headers), None)
|
||||||
else:
|
else:
|
||||||
@ -144,12 +144,12 @@ def get_response(url, faker = False):
|
|||||||
return response
|
return response
|
||||||
|
|
||||||
# DEPRECATED in favor of get_content()
|
# DEPRECATED in favor of get_content()
|
||||||
def get_html(url, encoding = None, faker = False):
|
def get_html(url, encoding = None, faker = True):
|
||||||
content = get_response(url, faker).data
|
content = get_response(url, faker).data
|
||||||
return str(content, 'utf-8', 'ignore')
|
return str(content, 'utf-8', 'ignore')
|
||||||
|
|
||||||
# DEPRECATED in favor of get_content()
|
# DEPRECATED in favor of get_content()
|
||||||
def get_decoded_html(url, faker = False):
|
def get_decoded_html(url, faker = True):
|
||||||
response = get_response(url, faker)
|
response = get_response(url, faker)
|
||||||
data = response.data
|
data = response.data
|
||||||
charset = r1(r'charset=([\w-]+)', response.headers['content-type'])
|
charset = r1(r'charset=([\w-]+)', response.headers['content-type'])
|
||||||
@ -194,7 +194,7 @@ def get_content(url, headers={}, decoded=True):
|
|||||||
|
|
||||||
return data
|
return data
|
||||||
|
|
||||||
def url_size(url, faker = False):
|
def url_size(url, faker = True):
|
||||||
if faker:
|
if faker:
|
||||||
response = request.urlopen(request.Request(url, headers = fake_headers), None)
|
response = request.urlopen(request.Request(url, headers = fake_headers), None)
|
||||||
else:
|
else:
|
||||||
@ -209,7 +209,7 @@ def url_size(url, faker = False):
|
|||||||
def urls_size(urls):
|
def urls_size(urls):
|
||||||
return sum(map(url_size, urls))
|
return sum(map(url_size, urls))
|
||||||
|
|
||||||
def url_info(url, faker = False):
|
def url_info(url, faker = True):
|
||||||
if faker:
|
if faker:
|
||||||
response = request.urlopen(request.Request(url, headers = fake_headers), None)
|
response = request.urlopen(request.Request(url, headers = fake_headers), None)
|
||||||
else:
|
else:
|
||||||
@ -253,7 +253,7 @@ def url_info(url, faker = False):
|
|||||||
|
|
||||||
return type, ext, size
|
return type, ext, size
|
||||||
|
|
||||||
def url_locations(urls, faker = False):
|
def url_locations(urls, faker = True):
|
||||||
locations = []
|
locations = []
|
||||||
for url in urls:
|
for url in urls:
|
||||||
if faker:
|
if faker:
|
||||||
@ -264,7 +264,7 @@ def url_locations(urls, faker = False):
|
|||||||
locations.append(response.url)
|
locations.append(response.url)
|
||||||
return locations
|
return locations
|
||||||
|
|
||||||
def url_save(url, filepath, bar, refer = None, is_part = False, faker = False):
|
def url_save(url, filepath, bar, refer = None, is_part = False, faker = True):
|
||||||
file_size = url_size(url, faker = faker)
|
file_size = url_size(url, faker = faker)
|
||||||
|
|
||||||
if os.path.exists(filepath):
|
if os.path.exists(filepath):
|
||||||
@ -342,7 +342,7 @@ def url_save(url, filepath, bar, refer = None, is_part = False, faker = False):
|
|||||||
os.remove(filepath) # on Windows rename could fail if destination filepath exists
|
os.remove(filepath) # on Windows rename could fail if destination filepath exists
|
||||||
os.rename(temp_filepath, filepath)
|
os.rename(temp_filepath, filepath)
|
||||||
|
|
||||||
def url_save_chunked(url, filepath, bar, refer = None, is_part = False, faker = False):
|
def url_save_chunked(url, filepath, bar, refer = None, is_part = False, faker = True):
|
||||||
if os.path.exists(filepath):
|
if os.path.exists(filepath):
|
||||||
if not force:
|
if not force:
|
||||||
if not is_part:
|
if not is_part:
|
||||||
@ -476,7 +476,7 @@ class DummyProgressBar:
|
|||||||
def done(self):
|
def done(self):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
def download_urls(urls, title, ext, total_size, output_dir='.', refer=None, merge=True, faker=False):
|
def download_urls(urls, title, ext, total_size, output_dir='.', refer=None, merge=True, faker=True):
|
||||||
assert urls
|
assert urls
|
||||||
if dry_run:
|
if dry_run:
|
||||||
dry_infos.clear()
|
dry_infos.clear()
|
||||||
@ -566,7 +566,7 @@ def download_urls(urls, title, ext, total_size, output_dir='.', refer=None, merg
|
|||||||
|
|
||||||
print()
|
print()
|
||||||
|
|
||||||
def download_urls_chunked(urls, title, ext, total_size, output_dir='.', refer=None, merge=True, faker=False):
|
def download_urls_chunked(urls, title, ext, total_size, output_dir='.', refer=None, merge=True, faker=True):
|
||||||
assert urls
|
assert urls
|
||||||
if dry_run:
|
if dry_run:
|
||||||
print('Real URLs:\n%s\n' % urls)
|
print('Real URLs:\n%s\n' % urls)
|
||||||
@ -647,7 +647,7 @@ def download_urls_chunked(urls, title, ext, total_size, output_dir='.', refer=No
|
|||||||
|
|
||||||
print()
|
print()
|
||||||
|
|
||||||
def download_rtmp_url(url,title, ext,params={}, total_size=0, output_dir='.', refer=None, merge=True, faker=False):
|
def download_rtmp_url(url,title, ext,params={}, total_size=0, output_dir='.', refer=None, merge=True, faker=True):
|
||||||
assert url
|
assert url
|
||||||
if dry_run:
|
if dry_run:
|
||||||
print('Real URL:\n%s\n' % [url])
|
print('Real URL:\n%s\n' % [url])
|
||||||
|
Loading…
Reference in New Issue
Block a user