faker is true

This commit is contained in:
邓景 2014-12-22 20:09:40 +08:00
parent 0e8d9e54e5
commit 4ebb9611bb

View File

@ -129,7 +129,7 @@ def undeflate(data):
return decompressobj.decompress(data)+decompressobj.flush() return decompressobj.decompress(data)+decompressobj.flush()
# DEPRECATED in favor of get_content() # DEPRECATED in favor of get_content()
def get_response(url, faker = False): def get_response(url, faker = True):
if faker: if faker:
response = request.urlopen(request.Request(url, headers = fake_headers), None) response = request.urlopen(request.Request(url, headers = fake_headers), None)
else: else:
@ -144,12 +144,12 @@ def get_response(url, faker = False):
return response return response
# DEPRECATED in favor of get_content() # DEPRECATED in favor of get_content()
def get_html(url, encoding = None, faker = False): def get_html(url, encoding = None, faker = True):
content = get_response(url, faker).data content = get_response(url, faker).data
return str(content, 'utf-8', 'ignore') return str(content, 'utf-8', 'ignore')
# DEPRECATED in favor of get_content() # DEPRECATED in favor of get_content()
def get_decoded_html(url, faker = False): def get_decoded_html(url, faker = True):
response = get_response(url, faker) response = get_response(url, faker)
data = response.data data = response.data
charset = r1(r'charset=([\w-]+)', response.headers['content-type']) charset = r1(r'charset=([\w-]+)', response.headers['content-type'])
@ -194,7 +194,7 @@ def get_content(url, headers={}, decoded=True):
return data return data
def url_size(url, faker = False): def url_size(url, faker = True):
if faker: if faker:
response = request.urlopen(request.Request(url, headers = fake_headers), None) response = request.urlopen(request.Request(url, headers = fake_headers), None)
else: else:
@ -209,7 +209,7 @@ def url_size(url, faker = False):
def urls_size(urls): def urls_size(urls):
return sum(map(url_size, urls)) return sum(map(url_size, urls))
def url_info(url, faker = False): def url_info(url, faker = True):
if faker: if faker:
response = request.urlopen(request.Request(url, headers = fake_headers), None) response = request.urlopen(request.Request(url, headers = fake_headers), None)
else: else:
@ -253,7 +253,7 @@ def url_info(url, faker = False):
return type, ext, size return type, ext, size
def url_locations(urls, faker = False): def url_locations(urls, faker = True):
locations = [] locations = []
for url in urls: for url in urls:
if faker: if faker:
@ -264,7 +264,7 @@ def url_locations(urls, faker = False):
locations.append(response.url) locations.append(response.url)
return locations return locations
def url_save(url, filepath, bar, refer = None, is_part = False, faker = False): def url_save(url, filepath, bar, refer = None, is_part = False, faker = True):
file_size = url_size(url, faker = faker) file_size = url_size(url, faker = faker)
if os.path.exists(filepath): if os.path.exists(filepath):
@ -342,7 +342,7 @@ def url_save(url, filepath, bar, refer = None, is_part = False, faker = False):
os.remove(filepath) # on Windows rename could fail if destination filepath exists os.remove(filepath) # on Windows rename could fail if destination filepath exists
os.rename(temp_filepath, filepath) os.rename(temp_filepath, filepath)
def url_save_chunked(url, filepath, bar, refer = None, is_part = False, faker = False): def url_save_chunked(url, filepath, bar, refer = None, is_part = False, faker = True):
if os.path.exists(filepath): if os.path.exists(filepath):
if not force: if not force:
if not is_part: if not is_part:
@ -476,7 +476,7 @@ class DummyProgressBar:
def done(self): def done(self):
pass pass
def download_urls(urls, title, ext, total_size, output_dir='.', refer=None, merge=True, faker=False): def download_urls(urls, title, ext, total_size, output_dir='.', refer=None, merge=True, faker=True):
assert urls assert urls
if dry_run: if dry_run:
dry_infos.clear() dry_infos.clear()
@ -566,7 +566,7 @@ def download_urls(urls, title, ext, total_size, output_dir='.', refer=None, merg
print() print()
def download_urls_chunked(urls, title, ext, total_size, output_dir='.', refer=None, merge=True, faker=False): def download_urls_chunked(urls, title, ext, total_size, output_dir='.', refer=None, merge=True, faker=True):
assert urls assert urls
if dry_run: if dry_run:
print('Real URLs:\n%s\n' % urls) print('Real URLs:\n%s\n' % urls)
@ -647,7 +647,7 @@ def download_urls_chunked(urls, title, ext, total_size, output_dir='.', refer=No
print() print()
def download_rtmp_url(url,title, ext,params={}, total_size=0, output_dir='.', refer=None, merge=True, faker=False): def download_rtmp_url(url,title, ext,params={}, total_size=0, output_dir='.', refer=None, merge=True, faker=True):
assert url assert url
if dry_run: if dry_run:
print('Real URL:\n%s\n' % [url]) print('Real URL:\n%s\n' % [url])