This commit is contained in:
perror 2017-12-10 13:14:32 +08:00
parent 219ac3ce62
commit 6c3bdfd90c
3 changed files with 87 additions and 0 deletions

View File

@ -413,6 +413,7 @@ Use `--url`/`-u` to get a list of downloadable resource URLs extracted from the
| 火猫TV | <http://www.huomao.com/> |✓| | |
| 全民直播 | <http://www.quanmin.tv/> |✓| | |
| 阳光宽频网 | <http://www.365yg.com/> |✓| | |
| 西瓜视频 | <https://www.ixigua.com/> |✓| | |
For all other sites not on the list, the universal extractor will take care of finding and downloading interesting resources from the page.

View File

@ -61,6 +61,7 @@ SITES = {
'interest' : 'interest',
'iqilu' : 'iqilu',
'iqiyi' : 'iqiyi',
'ixigua' : 'ixigua',
'isuntv' : 'suntv',
'joy' : 'joy',
'kankanews' : 'bilibili',

View File

@ -0,0 +1,85 @@
#!/usr/bin/env python
__all__ = ['ixigua_download', 'ixigua_download_playlist']
import base64
import random
import binascii
from ..common import *
def get_video_id(text):
re_id = r"videoId: '(.*?)'"
return re.findall(re_id, text)[0]
def get_r():
return str(random.random())[2:]
def right_shift(val, n):
return val >> n if val >= 0 else (val + 0x100000000) >> n
def get_s(text):
"""get video info"""
id = get_video_id(text)
p = get_r()
url = 'http://i.snssdk.com/video/urls/v/1/toutiao/mp4/%s' % id
n = parse.urlparse(url).path + '?r=%s' % p
c = binascii.crc32(n.encode('utf-8'))
s = right_shift(c, 0)
title = ''.join(re.findall(r"title: '(.*?)',", text))
return url + '?r=%s&s=%s' % (p, s), title
def get_moment(url, user_id, base_url, video_list):
"""Recursively obtaining a video list"""
video_list_data = json.loads(get_content(url))
if not video_list_data['next']['max_behot_time']:
return video_list
[video_list.append(i["display_url"]) for i in video_list_data["data"]]
max_behot_time = video_list_data['next']['max_behot_time']
_param = {
'user_id': user_id,
'base_url': base_url,
'video_list': video_list,
'url': base_url.format(user_id=user_id, max_behot_time=max_behot_time),
}
return get_moment(**_param)
def ixigua_download(url, output_dir='.', info_only=False, **kwargs):
""" Download a single video
Sample URL: https://www.ixigua.com/a6487187567887254029/#mid=59051127876
"""
try:
video_info_url, title = get_s(get_content(url))
video_info = json.loads(get_content(video_info_url))
except Exception:
raise NotImplementedError(url)
try:
video_url = base64.b64decode(video_info["data"]["video_list"]["video_1"]["main_url"]).decode()
except Exception:
raise NotImplementedError(url)
filetype, ext, size = url_info(video_url)
print_info(site_info, title, filetype, size)
if not info_only:
download_urls([video_url], title, ext, size, output_dir=output_dir)
def ixigua_download_playlist(url, output_dir='.', info_only=False, **kwargs):
"""Download all video from the user's video list
Sample URL: https://www.ixigua.com/c/user/71141690831/
"""
if 'user' not in url:
raise NotImplementedError(url)
user_id = url.split('/')[-2]
max_behot_time = 0
if not user_id:
raise NotImplementedError(url)
base_url = "https://www.ixigua.com/c/user/article/?user_id={user_id}" \
"&max_behot_time={max_behot_time}&max_repin_time=0&count=20&page_type=0"
_param = {
'user_id': user_id,
'base_url': base_url,
'video_list': [],
'url': base_url.format(user_id=user_id, max_behot_time=max_behot_time),
}
for i in get_moment(**_param):
ixigua_download(i, output_dir, info_only, **kwargs)
site_info = "ixigua.com"
download = ixigua_download
download_playlist = ixigua_download_playlist