mirror of
https://github.com/soimort/you-get.git
synced 2025-02-02 16:24:00 +03:00
feat: add lrts extractor
This commit is contained in:
parent
b00860e4be
commit
5c2bb3fa34
@ -414,6 +414,7 @@ Use `--url`/`-u` to get a list of downloadable resource URLs extracted from the
|
||||
| 酷我音乐 | <http://www.kuwo.cn/> | | |✓|
|
||||
| 乐视网 | <http://www.le.com/> |✓| | |
|
||||
| 荔枝FM | <http://www.lizhi.fm/> | | |✓|
|
||||
| 懒人听书 | <http://www.lrts.me/> | | |✓|
|
||||
| 秒拍 | <http://www.miaopai.com/> |✓| | |
|
||||
| MioMio弹幕网 | <http://www.miomio.tv/> |✓| | |
|
||||
| MissEvan<br/>猫耳FM | <http://www.missevan.com/> | | |✓|
|
||||
|
@ -76,6 +76,7 @@ SITES = {
|
||||
'letv' : 'le',
|
||||
'lizhi' : 'lizhi',
|
||||
'longzhu' : 'longzhu',
|
||||
'lrts' : 'lrts',
|
||||
'magisto' : 'magisto',
|
||||
'metacafe' : 'metacafe',
|
||||
'mgtv' : 'mgtv',
|
||||
|
70
src/you_get/extractors/lrts.py
Normal file
70
src/you_get/extractors/lrts.py
Normal file
@ -0,0 +1,70 @@
|
||||
#!/usr/bin/env python
|
||||
|
||||
__all__ = ['lrts_download']
|
||||
|
||||
import logging
|
||||
from ..common import *
|
||||
|
||||
def lrts_download(url, output_dir='.', merge=True, info_only=False, **kwargs):
|
||||
html = get_html(url)
|
||||
matched = re.search(r"/book/(\d+)", url)
|
||||
if not matched:
|
||||
raise AssertionError("not found book number: %s" % url)
|
||||
book_no = matched.group(1)
|
||||
book_title = book_no
|
||||
matched = re.search(r"<title>(.*)-(.*)</title>", html)
|
||||
if matched:
|
||||
book_title = matched.group(1)
|
||||
|
||||
matched = re.search(r"var totalCount='(\d+)'", html)
|
||||
if not matched:
|
||||
raise AssertionError("not found total count in html")
|
||||
total_count = int(matched.group(1))
|
||||
logging.debug('total: %s' % total_count)
|
||||
page_size = 10
|
||||
logging.debug('total page count: %s' % ((total_count // page_size) + 1))
|
||||
headers = {
|
||||
'Referer': url
|
||||
}
|
||||
items = []
|
||||
if (total_count > page_size):
|
||||
for page in range((total_count // page_size) + 1):
|
||||
page_url = 'http://www.lrts.me/ajax/book/%s/%s/%s' % (book_no, page, page_size)
|
||||
response_content = json.loads(post_content(page_url, headers))
|
||||
if response_content['status'] != 'success':
|
||||
raise AssertionError("got the page failed: %s" % (page_url))
|
||||
data = response_content['data']['data']
|
||||
if data:
|
||||
for i in data:
|
||||
i['resName'] = parse.unquote(i['resName'])
|
||||
items.extend(data)
|
||||
else:
|
||||
break
|
||||
|
||||
headers = {
|
||||
'Referer': 'http://www.lrts.me/playlist'
|
||||
}
|
||||
|
||||
for item in items:
|
||||
i_url = 'http://www.lrts.me/ajax/path/4/%s/%s' % (item['fatherResId'], item['resId'])
|
||||
response_content = json.loads(post_content(i_url, headers))
|
||||
# logging.debug(response_content)
|
||||
if response_content['status'] == 'success' and response_content['data']:
|
||||
item['ok'] = True
|
||||
item['url'] = response_content['data']
|
||||
|
||||
items = list(filter(lambda i: 'ok' in i and i['ok'], items))
|
||||
print('Downloading %s: %s count ...' % (book_title, len(items)))
|
||||
|
||||
for item in items:
|
||||
title = item['resName']
|
||||
file_url = item['url']
|
||||
# if not file_url: continue
|
||||
_, _, size = url_info(file_url)
|
||||
print_info(site_info, title, 'mp3', size)
|
||||
if not info_only:
|
||||
download_urls([file_url], title, 'mp3', size, output_dir, merge=merge)
|
||||
|
||||
site_info = "lrts.me"
|
||||
download = lrts_download
|
||||
download_playlist = lrts_download
|
Loading…
Reference in New Issue
Block a user