2020-08-28 15:44:50 +00:00
|
|
|
# coding: utf-8
|
|
|
|
from __future__ import unicode_literals
|
|
|
|
|
|
|
|
import re
|
|
|
|
|
|
|
|
from .common import InfoExtractor
|
2020-08-29 07:04:16 +00:00
|
|
|
from ..compat import compat_urlparse
|
2020-08-29 08:25:42 +00:00
|
|
|
from ..utils import (
|
|
|
|
clean_html,
|
|
|
|
extract_attributes,
|
|
|
|
ExtractorError,
|
|
|
|
get_elements_by_class,
|
|
|
|
int_or_none,
|
|
|
|
js_to_json,
|
|
|
|
smuggle_url,
|
|
|
|
unescapeHTML,
|
|
|
|
)
|
2020-08-28 15:44:50 +00:00
|
|
|
|
|
|
|
|
2020-08-29 07:04:16 +00:00
|
|
|
def _get_elements_by_tag_and_attrib(html, tag=None, attribute=None, value=None, escape_value=True):
|
|
|
|
"""Return the content of the tag with the specified attribute in the passed HTML document"""
|
|
|
|
|
|
|
|
if tag is None:
|
|
|
|
tag = '[a-zA-Z0-9:._-]+'
|
|
|
|
if attribute is None:
|
|
|
|
attribute = ''
|
|
|
|
else:
|
|
|
|
attribute = r'\s+(?P<attribute>%s)' % re.escape(attribute)
|
|
|
|
if value is None:
|
|
|
|
value = ''
|
|
|
|
else:
|
|
|
|
value = re.escape(value) if escape_value else value
|
|
|
|
value = '=[\'"]?(?P<value>%s)[\'"]?' % value
|
|
|
|
|
|
|
|
retlist = []
|
|
|
|
for m in re.finditer(r'''(?xs)
|
|
|
|
<(?P<tag>%s)
|
|
|
|
(?:\s+[a-zA-Z0-9:._-]+(?:=[a-zA-Z0-9:._-]*|="[^"]*"|='[^']*'|))*?
|
|
|
|
%s%s
|
|
|
|
(?:\s+[a-zA-Z0-9:._-]+(?:=[a-zA-Z0-9:._-]*|="[^"]*"|='[^']*'|))*?
|
|
|
|
\s*>
|
|
|
|
(?P<content>.*?)
|
|
|
|
</\1>
|
|
|
|
''' % (tag, attribute, value), html):
|
|
|
|
retlist.append(m)
|
|
|
|
|
|
|
|
return retlist
|
|
|
|
|
|
|
|
|
|
|
|
def _get_element_by_tag_and_attrib(html, tag=None, attribute=None, value=None, escape_value=True):
|
|
|
|
retval = _get_elements_by_tag_and_attrib(html, tag, attribute, value, escape_value)
|
|
|
|
return retval[0] if retval else None
|
|
|
|
|
|
|
|
|
2020-08-28 15:44:50 +00:00
|
|
|
class DubokuIE(InfoExtractor):
|
2020-08-29 07:04:16 +00:00
|
|
|
IE_NAME = 'duboku'
|
|
|
|
IE_DESC = 'www.duboku.co'
|
|
|
|
|
2020-08-29 07:44:56 +00:00
|
|
|
_VALID_URL = r'(?:https?://[^/]+\.duboku\.co/vodplay/)(?P<id>[0-9]+-[0-9-]+)\.html.*'
|
2020-08-28 15:44:50 +00:00
|
|
|
_TESTS = [{
|
|
|
|
'url': 'https://www.duboku.co/vodplay/1575-1-1.html',
|
|
|
|
'info_dict': {
|
|
|
|
'id': '1575-1-1',
|
2020-08-29 08:25:42 +00:00
|
|
|
'ext': 'ts',
|
|
|
|
'series': '白色月光',
|
|
|
|
'title': 'contains:白色月光',
|
|
|
|
'season_number': 1,
|
|
|
|
'episode_number': 1,
|
|
|
|
},
|
|
|
|
'params': {
|
|
|
|
'skip_download': 'm3u8 download',
|
|
|
|
},
|
|
|
|
}, {
|
|
|
|
'url': 'https://www.duboku.co/vodplay/1588-1-1.html',
|
|
|
|
'info_dict': {
|
|
|
|
'id': '1588-1-1',
|
|
|
|
'ext': 'ts',
|
|
|
|
'series': '亲爱的自己',
|
|
|
|
'title': 'contains:预告片',
|
|
|
|
'season_number': 1,
|
|
|
|
'episode_number': 1,
|
2020-08-28 15:44:50 +00:00
|
|
|
},
|
|
|
|
'params': {
|
|
|
|
'skip_download': 'm3u8 download',
|
|
|
|
},
|
|
|
|
}]
|
|
|
|
|
|
|
|
_PLAYER_DATA_PATTERN = r'player_data\s*=\s*(\{\s*(.*)})\s*;?\s*</script'
|
|
|
|
|
|
|
|
def _real_extract(self, url):
|
|
|
|
video_id = self._match_id(url)
|
|
|
|
temp = video_id.split('-')
|
|
|
|
series_id = temp[0]
|
|
|
|
season_id = temp[1]
|
|
|
|
episode_id = temp[2]
|
|
|
|
|
|
|
|
webpage_url = 'https://www.duboku.co/vodplay/%s.html' % video_id
|
|
|
|
webpage_html = self._download_webpage(webpage_url, video_id)
|
|
|
|
|
|
|
|
# extract video url
|
|
|
|
|
|
|
|
player_data = self._search_regex(
|
|
|
|
self._PLAYER_DATA_PATTERN, webpage_html, 'player_data')
|
|
|
|
player_data = self._parse_json(js_to_json(player_data), video_id)
|
|
|
|
|
|
|
|
# extract title
|
|
|
|
|
|
|
|
temp = get_elements_by_class('title', webpage_html)
|
|
|
|
series_title = None
|
|
|
|
title = None
|
|
|
|
for html in temp:
|
|
|
|
mobj = re.search(r'<a\s+.*>(.*)</a>', html)
|
|
|
|
if mobj:
|
|
|
|
href = extract_attributes(mobj.group(0)).get('href')
|
|
|
|
if href:
|
|
|
|
mobj1 = re.search(r'/(\d+)\.html', href)
|
|
|
|
if mobj1 and mobj1.group(1) == series_id:
|
|
|
|
series_title = clean_html(mobj.group(0))
|
|
|
|
series_title = re.sub(r'[\s\r\n\t]+', ' ', series_title)
|
|
|
|
title = clean_html(html)
|
|
|
|
title = re.sub(r'[\s\r\n\t]+', ' ', title)
|
|
|
|
break
|
|
|
|
|
|
|
|
data_url = player_data['url']
|
|
|
|
assert data_url
|
|
|
|
data_from = player_data.get('from')
|
|
|
|
|
|
|
|
# if it is an embedded iframe, maybe it's an external source
|
|
|
|
if data_from == 'iframe':
|
|
|
|
# use _type url_transparent to retain the meaningful details
|
|
|
|
# of the video.
|
|
|
|
return {
|
|
|
|
'_type': 'url_transparent',
|
|
|
|
'url': smuggle_url(data_url, {'http_headers': {'Referer': webpage_url}}),
|
|
|
|
'id': video_id,
|
|
|
|
'title': title,
|
|
|
|
'series': series_title,
|
|
|
|
'season_number': int_or_none(season_id),
|
|
|
|
'season_id': season_id,
|
|
|
|
'episode_number': int_or_none(episode_id),
|
|
|
|
'episode_id': episode_id,
|
|
|
|
}
|
|
|
|
|
|
|
|
formats = self._extract_m3u8_formats(data_url, video_id, 'ts')
|
|
|
|
|
|
|
|
return {
|
|
|
|
'id': video_id,
|
|
|
|
'title': title,
|
|
|
|
'series': series_title,
|
|
|
|
'season_number': int_or_none(season_id),
|
|
|
|
'season_id': season_id,
|
|
|
|
'episode_number': int_or_none(episode_id),
|
|
|
|
'episode_id': episode_id,
|
|
|
|
'formats': formats,
|
|
|
|
}
|
2020-08-29 07:04:16 +00:00
|
|
|
|
|
|
|
|
|
|
|
class DubokuPlaylistIE(InfoExtractor):
|
|
|
|
IE_NAME = 'duboku:list'
|
|
|
|
IE_DESC = 'www.duboku.co entire series'
|
|
|
|
|
|
|
|
_VALID_URL = r'(?:https?://[^/]+\.duboku\.co/voddetail/)(?P<id>[0-9]+)\.html.*'
|
2020-08-29 07:23:43 +00:00
|
|
|
_TESTS = [{
|
2020-08-29 08:25:42 +00:00
|
|
|
'url': 'https://www.duboku.co/voddetail/1575.html',
|
2020-08-29 07:23:43 +00:00
|
|
|
'info_dict': {
|
2020-08-29 08:25:42 +00:00
|
|
|
'id': 'startswith:1575',
|
2020-08-29 07:23:43 +00:00
|
|
|
'title': '白色月光',
|
|
|
|
},
|
|
|
|
'playlist_count': 12,
|
|
|
|
}, {
|
2020-08-29 08:25:42 +00:00
|
|
|
'url': 'https://www.duboku.co/voddetail/1554.html',
|
2020-08-29 07:23:43 +00:00
|
|
|
'info_dict': {
|
2020-08-29 08:25:42 +00:00
|
|
|
'id': 'startswith:1554',
|
2020-08-29 07:23:43 +00:00
|
|
|
'title': '以家人之名',
|
|
|
|
},
|
2020-08-29 08:25:42 +00:00
|
|
|
'playlist_mincount': 30,
|
2020-08-29 07:23:43 +00:00
|
|
|
}, {
|
2020-08-29 08:25:42 +00:00
|
|
|
'url': 'https://www.duboku.co/voddetail/1554.html#playlist2',
|
2020-08-29 07:23:43 +00:00
|
|
|
'info_dict': {
|
|
|
|
'id': '1554#playlist2',
|
|
|
|
'title': '以家人之名',
|
|
|
|
},
|
2020-08-29 08:25:42 +00:00
|
|
|
'playlist_mincount': 27,
|
2020-08-29 07:23:43 +00:00
|
|
|
}]
|
2020-08-29 07:04:16 +00:00
|
|
|
|
|
|
|
def _real_extract(self, url):
|
|
|
|
mobj = re.match(self._VALID_URL, url)
|
|
|
|
if mobj is None:
|
|
|
|
raise ExtractorError('Invalid URL: %s' % url)
|
|
|
|
series_id = mobj.group('id')
|
|
|
|
fragment = compat_urlparse.urlparse(url).fragment
|
|
|
|
|
|
|
|
webpage_url = 'https://www.duboku.co/voddetail/%s.html' % series_id
|
|
|
|
webpage_html = self._download_webpage(webpage_url, series_id)
|
|
|
|
|
|
|
|
# extract title
|
|
|
|
|
|
|
|
title = _get_element_by_tag_and_attrib(webpage_html, 'h1', 'class', 'title')
|
|
|
|
title = unescapeHTML(title.group('content')) if title else None
|
|
|
|
if not title:
|
|
|
|
title = self._html_search_meta('keywords', webpage_html)
|
|
|
|
if not title:
|
|
|
|
title = _get_element_by_tag_and_attrib(webpage_html, 'title')
|
|
|
|
title = unescapeHTML(title.group('content')) if title else None
|
|
|
|
|
|
|
|
# extract playlists
|
|
|
|
|
|
|
|
playlists = {}
|
|
|
|
for div in _get_elements_by_tag_and_attrib(
|
|
|
|
webpage_html, attribute='id', value='playlist\\d+', escape_value=False):
|
|
|
|
playlist_id = div.group('value')
|
|
|
|
playlist = []
|
|
|
|
for a in _get_elements_by_tag_and_attrib(
|
|
|
|
div.group('content'), 'a', 'href', value='[^\'"]+?', escape_value=False):
|
|
|
|
playlist.append({
|
|
|
|
'href': unescapeHTML(a.group('value')),
|
|
|
|
'title': unescapeHTML(a.group('content'))
|
|
|
|
})
|
|
|
|
playlists[playlist_id] = playlist
|
|
|
|
|
|
|
|
# select the specified playlist if url fragment exists
|
2020-08-29 07:23:43 +00:00
|
|
|
playlist = None
|
|
|
|
playlist_id = None
|
|
|
|
if fragment:
|
|
|
|
playlist = playlists.get(fragment)
|
|
|
|
playlist_id = fragment
|
|
|
|
else:
|
|
|
|
first = next(iter(playlists.items()))
|
|
|
|
if first:
|
|
|
|
(playlist_id, playlist) = first
|
2020-08-29 07:04:16 +00:00
|
|
|
if not playlist:
|
|
|
|
raise ExtractorError(
|
|
|
|
'Cannot find %s' % fragment if fragment else 'Cannot extract playlist')
|
|
|
|
|
|
|
|
# return url results
|
|
|
|
return self.playlist_result([
|
|
|
|
self.url_result(
|
2020-08-29 07:23:43 +00:00
|
|
|
'https://www.duboku.co' + x['href'], DubokuIE.IE_NAME, video_title=x.get('title'))
|
|
|
|
for x in playlist], series_id + '#' + playlist_id, title)
|