|
@@ -0,0 +1,146 @@
|
|
|
+# encoding: utf-8
|
|
|
+from __future__ import unicode_literals
|
|
|
+
|
|
|
+import re
|
|
|
+
|
|
|
+from .crunchyroll import CrunchyrollIE
|
|
|
+
|
|
|
+from .common import InfoExtractor
|
|
|
+from ..compat import compat_HTTPError
|
|
|
+from ..utils import (
|
|
|
+ ExtractorError,
|
|
|
+ int_or_none,
|
|
|
+ remove_start,
|
|
|
+ xpath_text,
|
|
|
+)
|
|
|
+
|
|
|
+
|
|
|
+class SoompiBaseIE(InfoExtractor):
|
|
|
+ def _get_episodes(self, webpage, episode_filter=None):
|
|
|
+ episodes = self._parse_json(
|
|
|
+ self._search_regex(
|
|
|
+ r'VIDEOS\s*=\s*(\[.+?\]);', webpage, 'episodes JSON'),
|
|
|
+ None)
|
|
|
+ return list(filter(episode_filter, episodes))
|
|
|
+
|
|
|
+
|
|
|
+class SoompiIE(SoompiBaseIE, CrunchyrollIE):
|
|
|
+ IE_NAME = 'soompi'
|
|
|
+ _VALID_URL = r'https?://tv\.soompi\.com/(?:en/)?watch/(?P<id>[0-9]+)'
|
|
|
+ _TESTS = [{
|
|
|
+ 'url': 'http://tv.soompi.com/en/watch/29235',
|
|
|
+ 'info_dict': {
|
|
|
+ 'id': '29235',
|
|
|
+ 'ext': 'mp4',
|
|
|
+ 'title': 'Episode 1096',
|
|
|
+ 'description': '2015-05-20'
|
|
|
+ },
|
|
|
+ 'params': {
|
|
|
+ 'skip_download': True,
|
|
|
+ },
|
|
|
+ }]
|
|
|
+
|
|
|
+ def _get_episode(self, webpage, video_id):
|
|
|
+ return self._get_episodes(webpage, lambda x: x['id'] == video_id)[0]
|
|
|
+
|
|
|
+ def _get_subtitles(self, config, video_id):
|
|
|
+ sub_langs = {}
|
|
|
+ for subtitle in config.findall('./{default}preload/subtitles/subtitle'):
|
|
|
+ sub_langs[subtitle.attrib['id']] = subtitle.attrib['title']
|
|
|
+
|
|
|
+ subtitles = {}
|
|
|
+ for s in config.findall('./{default}preload/subtitle'):
|
|
|
+ lang_code = sub_langs.get(s.attrib['id'])
|
|
|
+ if not lang_code:
|
|
|
+ continue
|
|
|
+ sub_id = s.get('id')
|
|
|
+ data = xpath_text(s, './data', 'data')
|
|
|
+ iv = xpath_text(s, './iv', 'iv')
|
|
|
+ if not id or not iv or not data:
|
|
|
+ continue
|
|
|
+ subtitle = self._decrypt_subtitles(data, iv, sub_id).decode('utf-8')
|
|
|
+ subtitles[lang_code] = self._extract_subtitles(subtitle)
|
|
|
+ return subtitles
|
|
|
+
|
|
|
+ def _real_extract(self, url):
|
|
|
+ video_id = self._match_id(url)
|
|
|
+
|
|
|
+ try:
|
|
|
+ webpage = self._download_webpage(
|
|
|
+ url, video_id, 'Downloading episode page')
|
|
|
+ except ExtractorError as ee:
|
|
|
+ if isinstance(ee.cause, compat_HTTPError) and ee.cause.code == 403:
|
|
|
+ webpage = ee.cause.read()
|
|
|
+ block_message = self._html_search_regex(
|
|
|
+ r'(?s)<div class="block-message">(.+?)</div>', webpage,
|
|
|
+ 'block message', default=None)
|
|
|
+ if block_message:
|
|
|
+ raise ExtractorError(block_message, expected=True)
|
|
|
+ raise
|
|
|
+
|
|
|
+ formats = []
|
|
|
+ config = None
|
|
|
+ for format_id in re.findall(r'\?quality=([0-9a-zA-Z]+)', webpage):
|
|
|
+ config = self._download_xml(
|
|
|
+ 'http://tv.soompi.com/en/show/_/%s-config.xml?mode=hls&quality=%s' % (video_id, format_id),
|
|
|
+ video_id, 'Downloading %s XML' % format_id)
|
|
|
+ m3u8_url = xpath_text(
|
|
|
+ config, './{default}preload/stream_info/file',
|
|
|
+ '%s m3u8 URL' % format_id)
|
|
|
+ if not m3u8_url:
|
|
|
+ continue
|
|
|
+ formats.extend(self._extract_m3u8_formats(
|
|
|
+ m3u8_url, video_id, 'mp4', m3u8_id=format_id))
|
|
|
+ self._sort_formats(formats)
|
|
|
+
|
|
|
+ episode = self._get_episode(webpage, video_id)
|
|
|
+
|
|
|
+ title = episode['name']
|
|
|
+ description = episode.get('description')
|
|
|
+ duration = int_or_none(episode.get('duration'))
|
|
|
+
|
|
|
+ thumbnails = [{
|
|
|
+ 'id': thumbnail_id,
|
|
|
+ 'url': thumbnail_url,
|
|
|
+ } for thumbnail_id, thumbnail_url in episode.get('img_url', {}).items()]
|
|
|
+
|
|
|
+ subtitles = self.extract_subtitles(config, video_id)
|
|
|
+
|
|
|
+ return {
|
|
|
+ 'id': video_id,
|
|
|
+ 'title': title,
|
|
|
+ 'description': description,
|
|
|
+ 'thumbnails': thumbnails,
|
|
|
+ 'duration': duration,
|
|
|
+ 'formats': formats,
|
|
|
+ 'subtitles': subtitles
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+class SoompiShowIE(SoompiBaseIE):
|
|
|
+ IE_NAME = 'soompi:show'
|
|
|
+ _VALID_URL = r'https?://tv\.soompi\.com/en/shows/(?P<id>[0-9a-zA-Z\-_]+)'
|
|
|
+ _TESTS = [{
|
|
|
+ 'url': 'http://tv.soompi.com/en/shows/liar-game',
|
|
|
+ 'info_dict': {
|
|
|
+ 'id': 'liar-game',
|
|
|
+ 'title': 'Liar Game',
|
|
|
+ 'description': 'md5:52c02bce0c1a622a95823591d0589b66',
|
|
|
+ },
|
|
|
+ 'playlist_count': 14,
|
|
|
+ }]
|
|
|
+
|
|
|
+ def _real_extract(self, url):
|
|
|
+ show_id = self._match_id(url)
|
|
|
+
|
|
|
+ webpage = self._download_webpage(
|
|
|
+ url, show_id, 'Downloading show page')
|
|
|
+
|
|
|
+ title = remove_start(self._og_search_title(webpage), 'SoompiTV | ')
|
|
|
+ description = self._og_search_description(webpage)
|
|
|
+
|
|
|
+ entries = [
|
|
|
+ self.url_result('http://tv.soompi.com/en/watch/%s' % episode['id'], 'Soompi')
|
|
|
+ for episode in self._get_episodes(webpage)]
|
|
|
+
|
|
|
+ return self.playlist_result(entries, show_id, title, description)
|