revision3.py 4.3 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103
  1. # coding: utf-8
  2. from __future__ import unicode_literals
  3. import re
  4. from .common import InfoExtractor
  5. from ..compat import compat_str
  6. from ..utils import (
  7. int_or_none,
  8. parse_iso8601,
  9. unescapeHTML,
  10. )
  11. class Revision3IE(InfoExtractor):
  12. _VALID_URL = r'https?://(?:www\.)?(?P<domain>(?:revision3|testtube|animalist)\.com)/(?P<id>[^/]+(?:/[^/?#]+)?)'
  13. _TESTS = [{
  14. 'url': 'http://www.revision3.com/technobuffalo/5-google-predictions-for-2016',
  15. 'md5': 'd94a72d85d0a829766de4deb8daaf7df',
  16. 'info_dict': {
  17. 'id': '73034',
  18. 'ext': 'webm',
  19. 'title': '5 Google Predictions for 2016',
  20. 'description': 'Google had a great 2015, but it\'s already time to look ahead. Here are our five predictions for 2016.',
  21. 'upload_date': '20151228',
  22. 'timestamp': 1451325600,
  23. 'duration': 187,
  24. }
  25. }, {
  26. 'url': 'http://testtube.com/brainstuff',
  27. 'info_dict': {
  28. 'id': '251',
  29. 'title': 'BrainStuff',
  30. 'description': 'Whether the topic is popcorn or particle physics, you can count on the HowStuffWorks team to explore-and explain-the everyday science in the world around us on BrainStuff.',
  31. },
  32. 'playlist_mincount': 93,
  33. }]
  34. _PAGE_DATA_TEMPLATE = 'http://www.%s/apiProxy/ddn/%s?domain=%s'
  35. _API_KEY = 'ba9c741bce1b9d8e3defcc22193f3651b8867e62'
  36. def _real_extract(self, url):
  37. domain, display_id = re.match(self._VALID_URL, url).groups()
  38. page_info = self._download_json(
  39. self._PAGE_DATA_TEMPLATE % (domain, display_id, domain), display_id)
  40. if page_info['data']['type'] == 'episode':
  41. episode_data = page_info['data']
  42. video_id = compat_str(episode_data['video']['data']['id'])
  43. video_data = self._download_json(
  44. 'http://revision3.com/api/getPlaylist.json?api_key=%s&codecs=h264,vp8,theora&video_id=%s' % (self._API_KEY, video_id),
  45. video_id)['items'][0]
  46. formats = []
  47. for media_type, media in video_data['media'].items():
  48. for quality_id, quality in media.items():
  49. if quality_id == 'hls':
  50. formats.extend(self._extract_m3u8_formats(
  51. quality['url'], video_id, 'mp4',
  52. 'm3u8_native', m3u8_id='hls', fatal=False))
  53. else:
  54. formats.append({
  55. 'url': quality['url'],
  56. 'format_id': '%s-%s' % (media_type, quality_id),
  57. 'tbr': int_or_none(quality.get('bitrate')),
  58. })
  59. self._sort_formats(formats)
  60. thumbnails = [{
  61. 'url': image_url,
  62. 'id': image_id,
  63. } for image_id, image_url in video_data.get('images', {}).items()]
  64. return {
  65. 'id': video_id,
  66. 'title': unescapeHTML(video_data['title']),
  67. 'description': unescapeHTML(video_data.get('summary')),
  68. 'timestamp': parse_iso8601(episode_data.get('publishTime'), ' '),
  69. 'author': episode_data.get('author'),
  70. 'duration': int_or_none(video_data.get('duration')),
  71. 'thumbnails': thumbnails,
  72. 'formats': formats,
  73. }
  74. else:
  75. show_data = page_info['show']['data']
  76. episodes_data = page_info['episodes']['data']
  77. num_episodes = page_info['meta']['totalEpisodes']
  78. processed_episodes = 0
  79. entries = []
  80. page_num = 1
  81. while True:
  82. entries.extend([self.url_result(
  83. url + '/%s' % episode['slug']) for episode in episodes_data])
  84. processed_episodes += len(episodes_data)
  85. if processed_episodes == num_episodes:
  86. break
  87. page_num += 1
  88. episodes_data = self._download_json(self._PAGE_DATA_TEMPLATE % (
  89. domain, display_id + '/' + compat_str(page_num), domain),
  90. display_id)['episodes']['data']
  91. return self.playlist_result(
  92. entries, compat_str(show_data['id']),
  93. show_data.get('name'), show_data.get('summary'))