|
@@ -1,13 +1,11 @@
|
|
# coding: utf-8
|
|
# coding: utf-8
|
|
from __future__ import unicode_literals
|
|
from __future__ import unicode_literals
|
|
|
|
|
|
-import datetime
|
|
|
|
import re
|
|
import re
|
|
|
|
+import codecs
|
|
|
|
|
|
from .common import InfoExtractor
|
|
from .common import InfoExtractor
|
|
-from ..utils import (
|
|
|
|
- ExtractorError,
|
|
|
|
-)
|
|
|
|
|
|
+from ..utils import unified_strdate
|
|
|
|
|
|
|
|
|
|
class GooglePlusIE(InfoExtractor):
|
|
class GooglePlusIE(InfoExtractor):
|
|
@@ -19,74 +17,57 @@ class GooglePlusIE(InfoExtractor):
|
|
'info_dict': {
|
|
'info_dict': {
|
|
'id': 'ZButuJc6CtH',
|
|
'id': 'ZButuJc6CtH',
|
|
'ext': 'flv',
|
|
'ext': 'flv',
|
|
|
|
+ 'title': '嘆きの天使 降臨',
|
|
'upload_date': '20120613',
|
|
'upload_date': '20120613',
|
|
'uploader': '井上ヨシマサ',
|
|
'uploader': '井上ヨシマサ',
|
|
- 'title': '嘆きの天使 降臨',
|
|
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
def _real_extract(self, url):
|
|
def _real_extract(self, url):
|
|
- # Extract id from URL
|
|
|
|
- mobj = re.match(self._VALID_URL, url)
|
|
|
|
-
|
|
|
|
- video_id = mobj.group('id')
|
|
|
|
|
|
+ video_id = self._match_id(url)
|
|
|
|
|
|
# Step 1, Retrieve post webpage to extract further information
|
|
# Step 1, Retrieve post webpage to extract further information
|
|
webpage = self._download_webpage(url, video_id, 'Downloading entry webpage')
|
|
webpage = self._download_webpage(url, video_id, 'Downloading entry webpage')
|
|
|
|
|
|
- self.report_extraction(video_id)
|
|
|
|
-
|
|
|
|
- # Extract update date
|
|
|
|
- upload_date = self._html_search_regex(
|
|
|
|
|
|
+ title = self._og_search_description(webpage).splitlines()[0]
|
|
|
|
+ upload_date = unified_strdate(self._html_search_regex(
|
|
r'''(?x)<a.+?class="o-U-s\s[^"]+"\s+style="display:\s*none"\s*>
|
|
r'''(?x)<a.+?class="o-U-s\s[^"]+"\s+style="display:\s*none"\s*>
|
|
([0-9]{4}-[0-9]{2}-[0-9]{2})</a>''',
|
|
([0-9]{4}-[0-9]{2}-[0-9]{2})</a>''',
|
|
- webpage, 'upload date', fatal=False, flags=re.VERBOSE)
|
|
|
|
- if upload_date:
|
|
|
|
- # Convert timestring to a format suitable for filename
|
|
|
|
- upload_date = datetime.datetime.strptime(upload_date, "%Y-%m-%d")
|
|
|
|
- upload_date = upload_date.strftime('%Y%m%d')
|
|
|
|
-
|
|
|
|
- # Extract uploader
|
|
|
|
- uploader = self._html_search_regex(r'rel\="author".*?>(.*?)</a>',
|
|
|
|
- webpage, 'uploader', fatal=False)
|
|
|
|
-
|
|
|
|
- # Extract title
|
|
|
|
- # Get the first line for title
|
|
|
|
- video_title = self._og_search_description(webpage).splitlines()[0]
|
|
|
|
|
|
+ webpage, 'upload date', fatal=False, flags=re.VERBOSE))
|
|
|
|
+ uploader = self._html_search_regex(
|
|
|
|
+ r'rel="author".*?>(.*?)</a>', webpage, 'uploader', fatal=False)
|
|
|
|
|
|
# Step 2, Simulate clicking the image box to launch video
|
|
# Step 2, Simulate clicking the image box to launch video
|
|
DOMAIN = 'https://plus.google.com/'
|
|
DOMAIN = 'https://plus.google.com/'
|
|
- video_page = self._search_regex(r'<a href="((?:%s)?photos/.*?)"' % re.escape(DOMAIN),
|
|
|
|
|
|
+ video_page = self._search_regex(
|
|
|
|
+ r'<a href="((?:%s)?photos/.*?)"' % re.escape(DOMAIN),
|
|
webpage, 'video page URL')
|
|
webpage, 'video page URL')
|
|
if not video_page.startswith(DOMAIN):
|
|
if not video_page.startswith(DOMAIN):
|
|
video_page = DOMAIN + video_page
|
|
video_page = DOMAIN + video_page
|
|
|
|
|
|
webpage = self._download_webpage(video_page, video_id, 'Downloading video page')
|
|
webpage = self._download_webpage(video_page, video_id, 'Downloading video page')
|
|
|
|
|
|
- # Extract video links all sizes
|
|
|
|
- pattern = r'\d+,\d+,(\d+),"(https\://redirector\.googlevideo\.com.*?)"'
|
|
|
|
- mobj = re.findall(pattern, webpage)
|
|
|
|
- if len(mobj) == 0:
|
|
|
|
- raise ExtractorError('Unable to extract video links')
|
|
|
|
-
|
|
|
|
- # Sort in resolution
|
|
|
|
- links = sorted(mobj)
|
|
|
|
|
|
+ def unicode_escape(s):
|
|
|
|
+ decoder = codecs.getdecoder('unicode_escape')
|
|
|
|
+ return re.sub(
|
|
|
|
+ r'\\u[0-9a-fA-F]{4,}',
|
|
|
|
+ lambda m: decoder(m.group(0))[0],
|
|
|
|
+ s)
|
|
|
|
|
|
- # Choose the lowest of the sort, i.e. highest resolution
|
|
|
|
- video_url = links[-1]
|
|
|
|
- # Only get the url. The resolution part in the tuple has no use anymore
|
|
|
|
- video_url = video_url[-1]
|
|
|
|
- # Treat escaped \u0026 style hex
|
|
|
|
- try:
|
|
|
|
- video_url = video_url.decode("unicode_escape")
|
|
|
|
- except AttributeError: # Python 3
|
|
|
|
- video_url = bytes(video_url, 'ascii').decode('unicode-escape')
|
|
|
|
|
|
+ # Extract video links all sizes
|
|
|
|
+ formats = [{
|
|
|
|
+ 'url': unicode_escape(video_url),
|
|
|
|
+ 'ext': 'flv',
|
|
|
|
+ 'width': int(width),
|
|
|
|
+ 'height': int(height),
|
|
|
|
+ } for width, height, video_url in re.findall(
|
|
|
|
+ r'\d+,(\d+),(\d+),"(https?://redirector\.googlevideo\.com.*?)"', webpage)]
|
|
|
|
+ self._sort_formats(formats)
|
|
|
|
|
|
return {
|
|
return {
|
|
'id': video_id,
|
|
'id': video_id,
|
|
- 'url': video_url,
|
|
|
|
|
|
+ 'title': title,
|
|
'uploader': uploader,
|
|
'uploader': uploader,
|
|
'upload_date': upload_date,
|
|
'upload_date': upload_date,
|
|
- 'title': video_title,
|
|
|
|
- 'ext': 'flv',
|
|
|
|
|
|
+ 'formats': formats,
|
|
}
|
|
}
|