- from __future__ import unicode_literals
-
- import json
- import re
- import socket
-
- from .common import InfoExtractor
- from ..compat import (
- compat_http_client,
- compat_str,
- compat_urllib_error,
- compat_urllib_parse,
- compat_urllib_request,
- )
- from ..utils import (
- ExtractorError,
- int_or_none,
- limit_length,
- urlencode_postdata,
- )
-
-
- class FacebookIE(InfoExtractor):
- _VALID_URL = r'''(?x)
- https?://(?:\w+\.)?facebook\.com/
- (?:[^#]*?\#!/)?
- (?:
- (?:video/video\.php|photo\.php|video\.php|video/embed)\?(?:.*?)
- (?:v|video_id)=|
- [^/]+/videos/(?:[^/]+/)?
- )
- (?P<id>[0-9]+)
- (?:.*)'''
- _LOGIN_URL = 'https://www.facebook.com/login.php?next=http%3A%2F%2Ffacebook.com%2Fhome.php&login_attempt=1'
- _CHECKPOINT_URL = 'https://www.facebook.com/checkpoint/?next=http%3A%2F%2Ffacebook.com%2Fhome.php&_fb_noscript=1'
- _NETRC_MACHINE = 'facebook'
- IE_NAME = 'facebook'
- _TESTS = [{
- 'url': 'https://www.facebook.com/video.php?v=637842556329505&fref=nf',
- 'md5': '6a40d33c0eccbb1af76cf0485a052659',
- 'info_dict': {
- 'id': '637842556329505',
- 'ext': 'mp4',
- 'title': 're:Did you know Kei Nishikori is the first Asian man to ever reach a Grand Slam',
- }
- }, {
- 'note': 'Video without discernible title',
- 'url': 'https://www.facebook.com/video.php?v=274175099429670',
- 'info_dict': {
- 'id': '274175099429670',
- 'ext': 'mp4',
- 'title': 'Facebook video #274175099429670',
- },
- 'expected_warnings': [
- 'title'
- ]
- }, {
- 'url': 'https://www.facebook.com/video.php?v=10204634152394104',
- 'only_matching': True,
- }, {
- 'url': 'https://www.facebook.com/amogood/videos/1618742068337349/?fref=nf',
- 'only_matching': True,
- }, {
- 'url': 'https://www.facebook.com/ChristyClarkForBC/videos/vb.22819070941/10153870694020942/?type=2&theater',
- 'only_matching': True,
- }]
-
- def _login(self):
- (useremail, password) = self._get_login_info()
- if useremail is None:
- return
-
- login_page_req = compat_urllib_request.Request(self._LOGIN_URL)
- login_page_req.add_header('Cookie', 'locale=en_US')
- login_page = self._download_webpage(login_page_req, None,
- note='Downloading login page',
- errnote='Unable to download login page')
- lsd = self._search_regex(
- r'<input type="hidden" name="lsd" value="([^"]*)"',
- login_page, 'lsd')
- lgnrnd = self._search_regex(r'name="lgnrnd" value="([^"]*?)"', login_page, 'lgnrnd')
-
- login_form = {
- 'email': useremail,
- 'pass': password,
- 'lsd': lsd,
- 'lgnrnd': lgnrnd,
- 'next': 'http://facebook.com/home.php',
- 'default_persistent': '0',
- 'legacy_return': '1',
- 'timezone': '-60',
- 'trynum': '1',
- }
- request = compat_urllib_request.Request(self._LOGIN_URL, urlencode_postdata(login_form))
- request.add_header('Content-Type', 'application/x-www-form-urlencoded')
- try:
- login_results = self._download_webpage(request, None,
- note='Logging in', errnote='unable to fetch login page')
- if re.search(r'<form(.*)name="login"(.*)</form>', login_results) is not None:
- self._downloader.report_warning('unable to log in: bad username/password, or exceded login rate limit (~3/min). Check credentials or wait.')
- return
-
- check_form = {
- 'fb_dtsg': self._search_regex(r'name="fb_dtsg" value="(.+?)"', login_results, 'fb_dtsg'),
- 'h': self._search_regex(
- r'name="h"\s+(?:\w+="[^"]+"\s+)*?value="([^"]+)"', login_results, 'h'),
- 'name_action_selected': 'dont_save',
- }
- check_req = compat_urllib_request.Request(self._CHECKPOINT_URL, urlencode_postdata(check_form))
- check_req.add_header('Content-Type', 'application/x-www-form-urlencoded')
- check_response = self._download_webpage(check_req, None,
- note='Confirming login')
- if re.search(r'id="checkpointSubmitButton"', check_response) is not None:
- self._downloader.report_warning('Unable to confirm login, you have to login in your brower and authorize the login.')
- except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
- self._downloader.report_warning('unable to log in: %s' % compat_str(err))
- return
-
- def _real_initialize(self):
- self._login()
-
- def _real_extract(self, url):
- video_id = self._match_id(url)
- url = 'https://www.facebook.com/video/video.php?v=%s' % video_id
- webpage = self._download_webpage(url, video_id)
-
- BEFORE = '{swf.addParam(param[0], param[1]);});\n'
- AFTER = '.forEach(function(variable) {swf.addVariable(variable[0], variable[1]);});'
- m = re.search(re.escape(BEFORE) + '(.*?)' + re.escape(AFTER), webpage)
- if not m:
- m_msg = re.search(r'class="[^"]*uiInterstitialContent[^"]*"><div>(.*?)</div>', webpage)
- if m_msg is not None:
- raise ExtractorError(
- 'The video is not available, Facebook said: "%s"' % m_msg.group(1),
- expected=True)
- else:
- raise ExtractorError('Cannot parse data')
- data = dict(json.loads(m.group(1)))
- params_raw = compat_urllib_parse.unquote(data['params'])
- params = json.loads(params_raw)
- video_data = params['video_data'][0]
-
- formats = []
- for quality in ['sd', 'hd']:
- src = video_data.get('%s_src' % quality)
- if src is not None:
- formats.append({
- 'format_id': quality,
- 'url': src,
- })
- if not formats:
- raise ExtractorError('Cannot find video formats')
-
- video_title = self._html_search_regex(
- r'<h2\s+[^>]*class="uiHeaderTitle"[^>]*>([^<]*)</h2>', webpage, 'title',
- default=None)
- if not video_title:
- video_title = self._html_search_regex(
- r'(?s)<span class="fbPhotosPhotoCaption".*?id="fbPhotoPageCaption"><span class="hasCaption">(.*?)</span>',
- webpage, 'alternative title', fatal=False)
- video_title = limit_length(video_title, 80)
- if not video_title:
- video_title = 'Facebook video #%s' % video_id
-
- return {
- 'id': video_id,
- 'title': video_title,
- 'formats': formats,
- 'duration': int_or_none(video_data.get('video_duration')),
- 'thumbnail': video_data.get('thumbnail_src'),
- }
|