|
|
@ -2,43 +2,26 @@ from .common import InfoExtractor |
|
|
|
from math import ceil |
|
|
|
import re |
|
|
|
|
|
|
|
|
|
|
|
class ToypicsIE(InfoExtractor): |
|
|
|
_VALID_URL = r'(?:http://)?videos\.toypics\.net/.*' |
|
|
|
IE_DESC = 'Toypics user profile' |
|
|
|
_VALID_URL = r'http://videos\.toypics\.net/view/(?P<id>[0-9]+)/.*' |
|
|
|
_TEST = { |
|
|
|
'url': 'http://videos.toypics.net/view/514/chancebulged,-2-1/', |
|
|
|
#'md5': '8a8b546956bbd0e769dbe28f6e80abb3', == $head -c10K 12929646011616163504.mp4 |md5sum //no idea why it fails |
|
|
|
'md5': '16e806ad6d6f58079d210fe30985e08b', |
|
|
|
'info_dict': { |
|
|
|
'id': '514', |
|
|
|
'ext': 'mp4', |
|
|
|
'title': 'Chance-Bulge\'d, 2', |
|
|
|
'age_limit': 18 |
|
|
|
'age_limit': 18, |
|
|
|
'uploader': 'kidsune', |
|
|
|
} |
|
|
|
} |
|
|
|
PAGINATED=8 |
|
|
|
|
|
|
|
def _real_extract(self, url): |
|
|
|
mobj = re.match(r'(http://)?videos\.toypics\.net/(?P<username>[^/?]+)$', url) |
|
|
|
if not mobj: |
|
|
|
return self.extract_one(url) |
|
|
|
return [self.extract_one(u) for u in self.process_paginated(url, |
|
|
|
r'public/">Public Videos \((?P<videos_count>[0-9]+)\)</a></li>', |
|
|
|
r'<p class="video-entry-title">\n\s*<a href="(http://videos.toypics.net/view/[^"]+)">' |
|
|
|
)] |
|
|
|
|
|
|
|
def process_paginated(self, profile_url, re_total, re_video_page): |
|
|
|
profile_page = self._download_webpage(profile_url, 'profile' , 'getting profile page: '+profile_url) |
|
|
|
videos_count = self._html_search_regex(re_total, profile_page, 'videos count') |
|
|
|
lst = [] |
|
|
|
for n in xrange(1,int(ceil(float(videos_count)/self.PAGINATED)) +1): |
|
|
|
lpage_url = profile_url +'/public/%d'%n |
|
|
|
lpage = self._download_webpage(lpage_url, 'page %d'%n) |
|
|
|
lst.extend(re.findall(re_video_page, lpage)) |
|
|
|
return lst |
|
|
|
|
|
|
|
def extract_one(self,url): |
|
|
|
mobj = re.match(r'(http://)?videos\.toypics\.net/view/(?P<videoid>[0-9]+)/.*', url) |
|
|
|
video_id = mobj.group('videoid') |
|
|
|
page = self._download_webpage(url, video_id, 'getting page: '+url) |
|
|
|
mobj = re.match(self._VALID_URL, url) |
|
|
|
video_id = mobj.group('id') |
|
|
|
page = self._download_webpage(url, video_id) |
|
|
|
video_url = self._html_search_regex( |
|
|
|
r'src:\s+"(http://static[0-9]+\.toypics\.net/flvideo/[^"]+)"', page, 'video URL') |
|
|
|
title = self._html_search_regex( |
|
|
@ -48,8 +31,46 @@ class ToypicsIE(InfoExtractor): |
|
|
|
return { |
|
|
|
'id': video_id, |
|
|
|
'url': video_url, |
|
|
|
'ext': video_url[-3:], |
|
|
|
'title': title, |
|
|
|
'uploader': username, |
|
|
|
'age_limit': 18 |
|
|
|
'age_limit': 18, |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
class ToypicsUserIE(InfoExtractor): |
|
|
|
IE_DESC = 'Toypics user profile' |
|
|
|
_VALID_URL = r'http://videos\.toypics\.net/(?P<username>[^/?]+)(?:$|[?#])' |
|
|
|
|
|
|
|
def _real_extract(self, url): |
|
|
|
mobj = re.match(self._VALID_URL, url) |
|
|
|
username = mobj.group('username') |
|
|
|
|
|
|
|
profile_page = self._download_webpage( |
|
|
|
url, username, note='Retrieving profile page') |
|
|
|
|
|
|
|
video_count = int(self._search_regex( |
|
|
|
r'public/">Public Videos \(([0-9]+)\)</a></li>', profile_page, |
|
|
|
'video count')) |
|
|
|
|
|
|
|
PAGE_SIZE = 8 |
|
|
|
urls = [] |
|
|
|
page_count = (video_count + PAGE_SIZE + 1) // PAGE_SIZE |
|
|
|
for n in range(1, page_count + 1): |
|
|
|
lpage_url = url + '/public/%d' % n |
|
|
|
lpage = self._download_webpage( |
|
|
|
lpage_url, username, |
|
|
|
note='Downloading page %d/%d' % (n, page_count)) |
|
|
|
urls.extend( |
|
|
|
re.findall( |
|
|
|
r'<p class="video-entry-title">\n\s*<a href="(http://videos.toypics.net/view/[^"]+)">', |
|
|
|
lpage)) |
|
|
|
|
|
|
|
return { |
|
|
|
'_type': 'playlist', |
|
|
|
'id': username, |
|
|
|
'entries': [{ |
|
|
|
'_type': 'url', |
|
|
|
'url': url, |
|
|
|
'ie_key': 'Toypics', |
|
|
|
} for url in urls] |
|
|
|
} |