diff options
author | Dave Vasilevsky <dave@vasilevsky.ca> | 2012-12-16 04:05:39 -0500 |
---|---|---|
committer | Dave Vasilevsky <dave@vasilevsky.ca> | 2012-12-16 04:26:22 -0500 |
commit | 2ab1c5ed1a5bf3f63b3e7e6f09d59e431cbe783c (patch) | |
tree | f2c56849c7649eab666ca21842612faca56dcd17 | |
parent | 0b40544f290de329679aebf06e98056e707dd7e1 (diff) |
Support more than 100 videos for justin.tv
-rwxr-xr-x | youtube_dl/InfoExtractors.py | 63 |
1 files changed, 42 insertions, 21 deletions
diff --git a/youtube_dl/InfoExtractors.py b/youtube_dl/InfoExtractors.py index c5ab8907b..e5118a717 100755 --- a/youtube_dl/InfoExtractors.py +++ b/youtube_dl/InfoExtractors.py @@ -3637,37 +3637,27 @@ class NBAIE(InfoExtractor): class JustinTVIE(InfoExtractor): """Information extractor for justin.tv and twitch.tv""" - + # TODO: One broadcast may be split into multiple videos. The key + # 'broadcast_id' is the same for all parts, and 'broadcast_part' + # starts at 1 and increases. Can we treat all parts as one video? + # _VALID_URL = r"""^(?:http(?:s?)://)?www\.(?:justin|twitch)\.tv/ # ([^/]+)(?:/b/([^/]+))?/?(?:#.*)?$""" _VALID_URL = r'^http://www.twitch.tv/(.*)$' IE_NAME = u'justin.tv' + + _max_justin_results = 1000 + _justin_page_limit = 100 def report_extraction(self, file_id): """Report information extraction.""" self._downloader.to_screen(u'[%s] %s: Extracting information' % (self.IE_NAME, file_id)) - def _real_extract(self, url): - mobj = re.match(self._VALID_URL, url) - if mobj is None: - self._downloader.trouble(u'ERROR: invalid URL: %s' % url) - return - - api = 'http://api.justin.tv' - video_id = mobj.group(mobj.lastindex) - if mobj.lastindex == 1: - api += '/channel/archives/%s.json?limit=100' - else: - api += '/clip/show/%s.json' - api = api % (video_id,) - - self.report_extraction(video_id) - # TODO: multiple pages - # TODO: One broadcast may be split into multiple videos. The key - # 'broadcast_id' is the same for all parts, and 'broadcast_part' - # starts at 1 and increases. Can we treat all parts as one video? + # Return count of items, list of *valid* items + def _parse_page(self, url): + print url try: - urlh = compat_urllib_request.urlopen(api) + urlh = compat_urllib_request.urlopen(url) webpage_bytes = urlh.read() webpage = webpage_bytes.decode('utf-8', 'ignore') except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err: @@ -3689,4 +3679,35 @@ class JustinTVIE(InfoExtractor): 'upload_date': video_date, 'ext': video_extension, }) + print len(response) + return (len(response), info) + + def _real_extract(self, url): + mobj = re.match(self._VALID_URL, url) + if mobj is None: + self._downloader.trouble(u'ERROR: invalid URL: %s' % url) + return + + api = 'http://api.justin.tv' + video_id = mobj.group(mobj.lastindex) + paged = False + if mobj.lastindex == 1: + paged = True + api += '/channel/archives/%s.json' + else: + api += '/clip/show/%s.json' + api = api % (video_id,) + + self.report_extraction(video_id) + + info = [] + offset = 0 + limit = self._justin_page_limit + while offset < self._max_justin_results: + page_url = api + ('?offset=%d&limit=%d' % (offset, limit)) + page_count, page_info = self._parse_page(page_url) + info.extend(page_info) + if not paged or page_count != limit: + break + offset += limit return info |