diff options
author | Takuya Tsuchida <takuya0301@gmail.com> | 2013-11-23 18:19:44 +0900 |
---|---|---|
committer | Takuya Tsuchida <takuya0301@gmail.com> | 2013-11-23 18:19:44 +0900 |
commit | 52ad14aeb0178a187a861e7ce2259b7046702281 (patch) | |
tree | c9fb6aae6e234e7df2bcf4fd0c9957c660bd6b8c | |
parent | a87b0615aa311083923e607c3d1a5cdceab818f7 (diff) |
Add support for niconico
-rw-r--r-- | youtube_dl/extractor/__init__.py | 1 | ||||
-rw-r--r-- | youtube_dl/extractor/niconico.py | 190 |
2 files changed, 191 insertions, 0 deletions
diff --git a/youtube_dl/extractor/__init__.py b/youtube_dl/extractor/__init__.py index 02f9e2546..f443f11f6 100644 --- a/youtube_dl/extractor/__init__.py +++ b/youtube_dl/extractor/__init__.py @@ -98,6 +98,7 @@ from .nba import NBAIE from .nbc import NBCNewsIE from .newgrounds import NewgroundsIE from .nhl import NHLIE, NHLVideocenterIE +from .niconico import NiconicoIE from .nowvideo import NowVideoIE from .ooyala import OoyalaIE from .orf import ORFIE diff --git a/youtube_dl/extractor/niconico.py b/youtube_dl/extractor/niconico.py new file mode 100644 index 000000000..8638a8ee8 --- /dev/null +++ b/youtube_dl/extractor/niconico.py @@ -0,0 +1,190 @@ +# encoding: utf-8 + +import re +import socket +import xml.etree.ElementTree + +from .common import InfoExtractor +from ..utils import ( + compat_http_client, + compat_urllib_error, + compat_urllib_parse, + compat_urllib_request, + compat_urlparse, + compat_str, + + ExtractorError, + unified_strdate, +) + +class NiconicoIE(InfoExtractor): + IE_NAME = u'niconico' + IE_DESC = u'ニコニコ動画' + + _TEST = { + u'url': u'http://www.nicovideo.jp/watch/sm22312215', + u'file': u'sm22312215.mp4', + u'md5': u'd1a75c0823e2f629128c43e1212760f9', + u'info_dict': { + u'title': u'Big Buck Bunny', + u'uploader': u'takuya0301', + u'uploader_id': u'2698420', + u'upload_date': u'20131123', + u'description': u'(c) copyright 2008, Blender Foundation / www.bigbuckbunny.org', + }, + u'params': { + u'username': u'ydl.niconico@gmail.com', + u'password': u'youtube-dl', + }, + } + + _VALID_URL = r'^(?:https?://)?(?:www\.)?nicovideo\.jp/watch/([a-z][a-z][0-9]+)(?:.*)$' + _LOGIN_URL = 'https://secure.nicovideo.jp/secure/login' + _NETRC_MACHINE = 'niconico' + # If True it will raise an error if no login info is provided + _LOGIN_REQUIRED = True + + def _real_initialize(self): + self._login() + + def _login(self): + (username, password) = self._get_login_info() + # No authentication to be performed + if username is None: + if self._LOGIN_REQUIRED: + raise ExtractorError(u'No login info available, needed for using %s.' % self.IE_NAME, expected=True) + return False + + # Log in + login_form_strs = { + u'mail': username, + u'password': password, + } + # Convert to UTF-8 *before* urlencode because Python 2.x's urlencode + # chokes on unicode + login_form = dict((k.encode('utf-8'), v.encode('utf-8')) for k,v in login_form_strs.items()) + login_data = compat_urllib_parse.urlencode(login_form).encode('ascii') + request = compat_urllib_request.Request(self._LOGIN_URL, login_data) + try: + self.report_login() + login_results = compat_urllib_request.urlopen(request).read().decode('utf-8') + if re.search(r'(?i)<h1 class="mb8p4">Log in error</h1>', login_results) is not None: + self._downloader.report_warning(u'unable to log in: bad username or password') + return False + except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err: + self._downloader.report_warning(u'unable to log in: %s' % compat_str(err)) + return False + return True + + def _real_extract(self, url): + video_id = self._extract_id(url) + + # Get video webpage + self.report_video_webpage_download(video_id) + url = 'http://www.nicovideo.jp/watch/' + video_id + request = compat_urllib_request.Request(url) + try: + video_webpage = compat_urllib_request.urlopen(request).read() + except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err: + raise ExtractorError(u'Unable to download video webpage: %s' % compat_str(err)) + + # Get video info + self.report_video_info_webpage_download(video_id) + url = 'http://ext.nicovideo.jp/api/getthumbinfo/' + video_id + request = compat_urllib_request.Request(url) + try: + video_info_webpage = compat_urllib_request.urlopen(request).read() + except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err: + raise ExtractorError(u'Unable to download video info webpage: %s' % compat_str(err)) + + # Get flv info + self.report_flv_info_webpage_download(video_id) + url = 'http://flapi.nicovideo.jp/api/getflv?v=' + video_id + request = compat_urllib_request.Request(url) + try: + flv_info_webpage = compat_urllib_request.urlopen(request).read() + except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err: + raise ExtractorError(u'Unable to download flv info webpage: %s' % compat_str(err)) + + # Start extracting information + self.report_information_extraction(video_id) + video_info = xml.etree.ElementTree.fromstring(video_info_webpage) + + # url + video_real_url = compat_urlparse.parse_qs(flv_info_webpage.decode('utf-8'))['url'][0] + + # title + video_title = video_info.find('.//title').text + + # ext + video_extension = video_info.find('.//movie_type').text + + # format + video_format = video_extension.upper() + + # thumbnail + video_thumbnail = video_info.find('.//thumbnail_url').text + + # description + video_description = video_info.find('.//description').text + + # uploader_id + video_uploader_id = video_info.find('.//user_id').text + + # uploader + url = 'http://seiga.nicovideo.jp/api/user/info?id=' + video_uploader_id + request = compat_urllib_request.Request(url) + try: + user_info_webpage = compat_urllib_request.urlopen(request).read() + except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err: + self._downloader.report_warning(u'Unable to download user info webpage: %s' % compat_str(err)) + + user_info = xml.etree.ElementTree.fromstring(user_info_webpage) + video_uploader = user_info.find('.//nickname').text + + # uploder_date + video_upload_date = unified_strdate(video_info.find('.//first_retrieve').text.split('+')[0]) + + # view_count + video_view_count = video_info.find('.//view_counter').text + + # webpage_url + video_webpage_url = video_info.find('.//watch_url').text + + return { + 'id': video_id, + 'url': video_real_url, + 'title': video_title, + 'ext': video_extension, + 'format': video_format, + 'thumbnail': video_thumbnail, + 'description': video_description, + 'uploader': video_uploader, + 'upload_date': video_upload_date, + 'uploader_id': video_uploader_id, + 'view_count': video_view_count, + 'webpage_url': video_webpage_url, + } + + def _extract_id(self, url): + mobj = re.match(self._VALID_URL, url) + if mobj is None: + raise ExtractorError(u'Invalid URL: %s' % url) + video_id = mobj.group(1) + return video_id + + def report_video_webpage_download(self, video_id): + """Report attempt to download video webpage.""" + self.to_screen(u'%s: Downloading video webpage' % video_id) + + def report_video_info_webpage_download(self, video_id): + """Report attempt to download video info webpage.""" + self.to_screen(u'%s: Downloading video info webpage' % video_id) + + def report_flv_info_webpage_download(self, video_id): + """Report attempt to download flv info webpage.""" + self.to_screen(u'%s: Downloading flv info webpage' % video_id) + + def report_information_extraction(self, video_id): + """Report attempt to extract video information.""" + self.to_screen(u'%s: Extracting video information' % video_id) |