diff options
Diffstat (limited to 'youtube_dl/extractor')
-rw-r--r-- | youtube_dl/extractor/extractors.py | 5 | ||||
-rw-r--r-- | youtube_dl/extractor/nhk.py | 162 |
2 files changed, 118 insertions, 49 deletions
diff --git a/youtube_dl/extractor/extractors.py b/youtube_dl/extractor/extractors.py index 533c074b3..e664cc685 100644 --- a/youtube_dl/extractor/extractors.py +++ b/youtube_dl/extractor/extractors.py @@ -725,7 +725,10 @@ from .nexx import ( NexxEmbedIE, ) from .nfl import NFLIE -from .nhk import NhkVodIE +from .nhk import ( + NhkVodIE, + NhkVodProgramIE, +) from .nhl import NHLIE from .nick import ( NickIE, diff --git a/youtube_dl/extractor/nhk.py b/youtube_dl/extractor/nhk.py index 6a61a47d2..907db4de9 100644 --- a/youtube_dl/extractor/nhk.py +++ b/youtube_dl/extractor/nhk.py @@ -3,14 +3,96 @@ from __future__ import unicode_literals import re from .common import InfoExtractor +from ..utils import ExtractorError -class NhkVodIE(InfoExtractor): +class NhkBaseIE(InfoExtractor): + _API_URL_TEMPLATE = 'https://api.nhk.or.jp/nhkworld/%sod%slist/v7a/%s/%s/%s/all%s.json' + + def _get_clean_field(self, episode, key): + return episode.get(key + '_clean') or episode.get(key) + + def _list_episodes(self, m_id, lang, is_video, is_episode): + return self._download_json( + self._API_URL_TEMPLATE % ( + 'v' if is_video else 'r', + 'clip' if m_id[:4] == '9999' else 'esd', + 'episode' if is_episode else 'program', + m_id, lang, '/all' if is_video else ''), + m_id, query={'apikey': 'EJfK8jdS57GqlupFgAfAAwr573q01y6k'})['data']['episodes'] + + def _parse_episode_json(self, episode, lang, is_video): + title = episode.get('sub_title_clean') or episode['sub_title'] + + episode_id = None + if is_video: + pgm_id = episode.get('pgm_id') + pgm_no = episode.get('pgm_no') + + if not (pgm_id and pgm_no): + missing_field = 'pgm_id' if not pgm_id else 'pgm_no' + raise ExtractorError('Cannot download episode. Field %s is missing from episode JSON.' % missing_field) + + episode_id = pgm_id + pgm_no + else: + pgm_gr_id = episode.get('pgm_gr_id') + first_onair_date = episode.get('first_onair_date') + first_onair_no = episode.get('first_onair_no') + + if not (pgm_gr_id and first_onair_date and first_onair_no): + missing_field = 'pgm_gr_id' if not pgm_gr_id else 'first_onair_date' if not first_onair_date else 'first_onair_no' + raise ExtractorError('Cannot download episode. Field %s is missing from episode JSON.' % missing_field) + + episode_id = pgm_gr_id + '-' + first_onair_date + '-' + first_onair_no + + series = self._get_clean_field(episode, 'title') + + thumbnails = [] + for s, w, h in [('', 640, 360), ('_l', 1280, 720)]: + img_path = episode.get('image' + s) + if not img_path: + continue + thumbnails.append({ + 'id': '%dp' % h, + 'height': h, + 'width': w, + 'url': 'https://www3.nhk.or.jp' + img_path, + }) + + info = { + 'id': episode_id + '-' + lang, + 'title': '%s - %s' % (series, title) if series and title else title, + 'description': self._get_clean_field(episode, 'description'), + 'thumbnails': thumbnails, + 'series': series, + 'episode': title, + } + + if is_video: + info.update({ + '_type': 'url_transparent', + 'ie_key': 'Piksel', + 'url': 'https://player.piksel.com/v/refid/nhkworld/prefid/' + episode['vod_id'], + }) + else: + audio = episode['audio'] + audio_path = audio['audio'] + info['formats'] = self._extract_m3u8_formats( + 'https://nhkworld-vh.akamaihd.net/i%s/master.m3u8' % audio_path, + episode_id, 'm4a', entry_protocol='m3u8_native', + m3u8_id='hls', fatal=False) + for f in info['formats']: + f['language'] = lang + + return info + + +class NhkVodIE(NhkBaseIE): _VALID_URL = r'https?://www3\.nhk\.or\.jp/nhkworld/(?P<lang>[a-z]{2})/ondemand/(?P<type>video|audio)/(?P<id>\d{7}|[^/]+?-\d{8}-\d+)' # Content available only for a limited period of time. Visit # https://www3.nhk.or.jp/nhkworld/en/ondemand/ for working samples. _TESTS = [{ - # video clip + # clip 'url': 'https://www3.nhk.or.jp/nhkworld/en/ondemand/video/9999011/', 'md5': '256a1be14f48d960a7e61e2532d95ec3', 'info_dict': { @@ -47,60 +129,44 @@ class NhkVodIE(InfoExtractor): 'url': 'https://www3.nhk.or.jp/nhkworld/en/ondemand/audio/j_art-20150903-1/', 'only_matching': True, }] - _API_URL_TEMPLATE = 'https://api.nhk.or.jp/nhkworld/%sod%slist/v7a/episode/%s/%s/all%s.json' def _real_extract(self, url): lang, m_type, episode_id = re.match(self._VALID_URL, url).groups() + if episode_id.isdigit(): episode_id = episode_id[:4] + '-' + episode_id[4:] - is_video = m_type == 'video' - episode = self._download_json( - self._API_URL_TEMPLATE % ( - 'v' if is_video else 'r', - 'clip' if episode_id[:4] == '9999' else 'esd', - episode_id, lang, '/all' if is_video else ''), - episode_id, query={'apikey': 'EJfK8jdS57GqlupFgAfAAwr573q01y6k'})['data']['episodes'][0] - title = episode.get('sub_title_clean') or episode['sub_title'] + episode = self._list_episodes(episode_id, lang, m_type == 'video', True)[0] - def get_clean_field(key): - return episode.get(key + '_clean') or episode.get(key) + return self._parse_episode_json(episode, lang, m_type == 'video') - series = get_clean_field('title') - thumbnails = [] - for s, w, h in [('', 640, 360), ('_l', 1280, 720)]: - img_path = episode.get('image' + s) - if not img_path: - continue - thumbnails.append({ - 'id': '%dp' % h, - 'height': h, - 'width': w, - 'url': 'https://www3.nhk.or.jp' + img_path, - }) +class NhkVodProgramIE(NhkBaseIE): + _VALID_URL = r'https?://www3\.nhk\.or\.jp/nhkworld/(?P<lang>[a-z]{2})/ondemand/(program/video)/(?P<id>\w+)' + # Content available only for a limited period of time. Visit + # https://www3.nhk.or.jp/nhkworld/en/ondemand/ for working samples. + _TESTS = [{ + # video program + 'url': 'https://www3.nhk.or.jp/nhkworld/en/ondemand/program/video/japanrailway', + 'info_dict': { + 'id': 'japanrailway', + 'title': 'Japan Railway Journal', + }, + 'playlist_mincount': 1, + }, { + 'url': 'https://www3.nhk.or.jp/nhkworld/en/ondemand/program/video/10yearshayaomiyazaki/', + 'only_matching': True, + }] - info = { - 'id': episode_id + '-' + lang, - 'title': '%s - %s' % (series, title) if series and title else title, - 'description': get_clean_field('description'), - 'thumbnails': thumbnails, - 'series': series, - 'episode': title, - } - if is_video: - info.update({ - '_type': 'url_transparent', - 'ie_key': 'Piksel', - 'url': 'https://player.piksel.com/v/refid/nhkworld/prefid/' + episode['vod_id'], - }) + def _real_extract(self, url): + lang, m_type, program_id = re.match(self._VALID_URL, url).groups() + + episodes = self._list_episodes(program_id, lang, True, False) + + if episodes: + return self.playlist_result( + [self._parse_episode_json(episode, lang, True) + for episode in episodes], + self._get_clean_field(episodes[0], 'pgm_gr_id'), self._get_clean_field(episodes[0], 'title')) else: - audio = episode['audio'] - audio_path = audio['audio'] - info['formats'] = self._extract_m3u8_formats( - 'https://nhkworld-vh.akamaihd.net/i%s/master.m3u8' % audio_path, - episode_id, 'm4a', entry_protocol='m3u8_native', - m3u8_id='hls', fatal=False) - for f in info['formats']: - f['language'] = lang - return info + raise ExtractorError('No episodes returned for program with ID: %s' % program_id, expected=True) |