aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorMatthew Rayermann <matthew.rayermann@gmail.com>2020-12-13 12:57:19 -0800
committerGitHub <noreply@github.com>2020-12-13 20:57:19 +0000
commit2797c7be45665c60a35bc2dd390cdd897ed5fc87 (patch)
tree6fd3c5494c572a84eb9dd5751c6c2ec01d948e73
parent755f186e213f0976bf9a3ceb4eb99e29ad1253c8 (diff)
downloadyoutube-dl-2797c7be45665c60a35bc2dd390cdd897ed5fc87.tar.xz
[nhk] Add support for NHK video programs (#27230)
-rw-r--r--youtube_dl/extractor/extractors.py5
-rw-r--r--youtube_dl/extractor/nhk.py162
2 files changed, 118 insertions, 49 deletions
diff --git a/youtube_dl/extractor/extractors.py b/youtube_dl/extractor/extractors.py
index 533c074b3..e664cc685 100644
--- a/youtube_dl/extractor/extractors.py
+++ b/youtube_dl/extractor/extractors.py
@@ -725,7 +725,10 @@ from .nexx import (
NexxEmbedIE,
)
from .nfl import NFLIE
-from .nhk import NhkVodIE
+from .nhk import (
+ NhkVodIE,
+ NhkVodProgramIE,
+)
from .nhl import NHLIE
from .nick import (
NickIE,
diff --git a/youtube_dl/extractor/nhk.py b/youtube_dl/extractor/nhk.py
index 6a61a47d2..907db4de9 100644
--- a/youtube_dl/extractor/nhk.py
+++ b/youtube_dl/extractor/nhk.py
@@ -3,14 +3,96 @@ from __future__ import unicode_literals
import re
from .common import InfoExtractor
+from ..utils import ExtractorError
-class NhkVodIE(InfoExtractor):
+class NhkBaseIE(InfoExtractor):
+ _API_URL_TEMPLATE = 'https://api.nhk.or.jp/nhkworld/%sod%slist/v7a/%s/%s/%s/all%s.json'
+
+ def _get_clean_field(self, episode, key):
+ return episode.get(key + '_clean') or episode.get(key)
+
+ def _list_episodes(self, m_id, lang, is_video, is_episode):
+ return self._download_json(
+ self._API_URL_TEMPLATE % (
+ 'v' if is_video else 'r',
+ 'clip' if m_id[:4] == '9999' else 'esd',
+ 'episode' if is_episode else 'program',
+ m_id, lang, '/all' if is_video else ''),
+ m_id, query={'apikey': 'EJfK8jdS57GqlupFgAfAAwr573q01y6k'})['data']['episodes']
+
+ def _parse_episode_json(self, episode, lang, is_video):
+ title = episode.get('sub_title_clean') or episode['sub_title']
+
+ episode_id = None
+ if is_video:
+ pgm_id = episode.get('pgm_id')
+ pgm_no = episode.get('pgm_no')
+
+ if not (pgm_id and pgm_no):
+ missing_field = 'pgm_id' if not pgm_id else 'pgm_no'
+ raise ExtractorError('Cannot download episode. Field %s is missing from episode JSON.' % missing_field)
+
+ episode_id = pgm_id + pgm_no
+ else:
+ pgm_gr_id = episode.get('pgm_gr_id')
+ first_onair_date = episode.get('first_onair_date')
+ first_onair_no = episode.get('first_onair_no')
+
+ if not (pgm_gr_id and first_onair_date and first_onair_no):
+ missing_field = 'pgm_gr_id' if not pgm_gr_id else 'first_onair_date' if not first_onair_date else 'first_onair_no'
+ raise ExtractorError('Cannot download episode. Field %s is missing from episode JSON.' % missing_field)
+
+ episode_id = pgm_gr_id + '-' + first_onair_date + '-' + first_onair_no
+
+ series = self._get_clean_field(episode, 'title')
+
+ thumbnails = []
+ for s, w, h in [('', 640, 360), ('_l', 1280, 720)]:
+ img_path = episode.get('image' + s)
+ if not img_path:
+ continue
+ thumbnails.append({
+ 'id': '%dp' % h,
+ 'height': h,
+ 'width': w,
+ 'url': 'https://www3.nhk.or.jp' + img_path,
+ })
+
+ info = {
+ 'id': episode_id + '-' + lang,
+ 'title': '%s - %s' % (series, title) if series and title else title,
+ 'description': self._get_clean_field(episode, 'description'),
+ 'thumbnails': thumbnails,
+ 'series': series,
+ 'episode': title,
+ }
+
+ if is_video:
+ info.update({
+ '_type': 'url_transparent',
+ 'ie_key': 'Piksel',
+ 'url': 'https://player.piksel.com/v/refid/nhkworld/prefid/' + episode['vod_id'],
+ })
+ else:
+ audio = episode['audio']
+ audio_path = audio['audio']
+ info['formats'] = self._extract_m3u8_formats(
+ 'https://nhkworld-vh.akamaihd.net/i%s/master.m3u8' % audio_path,
+ episode_id, 'm4a', entry_protocol='m3u8_native',
+ m3u8_id='hls', fatal=False)
+ for f in info['formats']:
+ f['language'] = lang
+
+ return info
+
+
+class NhkVodIE(NhkBaseIE):
_VALID_URL = r'https?://www3\.nhk\.or\.jp/nhkworld/(?P<lang>[a-z]{2})/ondemand/(?P<type>video|audio)/(?P<id>\d{7}|[^/]+?-\d{8}-\d+)'
# Content available only for a limited period of time. Visit
# https://www3.nhk.or.jp/nhkworld/en/ondemand/ for working samples.
_TESTS = [{
- # video clip
+ # clip
'url': 'https://www3.nhk.or.jp/nhkworld/en/ondemand/video/9999011/',
'md5': '256a1be14f48d960a7e61e2532d95ec3',
'info_dict': {
@@ -47,60 +129,44 @@ class NhkVodIE(InfoExtractor):
'url': 'https://www3.nhk.or.jp/nhkworld/en/ondemand/audio/j_art-20150903-1/',
'only_matching': True,
}]
- _API_URL_TEMPLATE = 'https://api.nhk.or.jp/nhkworld/%sod%slist/v7a/episode/%s/%s/all%s.json'
def _real_extract(self, url):
lang, m_type, episode_id = re.match(self._VALID_URL, url).groups()
+
if episode_id.isdigit():
episode_id = episode_id[:4] + '-' + episode_id[4:]
- is_video = m_type == 'video'
- episode = self._download_json(
- self._API_URL_TEMPLATE % (
- 'v' if is_video else 'r',
- 'clip' if episode_id[:4] == '9999' else 'esd',
- episode_id, lang, '/all' if is_video else ''),
- episode_id, query={'apikey': 'EJfK8jdS57GqlupFgAfAAwr573q01y6k'})['data']['episodes'][0]
- title = episode.get('sub_title_clean') or episode['sub_title']
+ episode = self._list_episodes(episode_id, lang, m_type == 'video', True)[0]
- def get_clean_field(key):
- return episode.get(key + '_clean') or episode.get(key)
+ return self._parse_episode_json(episode, lang, m_type == 'video')
- series = get_clean_field('title')
- thumbnails = []
- for s, w, h in [('', 640, 360), ('_l', 1280, 720)]:
- img_path = episode.get('image' + s)
- if not img_path:
- continue
- thumbnails.append({
- 'id': '%dp' % h,
- 'height': h,
- 'width': w,
- 'url': 'https://www3.nhk.or.jp' + img_path,
- })
+class NhkVodProgramIE(NhkBaseIE):
+ _VALID_URL = r'https?://www3\.nhk\.or\.jp/nhkworld/(?P<lang>[a-z]{2})/ondemand/(program/video)/(?P<id>\w+)'
+ # Content available only for a limited period of time. Visit
+ # https://www3.nhk.or.jp/nhkworld/en/ondemand/ for working samples.
+ _TESTS = [{
+ # video program
+ 'url': 'https://www3.nhk.or.jp/nhkworld/en/ondemand/program/video/japanrailway',
+ 'info_dict': {
+ 'id': 'japanrailway',
+ 'title': 'Japan Railway Journal',
+ },
+ 'playlist_mincount': 1,
+ }, {
+ 'url': 'https://www3.nhk.or.jp/nhkworld/en/ondemand/program/video/10yearshayaomiyazaki/',
+ 'only_matching': True,
+ }]
- info = {
- 'id': episode_id + '-' + lang,
- 'title': '%s - %s' % (series, title) if series and title else title,
- 'description': get_clean_field('description'),
- 'thumbnails': thumbnails,
- 'series': series,
- 'episode': title,
- }
- if is_video:
- info.update({
- '_type': 'url_transparent',
- 'ie_key': 'Piksel',
- 'url': 'https://player.piksel.com/v/refid/nhkworld/prefid/' + episode['vod_id'],
- })
+ def _real_extract(self, url):
+ lang, m_type, program_id = re.match(self._VALID_URL, url).groups()
+
+ episodes = self._list_episodes(program_id, lang, True, False)
+
+ if episodes:
+ return self.playlist_result(
+ [self._parse_episode_json(episode, lang, True)
+ for episode in episodes],
+ self._get_clean_field(episodes[0], 'pgm_gr_id'), self._get_clean_field(episodes[0], 'title'))
else:
- audio = episode['audio']
- audio_path = audio['audio']
- info['formats'] = self._extract_m3u8_formats(
- 'https://nhkworld-vh.akamaihd.net/i%s/master.m3u8' % audio_path,
- episode_id, 'm4a', entry_protocol='m3u8_native',
- m3u8_id='hls', fatal=False)
- for f in info['formats']:
- f['language'] = lang
- return info
+ raise ExtractorError('No episodes returned for program with ID: %s' % program_id, expected=True)