diff options
| -rw-r--r-- | youtube_dl/extractor/__init__.py | 5 | ||||
| -rw-r--r-- | youtube_dl/extractor/aftenposten.py | 23 | ||||
| -rw-r--r-- | youtube_dl/extractor/bleacherreport.py | 2 | ||||
| -rw-r--r-- | youtube_dl/extractor/franceinter.py | 7 | ||||
| -rw-r--r-- | youtube_dl/extractor/imgur.py | 51 | ||||
| -rw-r--r-- | youtube_dl/extractor/soompi.py | 146 | ||||
| -rw-r--r-- | youtube_dl/extractor/vgtv.py | 166 | ||||
| -rw-r--r-- | youtube_dl/extractor/viki.py | 24 | ||||
| -rw-r--r-- | youtube_dl/extractor/xstream.py | 14 | 
9 files changed, 185 insertions, 253 deletions
| diff --git a/youtube_dl/extractor/__init__.py b/youtube_dl/extractor/__init__.py index 760b65441..fede5ff0d 100644 --- a/youtube_dl/extractor/__init__.py +++ b/youtube_dl/extractor/__init__.py @@ -15,7 +15,6 @@ from .adobetv import (      AdobeTVVideoIE,  )  from .adultswim import AdultSwimIE -from .aftenposten import AftenpostenIE  from .aftonbladet import AftonbladetIE  from .airmozilla import AirMozillaIE  from .aljazeera import AlJazeeraIE @@ -591,10 +590,6 @@ from .snagfilms import (  )  from .snotr import SnotrIE  from .sohu import SohuIE -from .soompi import ( -    SoompiIE, -    SoompiShowIE, -)  from .soundcloud import (      SoundcloudIE,      SoundcloudSetIE, diff --git a/youtube_dl/extractor/aftenposten.py b/youtube_dl/extractor/aftenposten.py deleted file mode 100644 index 0c00acfb5..000000000 --- a/youtube_dl/extractor/aftenposten.py +++ /dev/null @@ -1,23 +0,0 @@ -# coding: utf-8 -from __future__ import unicode_literals - -from .common import InfoExtractor - - -class AftenpostenIE(InfoExtractor): -    _VALID_URL = r'https?://(?:www\.)?aftenposten\.no/webtv/(?:#!/)?video/(?P<id>\d+)' -    _TEST = { -        'url': 'http://www.aftenposten.no/webtv/#!/video/21039/trailer-sweatshop-i-can-t-take-any-more', -        'md5': 'fd828cd29774a729bf4d4425fe192972', -        'info_dict': { -            'id': '21039', -            'ext': 'mov', -            'title': 'TRAILER: "Sweatshop" - I can´t take any more', -            'description': 'md5:21891f2b0dd7ec2f78d84a50e54f8238', -            'timestamp': 1416927969, -            'upload_date': '20141125', -        } -    } - -    def _real_extract(self, url): -        return self.url_result('xstream:ap:%s' % self._match_id(url), 'Xstream') diff --git a/youtube_dl/extractor/bleacherreport.py b/youtube_dl/extractor/bleacherreport.py index bd2a6340b..38bda3af5 100644 --- a/youtube_dl/extractor/bleacherreport.py +++ b/youtube_dl/extractor/bleacherreport.py @@ -90,7 +90,7 @@ class BleacherReportCMSIE(AMPIE):      _VALID_URL = r'https?://(?:www\.)?bleacherreport\.com/video_embed\?id=(?P<id>[0-9a-f-]{36})'      _TESTS = [{          'url': 'http://bleacherreport.com/video_embed?id=8fd44c2f-3dc5-4821-9118-2c825a98c0e1', -        'md5': 'f0ca220af012d4df857b54f792c586bb', +        'md5': '8c2c12e3af7805152675446c905d159b',          'info_dict': {              'id': '8fd44c2f-3dc5-4821-9118-2c825a98c0e1',              'ext': 'flv', diff --git a/youtube_dl/extractor/franceinter.py b/youtube_dl/extractor/franceinter.py index 6613ee17a..fdc51f44f 100644 --- a/youtube_dl/extractor/franceinter.py +++ b/youtube_dl/extractor/franceinter.py @@ -1,8 +1,6 @@  # coding: utf-8  from __future__ import unicode_literals -import re -  from .common import InfoExtractor  from ..utils import int_or_none @@ -23,8 +21,7 @@ class FranceInterIE(InfoExtractor):      }      def _real_extract(self, url): -        mobj = re.match(self._VALID_URL, url) -        video_id = mobj.group('id') +        video_id = self._match_id(url)          webpage = self._download_webpage(url, video_id) @@ -33,7 +30,7 @@ class FranceInterIE(InfoExtractor):          video_url = 'http://www.franceinter.fr/' + path          title = self._html_search_regex( -            r'<span class="title">(.+?)</span>', webpage, 'title') +            r'<span class="title-diffusion">(.+?)</span>', webpage, 'title')          description = self._html_search_regex(              r'<span class="description">(.*?)</span>',              webpage, 'description', fatal=False) diff --git a/youtube_dl/extractor/imgur.py b/youtube_dl/extractor/imgur.py index 70c8ca64e..85e9344aa 100644 --- a/youtube_dl/extractor/imgur.py +++ b/youtube_dl/extractor/imgur.py @@ -13,7 +13,7 @@ from ..utils import (  class ImgurIE(InfoExtractor): -    _VALID_URL = r'https?://(?:i\.)?imgur\.com/(?!gallery)(?P<id>[a-zA-Z0-9]+)' +    _VALID_URL = r'https?://(?:i\.)?imgur\.com/(?:(?:gallery|topic/[^/]+)/)?(?P<id>[a-zA-Z0-9]{6,})(?:[/?#&]+|\.[a-z]+)?$'      _TESTS = [{          'url': 'https://i.imgur.com/A61SaA1.gifv', @@ -21,7 +21,7 @@ class ImgurIE(InfoExtractor):              'id': 'A61SaA1',              'ext': 'mp4',              'title': 're:Imgur GIF$|MRW gifv is up and running without any bugs$', -            'description': 're:The origin of the Internet\'s most viral images$|The Internet\'s visual storytelling community\. Explore, share, and discuss the best visual stories the Internet has to offer\.$', +            'description': 'Imgur: The most awesome images on the Internet.',          },      }, {          'url': 'https://imgur.com/A61SaA1', @@ -29,8 +29,20 @@ class ImgurIE(InfoExtractor):              'id': 'A61SaA1',              'ext': 'mp4',              'title': 're:Imgur GIF$|MRW gifv is up and running without any bugs$', -            'description': 're:The origin of the Internet\'s most viral images$|The Internet\'s visual storytelling community\. Explore, share, and discuss the best visual stories the Internet has to offer\.$', +            'description': 'Imgur: The most awesome images on the Internet.',          }, +    }, { +        'url': 'https://imgur.com/gallery/YcAQlkx', +        'info_dict': { +            'id': 'YcAQlkx', +            'ext': 'mp4', +            'title': 'Classic Steve Carell gif...cracks me up everytime....damn the repost downvotes....', +            'description': 'Imgur: The most awesome images on the Internet.' + +        } +    }, { +        'url': 'http://imgur.com/topic/Funny/N8rOudd', +        'only_matching': True,      }]      def _real_extract(self, url): @@ -100,25 +112,38 @@ class ImgurIE(InfoExtractor):  class ImgurAlbumIE(InfoExtractor): -    _VALID_URL = r'https?://(?:i\.)?imgur\.com/gallery/(?P<id>[a-zA-Z0-9]+)' +    _VALID_URL = r'https?://(?:i\.)?imgur\.com/(?:(?:a|gallery|topic/[^/]+)/)?(?P<id>[a-zA-Z0-9]{5})(?:[/?#&]+)?$' -    _TEST = { +    _TESTS = [{          'url': 'http://imgur.com/gallery/Q95ko',          'info_dict': {              'id': 'Q95ko',          },          'playlist_count': 25, -    } +    }, { +        'url': 'http://imgur.com/a/j6Orj', +        'only_matching': True, +    }, { +        'url': 'http://imgur.com/topic/Aww/ll5Vk', +        'only_matching': True, +    }]      def _real_extract(self, url):          album_id = self._match_id(url)          album_images = self._download_json(              'http://imgur.com/gallery/%s/album_images/hit.json?all=true' % album_id, -            album_id)['data']['images'] - -        entries = [ -            self.url_result('http://imgur.com/%s' % image['hash']) -            for image in album_images if image.get('hash')] - -        return self.playlist_result(entries, album_id) +            album_id, fatal=False) + +        if album_images: +            data = album_images.get('data') +            if data and isinstance(data, dict): +                images = data.get('images') +                if images and isinstance(images, list): +                    entries = [ +                        self.url_result('http://imgur.com/%s' % image['hash']) +                        for image in images if image.get('hash')] +                    return self.playlist_result(entries, album_id) + +        # Fallback to single video +        return self.url_result('http://imgur.com/%s' % album_id, ImgurIE.ie_key()) diff --git a/youtube_dl/extractor/soompi.py b/youtube_dl/extractor/soompi.py deleted file mode 100644 index 5da66ca9e..000000000 --- a/youtube_dl/extractor/soompi.py +++ /dev/null @@ -1,146 +0,0 @@ -# encoding: utf-8 -from __future__ import unicode_literals - -import re - -from .crunchyroll import CrunchyrollIE - -from .common import InfoExtractor -from ..compat import compat_HTTPError -from ..utils import ( -    ExtractorError, -    int_or_none, -    remove_start, -    xpath_text, -) - - -class SoompiBaseIE(InfoExtractor): -    def _get_episodes(self, webpage, episode_filter=None): -        episodes = self._parse_json( -            self._search_regex( -                r'VIDEOS\s*=\s*(\[.+?\]);', webpage, 'episodes JSON'), -            None) -        return list(filter(episode_filter, episodes)) - - -class SoompiIE(SoompiBaseIE, CrunchyrollIE): -    IE_NAME = 'soompi' -    _VALID_URL = r'https?://tv\.soompi\.com/(?:en/)?watch/(?P<id>[0-9]+)' -    _TESTS = [{ -        'url': 'http://tv.soompi.com/en/watch/29235', -        'info_dict': { -            'id': '29235', -            'ext': 'mp4', -            'title': 'Episode 1096', -            'description': '2015-05-20' -        }, -        'params': { -            'skip_download': True, -        }, -    }] - -    def _get_episode(self, webpage, video_id): -        return self._get_episodes(webpage, lambda x: x['id'] == video_id)[0] - -    def _get_subtitles(self, config, video_id): -        sub_langs = {} -        for subtitle in config.findall('./{default}preload/subtitles/subtitle'): -            sub_langs[subtitle.attrib['id']] = subtitle.attrib['title'] - -        subtitles = {} -        for s in config.findall('./{default}preload/subtitle'): -            lang_code = sub_langs.get(s.attrib['id']) -            if not lang_code: -                continue -            sub_id = s.get('id') -            data = xpath_text(s, './data', 'data') -            iv = xpath_text(s, './iv', 'iv') -            if not id or not iv or not data: -                continue -            subtitle = self._decrypt_subtitles(data, iv, sub_id).decode('utf-8') -            subtitles[lang_code] = self._extract_subtitles(subtitle) -        return subtitles - -    def _real_extract(self, url): -        video_id = self._match_id(url) - -        try: -            webpage = self._download_webpage( -                url, video_id, 'Downloading episode page') -        except ExtractorError as ee: -            if isinstance(ee.cause, compat_HTTPError) and ee.cause.code == 403: -                webpage = ee.cause.read() -                block_message = self._html_search_regex( -                    r'(?s)<div class="block-message">(.+?)</div>', webpage, -                    'block message', default=None) -                if block_message: -                    raise ExtractorError(block_message, expected=True) -            raise - -        formats = [] -        config = None -        for format_id in re.findall(r'\?quality=([0-9a-zA-Z]+)', webpage): -            config = self._download_xml( -                'http://tv.soompi.com/en/show/_/%s-config.xml?mode=hls&quality=%s' % (video_id, format_id), -                video_id, 'Downloading %s XML' % format_id) -            m3u8_url = xpath_text( -                config, './{default}preload/stream_info/file', -                '%s m3u8 URL' % format_id) -            if not m3u8_url: -                continue -            formats.extend(self._extract_m3u8_formats( -                m3u8_url, video_id, 'mp4', m3u8_id=format_id)) -        self._sort_formats(formats) - -        episode = self._get_episode(webpage, video_id) - -        title = episode['name'] -        description = episode.get('description') -        duration = int_or_none(episode.get('duration')) - -        thumbnails = [{ -            'id': thumbnail_id, -            'url': thumbnail_url, -        } for thumbnail_id, thumbnail_url in episode.get('img_url', {}).items()] - -        subtitles = self.extract_subtitles(config, video_id) - -        return { -            'id': video_id, -            'title': title, -            'description': description, -            'thumbnails': thumbnails, -            'duration': duration, -            'formats': formats, -            'subtitles': subtitles -        } - - -class SoompiShowIE(SoompiBaseIE): -    IE_NAME = 'soompi:show' -    _VALID_URL = r'https?://tv\.soompi\.com/en/shows/(?P<id>[0-9a-zA-Z\-_]+)' -    _TESTS = [{ -        'url': 'http://tv.soompi.com/en/shows/liar-game', -        'info_dict': { -            'id': 'liar-game', -            'title': 'Liar Game', -            'description': 'md5:52c02bce0c1a622a95823591d0589b66', -        }, -        'playlist_count': 14, -    }] - -    def _real_extract(self, url): -        show_id = self._match_id(url) - -        webpage = self._download_webpage( -            url, show_id, 'Downloading show page') - -        title = remove_start(self._og_search_title(webpage), 'SoompiTV | ') -        description = self._og_search_description(webpage) - -        entries = [ -            self.url_result('http://tv.soompi.com/en/watch/%s' % episode['id'], 'Soompi') -            for episode in self._get_episodes(webpage)] - -        return self.playlist_result(entries, show_id, title, description) diff --git a/youtube_dl/extractor/vgtv.py b/youtube_dl/extractor/vgtv.py index f38a72fde..811ee197d 100644 --- a/youtube_dl/extractor/vgtv.py +++ b/youtube_dl/extractor/vgtv.py @@ -4,26 +4,48 @@ from __future__ import unicode_literals  import re  from .common import InfoExtractor +from .xstream import XstreamIE  from ..utils import (      ExtractorError,      float_or_none,  ) -class VGTVIE(InfoExtractor): -    IE_DESC = 'VGTV and BTTV' +class VGTVIE(XstreamIE): +    IE_DESC = 'VGTV, BTTV, FTV, Aftenposten and Aftonbladet' + +    _HOST_TO_APPNAME = { +        'vgtv.no': 'vgtv', +        'bt.no/tv': 'bttv', +        'aftenbladet.no/tv': 'satv', +        'fvn.no/fvntv': 'fvntv', +        'aftenposten.no/webtv': 'aptv', +    } + +    _APP_NAME_TO_VENDOR = { +        'vgtv': 'vgtv', +        'bttv': 'bt', +        'satv': 'sa', +        'fvntv': 'fvn', +        'aptv': 'ap', +    } +      _VALID_URL = r'''(?x) -                    (?: -                        vgtv:| -                        http://(?:www\.)? +                    (?:https?://(?:www\.)? +                    (?P<host> +                        %s                      ) -                    (?P<host>vgtv|bt) +                    /                      (?: -                        :| -                        \.no/(?:tv/)?\#!/(?:video|live)/ -                    ) -                    (?P<id>[0-9]+) -                    ''' +                        \#!/(?:video|live)/| +                        embed?.*id= +                    )| +                    (?P<appname> +                        %s +                    ):) +                    (?P<id>\d+) +                    ''' % ('|'.join(_HOST_TO_APPNAME.keys()), '|'.join(_APP_NAME_TO_VENDOR.keys())) +      _TESTS = [          {              # streamType: vod @@ -59,17 +81,18 @@ class VGTVIE(InfoExtractor):                  # m3u8 download                  'skip_download': True,              }, +            'skip': 'Video is no longer available',          },          { -            # streamType: live +            # streamType: wasLive              'url': 'http://www.vgtv.no/#!/live/113063/direkte-v75-fra-solvalla',              'info_dict': {                  'id': '113063', -                'ext': 'flv', -                'title': 're:^DIREKTE: V75 fra Solvalla [0-9]{4}-[0-9]{2}-[0-9]{2} [0-9]{2}:[0-9]{2}$', +                'ext': 'mp4', +                'title': 'V75 fra Solvalla 30.05.15',                  'description': 'md5:b3743425765355855f88e096acc93231',                  'thumbnail': 're:^https?://.*\.jpg', -                'duration': 0, +                'duration': 25966,                  'timestamp': 1432975582,                  'upload_date': '20150530',                  'view_count': int, @@ -80,6 +103,20 @@ class VGTVIE(InfoExtractor):              },          },          { +            'url': 'http://www.aftenposten.no/webtv/#!/video/21039/trailer-sweatshop-i-can-t-take-any-more', +            'md5': 'fd828cd29774a729bf4d4425fe192972', +            'info_dict': { +                'id': '21039', +                'ext': 'mov', +                'title': 'TRAILER: «SWEATSHOP» - I can´t take any more', +                'description': 'md5:21891f2b0dd7ec2f78d84a50e54f8238', +                'duration': 66, +                'timestamp': 1417002452, +                'upload_date': '20141126', +                'view_count': int, +            } +        }, +        {              'url': 'http://www.bt.no/tv/#!/video/100250/norling-dette-er-forskjellen-paa-1-divisjon-og-eliteserien',              'only_matching': True,          }, @@ -89,21 +126,27 @@ class VGTVIE(InfoExtractor):          mobj = re.match(self._VALID_URL, url)          video_id = mobj.group('id')          host = mobj.group('host') - -        HOST_WEBSITES = { -            'vgtv': 'vgtv', -            'bt': 'bttv', -        } +        appname = self._HOST_TO_APPNAME[host] if host else mobj.group('appname') +        vendor = self._APP_NAME_TO_VENDOR[appname]          data = self._download_json(              'http://svp.vg.no/svp/api/v1/%s/assets/%s?appName=%s-website' -            % (host, video_id, HOST_WEBSITES[host]), +            % (vendor, video_id, appname),              video_id, 'Downloading media JSON')          if data.get('status') == 'inactive':              raise ExtractorError(                  'Video %s is no longer available' % video_id, expected=True) +        info = { +            'formats': [], +        } +        if len(video_id) == 5: +            if appname == 'bttv': +                info = self._extract_video_info('btno', video_id) +            elif appname == 'aptv': +                info = self._extract_video_info('ap', video_id) +          streams = data['streamUrls']          stream_type = data.get('streamType') @@ -111,48 +154,53 @@ class VGTVIE(InfoExtractor):          hls_url = streams.get('hls')          if hls_url: -            formats.extend(self._extract_m3u8_formats( -                hls_url, video_id, 'mp4', m3u8_id='hls')) +            m3u8_formats = self._extract_m3u8_formats( +                hls_url, video_id, 'mp4', m3u8_id='hls', fatal=False) +            if m3u8_formats: +                formats.extend(m3u8_formats)          hds_url = streams.get('hds')          # wasLive hds are always 404          if hds_url and stream_type != 'wasLive': -            formats.extend(self._extract_f4m_formats( -                hds_url + '?hdcore=3.2.0&plugin=aasp-3.2.0.77.18', -                video_id, f4m_id='hds')) +            f4m_formats = self._extract_f4m_formats( +                hds_url + '?hdcore=3.2.0&plugin=aasp-3.2.0.77.18', video_id, f4m_id='hds', fatal=False) +            if f4m_formats: +                formats.extend(f4m_formats) +        mp4_urls = streams.get('pseudostreaming') or []          mp4_url = streams.get('mp4')          if mp4_url: -            _url = hls_url or hds_url -            MP4_URL_TEMPLATE = '%s/%%s.%s' % (mp4_url.rpartition('/')[0], mp4_url.rpartition('.')[-1]) -            for mp4_format in _url.split(','): -                m = re.search('(?P<width>\d+)_(?P<height>\d+)_(?P<vbr>\d+)', mp4_format) -                if not m: -                    continue -                width = int(m.group('width')) -                height = int(m.group('height')) -                vbr = int(m.group('vbr')) -                formats.append({ -                    'url': MP4_URL_TEMPLATE % mp4_format, -                    'format_id': 'mp4-%s' % vbr, -                    'width': width, -                    'height': height, -                    'vbr': vbr, -                    'preference': 1, +            mp4_urls.append(mp4_url) +        for mp4_url in mp4_urls: +            format_info = { +                'url': mp4_url, +            } +            mobj = re.search('(\d+)_(\d+)_(\d+)', mp4_url) +            if mobj: +                tbr = int(mobj.group(3)) +                format_info.update({ +                    'width': int(mobj.group(1)), +                    'height': int(mobj.group(2)), +                    'tbr': tbr, +                    'format_id': 'mp4-%s' % tbr,                  }) -        self._sort_formats(formats) +            formats.append(format_info) + +        info['formats'].extend(formats) + +        self._sort_formats(info['formats']) -        return { +        info.update({              'id': video_id, -            'title': self._live_title(data['title']), +            'title': self._live_title(data['title']) if stream_type == 'live' else data['title'],              'description': data['description'],              'thumbnail': data['images']['main'] + '?t[]=900x506q80',              'timestamp': data['published'],              'duration': float_or_none(data['duration'], 1000),              'view_count': data['displays'], -            'formats': formats,              'is_live': True if stream_type == 'live' else False, -        } +        }) +        return info  class BTArticleIE(InfoExtractor): @@ -161,7 +209,7 @@ class BTArticleIE(InfoExtractor):      _VALID_URL = 'http://(?:www\.)?bt\.no/(?:[^/]+/)+(?P<id>[^/]+)-\d+\.html'      _TEST = {          'url': 'http://www.bt.no/nyheter/lokalt/Kjemper-for-internatet-1788214.html', -        'md5': 'd055e8ee918ef2844745fcfd1a4175fb', +        'md5': '2acbe8ad129b3469d5ae51b1158878df',          'info_dict': {              'id': '23199',              'ext': 'mp4', @@ -178,15 +226,15 @@ class BTArticleIE(InfoExtractor):      def _real_extract(self, url):          webpage = self._download_webpage(url, self._match_id(url))          video_id = self._search_regex( -            r'SVP\.Player\.load\(\s*(\d+)', webpage, 'video id') -        return self.url_result('vgtv:bt:%s' % video_id, 'VGTV') +            r'<video[^>]+data-id="(\d+)"', webpage, 'video id') +        return self.url_result('bttv:%s' % video_id, 'VGTV')  class BTVestlendingenIE(InfoExtractor):      IE_NAME = 'bt:vestlendingen'      IE_DESC = 'Bergens Tidende - Vestlendingen'      _VALID_URL = 'http://(?:www\.)?bt\.no/spesial/vestlendingen/#!/(?P<id>\d+)' -    _TEST = { +    _TESTS = [{          'url': 'http://www.bt.no/spesial/vestlendingen/#!/86588',          'md5': 'd7d17e3337dc80de6d3a540aefbe441b',          'info_dict': { @@ -197,7 +245,19 @@ class BTVestlendingenIE(InfoExtractor):              'timestamp': 1430473209,              'upload_date': '20150501',          }, -    } +        'skip': '404 Error', +    }, { +        'url': 'http://www.bt.no/spesial/vestlendingen/#!/86255', +        'md5': 'a2893f8632e96389f4bdf36aa9463ceb', +        'info_dict': { +            'id': '86255', +            'ext': 'mov', +            'title': 'Du må tåle å fryse og være sulten', +            'description': 'md5:b8046f4d022d5830ddab04865791d063', +            'upload_date': '20150321', +            'timestamp': 1426942023, +        }, +    }]      def _real_extract(self, url): -        return self.url_result('xstream:btno:%s' % self._match_id(url), 'Xstream') +        return self.url_result('bttv:%s' % self._match_id(url), 'VGTV') diff --git a/youtube_dl/extractor/viki.py b/youtube_dl/extractor/viki.py index a63c23617..ca3f20a3d 100644 --- a/youtube_dl/extractor/viki.py +++ b/youtube_dl/extractor/viki.py @@ -30,6 +30,12 @@ class VikiBaseIE(InfoExtractor):      _token = None +    _ERRORS = { +        'geo': 'Sorry, this content is not available in your region.', +        'upcoming': 'Sorry, this content is not yet available.', +        # 'paywall': 'paywall', +    } +      def _prepare_call(self, path, timestamp=None, post_data=None):          path += '?' if '?' not in path else '&'          if not timestamp: @@ -67,6 +73,12 @@ class VikiBaseIE(InfoExtractor):              '%s returned error: %s' % (self.IE_NAME, error),              expected=True) +    def _check_errors(self, data): +        for reason, status in data.get('blocking', {}).items(): +            if status and reason in self._ERRORS: +                raise ExtractorError('%s said: %s' % ( +                    self.IE_NAME, self._ERRORS[reason]), expected=True) +      def _real_initialize(self):          self._login() @@ -193,6 +205,7 @@ class VikiIE(VikiBaseIE):              'timestamp': 1321985454,              'description': 'md5:44b1e46619df3a072294645c770cef36',              'title': 'Love In Magic', +            'age_limit': 13,          },      }] @@ -202,6 +215,8 @@ class VikiIE(VikiBaseIE):          video = self._call_api(              'videos/%s.json' % video_id, video_id, 'Downloading video JSON') +        self._check_errors(video) +          title = self.dict_selection(video.get('titles', {}), 'en')          if not title:              title = 'Episode %d' % video.get('number') if video.get('type') == 'episode' else video.get('id') or video_id @@ -262,8 +277,11 @@ class VikiIE(VikiBaseIE):                  r'^(\d+)[pP]$', format_id, 'height', default=None))              for protocol, format_dict in stream_dict.items():                  if format_id == 'm3u8': -                    formats = self._extract_m3u8_formats( -                        format_dict['url'], video_id, 'mp4', m3u8_id='m3u8-%s' % protocol) +                    m3u8_formats = self._extract_m3u8_formats( +                        format_dict['url'], video_id, 'mp4', 'm3u8_native', +                        m3u8_id='m3u8-%s' % protocol, fatal=None) +                    if m3u8_formats: +                        formats.extend(m3u8_formats)                  else:                      formats.append({                          'url': format_dict['url'], @@ -315,6 +333,8 @@ class VikiChannelIE(VikiBaseIE):              'containers/%s.json' % channel_id, channel_id,              'Downloading channel JSON') +        self._check_errors(channel) +          title = self.dict_selection(channel['titles'], 'en')          description = self.dict_selection(channel['descriptions'], 'en') diff --git a/youtube_dl/extractor/xstream.py b/youtube_dl/extractor/xstream.py index 71584c291..76c91bd92 100644 --- a/youtube_dl/extractor/xstream.py +++ b/youtube_dl/extractor/xstream.py @@ -42,11 +42,7 @@ class XstreamIE(InfoExtractor):          'only_matching': True,      }] -    def _real_extract(self, url): -        mobj = re.match(self._VALID_URL, url) -        partner_id = mobj.group('partner_id') -        video_id = mobj.group('id') - +    def _extract_video_info(self, partner_id, video_id):          data = self._download_xml(              'http://frontend.xstream.dk/%s/feed/video/?platform=web&id=%s'              % (partner_id, video_id), @@ -97,6 +93,7 @@ class XstreamIE(InfoExtractor):              formats.append({                  'url': link.get('href'),                  'format_id': link.get('rel'), +                'preference': 1,              })          thumbnails = [{ @@ -113,3 +110,10 @@ class XstreamIE(InfoExtractor):              'formats': formats,              'thumbnails': thumbnails,          } + +    def _real_extract(self, url): +        mobj = re.match(self._VALID_URL, url) +        partner_id = mobj.group('partner_id') +        video_id = mobj.group('id') + +        return self._extract_video_info(partner_id, video_id) | 
