diff options
| author | Remita Amine <remitamine@gmail.com> | 2018-07-28 06:55:18 +0100 | 
|---|---|---|
| committer | Remita Amine <remitamine@gmail.com> | 2018-07-28 06:55:42 +0100 | 
| commit | a098c99f0d0deb4ad0d5c9b67496582d89970368 (patch) | |
| tree | edcccfe1767e7c929b6e70e34e139f2c82f8a8c7 | |
| parent | 8e37a7e4cce7555c12b8d9f7a1d331476aba357c (diff) | |
[telecinco] fix extraction(closes #17080)
| -rw-r--r-- | youtube_dl/extractor/telecinco.py | 122 | 
1 files changed, 106 insertions, 16 deletions
| diff --git a/youtube_dl/extractor/telecinco.py b/youtube_dl/extractor/telecinco.py index fdcc7d573..d37e1b055 100644 --- a/youtube_dl/extractor/telecinco.py +++ b/youtube_dl/extractor/telecinco.py @@ -1,26 +1,43 @@  # coding: utf-8  from __future__ import unicode_literals -from .mitele import MiTeleBaseIE +import json +import re +from .common import InfoExtractor +from .ooyala import OoyalaIE +from ..utils import ( +    clean_html, +    determine_ext, +    int_or_none, +    str_or_none, +    urljoin, +) -class TelecincoIE(MiTeleBaseIE): + +class TelecincoIE(InfoExtractor):      IE_DESC = 'telecinco.es, cuatro.com and mediaset.es'      _VALID_URL = r'https?://(?:www\.)?(?:telecinco\.es|cuatro\.com|mediaset\.es)/(?:[^/]+/)+(?P<id>.+?)\.html'      _TESTS = [{          'url': 'http://www.telecinco.es/robinfood/temporada-01/t01xp14/Bacalao-cocochas-pil-pil_0_1876350223.html', -        'md5': '8d7b2d5f699ee2709d992a63d5cd1712',          'info_dict': { -            'id': 'JEA5ijCnF6p5W08A1rNKn7', -            'ext': 'mp4', +            'id': '1876350223',              'title': 'Bacalao con kokotxas al pil-pil',              'description': 'md5:1382dacd32dd4592d478cbdca458e5bb', -            'duration': 662,          }, +        'playlist': [{ +            'md5': 'adb28c37238b675dad0f042292f209a7', +            'info_dict': { +                'id': 'JEA5ijCnF6p5W08A1rNKn7', +                'ext': 'mp4', +                'title': 'Con Martín Berasategui, hacer un bacalao al pil-pil es fácil y divertido', +                'duration': 662, +            }, +        }]      }, {          'url': 'http://www.cuatro.com/deportes/futbol/barcelona/Leo_Messi-Champions-Roma_2_2052780128.html', -        'md5': '284393e5387b3b947b77c613ef04749a', +        'md5': '9468140ebc300fbb8b9d65dc6e5c4b43',          'info_dict': {              'id': 'jn24Od1zGLG4XUZcnUnZB6',              'ext': 'mp4', @@ -30,7 +47,7 @@ class TelecincoIE(MiTeleBaseIE):          },      }, {          'url': 'http://www.mediaset.es/12meses/campanas/doylacara/conlatratanohaytrato/Ayudame-dar-cara-trata-trato_2_1986630220.html', -        'md5': '749afab6ea5a136a8806855166ae46a2', +        'md5': 'ae2dc6b7b50b2392076a51c0f70e01f6',          'info_dict': {              'id': 'aywerkD2Sv1vGNqq9b85Q2',              'ext': 'mp4', @@ -50,17 +67,90 @@ class TelecincoIE(MiTeleBaseIE):          'only_matching': True,      }] +    def _parse_content(self, content, url): +        video_id = content['dataMediaId'] +        if content.get('dataCmsId') == 'ooyala': +            return self.url_result( +                'ooyala:%s' % video_id, OoyalaIE.ie_key(), video_id) +        config_url = urljoin(url, content['dataConfig']) +        config = self._download_json( +            config_url, video_id, 'Downloading config JSON') +        title = config['info']['title'] + +        def mmc_url(mmc_type): +            return re.sub( +                r'/(?:flash|html5)\.json', '/%s.json' % mmc_type, +                config['services']['mmc']) + +        duration = None +        formats = [] +        for mmc_type in ('flash', 'html5'): +            mmc = self._download_json( +                mmc_url(mmc_type), video_id, +                'Downloading %s mmc JSON' % mmc_type, fatal=False) +            if not mmc: +                continue +            if not duration: +                duration = int_or_none(mmc.get('duration')) +            for location in mmc['locations']: +                gat = self._proto_relative_url(location.get('gat'), 'http:') +                gcp = location.get('gcp') +                ogn = location.get('ogn') +                if None in (gat, gcp, ogn): +                    continue +                token_data = { +                    'gcp': gcp, +                    'ogn': ogn, +                    'sta': 0, +                } +                media = self._download_json( +                    gat, video_id, data=json.dumps(token_data).encode('utf-8'), +                    headers={ +                        'Content-Type': 'application/json;charset=utf-8', +                        'Referer': url, +                    }, fatal=False) or {} +                stream = media.get('stream') or media.get('file') +                if not stream: +                    continue +                ext = determine_ext(stream) +                if ext == 'f4m': +                    formats.extend(self._extract_f4m_formats( +                        stream + '&hdcore=3.2.0&plugin=aasp-3.2.0.77.18', +                        video_id, f4m_id='hds', fatal=False)) +                elif ext == 'm3u8': +                    formats.extend(self._extract_m3u8_formats( +                        stream, video_id, 'mp4', 'm3u8_native', +                        m3u8_id='hls', fatal=False)) +        self._sort_formats(formats) + +        return { +            'id': video_id, +            'title': title, +            'formats': formats, +            'thumbnail': content.get('dataPoster') or config.get('poster', {}).get('imageUrl'), +            'duration': duration, +        } +      def _real_extract(self, url):          display_id = self._match_id(url)          webpage = self._download_webpage(url, display_id) -        title = self._html_search_meta( -            ['og:title', 'twitter:title'], webpage, 'title') -        info = self._get_player_info(url, webpage) +        article = self._parse_json(self._search_regex( +            r'window\.\$REACTBASE_STATE\.article\s*=\s*({.+})', +            webpage, 'article'), display_id)['article'] +        title = article.get('title') +        description = clean_html(article.get('leadParagraph')) +        if article.get('editorialType') != 'VID': +            entries = [] +            for p in article.get('body', []): +                content = p.get('content') +                if p.get('type') != 'video' or not content: +                    continue +                entries.append(self._parse_content(content, url)) +            return self.playlist_result( +                entries, str_or_none(article.get('id')), title, description) +        content = article['opening']['content'] +        info = self._parse_content(content, url)          info.update({ -            'display_id': display_id, -            'title': title, -            'description': self._html_search_meta( -                ['og:description', 'twitter:description'], -                webpage, 'title', fatal=False), +            'description': description,          })          return info | 
