diff options
Diffstat (limited to 'youtube_dl/downloader/dash.py')
| -rw-r--r-- | youtube_dl/downloader/dash.py | 90 |
1 files changed, 46 insertions, 44 deletions
diff --git a/youtube_dl/downloader/dash.py b/youtube_dl/downloader/dash.py index 8bbab9dbc..f3c058879 100644 --- a/youtube_dl/downloader/dash.py +++ b/youtube_dl/downloader/dash.py @@ -1,13 +1,12 @@ from __future__ import unicode_literals -import os -import re +import itertools from .fragment import FragmentFD from ..compat import compat_urllib_error from ..utils import ( - sanitize_open, - encodeFilename, + DownloadError, + urljoin, ) @@ -19,63 +18,66 @@ class DashSegmentsFD(FragmentFD): FD_NAME = 'dashsegments' def real_download(self, filename, info_dict): - base_url = info_dict['url'] - segment_urls = [info_dict['segment_urls'][0]] if self.params.get('test', False) else info_dict['segment_urls'] - initialization_url = info_dict.get('initialization_url') + fragment_base_url = info_dict.get('fragment_base_url') + fragments = info_dict['fragments'][:1] if self.params.get( + 'test', False) else info_dict['fragments'] ctx = { 'filename': filename, - 'total_frags': len(segment_urls) + (1 if initialization_url else 0), + 'total_frags': len(fragments), } self._prepare_and_start_frag_download(ctx) - def combine_url(base_url, target_url): - if re.match(r'^https?://', target_url): - return target_url - return '%s%s%s' % (base_url, '' if base_url.endswith('/') else '/', target_url) - - segments_filenames = [] - fragment_retries = self.params.get('fragment_retries', 0) - - def append_url_to_file(target_url, tmp_filename, segment_name): - target_filename = '%s-%s' % (tmp_filename, segment_name) - count = 0 - while count <= fragment_retries: + skip_unavailable_fragments = self.params.get('skip_unavailable_fragments', True) + + for frag_index, fragment in enumerate(fragments, 1): + if frag_index <= ctx['fragment_index']: + continue + success = False + # In DASH, the first segment contains necessary headers to + # generate a valid MP4 file, so always abort for the first segment + fatal = frag_index == 1 or not skip_unavailable_fragments + fragment_url = fragment.get('url') + if not fragment_url: + assert fragment_base_url + fragment_url = urljoin(fragment_base_url, fragment['path']) + headers = info_dict.get('http_headers') + fragment_range = fragment.get('range') + if fragment_range: + headers = headers.copy() if headers else {} + headers['Range'] = 'bytes=%s' % (fragment_range,) + for count in itertools.count(): try: - success = ctx['dl'].download(target_filename, {'url': combine_url(base_url, target_url)}) + success, frag_content = self._download_fragment(ctx, fragment_url, info_dict, headers) if not success: return False - down, target_sanitized = sanitize_open(target_filename, 'rb') - ctx['dest_stream'].write(down.read()) - down.close() - segments_filenames.append(target_sanitized) - break - except (compat_urllib_error.HTTPError, ) as err: + self._append_fragment(ctx, frag_content) + except compat_urllib_error.HTTPError as err: # YouTube may often return 404 HTTP error for a fragment causing the # whole download to fail. However if the same fragment is immediately - # retried with the same request data this usually succeeds (1-2 attemps + # retried with the same request data this usually succeeds (1-2 attempts # is usually enough) thus allowing to download the whole file successfully. - # So, we will retry all fragments that fail with 404 HTTP error for now. - if err.code != 404: + # To be future-proof we will retry all fragments that fail with any + # HTTP error. + if count < fragment_retries: + self.report_retry_fragment(err, frag_index, count + 1, fragment_retries) + continue + except DownloadError: + # Don't retry fragment if error occurred during HTTP downloading + # itself since it has its own retry settings + if fatal: raise - # Retry fragment - count += 1 - if count <= fragment_retries: - self.report_retry_fragment(segment_name, count, fragment_retries) - if count > fragment_retries: - self.report_error('giving up after %s fragment retries' % fragment_retries) - return False + break - if initialization_url: - append_url_to_file(initialization_url, ctx['tmpfilename'], 'Init') - for i, segment_url in enumerate(segment_urls): - append_url_to_file(segment_url, ctx['tmpfilename'], 'Seg%d' % i) + if not success: + if not fatal: + self.report_skip_fragment(frag_index) + continue + self.report_error('giving up after %s fragment retries' % count) + return False self._finish_frag_download(ctx) - for segment_file in segments_filenames: - os.remove(encodeFilename(segment_file)) - return True |
