X-Git-Url: http://git.bitcoin.ninja/index.cgi?a=blobdiff_plain;f=youtube_dl%2Fextractor%2Fmixcloud.py;h=84f29155841007f3088a86470040407073726067;hb=d5c418f29fcce3d62bba7f6228b76b69b5b731ce;hp=125d81551c26ea67eff82f2d2189bd058d16b873;hpb=b2ae513586cdf66b36ef4acc665c5eae17ee54a2;p=youtube-dl diff --git a/youtube_dl/extractor/mixcloud.py b/youtube_dl/extractor/mixcloud.py index 125d81551..84f291558 100644 --- a/youtube_dl/extractor/mixcloud.py +++ b/youtube_dl/extractor/mixcloud.py @@ -1,76 +1,149 @@ -import json +from __future__ import unicode_literals + import re +import itertools from .common import InfoExtractor +from ..compat import ( + compat_urllib_parse, +) from ..utils import ( - unified_strdate, ExtractorError, + HEADRequest, + str_to_int, ) class MixcloudIE(InfoExtractor): - _VALID_URL = r'^(?:https?://)?(?:www\.)?mixcloud\.com/([\w\d-]+)/([\w\d-]+)' - IE_NAME = u'mixcloud' - - _TEST = { - u'url': u'http://www.mixcloud.com/dholbach/cryptkeeper/', - u'file': u'dholbach-cryptkeeper.mp3', - u'info_dict': { - u'title': u'Cryptkeeper', - u'description': u'After quite a long silence from myself, finally another Drum\'n\'Bass mix with my favourite current dance floor bangers.', - u'uploader': u'Daniel Holbach', - u'uploader_id': u'dholbach', - u'upload_date': u'20111115', + _VALID_URL = r'^(?:https?://)?(?:www\.)?mixcloud\.com/([^/]+)/([^/]+)' + IE_NAME = 'mixcloud' + + _TESTS = [{ + 'url': 'http://www.mixcloud.com/dholbach/cryptkeeper/', + 'info_dict': { + 'id': 'dholbach-cryptkeeper', + 'ext': 'mp3', + 'title': 'Cryptkeeper', + 'description': 'After quite a long silence from myself, finally another Drum\'n\'Bass mix with my favourite current dance floor bangers.', + 'uploader': 'Daniel Holbach', + 'uploader_id': 'dholbach', + 'thumbnail': 're:https?://.*\.jpg', + 'view_count': int, + 'like_count': int, + }, + }, { + 'url': 'http://www.mixcloud.com/gillespeterson/caribou-7-inch-vinyl-mix-chat/', + 'info_dict': { + 'id': 'gillespeterson-caribou-7-inch-vinyl-mix-chat', + 'ext': 'mp3', + 'title': 'Caribou 7 inch Vinyl Mix & Chat', + 'description': 'md5:2b8aec6adce69f9d41724647c65875e8', + 'uploader': 'Gilles Peterson Worldwide', + 'uploader_id': 'gillespeterson', + 'thumbnail': 're:https?://.*/images/', + 'view_count': int, + 'like_count': int, }, - } + }] - def check_urls(self, url_list): - """Returns 1st active url from list""" - for url in url_list: + def _get_url(self, track_id, template_url, server_number): + boundaries = (1, 30) + for nr in server_numbers(server_number, boundaries): + url = template_url % nr try: # We only want to know if the request succeed # don't download the whole file - self._request_webpage(url, None, False) + self._request_webpage( + HEADRequest(url), track_id, + 'Checking URL %d/%d ...' % (nr, boundaries[-1])) return url except ExtractorError: - url = None - + pass return None - def _get_url(self, template_url): - return self.check_urls(template_url % i for i in range(30)) - def _real_extract(self, url): mobj = re.match(self._VALID_URL, url) - uploader = mobj.group(1) cloudcast_name = mobj.group(2) - track_id = '-'.join((uploader, cloudcast_name)) - api_url = 'http://api.mixcloud.com/%s/%s/' % (uploader, cloudcast_name) + track_id = compat_urllib_parse.unquote('-'.join((uploader, cloudcast_name))) + webpage = self._download_webpage(url, track_id) - json_data = self._download_webpage(api_url, track_id, - u'Downloading cloudcast info') - info = json.loads(json_data) - preview_url = self._search_regex(r'data-preview-url="(.+?)"', webpage, u'preview url') - song_url = preview_url.replace('/previews/', '/cloudcasts/originals/') + preview_url = self._search_regex( + r'\s(?:data-preview-url|m-preview)="([^"]+)"', webpage, 'preview url') + song_url = preview_url.replace('/previews/', '/c/originals/') + server_number = int(self._search_regex(r'stream(\d+)', song_url, 'server number')) template_url = re.sub(r'(stream\d*)', 'stream%d', song_url) - final_song_url = self._get_url(template_url) + final_song_url = self._get_url(track_id, template_url, server_number) if final_song_url is None: self.to_screen('Trying with m4a extension') template_url = template_url.replace('.mp3', '.m4a').replace('originals/', 'm4a/64/') - final_song_url = self._get_url(template_url) + final_song_url = self._get_url(track_id, template_url, server_number) if final_song_url is None: - raise ExtractorError(u'Unable to extract track url') + raise ExtractorError('Unable to extract track url') + + PREFIX = ( + r'm-play-on-spacebar[^>]+' + r'(?:\s+[a-zA-Z0-9-]+(?:="[^"]+")?)*?\s+') + title = self._html_search_regex( + PREFIX + r'm-title="([^"]+)"', webpage, 'title') + thumbnail = self._proto_relative_url(self._html_search_regex( + PREFIX + r'm-thumbnail-url="([^"]+)"', webpage, 'thumbnail', + fatal=False)) + uploader = self._html_search_regex( + PREFIX + r'm-owner-name="([^"]+)"', + webpage, 'uploader', fatal=False) + uploader_id = self._search_regex( + r'\s+"profile": "([^"]+)",', webpage, 'uploader id', fatal=False) + description = self._og_search_description(webpage) + like_count = str_to_int(self._search_regex( + r'\bbutton-favorite\b[^>]+m-ajax-toggle-count="([^"]+)"', + webpage, 'like count', fatal=False)) + view_count = str_to_int(self._search_regex( + [r'([0-9,.]+)'], + webpage, 'play count', fatal=False)) return { 'id': track_id, - 'title': info['name'], + 'title': title, 'url': final_song_url, - 'description': info.get('description'), - 'thumbnail': info['pictures'].get('extra_large'), - 'uploader': info['user']['name'], - 'uploader_id': info['user']['username'], - 'upload_date': unified_strdate(info['created_time']), - 'view_count': info['play_count'], + 'description': description, + 'thumbnail': thumbnail, + 'uploader': uploader, + 'uploader_id': uploader_id, + 'view_count': view_count, + 'like_count': like_count, } + + +def server_numbers(first, boundaries): + """ Server numbers to try in descending order of probable availability. + Starting from first (i.e. the number of the server hosting the preview file) + and going further and further up to the higher boundary and down to the + lower one in an alternating fashion. Namely: + + server_numbers(2, (1, 5)) + + # Where the preview server is 2, min number is 1 and max is 5. + # Yields: 2, 3, 1, 4, 5 + + Why not random numbers or increasing sequences? Since from what I've seen, + full length files seem to be hosted on servers whose number is closer to + that of the preview; to be confirmed. + """ + zip_longest = getattr(itertools, 'zip_longest', None) + if zip_longest is None: + # python 2.x + zip_longest = itertools.izip_longest + + if len(boundaries) != 2: + raise ValueError("boundaries should be a two-element tuple") + min, max = boundaries + highs = range(first + 1, max + 1) + lows = range(first - 1, min - 1, -1) + rest = filter( + None, itertools.chain.from_iterable(zip_longest(highs, lows))) + yield first + for n in rest: + yield n