compat_urllib_parse,
)
from ..utils import (
+ encode_dict,
ExtractorError,
int_or_none,
unified_strdate,
class SoundcloudSearchIE(SearchInfoExtractor, SoundcloudIE):
IE_NAME = 'soundcloud:search'
IE_DESC = 'Soundcloud search'
- _MAX_RESULTS = 200
+ _MAX_RESULTS = float('inf')
_TESTS = [{
'url': 'scsearch15:post-avant jazzcore',
'info_dict': {
}]
_SEARCH_KEY = 'scsearch'
- _RESULTS_PER_PAGE = 50
+ _MAX_RESULTS_PER_PAGE = 200
+ _DEFAULT_RESULTS_PER_PAGE = 50
_API_V2_BASE = 'https://api-v2.soundcloud.com'
def _get_collection(self, endpoint, collection_id, **query):
- query['limit'] = self._RESULTS_PER_PAGE
+ limit = min(
+ query.get('limit', self._DEFAULT_RESULTS_PER_PAGE),
+ self._MAX_RESULTS_PER_PAGE)
+ query['limit'] = limit
query['client_id'] = self._CLIENT_ID
query['linked_partitioning'] = '1'
+ query['offset'] = 0
+ data = compat_urllib_parse.urlencode(encode_dict(query))
+ next_url = '{0}{1}?{2}'.format(self._API_V2_BASE, endpoint, data)
- total_results = self._MAX_RESULTS
collected_results = 0
- next_url = None
-
- for i in itertools.count():
- if not next_url:
- query['offset'] = i * self._RESULTS_PER_PAGE
- data = compat_urllib_parse.urlencode(query)
- next_url = '{0}{1}?{2}'.format(self._API_V2_BASE, endpoint, data)
-
- response = self._download_json(next_url,
- video_id=collection_id,
- note='Downloading page {0}'.format(i+1),
- errnote='Unable to download API page')
+ for i in itertools.count(1):
+ response = self._download_json(
+ next_url, collection_id, 'Downloading page {0}'.format(i),
+ 'Unable to download API page')
- total_results = int(response.get(
- 'total_results', total_results))
+ collection = response.get('collection', [])
+ if not collection:
+ break
- collection = response['collection']
+ collection = list(filter(bool, collection))
collected_results += len(collection)
- for item in filter(bool, collection):
- yield item
+ for item in collection:
+ yield self.url_result(item['uri'], SoundcloudIE.ie_key())
- if collected_results >= total_results or not collection:
+ if not collection or collected_results >= limit:
break
- next_url = response.get('next_href', None)
-
- def _get_n_results(self, query, n):
- results = []
-
- tracks = self._get_collection('/search/tracks',
- collection_id='Query "{0}"'.format(query),
- q=query.encode('utf-8'))
-
- for _ in range(n):
- try:
- track = next(tracks)
- except StopIteration:
+ next_url = response.get('next_href')
+ if not next_url:
break
- uri = track['uri']
- title = track['title']
- results.append(self.url_result(url=uri))
-
- if not results:
- raise ExtractorError(
- '[soundcloud] No track results', expected=True)
-
- return self.playlist_result(results[:n], playlist_title=query)
+ def _get_n_results(self, query, n):
+ tracks = self._get_collection('/search/tracks', query, limit=n, q=query)
+ return self.playlist_result(tracks, playlist_title=query)