mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-12-14 23:32:33 +00:00
[soundcloud] Add Soundcloud search extractor
This commit is contained in:
parent
4137196899
commit
2abf7cab80
2 changed files with 89 additions and 2 deletions
|
@ -576,7 +576,8 @@ from .soundcloud import (
|
||||||
SoundcloudIE,
|
SoundcloudIE,
|
||||||
SoundcloudSetIE,
|
SoundcloudSetIE,
|
||||||
SoundcloudUserIE,
|
SoundcloudUserIE,
|
||||||
SoundcloudPlaylistIE
|
SoundcloudPlaylistIE,
|
||||||
|
SoundcloudSearchIE
|
||||||
)
|
)
|
||||||
from .soundgasm import (
|
from .soundgasm import (
|
||||||
SoundgasmIE,
|
SoundgasmIE,
|
||||||
|
|
|
@ -4,7 +4,10 @@ from __future__ import unicode_literals
|
||||||
import re
|
import re
|
||||||
import itertools
|
import itertools
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import (
|
||||||
|
InfoExtractor,
|
||||||
|
SearchInfoExtractor
|
||||||
|
)
|
||||||
from ..compat import (
|
from ..compat import (
|
||||||
compat_str,
|
compat_str,
|
||||||
compat_urlparse,
|
compat_urlparse,
|
||||||
|
@ -469,3 +472,86 @@ class SoundcloudPlaylistIE(SoundcloudIE):
|
||||||
'description': data.get('description'),
|
'description': data.get('description'),
|
||||||
'entries': entries,
|
'entries': entries,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class SoundcloudSearchIE(SearchInfoExtractor, SoundcloudIE):
|
||||||
|
IE_NAME = 'soundcloud:search'
|
||||||
|
IE_DESC = 'Soundcloud search'
|
||||||
|
_MAX_RESULTS = 200
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'scsearch15:post-avant jazzcore',
|
||||||
|
'info_dict': {
|
||||||
|
'title': 'post-avant jazzcore',
|
||||||
|
},
|
||||||
|
'playlist_count': 15,
|
||||||
|
}]
|
||||||
|
|
||||||
|
_SEARCH_KEY = 'scsearch'
|
||||||
|
_RESULTS_PER_PAGE = 50
|
||||||
|
|
||||||
|
def _get_collection(self, endpoint, collection_id, **query):
|
||||||
|
import itertools
|
||||||
|
|
||||||
|
query['limit'] = self._RESULTS_PER_PAGE
|
||||||
|
query['client_id'] = self._CLIENT_ID
|
||||||
|
query['linked_partitioning'] = '1'
|
||||||
|
|
||||||
|
api_base_url = '{0}//api-v2.soundcloud.com'.format(self.http_scheme())
|
||||||
|
|
||||||
|
total_results = self._MAX_RESULTS
|
||||||
|
collected_results = 0
|
||||||
|
|
||||||
|
next_url = None
|
||||||
|
|
||||||
|
for i in itertools.count():
|
||||||
|
|
||||||
|
if not next_url:
|
||||||
|
query['offset'] = i * self._RESULTS_PER_PAGE
|
||||||
|
data = compat_urllib_parse.urlencode(query)
|
||||||
|
next_url = '{0}{1}?{2}'.format(api_base_url, endpoint, data)
|
||||||
|
|
||||||
|
response = self._download_json(next_url,
|
||||||
|
video_id=collection_id,
|
||||||
|
note='Downloading page {0}'.format(i+1),
|
||||||
|
errnote='Unable to download API page')
|
||||||
|
|
||||||
|
total_results = int(response.get(
|
||||||
|
u'total_results', total_results))
|
||||||
|
|
||||||
|
collection = response['collection']
|
||||||
|
collected_results += len(collection)
|
||||||
|
|
||||||
|
for item in filter(bool, collection):
|
||||||
|
yield item
|
||||||
|
|
||||||
|
if collected_results >= total_results or not collection:
|
||||||
|
break
|
||||||
|
|
||||||
|
next_url = response.get(u'next_href', None)
|
||||||
|
|
||||||
|
def _get_n_results(self, query, n):
|
||||||
|
|
||||||
|
results = []
|
||||||
|
|
||||||
|
tracks = self._get_collection('/search/tracks',
|
||||||
|
collection_id='Query "{}"'.format(query),
|
||||||
|
q=query.encode('utf-8'))
|
||||||
|
|
||||||
|
for _ in range(n):
|
||||||
|
try:
|
||||||
|
track = next(tracks)
|
||||||
|
except StopIteration:
|
||||||
|
break
|
||||||
|
uri = track[u'uri']
|
||||||
|
title = track[u'title']
|
||||||
|
username = track[u'user'][u'username']
|
||||||
|
results.append(self.url_result(
|
||||||
|
url=uri,
|
||||||
|
video_title='{0} - {1}'.format(username, title)))
|
||||||
|
|
||||||
|
if not results:
|
||||||
|
raise ExtractorError(
|
||||||
|
'[soundcloud] No track results', expected=True)
|
||||||
|
|
||||||
|
return self.playlist_result(results[:n], playlist_title=query)
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue