projects
/
youtube-dl
/ commitdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
| commitdiff |
tree
raw
|
patch
|
inline
| side by side (parent:
f012823
)
[pornhub] Use actual URL host for requests (closes #18359)
author
Sergey M․
<dstftw@gmail.com>
Sat, 1 Dec 2018 09:45:51 +0000
(16:45 +0700)
committer
Sergey M․
<dstftw@gmail.com>
Sat, 1 Dec 2018 09:45:51 +0000
(16:45 +0700)
youtube_dl/extractor/pornhub.py
patch
|
blob
|
history
diff --git
a/youtube_dl/extractor/pornhub.py
b/youtube_dl/extractor/pornhub.py
index c9c8840954a91f8be9f988955e3199f539d18526..e377de19664524844699e4046d3ebe48fcbc0167 100644
(file)
--- a/
youtube_dl/extractor/pornhub.py
+++ b/
youtube_dl/extractor/pornhub.py
@@
-27,7
+27,7
@@
class PornHubIE(InfoExtractor):
_VALID_URL = r'''(?x)
https?://
(?:
_VALID_URL = r'''(?x)
https?://
(?:
- (?:[^/]+\.)?
pornhub\.(?:com|net
)/(?:(?:view_video\.php|video/show)\?viewkey=|embed/)|
+ (?:[^/]+\.)?
(?P<host>pornhub\.(?:com|net)
)/(?:(?:view_video\.php|video/show)\?viewkey=|embed/)|
(?:www\.)?thumbzilla\.com/video/
)
(?P<id>[\da-z]+)
(?:www\.)?thumbzilla\.com/video/
)
(?P<id>[\da-z]+)
@@
-129,7
+129,7
@@
class PornHubIE(InfoExtractor):
@staticmethod
def _extract_urls(webpage):
return re.findall(
@staticmethod
def _extract_urls(webpage):
return re.findall(
- r'<iframe[^>]+?src=["\'](?P<url>(?:https?:)?//(?:www\.)?pornhub\.
com
/embed/[\da-z]+)',
+ r'<iframe[^>]+?src=["\'](?P<url>(?:https?:)?//(?:www\.)?pornhub\.
(?:com|net)
/embed/[\da-z]+)',
webpage)
def _extract_count(self, pattern, webpage, name):
webpage)
def _extract_count(self, pattern, webpage, name):
@@
-137,14
+137,16
@@
class PornHubIE(InfoExtractor):
pattern, webpage, '%s count' % name, fatal=False))
def _real_extract(self, url):
pattern, webpage, '%s count' % name, fatal=False))
def _real_extract(self, url):
- video_id = self._match_id(url)
+ mobj = re.match(self._VALID_URL, url)
+ host = mobj.group('host') or 'pornhub.com'
+ video_id = mobj.group('id')
- self._set_cookie(
'pornhub.com'
, 'age_verified', '1')
+ self._set_cookie(
host
, 'age_verified', '1')
def dl_webpage(platform):
def dl_webpage(platform):
- self._set_cookie(
'pornhub.com'
, 'platform', platform)
+ self._set_cookie(
host
, 'platform', platform)
return self._download_webpage(
return self._download_webpage(
- 'http://www.
pornhub.com/view_video.php?viewkey=%s' % video_id
,
+ 'http://www.
%s/view_video.php?viewkey=%s' % (host, video_id)
,
video_id, 'Downloading %s webpage' % platform)
webpage = dl_webpage('pc')
video_id, 'Downloading %s webpage' % platform)
webpage = dl_webpage('pc')
@@
-306,7
+308,7
@@
class PornHubIE(InfoExtractor):
class PornHubPlaylistBaseIE(InfoExtractor):
class PornHubPlaylistBaseIE(InfoExtractor):
- def _extract_entries(self, webpage):
+ def _extract_entries(self, webpage
, host
):
# Only process container div with main playlist content skipping
# drop-down menu that uses similar pattern for videos (see
# https://github.com/rg3/youtube-dl/issues/11594).
# Only process container div with main playlist content skipping
# drop-down menu that uses similar pattern for videos (see
# https://github.com/rg3/youtube-dl/issues/11594).
@@
-316,7
+318,7
@@
class PornHubPlaylistBaseIE(InfoExtractor):
return [
self.url_result(
return [
self.url_result(
- 'http://www.
pornhub.com/%s' % video_url
,
+ 'http://www.
%s/%s' % (host, video_url)
,
PornHubIE.ie_key(), video_title=title)
for video_url, title in orderedSet(re.findall(
r'href="/?(view_video\.php\?.*\bviewkey=[\da-z]+[^"]*)"[^>]*\s+title="([^"]+)"',
PornHubIE.ie_key(), video_title=title)
for video_url, title in orderedSet(re.findall(
r'href="/?(view_video\.php\?.*\bviewkey=[\da-z]+[^"]*)"[^>]*\s+title="([^"]+)"',
@@
-324,11
+326,13
@@
class PornHubPlaylistBaseIE(InfoExtractor):
]
def _real_extract(self, url):
]
def _real_extract(self, url):
- playlist_id = self._match_id(url)
+ mobj = re.match(self._VALID_URL, url)
+ host = mobj.group('host')
+ playlist_id = mobj.group('id')
webpage = self._download_webpage(url, playlist_id)
webpage = self._download_webpage(url, playlist_id)
- entries = self._extract_entries(webpage)
+ entries = self._extract_entries(webpage
, host
)
playlist = self._parse_json(
self._search_regex(
playlist = self._parse_json(
self._search_regex(
@@
-343,7
+347,7
@@
class PornHubPlaylistBaseIE(InfoExtractor):
class PornHubPlaylistIE(PornHubPlaylistBaseIE):
class PornHubPlaylistIE(PornHubPlaylistBaseIE):
- _VALID_URL = r'https?://(?:[^/]+\.)?
pornhub\.(?:com|net
)/playlist/(?P<id>\d+)'
+ _VALID_URL = r'https?://(?:[^/]+\.)?
(?P<host>pornhub\.(?:com|net)
)/playlist/(?P<id>\d+)'
_TESTS = [{
'url': 'http://www.pornhub.com/playlist/4667351',
'info_dict': {
_TESTS = [{
'url': 'http://www.pornhub.com/playlist/4667351',
'info_dict': {
@@
-358,7
+362,7
@@
class PornHubPlaylistIE(PornHubPlaylistBaseIE):
class PornHubUserVideosIE(PornHubPlaylistBaseIE):
class PornHubUserVideosIE(PornHubPlaylistBaseIE):
- _VALID_URL = r'https?://(?:[^/]+\.)?
pornhub\.(?:com|net
)/(?:(?:user|channel)s|model|pornstar)/(?P<id>[^/]+)/videos'
+ _VALID_URL = r'https?://(?:[^/]+\.)?
(?P<host>pornhub\.(?:com|net)
)/(?:(?:user|channel)s|model|pornstar)/(?P<id>[^/]+)/videos'
_TESTS = [{
'url': 'http://www.pornhub.com/users/zoe_ph/videos/public',
'info_dict': {
_TESTS = [{
'url': 'http://www.pornhub.com/users/zoe_ph/videos/public',
'info_dict': {
@@
-399,7
+403,9
@@
class PornHubUserVideosIE(PornHubPlaylistBaseIE):
}]
def _real_extract(self, url):
}]
def _real_extract(self, url):
- user_id = self._match_id(url)
+ mobj = re.match(self._VALID_URL, url)
+ host = mobj.group('host')
+ user_id = mobj.group('id')
entries = []
for page_num in itertools.count(1):
entries = []
for page_num in itertools.count(1):
@@
-411,7
+417,7
@@
class PornHubUserVideosIE(PornHubPlaylistBaseIE):
if isinstance(e.cause, compat_HTTPError) and e.cause.code == 404:
break
raise
if isinstance(e.cause, compat_HTTPError) and e.cause.code == 404:
break
raise
- page_entries = self._extract_entries(webpage)
+ page_entries = self._extract_entries(webpage
, host
)
if not page_entries:
break
entries.extend(page_entries)
if not page_entries:
break
entries.extend(page_entries)