projects
/
youtube-dl
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Fix W504 and disable W503 (closes #20863)
[youtube-dl]
/
youtube_dl
/
extractor
/
ooyala.py
diff --git
a/youtube_dl/extractor/ooyala.py
b/youtube_dl/extractor/ooyala.py
index 52580baed3c49242be55529f8c75f5d110ff41f7..995b24d1bd4758f0c51350d517a52a83053d584f 100644
(file)
--- a/
youtube_dl/extractor/ooyala.py
+++ b/
youtube_dl/extractor/ooyala.py
@@
-1,9
+1,13
@@
from __future__ import unicode_literals
from __future__ import unicode_literals
+
import re
import re
-import base64
from .common import InfoExtractor
from .common import InfoExtractor
-from ..compat import compat_str
+from ..compat import (
+ compat_b64decode,
+ compat_str,
+ compat_urllib_parse_urlencode,
+)
from ..utils import (
determine_ext,
ExtractorError,
from ..utils import (
determine_ext,
ExtractorError,
@@
-12,7
+16,6
@@
from ..utils import (
try_get,
unsmuggle_url,
)
try_get,
unsmuggle_url,
)
-from ..compat import compat_urllib_parse_urlencode
class OoyalaBaseIE(InfoExtractor):
class OoyalaBaseIE(InfoExtractor):
@@
-28,12
+31,12
@@
class OoyalaBaseIE(InfoExtractor):
title = metadata['title']
auth_data = self._download_json(
title = metadata['title']
auth_data = self._download_json(
- self._AUTHORIZATION_URL_TEMPLATE % (pcode, embed_code)
+
- compat_urllib_parse_urlencode({
+ self._AUTHORIZATION_URL_TEMPLATE % (pcode, embed_code)
+
+
compat_urllib_parse_urlencode({
'domain': domain,
'supportedFormats': supportedformats or 'mp4,rtmp,m3u8,hds,dash,smooth',
'embedToken': embed_token,
'domain': domain,
'supportedFormats': supportedformats or 'mp4,rtmp,m3u8,hds,dash,smooth',
'embedToken': embed_token,
- }), video_id)
+ }), video_id
, headers=self.geo_verification_headers()
)
cur_auth_data = auth_data['authorization_data'][embed_code]
cur_auth_data = auth_data['authorization_data'][embed_code]
@@
-44,7
+47,7
@@
class OoyalaBaseIE(InfoExtractor):
url_data = try_get(stream, lambda x: x['url']['data'], compat_str)
if not url_data:
continue
url_data = try_get(stream, lambda x: x['url']['data'], compat_str)
if not url_data:
continue
- s_url =
base64.b64decode(url_data.encode('ascii')
).decode('utf-8')
+ s_url =
compat_b64decode(url_data
).decode('utf-8')
if not s_url or s_url in urls:
continue
urls.append(s_url)
if not s_url or s_url in urls:
continue
urls.append(s_url)