projects
/
youtube-dl
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Merge branch 'ustream-embed-recorded2' of https://github.com/anovicecodemonkey/youtub...
[youtube-dl]
/
youtube_dl
/
extractor
/
academicearth.py
diff --git
a/youtube_dl/extractor/academicearth.py
b/youtube_dl/extractor/academicearth.py
index ac05f824610a1385b3d2403c3a91ffa8ccc45e8a..59d3bbba413c3c256a3f77917708fb171e337b14 100644
(file)
--- a/
youtube_dl/extractor/academicearth.py
+++ b/
youtube_dl/extractor/academicearth.py
@@
-1,11
+1,12
@@
+from __future__ import unicode_literals
import re
from .common import InfoExtractor
class AcademicEarthCourseIE(InfoExtractor):
import re
from .common import InfoExtractor
class AcademicEarthCourseIE(InfoExtractor):
- _VALID_URL = r'^https?://(?:www\.)?academicearth\.org/
(?:courses|playlists)
/(?P<id>[^?#/]+)'
- IE_NAME =
u
'AcademicEarth:Course'
+ _VALID_URL = r'^https?://(?:www\.)?academicearth\.org/
playlists
/(?P<id>[^?#/]+)'
+ IE_NAME = 'AcademicEarth:Course'
def _real_extract(self, url):
m = re.match(self._VALID_URL, url)
def _real_extract(self, url):
m = re.match(self._VALID_URL, url)
@@
-13,12
+14,12
@@
class AcademicEarthCourseIE(InfoExtractor):
webpage = self._download_webpage(url, playlist_id)
title = self._html_search_regex(
webpage = self._download_webpage(url, playlist_id)
title = self._html_search_regex(
- r'<h1 class="playlist-name">(.*?)</h1>', webpage, u'title')
+ r'<h1 class="playlist-name"
[^>]*?
>(.*?)</h1>', webpage, u'title')
description = self._html_search_regex(
description = self._html_search_regex(
- r'<p class="excerpt">(.*?)</p>',
+ r'<p class="excerpt"
[^>]*?
>(.*?)</p>',
webpage, u'description', fatal=False)
urls = re.findall(
webpage, u'description', fatal=False)
urls = re.findall(
- r'<
h3 class="lecture-title">
<a target="_blank" href="([^"]+)">',
+ r'<
li class="lecture-preview">\s*?
<a target="_blank" href="([^"]+)">',
webpage)
entries = [self.url_result(u) for u in urls]
webpage)
entries = [self.url_result(u) for u in urls]