X-Git-Url: http://git.bitcoin.ninja/index.cgi?a=blobdiff_plain;f=youtube_dl%2Fextractor%2Fustream.py;h=74c82587f802acc82e3025235a4bfe92725e8111;hb=a921f40799d2ecb4be53b3241d2dbfc80f804d73;hp=f69b27d4405d719e9c8d87807d2c1392753150eb;hpb=74ac9bdd82b8a625ea9782251258ab7da1463877;p=youtube-dl diff --git a/youtube_dl/extractor/ustream.py b/youtube_dl/extractor/ustream.py index f69b27d44..74c82587f 100644 --- a/youtube_dl/extractor/ustream.py +++ b/youtube_dl/extractor/ustream.py @@ -4,7 +4,7 @@ import re from .common import InfoExtractor from ..utils import ( compat_urlparse, - compat_html_parser, + get_meta_content, ) @@ -49,40 +49,6 @@ class UstreamIE(InfoExtractor): } return info -# More robust than regular expressions - -class ChannelParser(compat_html_parser.HTMLParser): - """ - - """ - channel_id = None - - def handle_starttag(self, tag, attrs): - if tag != 'meta': - return - values = dict(attrs) - if values.get('name') != 'ustream:channel_id': - return - value = values.get('content', '') - if value.isdigit(): - self.channel_id = value - -class SocialstreamParser(compat_html_parser.HTMLParser): - """ -
  • - """ - def __init__(self): - compat_html_parser.HTMLParser.__init__(self) - self.content_ids = [] - - def handle_starttag(self, tag, attrs): - if tag != 'li': - return - for (attr, value) in attrs: - if attr == 'data-content-id' and value.isdigit(): - self.content_ids.append(value) - class UstreamChannelIE(InfoExtractor): _VALID_URL = r'https?://www\.ustream\.tv/channel/(?P.+)' IE_NAME = u'ustream:channel' @@ -90,21 +56,16 @@ class UstreamChannelIE(InfoExtractor): def _real_extract(self, url): m = re.match(self._VALID_URL, url) slug = m.group('slug') + webpage = self._download_webpage(url, slug) + channel_id = get_meta_content('ustream:channel_id', webpage) - p = ChannelParser() - p.feed(self._download_webpage(url, slug)) - p.close() - channel_id = p.channel_id - - p = SocialstreamParser() BASE = 'http://www.ustream.tv' next_url = '/ajax/socialstream/videos/%s/1.json' % channel_id + video_ids = [] while next_url: reply = json.loads(self._download_webpage(compat_urlparse.urljoin(BASE, next_url), channel_id)) - p.feed(reply['data']) + video_ids.extend(re.findall(r'data-content-id="(\d.*)"', reply['data'])) next_url = reply['nextUrl'] - p.close() - video_ids = p.content_ids urls = ['http://www.ustream.tv/recorded/' + vid for vid in video_ids] url_entries = [self.url_result(eurl, 'Ustream') for eurl in urls]