Merge pull request #1413 from tewe/master
authorJaime Marquínez Ferrándiz <jaime.marquinez.ferrandiz@gmail.com>
Fri, 13 Sep 2013 17:55:49 +0000 (19:55 +0200)
committerJaime Marquínez Ferrándiz <jaime.marquinez.ferrandiz@gmail.com>
Fri, 13 Sep 2013 19:34:31 +0000 (21:34 +0200)
Add Ustream channel support

1  2 
test/test_playlists.py
youtube_dl/extractor/__init__.py
youtube_dl/extractor/ustream.py

index 65de3a55c8694dd764dafa10977448d9274727d2,65de3a55c8694dd764dafa10977448d9274727d2..4a2e00b01a5a368a8ef2b14ecae231e53491935c
@@@ -8,7 -8,7 +8,7 @@@ import jso
  import os
  sys.path.append(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
  
--from youtube_dl.extractor import DailymotionPlaylistIE, VimeoChannelIE
++from youtube_dl.extractor import DailymotionPlaylistIE, VimeoChannelIE, UstreamChannelIE
  from youtube_dl.utils import *
  
  from helper import FakeYDL
@@@ -34,5 -34,5 +34,13 @@@ class TestPlaylists(unittest.TestCase)
          self.assertEqual(result['title'], u'Vimeo Tributes')
          self.assertTrue(len(result['entries']) > 24)
  
++    def test_ustream_channel(self):
++        dl = FakeYDL()
++        ie = UstreamChannelIE(dl)
++        result = ie.extract('http://www.ustream.tv/channel/young-americans-for-liberty')
++        self.assertIsPlaylist(result)
++        self.assertEqual(result['id'], u'5124905')
++        self.assertTrue(len(result['entries']) >= 11)
++
  if __name__ == '__main__':
      unittest.main()
Simple merge
index 5f423870abb6c9e337f7e0c328b16ed3ef5049ab,16cdcc76592feb03ad9b30b962e66041eb391168..f69b27d4405d719e9c8d87807d2c1392753150eb
@@@ -1,6 -1,9 +1,11 @@@
 -from HTMLParser import HTMLParser
+ import json
  import re
 -from urlparse import urljoin
  
  from .common import InfoExtractor
++from ..utils import (
++    compat_urlparse,
++    compat_html_parser,
++)
  
  
  class UstreamIE(InfoExtractor):
                  'thumbnail': thumbnail,
                 }
          return info
 -class ChannelParser(HTMLParser):
+ # More robust than regular expressions
 -class SocialstreamParser(HTMLParser):
++class ChannelParser(compat_html_parser.HTMLParser):
+     """
+     <meta name="ustream:channel_id" content="1234">
+     """
+     channel_id = None
+     def handle_starttag(self, tag, attrs):
+         if tag != 'meta':
+             return
+         values = dict(attrs)
+         if values.get('name') != 'ustream:channel_id':
+             return
+         value = values.get('content', '')
+         if value.isdigit():
+             self.channel_id = value
 -        HTMLParser.__init__(self)
++class SocialstreamParser(compat_html_parser.HTMLParser):
+     """
+     <li class="content123 video" data-content-id="123" data-length="1452"
+         data-href="/recorded/123" data-og-url="/recorded/123">
+     """
+     def __init__(self):
 -        # Slugs can be non-ascii, but youtube-dl can't handle non-ascii command lines,
 -        # so if we got this far it's probably percent encoded and we needn't worry.
++        compat_html_parser.HTMLParser.__init__(self)
+         self.content_ids = []
+     def handle_starttag(self, tag, attrs):
+         if tag != 'li':
+             return
+         for (attr, value) in attrs:
+             if attr == 'data-content-id' and value.isdigit():
+                 self.content_ids.append(value)
+ class UstreamChannelIE(InfoExtractor):
+     _VALID_URL = r'https?://www\.ustream\.tv/channel/(?P<slug>.+)'
+     IE_NAME = u'ustream:channel'
+     def _real_extract(self, url):
+         m = re.match(self._VALID_URL, url)
+         slug = m.group('slug')
 -            reply = json.loads(self._download_webpage(urljoin(BASE, next_url), channel_id))
+         p = ChannelParser()
+         p.feed(self._download_webpage(url, slug))
+         p.close()
+         channel_id = p.channel_id
+         p = SocialstreamParser()
+         BASE = 'http://www.ustream.tv'
+         next_url = '/ajax/socialstream/videos/%s/1.json' % channel_id
+         while next_url:
 -        # From YoutubeChannelIE
 -
 -        self._downloader.to_screen(u'[ustream] Channel %s: Found %i videos' % (channel_id, len(video_ids)))
 -
++            reply = json.loads(self._download_webpage(compat_urlparse.urljoin(BASE, next_url), channel_id))
+             p.feed(reply['data'])
+             next_url = reply['nextUrl']
+         p.close()
+         video_ids = p.content_ids
 -        return [self.playlist_result(url_entries, channel_id)]
+         urls = ['http://www.ustream.tv/recorded/' + vid for vid in video_ids]
+         url_entries = [self.url_result(eurl, 'Ustream') for eurl in urls]
++        return self.playlist_result(url_entries, channel_id)