1 # -*- coding: utf-8 -*-
3 # gPodder - A media aggregator and podcast client
4 # Copyright (c) 2005-2009 Thomas Perl and the gPodder Team
6 # gPodder is free software; you can redistribute it and/or modify
7 # it under the terms of the GNU General Public License as published by
8 # the Free Software Foundation; either version 3 of the License, or
9 # (at your option) any later version.
11 # gPodder is distributed in the hope that it will be useful,
12 # but WITHOUT ANY WARRANTY; without even the implied warranty of
13 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 # GNU General Public License for more details.
16 # You should have received a copy of the GNU General Public License
17 # along with this program. If not, see <http://www.gnu.org/licenses/>.
19 # resolver.py -- YouTube and related magic
20 # Justin Forest <justin.forest@gmail.com> 2008-10-13
25 # * Support for Vimeo, maybe blip.tv and others.
34 from xml
.sax
import saxutils
35 from gpodder
.liblogger
import log
36 from gpodder
.util
import proxy_request
38 def get_real_download_url(url
, proxy
=None):
39 # IDs from http://forum.videohelp.com/topic336882-1800.html#1912972
40 if gpodder
.interface
== gpodder
.MAEMO
:
41 # Use 3GP with AAC on Maemo
44 # Use MP4 with AAC by default
47 r1
= re
.compile('http://(?:[a-z]+\.)?youtube\.com/v/(.*)\.swf', re
.IGNORECASE
).match(url
)
49 page
= proxy_request('http://www.youtube.com/watch?v=' + r1
.group(1), proxy
, method
='GET').read()
51 r2
= re
.compile('.*"t"\:\s+"([^"]+)".*').search(page
)
53 next
= 'http://www.youtube.com/get_video?video_id=' + r1
.group(1) + '&t=' + r2
.group(1) + '&fmt=%d' % fmt_id
54 log('YouTube link resolved: %s => %s', url
, next
)
59 def get_real_channel_url(url
):
60 r
= re
.compile('http://(?:[a-z]+\.)?youtube\.com/user/([a-z0-9]+)', re
.IGNORECASE
)
64 next
= 'http://www.youtube.com/rss/user/'+ m
.group(1) +'/videos.rss'
65 log('YouTube link resolved: %s => %s', url
, next
)
68 r
= re
.compile('http://(?:[a-z]+\.)?youtube\.com/profile?user=([a-z0-9]+)', re
.IGNORECASE
)
72 next
= 'http://www.youtube.com/rss/user/'+ m
.group(1) +'/videos.rss'
73 log('YouTube link resolved: %s => %s', url
, next
)
78 def get_real_cover(url
):
81 r
= re
.compile('http://www\.youtube\.com/rss/user/([a-z0-9]+)/videos\.rss', re
.IGNORECASE
)
85 data
= urllib2
.urlopen('http://www.youtube.com/user/'+ m
.group(1)).read()
86 data
= data
[data
.find('id="user-profile-image"'):]
87 data
= data
[data
.find('src="') + 5:]
89 next
= data
[:data
.find('"')]
91 if next
.strip() == '':
94 log('YouTube userpic for %s is: %s', url
, next
)
99 def get_real_episode_length(episode
):
100 url
= get_real_download_url(episode
.url
)
102 if url
!= episode
.url
:
104 info
= urllib2
.urlopen(url
).info()
105 if 'content-length' in info
:
106 return info
['content-length']
107 except urllib2
.HTTPError
:
112 def find_youtube_channels(string
):
113 # FIXME: Make proper use of the YouTube API instead
114 # of screen-scraping the YouTube website
115 url
= 'http://www.youtube.com/results?search_query='+ urllib
.quote(string
, '') +'&search_type=search_users&aq=f'
117 r
= re
.compile('>\s+<')
118 data
= r
.sub('><', urllib
.urlopen(url
).read())
120 r1
= re
.compile('<a href="/user/([^"]+)"[^>]*>([^<]+)</a>')
121 m1
= r1
.findall(data
)
123 r2
= re
.compile('\s+')
125 model
= gtk
.ListStore(gobject
.TYPE_BOOLEAN
, gobject
.TYPE_STRING
, gobject
.TYPE_STRING
)
128 for (name
, title
) in m1
:
129 if name
not in found_users
:
130 found_users
.append(name
)
131 link
= 'http://www.youtube.com/rss/user/'+ name
+'/videos.rss'
132 model
.append([False, name
, link
])