1 """Open an arbitrary URL.
3 See the following document for more info on URLs:
4 "Names and Addresses, URIs, URLs, URNs, URCs", at
5 http://www.w3.org/pub/WWW/Addressing/Overview.html
7 See also the HTTP spec (from which the error codes are derived):
8 "HTTP - Hypertext Transfer Protocol", at
9 http://www.w3.org/pub/WWW/Protocols/
11 Related standards and specs:
12 - RFC1808: the "relative URL" spec. (authoritative status)
13 - RFC1738 - the "URL standard". (authoritative status)
14 - RFC1630 - the "URI spec". (informational status)
16 The object returned by URLopener().open(file) will differ per
17 protocol. All you know is that is has methods read(), readline(),
18 readlines(), fileno(), close() and info(). The read*(), fileno()
19 and close() methods work like those of open files.
20 The info() method returns a mimetools.Message object which can be
21 used to query various info about the object, if available.
22 (mimetools.Message objects are queried with the getheader() method.)
30 from urlparse
import urljoin
as basejoin
32 __all__
= ["urlopen", "URLopener", "FancyURLopener", "urlretrieve",
33 "urlcleanup", "quote", "quote_plus", "unquote", "unquote_plus",
34 "urlencode", "url2pathname", "pathname2url", "splittag",
35 "localhost", "thishost", "ftperrors", "basejoin", "unwrap",
36 "splittype", "splithost", "splituser", "splitpasswd", "splitport",
37 "splitnport", "splitquery", "splitattr", "splitvalue",
38 "splitgophertype", "getproxies"]
40 __version__
= '1.17' # XXX This version is not always updated :-(
42 MAXFTPCACHE
= 10 # Trim the ftp cache beyond this size
44 # Helper for non-unix systems
46 from macurl2path
import url2pathname
, pathname2url
48 from nturl2path
import url2pathname
, pathname2url
49 elif os
.name
== 'riscos':
50 from rourl2path
import url2pathname
, pathname2url
52 def url2pathname(pathname
):
53 """OS-specific conversion from a relative URL of the 'file' scheme
54 to a file system path; not recommended for general use."""
55 return unquote(pathname
)
57 def pathname2url(pathname
):
58 """OS-specific conversion from a file system path to a relative URL
59 of the 'file' scheme; not recommended for general use."""
60 return quote(pathname
)
62 # This really consists of two pieces:
63 # (1) a class which handles opening of all sorts of URLs
64 # (plus assorted utilities etc.)
65 # (2) a set of functions for parsing URLs
66 # XXX Should these be separated out into different modules?
69 # Shortcut for basic usage
71 def urlopen(url
, data
=None, proxies
=None):
72 """urlopen(url [, data]) -> open file-like object"""
74 if proxies
is not None:
75 opener
= FancyURLopener(proxies
=proxies
)
77 opener
= FancyURLopener()
82 return opener
.open(url
)
84 return opener
.open(url
, data
)
85 def urlretrieve(url
, filename
=None, reporthook
=None, data
=None):
88 _urlopener
= FancyURLopener()
89 return _urlopener
.retrieve(url
, filename
, reporthook
, data
)
94 # exception raised when downloaded size does not match content-length
95 class ContentTooShortError(IOError):
96 def __init__(self
, message
, content
):
97 IOError.__init
__(self
, message
)
98 self
.content
= content
102 """Class to open URLs.
103 This is a class rather than just a subroutine because we may need
104 more than one set of global protocol-specific options.
105 Note -- this is a base class for those who don't want the
106 automatic handling of errors type 302 (relocated) and 401
107 (authorization needed)."""
111 version
= "Python-urllib/%s" % __version__
114 def __init__(self
, proxies
=None, **x509
):
116 proxies
= getproxies()
117 assert hasattr(proxies
, 'has_key'), "proxies must be a mapping"
118 self
.proxies
= proxies
119 self
.key_file
= x509
.get('key_file')
120 self
.cert_file
= x509
.get('cert_file')
121 self
.addheaders
= [('User-Agent', self
.version
)]
122 self
.__tempfiles
= []
123 self
.__unlink
= os
.unlink
# See cleanup()
124 self
.tempcache
= None
125 # Undocumented feature: if you assign {} to tempcache,
126 # it is used to cache files retrieved with
127 # self.retrieve(). This is not enabled by default
128 # since it does not work for changing documents (and I
129 # haven't got the logic to check expiration headers
131 self
.ftpcache
= ftpcache
132 # Undocumented feature: you can use a different
133 # ftp cache by assigning to the .ftpcache member;
134 # in case you want logically independent URL openers
135 # XXX This is not threadsafe. Bah.
144 # This code sometimes runs when the rest of this module
145 # has already been deleted, so it can't use any globals
146 # or import anything.
148 for file in self
.__tempfiles
:
153 del self
.__tempfiles
[:]
155 self
.tempcache
.clear()
157 def addheader(self
, *args
):
158 """Add a header to be used by the HTTP interface only
159 e.g. u.addheader('Accept', 'sound/basic')"""
160 self
.addheaders
.append(args
)
163 def open(self
, fullurl
, data
=None):
164 """Use URLopener().open(file) instead of open(file, 'r')."""
165 fullurl
= unwrap(toBytes(fullurl
))
166 if self
.tempcache
and fullurl
in self
.tempcache
:
167 filename
, headers
= self
.tempcache
[fullurl
]
168 fp
= open(filename
, 'rb')
169 return addinfourl(fp
, headers
, fullurl
)
170 urltype
, url
= splittype(fullurl
)
173 if urltype
in self
.proxies
:
174 proxy
= self
.proxies
[urltype
]
175 urltype
, proxyhost
= splittype(proxy
)
176 host
, selector
= splithost(proxyhost
)
177 url
= (host
, fullurl
) # Signal special case to open_*()
180 name
= 'open_' + urltype
182 name
= name
.replace('-', '_')
183 if not hasattr(self
, name
):
185 return self
.open_unknown_proxy(proxy
, fullurl
, data
)
187 return self
.open_unknown(fullurl
, data
)
190 return getattr(self
, name
)(url
)
192 return getattr(self
, name
)(url
, data
)
193 except socket
.error
, msg
:
194 raise IOError, ('socket error', msg
), sys
.exc_info()[2]
196 def open_unknown(self
, fullurl
, data
=None):
197 """Overridable interface to open unknown URL type."""
198 type, url
= splittype(fullurl
)
199 raise IOError, ('url error', 'unknown url type', type)
201 def open_unknown_proxy(self
, proxy
, fullurl
, data
=None):
202 """Overridable interface to open unknown URL type."""
203 type, url
= splittype(fullurl
)
204 raise IOError, ('url error', 'invalid proxy for %s' % type, proxy
)
207 def retrieve(self
, url
, filename
=None, reporthook
=None, data
=None):
208 """retrieve(url) returns (filename, headers) for a local object
209 or (tempfilename, headers) for a remote object."""
210 url
= unwrap(toBytes(url
))
211 if self
.tempcache
and url
in self
.tempcache
:
212 return self
.tempcache
[url
]
213 type, url1
= splittype(url
)
214 if filename
is None and (not type or type == 'file'):
216 fp
= self
.open_local_file(url1
)
219 return url2pathname(splithost(url1
)[1]), hdrs
222 fp
= self
.open(url
, data
)
225 tfp
= open(filename
, 'wb')
228 garbage
, path
= splittype(url
)
229 garbage
, path
= splithost(path
or "")
230 path
, garbage
= splitquery(path
or "")
231 path
, garbage
= splitattr(path
or "")
232 suffix
= os
.path
.splitext(path
)[1]
233 (fd
, filename
) = tempfile
.mkstemp(suffix
)
234 self
.__tempfiles
.append(filename
)
235 tfp
= os
.fdopen(fd
, 'wb')
236 result
= filename
, headers
237 if self
.tempcache
is not None:
238 self
.tempcache
[url
] = result
244 if "content-length" in headers
:
245 size
= int(headers
["Content-Length"])
246 reporthook(blocknum
, bs
, size
)
255 reporthook(blocknum
, bs
, size
)
261 # raise exception if actual size does not match content-length header
262 if size
>= 0 and read
< size
:
263 raise ContentTooShortError("retrieval incomplete: got only %i out "
264 "of %i bytes" % (read
, size
), result
)
268 # Each method named open_<type> knows how to open that type of URL
270 def open_http(self
, url
, data
=None):
271 """Use HTTP protocol."""
275 if isinstance(url
, str):
276 host
, selector
= splithost(url
)
278 user_passwd
, host
= splituser(host
)
283 # check whether the proxy contains authorization information
284 proxy_passwd
, host
= splituser(host
)
285 # now we proceed with the url we want to obtain
286 urltype
, rest
= splittype(selector
)
289 if urltype
.lower() != 'http':
292 realhost
, rest
= splithost(rest
)
294 user_passwd
, realhost
= splituser(realhost
)
296 selector
= "%s://%s%s" % (urltype
, realhost
, rest
)
297 if proxy_bypass(realhost
):
300 #print "proxy via http:", host, selector
301 if not host
: raise IOError, ('http error', 'no host given')
305 proxy_auth
= base64
.b64encode(proxy_passwd
).strip()
311 auth
= base64
.b64encode(user_passwd
).strip()
314 h
= httplib
.HTTP(host
)
316 h
.putrequest('POST', selector
)
317 h
.putheader('Content-Type', 'application/x-www-form-urlencoded')
318 h
.putheader('Content-Length', '%d' % len(data
))
320 h
.putrequest('GET', selector
)
321 if proxy_auth
: h
.putheader('Proxy-Authorization', 'Basic %s' % proxy_auth
)
322 if auth
: h
.putheader('Authorization', 'Basic %s' % auth
)
323 if realhost
: h
.putheader('Host', realhost
)
324 for args
in self
.addheaders
: h
.putheader(*args
)
328 errcode
, errmsg
, headers
= h
.getreply()
332 # something went wrong with the HTTP status line
333 raise IOError, ('http protocol error', 0,
334 'got a bad status line', None)
336 return addinfourl(fp
, headers
, "http:" + url
)
339 return self
.http_error(url
, fp
, errcode
, errmsg
, headers
)
341 return self
.http_error(url
, fp
, errcode
, errmsg
, headers
, data
)
343 def http_error(self
, url
, fp
, errcode
, errmsg
, headers
, data
=None):
344 """Handle http errors.
345 Derived class can override this, or provide specific handlers
346 named http_error_DDD where DDD is the 3-digit error code."""
347 # First check if there's a specific handler for this error
348 name
= 'http_error_%d' % errcode
349 if hasattr(self
, name
):
350 method
= getattr(self
, name
)
352 result
= method(url
, fp
, errcode
, errmsg
, headers
)
354 result
= method(url
, fp
, errcode
, errmsg
, headers
, data
)
355 if result
: return result
356 return self
.http_error_default(url
, fp
, errcode
, errmsg
, headers
)
358 def http_error_default(self
, url
, fp
, errcode
, errmsg
, headers
):
359 """Default error handler: close the connection and raise IOError."""
362 raise IOError, ('http error', errcode
, errmsg
, headers
)
364 if hasattr(socket
, "ssl"):
365 def open_https(self
, url
, data
=None):
366 """Use HTTPS protocol."""
370 if isinstance(url
, str):
371 host
, selector
= splithost(url
)
373 user_passwd
, host
= splituser(host
)
378 # here, we determine, whether the proxy contains authorization information
379 proxy_passwd
, host
= splituser(host
)
380 urltype
, rest
= splittype(selector
)
383 if urltype
.lower() != 'https':
386 realhost
, rest
= splithost(rest
)
388 user_passwd
, realhost
= splituser(realhost
)
390 selector
= "%s://%s%s" % (urltype
, realhost
, rest
)
391 #print "proxy via https:", host, selector
392 if not host
: raise IOError, ('https error', 'no host given')
395 proxy_auth
= base64
.b64encode(proxy_passwd
).strip()
400 auth
= base64
.b64encode(user_passwd
).strip()
403 h
= httplib
.HTTPS(host
, 0,
404 key_file
=self
.key_file
,
405 cert_file
=self
.cert_file
)
407 h
.putrequest('POST', selector
)
408 h
.putheader('Content-Type',
409 'application/x-www-form-urlencoded')
410 h
.putheader('Content-Length', '%d' % len(data
))
412 h
.putrequest('GET', selector
)
413 if proxy_auth
: h
.putheader('Proxy-Authorization', 'Basic %s' % proxy_auth
)
414 if auth
: h
.putheader('Authorization', 'Basic %s' % auth
)
415 if realhost
: h
.putheader('Host', realhost
)
416 for args
in self
.addheaders
: h
.putheader(*args
)
420 errcode
, errmsg
, headers
= h
.getreply()
424 # something went wrong with the HTTP status line
425 raise IOError, ('http protocol error', 0,
426 'got a bad status line', None)
428 return addinfourl(fp
, headers
, "https:" + url
)
431 return self
.http_error(url
, fp
, errcode
, errmsg
, headers
)
433 return self
.http_error(url
, fp
, errcode
, errmsg
, headers
,
436 def open_gopher(self
, url
):
437 """Use Gopher protocol."""
438 if not isinstance(url
, str):
439 raise IOError, ('gopher error', 'proxy support for gopher protocol currently not implemented')
441 host
, selector
= splithost(url
)
442 if not host
: raise IOError, ('gopher error', 'no host given')
444 type, selector
= splitgophertype(selector
)
445 selector
, query
= splitquery(selector
)
446 selector
= unquote(selector
)
448 query
= unquote(query
)
449 fp
= gopherlib
.send_query(selector
, query
, host
)
451 fp
= gopherlib
.send_selector(selector
, host
)
452 return addinfourl(fp
, noheaders(), "gopher:" + url
)
454 def open_file(self
, url
):
455 """Use local file or FTP depending on form of URL."""
456 if not isinstance(url
, str):
457 raise IOError, ('file error', 'proxy support for file protocol currently not implemented')
458 if url
[:2] == '//' and url
[2:3] != '/' and url
[2:12].lower() != 'localhost/':
459 return self
.open_ftp(url
)
461 return self
.open_local_file(url
)
463 def open_local_file(self
, url
):
464 """Use local file."""
465 import mimetypes
, mimetools
, email
.utils
467 from cStringIO
import StringIO
469 from StringIO
import StringIO
470 host
, file = splithost(url
)
471 localname
= url2pathname(file)
473 stats
= os
.stat(localname
)
475 raise IOError(e
.errno
, e
.strerror
, e
.filename
)
477 modified
= email
.utils
.formatdate(stats
.st_mtime
, usegmt
=True)
478 mtype
= mimetypes
.guess_type(url
)[0]
479 headers
= mimetools
.Message(StringIO(
480 'Content-Type: %s\nContent-Length: %d\nLast-modified: %s\n' %
481 (mtype
or 'text/plain', size
, modified
)))
485 urlfile
= 'file://' + file
486 return addinfourl(open(localname
, 'rb'),
488 host
, port
= splitport(host
)
490 and socket
.gethostbyname(host
) in (localhost(), thishost()):
493 urlfile
= 'file://' + file
494 return addinfourl(open(localname
, 'rb'),
496 raise IOError, ('local file error', 'not on local host')
498 def open_ftp(self
, url
):
499 """Use FTP protocol."""
500 if not isinstance(url
, str):
501 raise IOError, ('ftp error', 'proxy support for ftp protocol currently not implemented')
502 import mimetypes
, mimetools
504 from cStringIO
import StringIO
506 from StringIO
import StringIO
507 host
, path
= splithost(url
)
508 if not host
: raise IOError, ('ftp error', 'no host given')
509 host
, port
= splitport(host
)
510 user
, host
= splituser(host
)
511 if user
: user
, passwd
= splitpasswd(user
)
514 user
= unquote(user
or '')
515 passwd
= unquote(passwd
or '')
516 host
= socket
.gethostbyname(host
)
519 port
= ftplib
.FTP_PORT
522 path
, attrs
= splitattr(path
)
524 dirs
= path
.split('/')
525 dirs
, file = dirs
[:-1], dirs
[-1]
526 if dirs
and not dirs
[0]: dirs
= dirs
[1:]
527 if dirs
and not dirs
[0]: dirs
[0] = '/'
528 key
= user
, host
, port
, '/'.join(dirs
)
530 if len(self
.ftpcache
) > MAXFTPCACHE
:
531 # Prune the cache, rather arbitrarily
532 for k
in self
.ftpcache
.keys():
538 if not key
in self
.ftpcache
:
539 self
.ftpcache
[key
] = \
540 ftpwrapper(user
, passwd
, host
, port
, dirs
)
541 if not file: type = 'D'
544 attr
, value
= splitvalue(attr
)
545 if attr
.lower() == 'type' and \
546 value
in ('a', 'A', 'i', 'I', 'd', 'D'):
548 (fp
, retrlen
) = self
.ftpcache
[key
].retrfile(file, type)
549 mtype
= mimetypes
.guess_type("ftp:" + url
)[0]
552 headers
+= "Content-Type: %s\n" % mtype
553 if retrlen
is not None and retrlen
>= 0:
554 headers
+= "Content-Length: %d\n" % retrlen
555 headers
= mimetools
.Message(StringIO(headers
))
556 return addinfourl(fp
, headers
, "ftp:" + url
)
557 except ftperrors(), msg
:
558 raise IOError, ('ftp error', msg
), sys
.exc_info()[2]
560 def open_data(self
, url
, data
=None):
561 """Use "data" URL."""
562 if not isinstance(url
, str):
563 raise IOError, ('data error', 'proxy support for data protocol currently not implemented')
566 # syntax of data URLs:
567 # dataurl := "data:" [ mediatype ] [ ";base64" ] "," data
568 # mediatype := [ type "/" subtype ] *( ";" parameter )
570 # parameter := attribute "=" value
573 from cStringIO
import StringIO
575 from StringIO
import StringIO
577 [type, data
] = url
.split(',', 1)
579 raise IOError, ('data error', 'bad data URL')
581 type = 'text/plain;charset=US-ASCII'
582 semi
= type.rfind(';')
583 if semi
>= 0 and '=' not in type[semi
:]:
584 encoding
= type[semi
+1:]
589 msg
.append('Date: %s'%time
.strftime('%a, %d %b %Y %T GMT',
590 time
.gmtime(time
.time())))
591 msg
.append('Content-type: %s' % type)
592 if encoding
== 'base64':
594 data
= base64
.decodestring(data
)
597 msg
.append('Content-Length: %d' % len(data
))
602 headers
= mimetools
.Message(f
, 0)
603 #f.fileno = None # needed for addinfourl
604 return addinfourl(f
, headers
, url
)
607 class FancyURLopener(URLopener
):
608 """Derived class with handlers for errors we can handle (perhaps)."""
610 def __init__(self
, *args
, **kwargs
):
611 URLopener
.__init
__(self
, *args
, **kwargs
)
616 def http_error_default(self
, url
, fp
, errcode
, errmsg
, headers
):
617 """Default error handling -- don't raise an exception."""
618 return addinfourl(fp
, headers
, "http:" + url
)
620 def http_error_302(self
, url
, fp
, errcode
, errmsg
, headers
, data
=None):
621 """Error 302 -- relocated (temporarily)."""
623 if self
.maxtries
and self
.tries
>= self
.maxtries
:
624 if hasattr(self
, "http_error_500"):
625 meth
= self
.http_error_500
627 meth
= self
.http_error_default
629 return meth(url
, fp
, 500,
630 "Internal Server Error: Redirect Recursion", headers
)
631 result
= self
.redirect_internal(url
, fp
, errcode
, errmsg
, headers
,
636 def redirect_internal(self
, url
, fp
, errcode
, errmsg
, headers
, data
):
637 if 'location' in headers
:
638 newurl
= headers
['location']
639 elif 'uri' in headers
:
640 newurl
= headers
['uri']
645 # In case the server sent a relative URL, join with original:
646 newurl
= basejoin(self
.type + ":" + url
, newurl
)
647 return self
.open(newurl
)
649 def http_error_301(self
, url
, fp
, errcode
, errmsg
, headers
, data
=None):
650 """Error 301 -- also relocated (permanently)."""
651 return self
.http_error_302(url
, fp
, errcode
, errmsg
, headers
, data
)
653 def http_error_303(self
, url
, fp
, errcode
, errmsg
, headers
, data
=None):
654 """Error 303 -- also relocated (essentially identical to 302)."""
655 return self
.http_error_302(url
, fp
, errcode
, errmsg
, headers
, data
)
657 def http_error_307(self
, url
, fp
, errcode
, errmsg
, headers
, data
=None):
658 """Error 307 -- relocated, but turn POST into error."""
660 return self
.http_error_302(url
, fp
, errcode
, errmsg
, headers
, data
)
662 return self
.http_error_default(url
, fp
, errcode
, errmsg
, headers
)
664 def http_error_401(self
, url
, fp
, errcode
, errmsg
, headers
, data
=None):
665 """Error 401 -- authentication required.
666 This function supports Basic authentication only."""
667 if not 'www-authenticate' in headers
:
668 URLopener
.http_error_default(self
, url
, fp
,
669 errcode
, errmsg
, headers
)
670 stuff
= headers
['www-authenticate']
672 match
= re
.match('[ \t]*([^ \t]+)[ \t]+realm="([^"]*)"', stuff
)
674 URLopener
.http_error_default(self
, url
, fp
,
675 errcode
, errmsg
, headers
)
676 scheme
, realm
= match
.groups()
677 if scheme
.lower() != 'basic':
678 URLopener
.http_error_default(self
, url
, fp
,
679 errcode
, errmsg
, headers
)
680 name
= 'retry_' + self
.type + '_basic_auth'
682 return getattr(self
,name
)(url
, realm
)
684 return getattr(self
,name
)(url
, realm
, data
)
686 def http_error_407(self
, url
, fp
, errcode
, errmsg
, headers
, data
=None):
687 """Error 407 -- proxy authentication required.
688 This function supports Basic authentication only."""
689 if not 'proxy-authenticate' in headers
:
690 URLopener
.http_error_default(self
, url
, fp
,
691 errcode
, errmsg
, headers
)
692 stuff
= headers
['proxy-authenticate']
694 match
= re
.match('[ \t]*([^ \t]+)[ \t]+realm="([^"]*)"', stuff
)
696 URLopener
.http_error_default(self
, url
, fp
,
697 errcode
, errmsg
, headers
)
698 scheme
, realm
= match
.groups()
699 if scheme
.lower() != 'basic':
700 URLopener
.http_error_default(self
, url
, fp
,
701 errcode
, errmsg
, headers
)
702 name
= 'retry_proxy_' + self
.type + '_basic_auth'
704 return getattr(self
,name
)(url
, realm
)
706 return getattr(self
,name
)(url
, realm
, data
)
708 def retry_proxy_http_basic_auth(self
, url
, realm
, data
=None):
709 host
, selector
= splithost(url
)
710 newurl
= 'http://' + host
+ selector
711 proxy
= self
.proxies
['http']
712 urltype
, proxyhost
= splittype(proxy
)
713 proxyhost
, proxyselector
= splithost(proxyhost
)
714 i
= proxyhost
.find('@') + 1
715 proxyhost
= proxyhost
[i
:]
716 user
, passwd
= self
.get_user_passwd(proxyhost
, realm
, i
)
717 if not (user
or passwd
): return None
718 proxyhost
= quote(user
, safe
='') + ':' + quote(passwd
, safe
='') + '@' + proxyhost
719 self
.proxies
['http'] = 'http://' + proxyhost
+ proxyselector
721 return self
.open(newurl
)
723 return self
.open(newurl
, data
)
725 def retry_proxy_https_basic_auth(self
, url
, realm
, data
=None):
726 host
, selector
= splithost(url
)
727 newurl
= 'https://' + host
+ selector
728 proxy
= self
.proxies
['https']
729 urltype
, proxyhost
= splittype(proxy
)
730 proxyhost
, proxyselector
= splithost(proxyhost
)
731 i
= proxyhost
.find('@') + 1
732 proxyhost
= proxyhost
[i
:]
733 user
, passwd
= self
.get_user_passwd(proxyhost
, realm
, i
)
734 if not (user
or passwd
): return None
735 proxyhost
= quote(user
, safe
='') + ':' + quote(passwd
, safe
='') + '@' + proxyhost
736 self
.proxies
['https'] = 'https://' + proxyhost
+ proxyselector
738 return self
.open(newurl
)
740 return self
.open(newurl
, data
)
742 def retry_http_basic_auth(self
, url
, realm
, data
=None):
743 host
, selector
= splithost(url
)
744 i
= host
.find('@') + 1
746 user
, passwd
= self
.get_user_passwd(host
, realm
, i
)
747 if not (user
or passwd
): return None
748 host
= quote(user
, safe
='') + ':' + quote(passwd
, safe
='') + '@' + host
749 newurl
= 'http://' + host
+ selector
751 return self
.open(newurl
)
753 return self
.open(newurl
, data
)
755 def retry_https_basic_auth(self
, url
, realm
, data
=None):
756 host
, selector
= splithost(url
)
757 i
= host
.find('@') + 1
759 user
, passwd
= self
.get_user_passwd(host
, realm
, i
)
760 if not (user
or passwd
): return None
761 host
= quote(user
, safe
='') + ':' + quote(passwd
, safe
='') + '@' + host
762 newurl
= 'https://' + host
+ selector
764 return self
.open(newurl
)
766 return self
.open(newurl
, data
)
768 def get_user_passwd(self
, host
, realm
, clear_cache
= 0):
769 key
= realm
+ '@' + host
.lower()
770 if key
in self
.auth_cache
:
772 del self
.auth_cache
[key
]
774 return self
.auth_cache
[key
]
775 user
, passwd
= self
.prompt_user_passwd(host
, realm
)
776 if user
or passwd
: self
.auth_cache
[key
] = (user
, passwd
)
779 def prompt_user_passwd(self
, host
, realm
):
780 """Override this in a GUI environment!"""
783 user
= raw_input("Enter username for %s at %s: " % (realm
,
785 passwd
= getpass
.getpass("Enter password for %s in %s at %s: " %
788 except KeyboardInterrupt:
797 """Return the IP address of the magic hostname 'localhost'."""
799 if _localhost
is None:
800 _localhost
= socket
.gethostbyname('localhost')
805 """Return the IP address of the current host."""
807 if _thishost
is None:
808 _thishost
= socket
.gethostbyname(socket
.gethostname())
813 """Return the set of errors raised by the FTP class."""
815 if _ftperrors
is None:
817 _ftperrors
= ftplib
.all_errors
822 """Return an empty mimetools.Message object."""
824 if _noheaders
is None:
827 from cStringIO
import StringIO
829 from StringIO
import StringIO
830 _noheaders
= mimetools
.Message(StringIO(), 0)
831 _noheaders
.fp
.close() # Recycle file descriptor
838 """Class used by open_ftp() for cache of open FTP connections."""
840 def __init__(self
, user
, passwd
, host
, port
, dirs
):
851 self
.ftp
= ftplib
.FTP()
852 self
.ftp
.connect(self
.host
, self
.port
)
853 self
.ftp
.login(self
.user
, self
.passwd
)
854 for dir in self
.dirs
:
857 def retrfile(self
, file, type):
860 if type in ('d', 'D'): cmd
= 'TYPE A'; isdir
= 1
861 else: cmd
= 'TYPE ' + type; isdir
= 0
863 self
.ftp
.voidcmd(cmd
)
864 except ftplib
.all_errors
:
866 self
.ftp
.voidcmd(cmd
)
868 if file and not isdir
:
869 # Try to retrieve as a file
872 conn
= self
.ftp
.ntransfercmd(cmd
)
873 except ftplib
.error_perm
, reason
:
874 if str(reason
)[:3] != '550':
875 raise IOError, ('ftp error', reason
), sys
.exc_info()[2]
877 # Set transfer mode to ASCII!
878 self
.ftp
.voidcmd('TYPE A')
879 # Try a directory listing
880 if file: cmd
= 'LIST ' + file
882 conn
= self
.ftp
.ntransfercmd(cmd
)
884 # Pass back both a suitably decorated object and a retrieval length
885 return (addclosehook(conn
[0].makefile('rb'),
886 self
.endtransfer
), conn
[1])
887 def endtransfer(self
):
904 """Base class for addinfo and addclosehook."""
906 def __init__(self
, fp
):
908 self
.read
= self
.fp
.read
909 self
.readline
= self
.fp
.readline
910 if hasattr(self
.fp
, "readlines"): self
.readlines
= self
.fp
.readlines
911 if hasattr(self
.fp
, "fileno"):
912 self
.fileno
= self
.fp
.fileno
914 self
.fileno
= lambda: None
915 if hasattr(self
.fp
, "__iter__"):
916 self
.__iter
__ = self
.fp
.__iter
__
917 if hasattr(self
.fp
, "next"):
918 self
.next
= self
.fp
.next
921 return '<%s at %r whose fp = %r>' % (self
.__class
__.__name
__,
927 self
.readlines
= None
929 if self
.fp
: self
.fp
.close()
932 class addclosehook(addbase
):
933 """Class to add a close hook to an open file."""
935 def __init__(self
, fp
, closehook
, *hookargs
):
936 addbase
.__init
__(self
, fp
)
937 self
.closehook
= closehook
938 self
.hookargs
= hookargs
943 self
.closehook(*self
.hookargs
)
944 self
.closehook
= None
947 class addinfo(addbase
):
948 """class to add an info() method to an open file."""
950 def __init__(self
, fp
, headers
):
951 addbase
.__init
__(self
, fp
)
952 self
.headers
= headers
957 class addinfourl(addbase
):
958 """class to add info() and geturl() methods to an open file."""
960 def __init__(self
, fp
, headers
, url
):
961 addbase
.__init
__(self
, fp
)
962 self
.headers
= headers
972 # Utilities to parse URLs (most of these return None for missing parts):
973 # unwrap('<URL:type://host/path>') --> 'type://host/path'
974 # splittype('type:opaquestring') --> 'type', 'opaquestring'
975 # splithost('//host[:port]/path') --> 'host[:port]', '/path'
976 # splituser('user[:passwd]@host[:port]') --> 'user[:passwd]', 'host[:port]'
977 # splitpasswd('user:passwd') -> 'user', 'passwd'
978 # splitport('host:port') --> 'host', 'port'
979 # splitquery('/path?query') --> '/path', 'query'
980 # splittag('/path#tag') --> '/path', 'tag'
981 # splitattr('/path;attr1=value1;attr2=value2;...') ->
982 # '/path', ['attr1=value1', 'attr2=value2', ...]
983 # splitvalue('attr=value') --> 'attr', 'value'
984 # splitgophertype('/Xselector') --> 'X', 'selector'
985 # unquote('abc%20def') -> 'abc def'
986 # quote('abc def') -> 'abc%20def')
995 return isinstance(x
, unicode)
998 """toBytes(u"URL") --> 'URL'."""
999 # Most URL schemes require ASCII. If that changes, the conversion
1001 if _is_unicode(url
):
1003 url
= url
.encode("ASCII")
1004 except UnicodeError:
1005 raise UnicodeError("URL " + repr(url
) +
1006 " contains non-ASCII characters")
1010 """unwrap('<URL:type://host/path>') --> 'type://host/path'."""
1012 if url
[:1] == '<' and url
[-1:] == '>':
1013 url
= url
[1:-1].strip()
1014 if url
[:4] == 'URL:': url
= url
[4:].strip()
1019 """splittype('type:opaquestring') --> 'type', 'opaquestring'."""
1021 if _typeprog
is None:
1023 _typeprog
= re
.compile('^([^/:]+):')
1025 match
= _typeprog
.match(url
)
1027 scheme
= match
.group(1)
1028 return scheme
.lower(), url
[len(scheme
) + 1:]
1033 """splithost('//host[:port]/path') --> 'host[:port]', '/path'."""
1035 if _hostprog
is None:
1037 _hostprog
= re
.compile('^//([^/?]*)(.*)$')
1039 match
= _hostprog
.match(url
)
1040 if match
: return match
.group(1, 2)
1044 def splituser(host
):
1045 """splituser('user[:passwd]@host[:port]') --> 'user[:passwd]', 'host[:port]'."""
1047 if _userprog
is None:
1049 _userprog
= re
.compile('^(.*)@(.*)$')
1051 match
= _userprog
.match(host
)
1052 if match
: return map(unquote
, match
.group(1, 2))
1056 def splitpasswd(user
):
1057 """splitpasswd('user:passwd') -> 'user', 'passwd'."""
1059 if _passwdprog
is None:
1061 _passwdprog
= re
.compile('^([^:]*):(.*)$')
1063 match
= _passwdprog
.match(user
)
1064 if match
: return match
.group(1, 2)
1067 # splittag('/path#tag') --> '/path', 'tag'
1069 def splitport(host
):
1070 """splitport('host:port') --> 'host', 'port'."""
1072 if _portprog
is None:
1074 _portprog
= re
.compile('^(.*):([0-9]+)$')
1076 match
= _portprog
.match(host
)
1077 if match
: return match
.group(1, 2)
1081 def splitnport(host
, defport
=-1):
1082 """Split host and port, returning numeric port.
1083 Return given default port if no ':' found; defaults to -1.
1084 Return numerical port if a valid number are found after ':'.
1085 Return None if ':' but not a valid number."""
1087 if _nportprog
is None:
1089 _nportprog
= re
.compile('^(.*):(.*)$')
1091 match
= _nportprog
.match(host
)
1093 host
, port
= match
.group(1, 2)
1095 if not port
: raise ValueError, "no digits"
1100 return host
, defport
1103 def splitquery(url
):
1104 """splitquery('/path?query') --> '/path', 'query'."""
1106 if _queryprog
is None:
1108 _queryprog
= re
.compile('^(.*)\?([^?]*)$')
1110 match
= _queryprog
.match(url
)
1111 if match
: return match
.group(1, 2)
1116 """splittag('/path#tag') --> '/path', 'tag'."""
1118 if _tagprog
is None:
1120 _tagprog
= re
.compile('^(.*)#([^#]*)$')
1122 match
= _tagprog
.match(url
)
1123 if match
: return match
.group(1, 2)
1127 """splitattr('/path;attr1=value1;attr2=value2;...') ->
1128 '/path', ['attr1=value1', 'attr2=value2', ...]."""
1129 words
= url
.split(';')
1130 return words
[0], words
[1:]
1133 def splitvalue(attr
):
1134 """splitvalue('attr=value') --> 'attr', 'value'."""
1136 if _valueprog
is None:
1138 _valueprog
= re
.compile('^([^=]*)=(.*)$')
1140 match
= _valueprog
.match(attr
)
1141 if match
: return match
.group(1, 2)
1144 def splitgophertype(selector
):
1145 """splitgophertype('/Xselector') --> 'X', 'selector'."""
1146 if selector
[:1] == '/' and selector
[1:2]:
1147 return selector
[1], selector
[2:]
1148 return None, selector
1150 _hextochr
= dict(('%02x' % i
, chr(i
)) for i
in range(256))
1151 _hextochr
.update(('%02X' % i
, chr(i
)) for i
in range(256))
1154 """unquote('abc%20def') -> 'abc def'."""
1156 for i
in xrange(1, len(res
)):
1159 res
[i
] = _hextochr
[item
[:2]] + item
[2:]
1162 except UnicodeDecodeError:
1163 res
[i
] = unichr(int(item
[:2], 16)) + item
[2:]
1166 def unquote_plus(s
):
1167 """unquote('%7e/abc+def') -> '~/abc def'"""
1168 s
= s
.replace('+', ' ')
1171 always_safe
= ('ABCDEFGHIJKLMNOPQRSTUVWXYZ'
1172 'abcdefghijklmnopqrstuvwxyz'
1176 def quote(s
, safe
= '/'):
1177 """quote('abc def') -> 'abc%20def'
1179 Each part of a URL, e.g. the path info, the query, etc., has a
1180 different set of reserved characters that must be quoted.
1182 RFC 2396 Uniform Resource Identifiers (URI): Generic Syntax lists
1183 the following reserved characters.
1185 reserved = ";" | "/" | "?" | ":" | "@" | "&" | "=" | "+" |
1188 Each of these characters is reserved in some component of a URL,
1189 but not necessarily in all of them.
1191 By default, the quote function is intended for quoting the path
1192 section of a URL. Thus, it will not encode '/'. This character
1193 is reserved, but in typical usage the quote function is being
1194 called on a path where the existing slash characters are used as
1195 reserved characters.
1197 cachekey
= (safe
, always_safe
)
1199 safe_map
= _safemaps
[cachekey
]
1203 for i
in range(256):
1205 safe_map
[c
] = (c
in safe
) and c
or ('%%%02X' % i
)
1206 _safemaps
[cachekey
] = safe_map
1207 res
= map(safe_map
.__getitem
__, s
)
1210 def quote_plus(s
, safe
= ''):
1211 """Quote the query fragment of a URL; replacing ' ' with '+'"""
1213 s
= quote(s
, safe
+ ' ')
1214 return s
.replace(' ', '+')
1215 return quote(s
, safe
)
1217 def urlencode(query
,doseq
=0):
1218 """Encode a sequence of two-element tuples or dictionary into a URL query string.
1220 If any values in the query arg are sequences and doseq is true, each
1221 sequence element is converted to a separate parameter.
1223 If the query arg is a sequence of two-element tuples, the order of the
1224 parameters in the output will match the order of parameters in the
1228 if hasattr(query
,"items"):
1230 query
= query
.items()
1232 # it's a bother at times that strings and string-like objects are
1235 # non-sequence items should not work with len()
1236 # non-empty strings will fail this
1237 if len(query
) and not isinstance(query
[0], tuple):
1239 # zero-length sequences of all types will get here and succeed,
1240 # but that's a minor nit - since the original implementation
1241 # allowed empty dicts that type of behavior probably should be
1242 # preserved for consistency
1244 ty
,va
,tb
= sys
.exc_info()
1245 raise TypeError, "not a valid non-string sequence or mapping object", tb
1249 # preserve old behavior
1251 k
= quote_plus(str(k
))
1252 v
= quote_plus(str(v
))
1253 l
.append(k
+ '=' + v
)
1256 k
= quote_plus(str(k
))
1257 if isinstance(v
, str):
1259 l
.append(k
+ '=' + v
)
1260 elif _is_unicode(v
):
1261 # is there a reasonable way to convert to ASCII?
1262 # encode generates a string, but "replace" or "ignore"
1263 # lose information and "strict" can raise UnicodeError
1264 v
= quote_plus(v
.encode("ASCII","replace"))
1265 l
.append(k
+ '=' + v
)
1268 # is this a sufficient test for sequence-ness?
1272 v
= quote_plus(str(v
))
1273 l
.append(k
+ '=' + v
)
1275 # loop over the sequence
1277 l
.append(k
+ '=' + quote_plus(str(elt
)))
1281 def getproxies_environment():
1282 """Return a dictionary of scheme -> proxy server URL mappings.
1284 Scan the environment for variables named <scheme>_proxy;
1285 this seems to be the standard convention. If you need a
1286 different way, you can pass a proxies dictionary to the
1287 [Fancy]URLopener constructor.
1291 for name
, value
in os
.environ
.items():
1293 if value
and name
[-6:] == '_proxy':
1294 proxies
[name
[:-6]] = value
1297 if sys
.platform
== 'darwin':
1298 def getproxies_internetconfig():
1299 """Return a dictionary of scheme -> proxy server URL mappings.
1301 By convention the mac uses Internet Config to store
1302 proxies. An HTTP proxy, for instance, is stored under
1317 if 'UseHTTPProxy' in config
and config
['UseHTTPProxy']:
1319 value
= config
['HTTPProxyHost']
1323 proxies
['http'] = 'http://%s' % value
1324 # FTP: XXXX To be done.
1325 # Gopher: XXXX To be done.
1328 def proxy_bypass(x
):
1332 return getproxies_environment() or getproxies_internetconfig()
1334 elif os
.name
== 'nt':
1335 def getproxies_registry():
1336 """Return a dictionary of scheme -> proxy server URL mappings.
1338 Win32 uses the registry to store proxies.
1345 # Std module, so should be around - but you never know!
1348 internetSettings
= _winreg
.OpenKey(_winreg
.HKEY_CURRENT_USER
,
1349 r
'Software\Microsoft\Windows\CurrentVersion\Internet Settings')
1350 proxyEnable
= _winreg
.QueryValueEx(internetSettings
,
1353 # Returned as Unicode but problems if not converted to ASCII
1354 proxyServer
= str(_winreg
.QueryValueEx(internetSettings
,
1356 if '=' in proxyServer
:
1357 # Per-protocol settings
1358 for p
in proxyServer
.split(';'):
1359 protocol
, address
= p
.split('=', 1)
1360 # See if address has a type:// prefix
1362 if not re
.match('^([^/:]+)://', address
):
1363 address
= '%s://%s' % (protocol
, address
)
1364 proxies
[protocol
] = address
1366 # Use one setting for all protocols
1367 if proxyServer
[:5] == 'http:':
1368 proxies
['http'] = proxyServer
1370 proxies
['http'] = 'http://%s' % proxyServer
1371 proxies
['ftp'] = 'ftp://%s' % proxyServer
1372 internetSettings
.Close()
1373 except (WindowsError, ValueError, TypeError):
1374 # Either registry key not found etc, or the value in an
1375 # unexpected format.
1376 # proxies already set up to be empty so nothing to do
1381 """Return a dictionary of scheme -> proxy server URL mappings.
1383 Returns settings gathered from the environment, if specified,
1387 return getproxies_environment() or getproxies_registry()
1389 def proxy_bypass(host
):
1394 # Std modules, so should be around - but you never know!
1397 internetSettings
= _winreg
.OpenKey(_winreg
.HKEY_CURRENT_USER
,
1398 r
'Software\Microsoft\Windows\CurrentVersion\Internet Settings')
1399 proxyEnable
= _winreg
.QueryValueEx(internetSettings
,
1401 proxyOverride
= str(_winreg
.QueryValueEx(internetSettings
,
1402 'ProxyOverride')[0])
1403 # ^^^^ Returned as Unicode but problems if not converted to ASCII
1404 except WindowsError:
1406 if not proxyEnable
or not proxyOverride
:
1408 # try to make a host list from name and IP address.
1409 rawHost
, port
= splitport(host
)
1412 addr
= socket
.gethostbyname(rawHost
)
1415 except socket
.error
:
1418 fqdn
= socket
.getfqdn(rawHost
)
1421 except socket
.error
:
1423 # make a check value list from the registry entry: replace the
1424 # '<local>' string by the localhost entry and the corresponding
1426 proxyOverride
= proxyOverride
.split(';')
1428 while i
< len(proxyOverride
):
1429 if proxyOverride
[i
] == '<local>':
1430 proxyOverride
[i
:i
+1] = ['localhost',
1432 socket
.gethostname(),
1433 socket
.gethostbyname(
1434 socket
.gethostname())]
1436 # print proxyOverride
1437 # now check if we match one of the registry values.
1438 for test
in proxyOverride
:
1439 test
= test
.replace(".", r
"\.") # mask dots
1440 test
= test
.replace("*", r
".*") # change glob sequence
1441 test
= test
.replace("?", r
".") # change glob char
1443 # print "%s <--> %s" %( test, val )
1444 if re
.match(test
, val
, re
.I
):
1449 # By default use environment variables
1450 getproxies
= getproxies_environment
1452 def proxy_bypass(host
):
1455 # Test and time quote() and unquote()
1458 for i
in range(256): s
= s
+ chr(i
)
1469 print round(t1
- t0
, 3), 'sec'
1472 def reporthook(blocknum
, blocksize
, totalsize
):
1473 # Report during remote transfers
1474 print "Block number: %d, Block size: %d, Total size: %d" % (
1475 blocknum
, blocksize
, totalsize
)
1483 'file://localhost/etc/passwd',
1484 'ftp://ftp.gnu.org/pub/README',
1485 ## 'gopher://gopher.micro.umn.edu/1/',
1486 'http://www.python.org/index.html',
1488 if hasattr(URLopener
, "open_https"):
1489 args
.append('https://synergy.as.cmu.edu/~geek/')
1492 print '-'*10, url
, '-'*10
1493 fn
, h
= urlretrieve(url
, None, reporthook
)
1497 for k
in h
.keys(): print k
+ ':', h
[k
]
1503 table
= string
.maketrans("", "")
1504 data
= data
.translate(table
, "\r")
1514 opts
, args
= getopt
.getopt(sys
.argv
[1:], "th")
1515 except getopt
.error
, msg
:
1517 print "Use -h for help"
1524 print "Usage: python urllib.py [-t] [url ...]"
1525 print "-t runs self-test;",
1526 print "otherwise, contents of urls are printed"
1534 print "Use -h for help"
1536 print urlopen(url
).read(),
1538 # Run test program when run as a script
1539 if __name__
== '__main__':