mirror of
				https://github.com/python/cpython.git
				synced 2025-10-31 13:41:24 +00:00 
			
		
		
		
	 4ea9c5ba78
			
		
	
	
		4ea9c5ba78
		
			
		
	
	
	
	
		
			
			gh-85110: Preserve relative path in URL without netloc in urllib.parse.urlunsplit() (GH-123179)
(cherry picked from commit 90c892efea)
Co-authored-by: Serhiy Storchaka <storchaka@gmail.com>
		
	
			
		
			
				
	
	
		
			1618 lines
		
	
	
	
		
			78 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			1618 lines
		
	
	
	
		
			78 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| import sys
 | |
| import unicodedata
 | |
| import unittest
 | |
| import urllib.parse
 | |
| 
 | |
| RFC1808_BASE = "http://a/b/c/d;p?q#f"
 | |
| RFC2396_BASE = "http://a/b/c/d;p?q"
 | |
| RFC3986_BASE = 'http://a/b/c/d;p?q'
 | |
| SIMPLE_BASE  = 'http://a/b/c/d'
 | |
| 
 | |
| # Each parse_qsl testcase is a two-tuple that contains
 | |
| # a string with the query and a list with the expected result.
 | |
| 
 | |
| parse_qsl_test_cases = [
 | |
|     ("", []),
 | |
|     ("&", []),
 | |
|     ("&&", []),
 | |
|     ("=", [('', '')]),
 | |
|     ("=a", [('', 'a')]),
 | |
|     ("a", [('a', '')]),
 | |
|     ("a=", [('a', '')]),
 | |
|     ("a=b=c", [('a', 'b=c')]),
 | |
|     ("a%3Db=c", [('a=b', 'c')]),
 | |
|     ("a=b&c=d", [('a', 'b'), ('c', 'd')]),
 | |
|     ("a=b%26c=d", [('a', 'b&c=d')]),
 | |
|     ("&a=b", [('a', 'b')]),
 | |
|     ("a=a+b&b=b+c", [('a', 'a b'), ('b', 'b c')]),
 | |
|     ("a=1&a=2", [('a', '1'), ('a', '2')]),
 | |
|     (b"", []),
 | |
|     (b"&", []),
 | |
|     (b"&&", []),
 | |
|     (b"=", [(b'', b'')]),
 | |
|     (b"=a", [(b'', b'a')]),
 | |
|     (b"a", [(b'a', b'')]),
 | |
|     (b"a=", [(b'a', b'')]),
 | |
|     (b"a=b=c", [(b'a', b'b=c')]),
 | |
|     (b"a%3Db=c", [(b'a=b', b'c')]),
 | |
|     (b"a=b&c=d", [(b'a', b'b'), (b'c', b'd')]),
 | |
|     (b"a=b%26c=d", [(b'a', b'b&c=d')]),
 | |
|     (b"&a=b", [(b'a', b'b')]),
 | |
|     (b"a=a+b&b=b+c", [(b'a', b'a b'), (b'b', b'b c')]),
 | |
|     (b"a=1&a=2", [(b'a', b'1'), (b'a', b'2')]),
 | |
|     (";a=b", [(';a', 'b')]),
 | |
|     ("a=a+b;b=b+c", [('a', 'a b;b=b c')]),
 | |
|     (b";a=b", [(b';a', b'b')]),
 | |
|     (b"a=a+b;b=b+c", [(b'a', b'a b;b=b c')]),
 | |
| 
 | |
|     ("\u0141=\xE9", [('\u0141', '\xE9')]),
 | |
|     ("%C5%81=%C3%A9", [('\u0141', '\xE9')]),
 | |
|     ("%81=%A9", [('\ufffd', '\ufffd')]),
 | |
|     (b"\xc5\x81=\xc3\xa9", [(b'\xc5\x81', b'\xc3\xa9')]),
 | |
|     (b"%C5%81=%C3%A9", [(b'\xc5\x81', b'\xc3\xa9')]),
 | |
|     (b"\x81=\xA9", [(b'\x81', b'\xa9')]),
 | |
|     (b"%81=%A9", [(b'\x81', b'\xa9')]),
 | |
| ]
 | |
| 
 | |
| # Each parse_qs testcase is a two-tuple that contains
 | |
| # a string with the query and a dictionary with the expected result.
 | |
| 
 | |
| parse_qs_test_cases = [
 | |
|     ("", {}),
 | |
|     ("&", {}),
 | |
|     ("&&", {}),
 | |
|     ("=", {'': ['']}),
 | |
|     ("=a", {'': ['a']}),
 | |
|     ("a", {'a': ['']}),
 | |
|     ("a=", {'a': ['']}),
 | |
|     ("a=b=c", {'a': ['b=c']}),
 | |
|     ("a%3Db=c", {'a=b': ['c']}),
 | |
|     ("a=b&c=d", {'a': ['b'], 'c': ['d']}),
 | |
|     ("a=b%26c=d", {'a': ['b&c=d']}),
 | |
|     ("&a=b", {'a': ['b']}),
 | |
|     ("a=a+b&b=b+c", {'a': ['a b'], 'b': ['b c']}),
 | |
|     ("a=1&a=2", {'a': ['1', '2']}),
 | |
|     (b"", {}),
 | |
|     (b"&", {}),
 | |
|     (b"&&", {}),
 | |
|     (b"=", {b'': [b'']}),
 | |
|     (b"=a", {b'': [b'a']}),
 | |
|     (b"a", {b'a': [b'']}),
 | |
|     (b"a=", {b'a': [b'']}),
 | |
|     (b"a=b=c", {b'a': [b'b=c']}),
 | |
|     (b"a%3Db=c", {b'a=b': [b'c']}),
 | |
|     (b"a=b&c=d", {b'a': [b'b'], b'c': [b'd']}),
 | |
|     (b"a=b%26c=d", {b'a': [b'b&c=d']}),
 | |
|     (b"&a=b", {b'a': [b'b']}),
 | |
|     (b"a=a+b&b=b+c", {b'a': [b'a b'], b'b': [b'b c']}),
 | |
|     (b"a=1&a=2", {b'a': [b'1', b'2']}),
 | |
|     (";a=b", {';a': ['b']}),
 | |
|     ("a=a+b;b=b+c", {'a': ['a b;b=b c']}),
 | |
|     (b";a=b", {b';a': [b'b']}),
 | |
|     (b"a=a+b;b=b+c", {b'a':[ b'a b;b=b c']}),
 | |
|     (b"a=a%E2%80%99b", {b'a': [b'a\xe2\x80\x99b']}),
 | |
| 
 | |
|     ("\u0141=\xE9", {'\u0141': ['\xE9']}),
 | |
|     ("%C5%81=%C3%A9", {'\u0141': ['\xE9']}),
 | |
|     ("%81=%A9", {'\ufffd': ['\ufffd']}),
 | |
|     (b"\xc5\x81=\xc3\xa9", {b'\xc5\x81': [b'\xc3\xa9']}),
 | |
|     (b"%C5%81=%C3%A9", {b'\xc5\x81': [b'\xc3\xa9']}),
 | |
|     (b"\x81=\xA9", {b'\x81': [b'\xa9']}),
 | |
|     (b"%81=%A9", {b'\x81': [b'\xa9']}),
 | |
| ]
 | |
| 
 | |
| class UrlParseTestCase(unittest.TestCase):
 | |
| 
 | |
|     def checkRoundtrips(self, url, parsed, split, url2=None):
 | |
|         if url2 is None:
 | |
|             url2 = url
 | |
|         result = urllib.parse.urlparse(url)
 | |
|         self.assertSequenceEqual(result, parsed)
 | |
|         t = (result.scheme, result.netloc, result.path,
 | |
|              result.params, result.query, result.fragment)
 | |
|         self.assertSequenceEqual(t, parsed)
 | |
|         # put it back together and it should be the same
 | |
|         result2 = urllib.parse.urlunparse(result)
 | |
|         self.assertSequenceEqual(result2, url2)
 | |
|         self.assertSequenceEqual(result2, result.geturl())
 | |
| 
 | |
|         # the result of geturl() is a fixpoint; we can always parse it
 | |
|         # again to get the same result:
 | |
|         result3 = urllib.parse.urlparse(result.geturl())
 | |
|         self.assertEqual(result3.geturl(), result.geturl())
 | |
|         self.assertSequenceEqual(result3, result)
 | |
|         self.assertEqual(result3.scheme,   result.scheme)
 | |
|         self.assertEqual(result3.netloc,   result.netloc)
 | |
|         self.assertEqual(result3.path,     result.path)
 | |
|         self.assertEqual(result3.params,   result.params)
 | |
|         self.assertEqual(result3.query,    result.query)
 | |
|         self.assertEqual(result3.fragment, result.fragment)
 | |
|         self.assertEqual(result3.username, result.username)
 | |
|         self.assertEqual(result3.password, result.password)
 | |
|         self.assertEqual(result3.hostname, result.hostname)
 | |
|         self.assertEqual(result3.port,     result.port)
 | |
| 
 | |
|         # check the roundtrip using urlsplit() as well
 | |
|         result = urllib.parse.urlsplit(url)
 | |
|         self.assertSequenceEqual(result, split)
 | |
|         t = (result.scheme, result.netloc, result.path,
 | |
|              result.query, result.fragment)
 | |
|         self.assertSequenceEqual(t, split)
 | |
|         result2 = urllib.parse.urlunsplit(result)
 | |
|         self.assertSequenceEqual(result2, url2)
 | |
|         self.assertSequenceEqual(result2, result.geturl())
 | |
| 
 | |
|         # check the fixpoint property of re-parsing the result of geturl()
 | |
|         result3 = urllib.parse.urlsplit(result.geturl())
 | |
|         self.assertEqual(result3.geturl(), result.geturl())
 | |
|         self.assertSequenceEqual(result3, result)
 | |
|         self.assertEqual(result3.scheme,   result.scheme)
 | |
|         self.assertEqual(result3.netloc,   result.netloc)
 | |
|         self.assertEqual(result3.path,     result.path)
 | |
|         self.assertEqual(result3.query,    result.query)
 | |
|         self.assertEqual(result3.fragment, result.fragment)
 | |
|         self.assertEqual(result3.username, result.username)
 | |
|         self.assertEqual(result3.password, result.password)
 | |
|         self.assertEqual(result3.hostname, result.hostname)
 | |
|         self.assertEqual(result3.port,     result.port)
 | |
| 
 | |
|     def test_qsl(self):
 | |
|         for orig, expect in parse_qsl_test_cases:
 | |
|             result = urllib.parse.parse_qsl(orig, keep_blank_values=True)
 | |
|             self.assertEqual(result, expect, "Error parsing %r" % orig)
 | |
|             expect_without_blanks = [v for v in expect if len(v[1])]
 | |
|             result = urllib.parse.parse_qsl(orig, keep_blank_values=False)
 | |
|             self.assertEqual(result, expect_without_blanks,
 | |
|                             "Error parsing %r" % orig)
 | |
| 
 | |
|     def test_qs(self):
 | |
|         for orig, expect in parse_qs_test_cases:
 | |
|             result = urllib.parse.parse_qs(orig, keep_blank_values=True)
 | |
|             self.assertEqual(result, expect, "Error parsing %r" % orig)
 | |
|             expect_without_blanks = {v: expect[v]
 | |
|                                      for v in expect if len(expect[v][0])}
 | |
|             result = urllib.parse.parse_qs(orig, keep_blank_values=False)
 | |
|             self.assertEqual(result, expect_without_blanks,
 | |
|                             "Error parsing %r" % orig)
 | |
| 
 | |
|     def test_roundtrips(self):
 | |
|         str_cases = [
 | |
|             ('path/to/file',
 | |
|              ('', '', 'path/to/file', '', '', ''),
 | |
|              ('', '', 'path/to/file', '', '')),
 | |
|             ('/path/to/file',
 | |
|              ('', '', '/path/to/file', '', '', ''),
 | |
|              ('', '', '/path/to/file', '', '')),
 | |
|             ('//path/to/file',
 | |
|              ('', 'path', '/to/file', '', '', ''),
 | |
|              ('', 'path', '/to/file', '', '')),
 | |
|             ('////path/to/file',
 | |
|              ('', '', '//path/to/file', '', '', ''),
 | |
|              ('', '', '//path/to/file', '', '')),
 | |
|             ('/////path/to/file',
 | |
|              ('', '', '///path/to/file', '', '', ''),
 | |
|              ('', '', '///path/to/file', '', '')),
 | |
|             ('scheme:path/to/file',
 | |
|              ('scheme', '', 'path/to/file', '', '', ''),
 | |
|              ('scheme', '', 'path/to/file', '', '')),
 | |
|             ('scheme:/path/to/file',
 | |
|              ('scheme', '', '/path/to/file', '', '', ''),
 | |
|              ('scheme', '', '/path/to/file', '', '')),
 | |
|             ('scheme://path/to/file',
 | |
|              ('scheme', 'path', '/to/file', '', '', ''),
 | |
|              ('scheme', 'path', '/to/file', '', '')),
 | |
|             ('scheme:////path/to/file',
 | |
|              ('scheme', '', '//path/to/file', '', '', ''),
 | |
|              ('scheme', '', '//path/to/file', '', '')),
 | |
|             ('scheme://///path/to/file',
 | |
|              ('scheme', '', '///path/to/file', '', '', ''),
 | |
|              ('scheme', '', '///path/to/file', '', '')),
 | |
|             ('file:tmp/junk.txt',
 | |
|              ('file', '', 'tmp/junk.txt', '', '', ''),
 | |
|              ('file', '', 'tmp/junk.txt', '', '')),
 | |
|             ('file:///tmp/junk.txt',
 | |
|              ('file', '', '/tmp/junk.txt', '', '', ''),
 | |
|              ('file', '', '/tmp/junk.txt', '', '')),
 | |
|             ('file:////tmp/junk.txt',
 | |
|              ('file', '', '//tmp/junk.txt', '', '', ''),
 | |
|              ('file', '', '//tmp/junk.txt', '', '')),
 | |
|             ('file://///tmp/junk.txt',
 | |
|              ('file', '', '///tmp/junk.txt', '', '', ''),
 | |
|              ('file', '', '///tmp/junk.txt', '', '')),
 | |
|             ('http:tmp/junk.txt',
 | |
|              ('http', '', 'tmp/junk.txt', '', '', ''),
 | |
|              ('http', '', 'tmp/junk.txt', '', '')),
 | |
|             ('http://example.com/tmp/junk.txt',
 | |
|              ('http', 'example.com', '/tmp/junk.txt', '', '', ''),
 | |
|              ('http', 'example.com', '/tmp/junk.txt', '', '')),
 | |
|             ('http:///example.com/tmp/junk.txt',
 | |
|              ('http', '', '/example.com/tmp/junk.txt', '', '', ''),
 | |
|              ('http', '', '/example.com/tmp/junk.txt', '', '')),
 | |
|             ('http:////example.com/tmp/junk.txt',
 | |
|              ('http', '', '//example.com/tmp/junk.txt', '', '', ''),
 | |
|              ('http', '', '//example.com/tmp/junk.txt', '', '')),
 | |
|             ('imap://mail.python.org/mbox1',
 | |
|              ('imap', 'mail.python.org', '/mbox1', '', '', ''),
 | |
|              ('imap', 'mail.python.org', '/mbox1', '', '')),
 | |
|             ('mms://wms.sys.hinet.net/cts/Drama/09006251100.asf',
 | |
|              ('mms', 'wms.sys.hinet.net', '/cts/Drama/09006251100.asf',
 | |
|               '', '', ''),
 | |
|              ('mms', 'wms.sys.hinet.net', '/cts/Drama/09006251100.asf',
 | |
|               '', '')),
 | |
|             ('nfs://server/path/to/file.txt',
 | |
|              ('nfs', 'server', '/path/to/file.txt', '', '', ''),
 | |
|              ('nfs', 'server', '/path/to/file.txt', '', '')),
 | |
|             ('svn+ssh://svn.zope.org/repos/main/ZConfig/trunk/',
 | |
|              ('svn+ssh', 'svn.zope.org', '/repos/main/ZConfig/trunk/',
 | |
|               '', '', ''),
 | |
|              ('svn+ssh', 'svn.zope.org', '/repos/main/ZConfig/trunk/',
 | |
|               '', '')),
 | |
|             ('git+ssh://git@github.com/user/project.git',
 | |
|              ('git+ssh', 'git@github.com','/user/project.git',
 | |
|               '','',''),
 | |
|              ('git+ssh', 'git@github.com','/user/project.git',
 | |
|               '', '')),
 | |
|             ('itms-services://?action=download-manifest&url=https://example.com/app',
 | |
|              ('itms-services', '', '', '',
 | |
|               'action=download-manifest&url=https://example.com/app', ''),
 | |
|              ('itms-services', '', '',
 | |
|               'action=download-manifest&url=https://example.com/app', '')),
 | |
|             ('+scheme:path/to/file',
 | |
|              ('', '', '+scheme:path/to/file', '', '', ''),
 | |
|              ('', '', '+scheme:path/to/file', '', '')),
 | |
|             ('sch_me:path/to/file',
 | |
|              ('', '', 'sch_me:path/to/file', '', '', ''),
 | |
|              ('', '', 'sch_me:path/to/file', '', '')),
 | |
|             ]
 | |
|         def _encode(t):
 | |
|             return (t[0].encode('ascii'),
 | |
|                     tuple(x.encode('ascii') for x in t[1]),
 | |
|                     tuple(x.encode('ascii') for x in t[2]))
 | |
|         bytes_cases = [_encode(x) for x in str_cases]
 | |
|         str_cases += [
 | |
|             ('schème:path/to/file',
 | |
|              ('', '', 'schème:path/to/file', '', '', ''),
 | |
|              ('', '', 'schème:path/to/file', '', '')),
 | |
|             ]
 | |
|         for url, parsed, split in str_cases + bytes_cases:
 | |
|             with self.subTest(url):
 | |
|                 self.checkRoundtrips(url, parsed, split)
 | |
| 
 | |
|     def test_roundtrips_normalization(self):
 | |
|         str_cases = [
 | |
|             ('///path/to/file',
 | |
|              '/path/to/file',
 | |
|              ('', '', '/path/to/file', '', '', ''),
 | |
|              ('', '', '/path/to/file', '', '')),
 | |
|             ('scheme:///path/to/file',
 | |
|              'scheme:/path/to/file',
 | |
|              ('scheme', '', '/path/to/file', '', '', ''),
 | |
|              ('scheme', '', '/path/to/file', '', '')),
 | |
|             ('file:/tmp/junk.txt',
 | |
|              'file:///tmp/junk.txt',
 | |
|              ('file', '', '/tmp/junk.txt', '', '', ''),
 | |
|              ('file', '', '/tmp/junk.txt', '', '')),
 | |
|             ('http:/tmp/junk.txt',
 | |
|              'http:///tmp/junk.txt',
 | |
|              ('http', '', '/tmp/junk.txt', '', '', ''),
 | |
|              ('http', '', '/tmp/junk.txt', '', '')),
 | |
|             ('https:/tmp/junk.txt',
 | |
|              'https:///tmp/junk.txt',
 | |
|              ('https', '', '/tmp/junk.txt', '', '', ''),
 | |
|              ('https', '', '/tmp/junk.txt', '', '')),
 | |
|         ]
 | |
|         def _encode(t):
 | |
|             return (t[0].encode('ascii'),
 | |
|                     t[1].encode('ascii'),
 | |
|                     tuple(x.encode('ascii') for x in t[2]),
 | |
|                     tuple(x.encode('ascii') for x in t[3]))
 | |
|         bytes_cases = [_encode(x) for x in str_cases]
 | |
|         for url, url2, parsed, split in str_cases + bytes_cases:
 | |
|             with self.subTest(url):
 | |
|                 self.checkRoundtrips(url, parsed, split, url2)
 | |
| 
 | |
|     def test_http_roundtrips(self):
 | |
|         # urllib.parse.urlsplit treats 'http:' as an optimized special case,
 | |
|         # so we test both 'http:' and 'https:' in all the following.
 | |
|         # Three cheers for white box knowledge!
 | |
|         str_cases = [
 | |
|             ('://www.python.org',
 | |
|              ('www.python.org', '', '', '', ''),
 | |
|              ('www.python.org', '', '', '')),
 | |
|             ('://www.python.org#abc',
 | |
|              ('www.python.org', '', '', '', 'abc'),
 | |
|              ('www.python.org', '', '', 'abc')),
 | |
|             ('://www.python.org?q=abc',
 | |
|              ('www.python.org', '', '', 'q=abc', ''),
 | |
|              ('www.python.org', '', 'q=abc', '')),
 | |
|             ('://www.python.org/#abc',
 | |
|              ('www.python.org', '/', '', '', 'abc'),
 | |
|              ('www.python.org', '/', '', 'abc')),
 | |
|             ('://a/b/c/d;p?q#f',
 | |
|              ('a', '/b/c/d', 'p', 'q', 'f'),
 | |
|              ('a', '/b/c/d;p', 'q', 'f')),
 | |
|             ]
 | |
|         def _encode(t):
 | |
|             return (t[0].encode('ascii'),
 | |
|                     tuple(x.encode('ascii') for x in t[1]),
 | |
|                     tuple(x.encode('ascii') for x in t[2]))
 | |
|         bytes_cases = [_encode(x) for x in str_cases]
 | |
|         str_schemes = ('http', 'https')
 | |
|         bytes_schemes = (b'http', b'https')
 | |
|         str_tests = str_schemes, str_cases
 | |
|         bytes_tests = bytes_schemes, bytes_cases
 | |
|         for schemes, test_cases in (str_tests, bytes_tests):
 | |
|             for scheme in schemes:
 | |
|                 for url, parsed, split in test_cases:
 | |
|                     url = scheme + url
 | |
|                     parsed = (scheme,) + parsed
 | |
|                     split = (scheme,) + split
 | |
|                     self.checkRoundtrips(url, parsed, split)
 | |
| 
 | |
|     def checkJoin(self, base, relurl, expected):
 | |
|         with self.subTest(base=base, relurl=relurl):
 | |
|             self.assertEqual(urllib.parse.urljoin(base, relurl), expected)
 | |
|             baseb = base.encode('ascii')
 | |
|             relurlb = relurl.encode('ascii')
 | |
|             expectedb = expected.encode('ascii')
 | |
|             self.assertEqual(urllib.parse.urljoin(baseb, relurlb), expectedb)
 | |
| 
 | |
|             relurl = urllib.parse.urlunsplit(urllib.parse.urlsplit(relurl))
 | |
|             self.assertEqual(urllib.parse.urljoin(base, relurl), expected)
 | |
|             relurlb = urllib.parse.urlunsplit(urllib.parse.urlsplit(relurlb))
 | |
|             self.assertEqual(urllib.parse.urljoin(baseb, relurlb), expectedb)
 | |
| 
 | |
|     def test_unparse_parse(self):
 | |
|         str_cases = ['Python', './Python','x-newscheme://foo.com/stuff','x://y','x:/y','x:/','/',]
 | |
|         bytes_cases = [x.encode('ascii') for x in str_cases]
 | |
|         for u in str_cases + bytes_cases:
 | |
|             self.assertEqual(urllib.parse.urlunsplit(urllib.parse.urlsplit(u)), u)
 | |
|             self.assertEqual(urllib.parse.urlunparse(urllib.parse.urlparse(u)), u)
 | |
| 
 | |
|     def test_RFC1808(self):
 | |
|         # "normal" cases from RFC 1808:
 | |
|         self.checkJoin(RFC1808_BASE, 'g:h', 'g:h')
 | |
|         self.checkJoin(RFC1808_BASE, 'g', 'http://a/b/c/g')
 | |
|         self.checkJoin(RFC1808_BASE, './g', 'http://a/b/c/g')
 | |
|         self.checkJoin(RFC1808_BASE, 'g/', 'http://a/b/c/g/')
 | |
|         self.checkJoin(RFC1808_BASE, '/g', 'http://a/g')
 | |
|         self.checkJoin(RFC1808_BASE, '//g', 'http://g')
 | |
|         self.checkJoin(RFC1808_BASE, 'g?y', 'http://a/b/c/g?y')
 | |
|         self.checkJoin(RFC1808_BASE, 'g?y/./x', 'http://a/b/c/g?y/./x')
 | |
|         self.checkJoin(RFC1808_BASE, '#s', 'http://a/b/c/d;p?q#s')
 | |
|         self.checkJoin(RFC1808_BASE, 'g#s', 'http://a/b/c/g#s')
 | |
|         self.checkJoin(RFC1808_BASE, 'g#s/./x', 'http://a/b/c/g#s/./x')
 | |
|         self.checkJoin(RFC1808_BASE, 'g?y#s', 'http://a/b/c/g?y#s')
 | |
|         self.checkJoin(RFC1808_BASE, 'g;x', 'http://a/b/c/g;x')
 | |
|         self.checkJoin(RFC1808_BASE, 'g;x?y#s', 'http://a/b/c/g;x?y#s')
 | |
|         self.checkJoin(RFC1808_BASE, '.', 'http://a/b/c/')
 | |
|         self.checkJoin(RFC1808_BASE, './', 'http://a/b/c/')
 | |
|         self.checkJoin(RFC1808_BASE, '..', 'http://a/b/')
 | |
|         self.checkJoin(RFC1808_BASE, '../', 'http://a/b/')
 | |
|         self.checkJoin(RFC1808_BASE, '../g', 'http://a/b/g')
 | |
|         self.checkJoin(RFC1808_BASE, '../..', 'http://a/')
 | |
|         self.checkJoin(RFC1808_BASE, '../../', 'http://a/')
 | |
|         self.checkJoin(RFC1808_BASE, '../../g', 'http://a/g')
 | |
| 
 | |
|         # "abnormal" cases from RFC 1808:
 | |
|         self.checkJoin(RFC1808_BASE, '', 'http://a/b/c/d;p?q#f')
 | |
|         self.checkJoin(RFC1808_BASE, 'g.', 'http://a/b/c/g.')
 | |
|         self.checkJoin(RFC1808_BASE, '.g', 'http://a/b/c/.g')
 | |
|         self.checkJoin(RFC1808_BASE, 'g..', 'http://a/b/c/g..')
 | |
|         self.checkJoin(RFC1808_BASE, '..g', 'http://a/b/c/..g')
 | |
|         self.checkJoin(RFC1808_BASE, './../g', 'http://a/b/g')
 | |
|         self.checkJoin(RFC1808_BASE, './g/.', 'http://a/b/c/g/')
 | |
|         self.checkJoin(RFC1808_BASE, 'g/./h', 'http://a/b/c/g/h')
 | |
|         self.checkJoin(RFC1808_BASE, 'g/../h', 'http://a/b/c/h')
 | |
| 
 | |
|         # RFC 1808 and RFC 1630 disagree on these (according to RFC 1808),
 | |
|         # so we'll not actually run these tests (which expect 1808 behavior).
 | |
|         #self.checkJoin(RFC1808_BASE, 'http:g', 'http:g')
 | |
|         #self.checkJoin(RFC1808_BASE, 'http:', 'http:')
 | |
| 
 | |
|         # XXX: The following tests are no longer compatible with RFC3986
 | |
|         # self.checkJoin(RFC1808_BASE, '../../../g', 'http://a/../g')
 | |
|         # self.checkJoin(RFC1808_BASE, '../../../../g', 'http://a/../../g')
 | |
|         # self.checkJoin(RFC1808_BASE, '/./g', 'http://a/./g')
 | |
|         # self.checkJoin(RFC1808_BASE, '/../g', 'http://a/../g')
 | |
| 
 | |
| 
 | |
|     def test_RFC2368(self):
 | |
|         # Issue 11467: path that starts with a number is not parsed correctly
 | |
|         self.assertEqual(urllib.parse.urlparse('mailto:1337@example.org'),
 | |
|                 ('mailto', '', '1337@example.org', '', '', ''))
 | |
| 
 | |
|     def test_RFC2396(self):
 | |
|         # cases from RFC 2396
 | |
| 
 | |
|         self.checkJoin(RFC2396_BASE, 'g:h', 'g:h')
 | |
|         self.checkJoin(RFC2396_BASE, 'g', 'http://a/b/c/g')
 | |
|         self.checkJoin(RFC2396_BASE, './g', 'http://a/b/c/g')
 | |
|         self.checkJoin(RFC2396_BASE, 'g/', 'http://a/b/c/g/')
 | |
|         self.checkJoin(RFC2396_BASE, '/g', 'http://a/g')
 | |
|         self.checkJoin(RFC2396_BASE, '//g', 'http://g')
 | |
|         self.checkJoin(RFC2396_BASE, 'g?y', 'http://a/b/c/g?y')
 | |
|         self.checkJoin(RFC2396_BASE, '#s', 'http://a/b/c/d;p?q#s')
 | |
|         self.checkJoin(RFC2396_BASE, 'g#s', 'http://a/b/c/g#s')
 | |
|         self.checkJoin(RFC2396_BASE, 'g?y#s', 'http://a/b/c/g?y#s')
 | |
|         self.checkJoin(RFC2396_BASE, 'g;x', 'http://a/b/c/g;x')
 | |
|         self.checkJoin(RFC2396_BASE, 'g;x?y#s', 'http://a/b/c/g;x?y#s')
 | |
|         self.checkJoin(RFC2396_BASE, '.', 'http://a/b/c/')
 | |
|         self.checkJoin(RFC2396_BASE, './', 'http://a/b/c/')
 | |
|         self.checkJoin(RFC2396_BASE, '..', 'http://a/b/')
 | |
|         self.checkJoin(RFC2396_BASE, '../', 'http://a/b/')
 | |
|         self.checkJoin(RFC2396_BASE, '../g', 'http://a/b/g')
 | |
|         self.checkJoin(RFC2396_BASE, '../..', 'http://a/')
 | |
|         self.checkJoin(RFC2396_BASE, '../../', 'http://a/')
 | |
|         self.checkJoin(RFC2396_BASE, '../../g', 'http://a/g')
 | |
|         self.checkJoin(RFC2396_BASE, '', RFC2396_BASE)
 | |
|         self.checkJoin(RFC2396_BASE, 'g.', 'http://a/b/c/g.')
 | |
|         self.checkJoin(RFC2396_BASE, '.g', 'http://a/b/c/.g')
 | |
|         self.checkJoin(RFC2396_BASE, 'g..', 'http://a/b/c/g..')
 | |
|         self.checkJoin(RFC2396_BASE, '..g', 'http://a/b/c/..g')
 | |
|         self.checkJoin(RFC2396_BASE, './../g', 'http://a/b/g')
 | |
|         self.checkJoin(RFC2396_BASE, './g/.', 'http://a/b/c/g/')
 | |
|         self.checkJoin(RFC2396_BASE, 'g/./h', 'http://a/b/c/g/h')
 | |
|         self.checkJoin(RFC2396_BASE, 'g/../h', 'http://a/b/c/h')
 | |
|         self.checkJoin(RFC2396_BASE, 'g;x=1/./y', 'http://a/b/c/g;x=1/y')
 | |
|         self.checkJoin(RFC2396_BASE, 'g;x=1/../y', 'http://a/b/c/y')
 | |
|         self.checkJoin(RFC2396_BASE, 'g?y/./x', 'http://a/b/c/g?y/./x')
 | |
|         self.checkJoin(RFC2396_BASE, 'g?y/../x', 'http://a/b/c/g?y/../x')
 | |
|         self.checkJoin(RFC2396_BASE, 'g#s/./x', 'http://a/b/c/g#s/./x')
 | |
|         self.checkJoin(RFC2396_BASE, 'g#s/../x', 'http://a/b/c/g#s/../x')
 | |
| 
 | |
|         # XXX: The following tests are no longer compatible with RFC3986
 | |
|         # self.checkJoin(RFC2396_BASE, '../../../g', 'http://a/../g')
 | |
|         # self.checkJoin(RFC2396_BASE, '../../../../g', 'http://a/../../g')
 | |
|         # self.checkJoin(RFC2396_BASE, '/./g', 'http://a/./g')
 | |
|         # self.checkJoin(RFC2396_BASE, '/../g', 'http://a/../g')
 | |
| 
 | |
|     def test_RFC3986(self):
 | |
|         self.checkJoin(RFC3986_BASE, '?y','http://a/b/c/d;p?y')
 | |
|         self.checkJoin(RFC3986_BASE, ';x', 'http://a/b/c/;x')
 | |
|         self.checkJoin(RFC3986_BASE, 'g:h','g:h')
 | |
|         self.checkJoin(RFC3986_BASE, 'g','http://a/b/c/g')
 | |
|         self.checkJoin(RFC3986_BASE, './g','http://a/b/c/g')
 | |
|         self.checkJoin(RFC3986_BASE, 'g/','http://a/b/c/g/')
 | |
|         self.checkJoin(RFC3986_BASE, '/g','http://a/g')
 | |
|         self.checkJoin(RFC3986_BASE, '//g','http://g')
 | |
|         self.checkJoin(RFC3986_BASE, '?y','http://a/b/c/d;p?y')
 | |
|         self.checkJoin(RFC3986_BASE, 'g?y','http://a/b/c/g?y')
 | |
|         self.checkJoin(RFC3986_BASE, '#s','http://a/b/c/d;p?q#s')
 | |
|         self.checkJoin(RFC3986_BASE, 'g#s','http://a/b/c/g#s')
 | |
|         self.checkJoin(RFC3986_BASE, 'g?y#s','http://a/b/c/g?y#s')
 | |
|         self.checkJoin(RFC3986_BASE, ';x','http://a/b/c/;x')
 | |
|         self.checkJoin(RFC3986_BASE, 'g;x','http://a/b/c/g;x')
 | |
|         self.checkJoin(RFC3986_BASE, 'g;x?y#s','http://a/b/c/g;x?y#s')
 | |
|         self.checkJoin(RFC3986_BASE, '','http://a/b/c/d;p?q')
 | |
|         self.checkJoin(RFC3986_BASE, '.','http://a/b/c/')
 | |
|         self.checkJoin(RFC3986_BASE, './','http://a/b/c/')
 | |
|         self.checkJoin(RFC3986_BASE, '..','http://a/b/')
 | |
|         self.checkJoin(RFC3986_BASE, '../','http://a/b/')
 | |
|         self.checkJoin(RFC3986_BASE, '../g','http://a/b/g')
 | |
|         self.checkJoin(RFC3986_BASE, '../..','http://a/')
 | |
|         self.checkJoin(RFC3986_BASE, '../../','http://a/')
 | |
|         self.checkJoin(RFC3986_BASE, '../../g','http://a/g')
 | |
|         self.checkJoin(RFC3986_BASE, '../../../g', 'http://a/g')
 | |
| 
 | |
|         # Abnormal Examples
 | |
| 
 | |
|         # The 'abnormal scenarios' are incompatible with RFC2986 parsing
 | |
|         # Tests are here for reference.
 | |
| 
 | |
|         self.checkJoin(RFC3986_BASE, '../../../g','http://a/g')
 | |
|         self.checkJoin(RFC3986_BASE, '../../../../g','http://a/g')
 | |
|         self.checkJoin(RFC3986_BASE, '/./g','http://a/g')
 | |
|         self.checkJoin(RFC3986_BASE, '/../g','http://a/g')
 | |
|         self.checkJoin(RFC3986_BASE, 'g.','http://a/b/c/g.')
 | |
|         self.checkJoin(RFC3986_BASE, '.g','http://a/b/c/.g')
 | |
|         self.checkJoin(RFC3986_BASE, 'g..','http://a/b/c/g..')
 | |
|         self.checkJoin(RFC3986_BASE, '..g','http://a/b/c/..g')
 | |
|         self.checkJoin(RFC3986_BASE, './../g','http://a/b/g')
 | |
|         self.checkJoin(RFC3986_BASE, './g/.','http://a/b/c/g/')
 | |
|         self.checkJoin(RFC3986_BASE, 'g/./h','http://a/b/c/g/h')
 | |
|         self.checkJoin(RFC3986_BASE, 'g/../h','http://a/b/c/h')
 | |
|         self.checkJoin(RFC3986_BASE, 'g;x=1/./y','http://a/b/c/g;x=1/y')
 | |
|         self.checkJoin(RFC3986_BASE, 'g;x=1/../y','http://a/b/c/y')
 | |
|         self.checkJoin(RFC3986_BASE, 'g?y/./x','http://a/b/c/g?y/./x')
 | |
|         self.checkJoin(RFC3986_BASE, 'g?y/../x','http://a/b/c/g?y/../x')
 | |
|         self.checkJoin(RFC3986_BASE, 'g#s/./x','http://a/b/c/g#s/./x')
 | |
|         self.checkJoin(RFC3986_BASE, 'g#s/../x','http://a/b/c/g#s/../x')
 | |
|         #self.checkJoin(RFC3986_BASE, 'http:g','http:g') # strict parser
 | |
|         self.checkJoin(RFC3986_BASE, 'http:g','http://a/b/c/g') #relaxed parser
 | |
| 
 | |
|         # Test for issue9721
 | |
|         self.checkJoin('http://a/b/c/de', ';x','http://a/b/c/;x')
 | |
| 
 | |
|     def test_urljoins(self):
 | |
|         self.checkJoin(SIMPLE_BASE, 'g:h','g:h')
 | |
|         self.checkJoin(SIMPLE_BASE, 'http:g','http://a/b/c/g')
 | |
|         self.checkJoin(SIMPLE_BASE, 'http:','http://a/b/c/d')
 | |
|         self.checkJoin(SIMPLE_BASE, 'g','http://a/b/c/g')
 | |
|         self.checkJoin(SIMPLE_BASE, './g','http://a/b/c/g')
 | |
|         self.checkJoin(SIMPLE_BASE, 'g/','http://a/b/c/g/')
 | |
|         self.checkJoin(SIMPLE_BASE, '/g','http://a/g')
 | |
|         self.checkJoin(SIMPLE_BASE, '//g','http://g')
 | |
|         self.checkJoin(SIMPLE_BASE, '?y','http://a/b/c/d?y')
 | |
|         self.checkJoin(SIMPLE_BASE, 'g?y','http://a/b/c/g?y')
 | |
|         self.checkJoin(SIMPLE_BASE, 'g?y/./x','http://a/b/c/g?y/./x')
 | |
|         self.checkJoin(SIMPLE_BASE, '.','http://a/b/c/')
 | |
|         self.checkJoin(SIMPLE_BASE, './','http://a/b/c/')
 | |
|         self.checkJoin(SIMPLE_BASE, '..','http://a/b/')
 | |
|         self.checkJoin(SIMPLE_BASE, '../','http://a/b/')
 | |
|         self.checkJoin(SIMPLE_BASE, '../g','http://a/b/g')
 | |
|         self.checkJoin(SIMPLE_BASE, '../..','http://a/')
 | |
|         self.checkJoin(SIMPLE_BASE, '../../g','http://a/g')
 | |
|         self.checkJoin(SIMPLE_BASE, './../g','http://a/b/g')
 | |
|         self.checkJoin(SIMPLE_BASE, './g/.','http://a/b/c/g/')
 | |
|         self.checkJoin(SIMPLE_BASE, 'g/./h','http://a/b/c/g/h')
 | |
|         self.checkJoin(SIMPLE_BASE, 'g/../h','http://a/b/c/h')
 | |
|         self.checkJoin(SIMPLE_BASE, 'http:g','http://a/b/c/g')
 | |
|         self.checkJoin(SIMPLE_BASE, 'http:','http://a/b/c/d')
 | |
|         self.checkJoin(SIMPLE_BASE, 'http:?y','http://a/b/c/d?y')
 | |
|         self.checkJoin(SIMPLE_BASE, 'http:g?y','http://a/b/c/g?y')
 | |
|         self.checkJoin(SIMPLE_BASE, 'http:g?y/./x','http://a/b/c/g?y/./x')
 | |
|         self.checkJoin('http:///', '..','http:///')
 | |
|         self.checkJoin('', 'http://a/b/c/g?y/./x','http://a/b/c/g?y/./x')
 | |
|         self.checkJoin('', 'http://a/./g', 'http://a/./g')
 | |
|         self.checkJoin('svn://pathtorepo/dir1', 'dir2', 'svn://pathtorepo/dir2')
 | |
|         self.checkJoin('svn+ssh://pathtorepo/dir1', 'dir2', 'svn+ssh://pathtorepo/dir2')
 | |
|         self.checkJoin('ws://a/b','g','ws://a/g')
 | |
|         self.checkJoin('wss://a/b','g','wss://a/g')
 | |
| 
 | |
|         # XXX: The following tests are no longer compatible with RFC3986
 | |
|         # self.checkJoin(SIMPLE_BASE, '../../../g','http://a/../g')
 | |
|         # self.checkJoin(SIMPLE_BASE, '/./g','http://a/./g')
 | |
| 
 | |
|         # test for issue22118 duplicate slashes
 | |
|         self.checkJoin(SIMPLE_BASE + '/', 'foo', SIMPLE_BASE + '/foo')
 | |
| 
 | |
|         # Non-RFC-defined tests, covering variations of base and trailing
 | |
|         # slashes
 | |
|         self.checkJoin('http://a/b/c/d/e/', '../../f/g/', 'http://a/b/c/f/g/')
 | |
|         self.checkJoin('http://a/b/c/d/e', '../../f/g/', 'http://a/b/f/g/')
 | |
|         self.checkJoin('http://a/b/c/d/e/', '/../../f/g/', 'http://a/f/g/')
 | |
|         self.checkJoin('http://a/b/c/d/e', '/../../f/g/', 'http://a/f/g/')
 | |
|         self.checkJoin('http://a/b/c/d/e/', '../../f/g', 'http://a/b/c/f/g')
 | |
|         self.checkJoin('http://a/b/', '../../f/g/', 'http://a/f/g/')
 | |
| 
 | |
|         # issue 23703: don't duplicate filename
 | |
|         self.checkJoin('a', 'b', 'b')
 | |
| 
 | |
|     def test_RFC2732(self):
 | |
|         str_cases = [
 | |
|             ('http://Test.python.org:5432/foo/', 'test.python.org', 5432),
 | |
|             ('http://12.34.56.78:5432/foo/', '12.34.56.78', 5432),
 | |
|             ('http://[::1]:5432/foo/', '::1', 5432),
 | |
|             ('http://[dead:beef::1]:5432/foo/', 'dead:beef::1', 5432),
 | |
|             ('http://[dead:beef::]:5432/foo/', 'dead:beef::', 5432),
 | |
|             ('http://[dead:beef:cafe:5417:affe:8FA3:deaf:feed]:5432/foo/',
 | |
|              'dead:beef:cafe:5417:affe:8fa3:deaf:feed', 5432),
 | |
|             ('http://[::12.34.56.78]:5432/foo/', '::12.34.56.78', 5432),
 | |
|             ('http://[::ffff:12.34.56.78]:5432/foo/',
 | |
|              '::ffff:12.34.56.78', 5432),
 | |
|             ('http://Test.python.org/foo/', 'test.python.org', None),
 | |
|             ('http://12.34.56.78/foo/', '12.34.56.78', None),
 | |
|             ('http://[::1]/foo/', '::1', None),
 | |
|             ('http://[dead:beef::1]/foo/', 'dead:beef::1', None),
 | |
|             ('http://[dead:beef::]/foo/', 'dead:beef::', None),
 | |
|             ('http://[dead:beef:cafe:5417:affe:8FA3:deaf:feed]/foo/',
 | |
|              'dead:beef:cafe:5417:affe:8fa3:deaf:feed', None),
 | |
|             ('http://[::12.34.56.78]/foo/', '::12.34.56.78', None),
 | |
|             ('http://[::ffff:12.34.56.78]/foo/',
 | |
|              '::ffff:12.34.56.78', None),
 | |
|             ('http://Test.python.org:/foo/', 'test.python.org', None),
 | |
|             ('http://12.34.56.78:/foo/', '12.34.56.78', None),
 | |
|             ('http://[::1]:/foo/', '::1', None),
 | |
|             ('http://[dead:beef::1]:/foo/', 'dead:beef::1', None),
 | |
|             ('http://[dead:beef::]:/foo/', 'dead:beef::', None),
 | |
|             ('http://[dead:beef:cafe:5417:affe:8FA3:deaf:feed]:/foo/',
 | |
|              'dead:beef:cafe:5417:affe:8fa3:deaf:feed', None),
 | |
|             ('http://[::12.34.56.78]:/foo/', '::12.34.56.78', None),
 | |
|             ('http://[::ffff:12.34.56.78]:/foo/',
 | |
|              '::ffff:12.34.56.78', None),
 | |
|             ]
 | |
|         def _encode(t):
 | |
|             return t[0].encode('ascii'), t[1].encode('ascii'), t[2]
 | |
|         bytes_cases = [_encode(x) for x in str_cases]
 | |
|         for url, hostname, port in str_cases + bytes_cases:
 | |
|             urlparsed = urllib.parse.urlparse(url)
 | |
|             self.assertEqual((urlparsed.hostname, urlparsed.port) , (hostname, port))
 | |
| 
 | |
|         str_cases = [
 | |
|                 'http://::12.34.56.78]/',
 | |
|                 'http://[::1/foo/',
 | |
|                 'ftp://[::1/foo/bad]/bad',
 | |
|                 'http://[::1/foo/bad]/bad',
 | |
|                 'http://[::ffff:12.34.56.78']
 | |
|         bytes_cases = [x.encode('ascii') for x in str_cases]
 | |
|         for invalid_url in str_cases + bytes_cases:
 | |
|             self.assertRaises(ValueError, urllib.parse.urlparse, invalid_url)
 | |
| 
 | |
|     def test_urldefrag(self):
 | |
|         str_cases = [
 | |
|             ('http://python.org#frag', 'http://python.org', 'frag'),
 | |
|             ('http://python.org', 'http://python.org', ''),
 | |
|             ('http://python.org/#frag', 'http://python.org/', 'frag'),
 | |
|             ('http://python.org/', 'http://python.org/', ''),
 | |
|             ('http://python.org/?q#frag', 'http://python.org/?q', 'frag'),
 | |
|             ('http://python.org/?q', 'http://python.org/?q', ''),
 | |
|             ('http://python.org/p#frag', 'http://python.org/p', 'frag'),
 | |
|             ('http://python.org/p?q', 'http://python.org/p?q', ''),
 | |
|             (RFC1808_BASE, 'http://a/b/c/d;p?q', 'f'),
 | |
|             (RFC2396_BASE, 'http://a/b/c/d;p?q', ''),
 | |
|         ]
 | |
|         def _encode(t):
 | |
|             return type(t)(x.encode('ascii') for x in t)
 | |
|         bytes_cases = [_encode(x) for x in str_cases]
 | |
|         for url, defrag, frag in str_cases + bytes_cases:
 | |
|             result = urllib.parse.urldefrag(url)
 | |
|             self.assertEqual(result.geturl(), url)
 | |
|             self.assertEqual(result, (defrag, frag))
 | |
|             self.assertEqual(result.url, defrag)
 | |
|             self.assertEqual(result.fragment, frag)
 | |
| 
 | |
|     def test_urlsplit_scoped_IPv6(self):
 | |
|         p = urllib.parse.urlsplit('http://[FE80::822a:a8ff:fe49:470c%tESt]:1234')
 | |
|         self.assertEqual(p.hostname, "fe80::822a:a8ff:fe49:470c%tESt")
 | |
|         self.assertEqual(p.netloc, '[FE80::822a:a8ff:fe49:470c%tESt]:1234')
 | |
| 
 | |
|         p = urllib.parse.urlsplit(b'http://[FE80::822a:a8ff:fe49:470c%tESt]:1234')
 | |
|         self.assertEqual(p.hostname, b"fe80::822a:a8ff:fe49:470c%tESt")
 | |
|         self.assertEqual(p.netloc, b'[FE80::822a:a8ff:fe49:470c%tESt]:1234')
 | |
| 
 | |
|     def test_urlsplit_attributes(self):
 | |
|         url = "HTTP://WWW.PYTHON.ORG/doc/#frag"
 | |
|         p = urllib.parse.urlsplit(url)
 | |
|         self.assertEqual(p.scheme, "http")
 | |
|         self.assertEqual(p.netloc, "WWW.PYTHON.ORG")
 | |
|         self.assertEqual(p.path, "/doc/")
 | |
|         self.assertEqual(p.query, "")
 | |
|         self.assertEqual(p.fragment, "frag")
 | |
|         self.assertEqual(p.username, None)
 | |
|         self.assertEqual(p.password, None)
 | |
|         self.assertEqual(p.hostname, "www.python.org")
 | |
|         self.assertEqual(p.port, None)
 | |
|         # geturl() won't return exactly the original URL in this case
 | |
|         # since the scheme is always case-normalized
 | |
|         # We handle this by ignoring the first 4 characters of the URL
 | |
|         self.assertEqual(p.geturl()[4:], url[4:])
 | |
| 
 | |
|         url = "http://User:Pass@www.python.org:080/doc/?query=yes#frag"
 | |
|         p = urllib.parse.urlsplit(url)
 | |
|         self.assertEqual(p.scheme, "http")
 | |
|         self.assertEqual(p.netloc, "User:Pass@www.python.org:080")
 | |
|         self.assertEqual(p.path, "/doc/")
 | |
|         self.assertEqual(p.query, "query=yes")
 | |
|         self.assertEqual(p.fragment, "frag")
 | |
|         self.assertEqual(p.username, "User")
 | |
|         self.assertEqual(p.password, "Pass")
 | |
|         self.assertEqual(p.hostname, "www.python.org")
 | |
|         self.assertEqual(p.port, 80)
 | |
|         self.assertEqual(p.geturl(), url)
 | |
| 
 | |
|         # Addressing issue1698, which suggests Username can contain
 | |
|         # "@" characters.  Though not RFC compliant, many ftp sites allow
 | |
|         # and request email addresses as usernames.
 | |
| 
 | |
|         url = "http://User@example.com:Pass@www.python.org:080/doc/?query=yes#frag"
 | |
|         p = urllib.parse.urlsplit(url)
 | |
|         self.assertEqual(p.scheme, "http")
 | |
|         self.assertEqual(p.netloc, "User@example.com:Pass@www.python.org:080")
 | |
|         self.assertEqual(p.path, "/doc/")
 | |
|         self.assertEqual(p.query, "query=yes")
 | |
|         self.assertEqual(p.fragment, "frag")
 | |
|         self.assertEqual(p.username, "User@example.com")
 | |
|         self.assertEqual(p.password, "Pass")
 | |
|         self.assertEqual(p.hostname, "www.python.org")
 | |
|         self.assertEqual(p.port, 80)
 | |
|         self.assertEqual(p.geturl(), url)
 | |
| 
 | |
|         # And check them all again, only with bytes this time
 | |
|         url = b"HTTP://WWW.PYTHON.ORG/doc/#frag"
 | |
|         p = urllib.parse.urlsplit(url)
 | |
|         self.assertEqual(p.scheme, b"http")
 | |
|         self.assertEqual(p.netloc, b"WWW.PYTHON.ORG")
 | |
|         self.assertEqual(p.path, b"/doc/")
 | |
|         self.assertEqual(p.query, b"")
 | |
|         self.assertEqual(p.fragment, b"frag")
 | |
|         self.assertEqual(p.username, None)
 | |
|         self.assertEqual(p.password, None)
 | |
|         self.assertEqual(p.hostname, b"www.python.org")
 | |
|         self.assertEqual(p.port, None)
 | |
|         self.assertEqual(p.geturl()[4:], url[4:])
 | |
| 
 | |
|         url = b"http://User:Pass@www.python.org:080/doc/?query=yes#frag"
 | |
|         p = urllib.parse.urlsplit(url)
 | |
|         self.assertEqual(p.scheme, b"http")
 | |
|         self.assertEqual(p.netloc, b"User:Pass@www.python.org:080")
 | |
|         self.assertEqual(p.path, b"/doc/")
 | |
|         self.assertEqual(p.query, b"query=yes")
 | |
|         self.assertEqual(p.fragment, b"frag")
 | |
|         self.assertEqual(p.username, b"User")
 | |
|         self.assertEqual(p.password, b"Pass")
 | |
|         self.assertEqual(p.hostname, b"www.python.org")
 | |
|         self.assertEqual(p.port, 80)
 | |
|         self.assertEqual(p.geturl(), url)
 | |
| 
 | |
|         url = b"http://User@example.com:Pass@www.python.org:080/doc/?query=yes#frag"
 | |
|         p = urllib.parse.urlsplit(url)
 | |
|         self.assertEqual(p.scheme, b"http")
 | |
|         self.assertEqual(p.netloc, b"User@example.com:Pass@www.python.org:080")
 | |
|         self.assertEqual(p.path, b"/doc/")
 | |
|         self.assertEqual(p.query, b"query=yes")
 | |
|         self.assertEqual(p.fragment, b"frag")
 | |
|         self.assertEqual(p.username, b"User@example.com")
 | |
|         self.assertEqual(p.password, b"Pass")
 | |
|         self.assertEqual(p.hostname, b"www.python.org")
 | |
|         self.assertEqual(p.port, 80)
 | |
|         self.assertEqual(p.geturl(), url)
 | |
| 
 | |
|         # Verify an illegal port raises ValueError
 | |
|         url = b"HTTP://WWW.PYTHON.ORG:65536/doc/#frag"
 | |
|         p = urllib.parse.urlsplit(url)
 | |
|         with self.assertRaisesRegex(ValueError, "out of range"):
 | |
|             p.port
 | |
| 
 | |
|     def test_urlsplit_remove_unsafe_bytes(self):
 | |
|         # Remove ASCII tabs and newlines from input
 | |
|         url = "http\t://www.python\n.org\t/java\nscript:\talert('msg\r\n')/?query\n=\tsomething#frag\nment"
 | |
|         p = urllib.parse.urlsplit(url)
 | |
|         self.assertEqual(p.scheme, "http")
 | |
|         self.assertEqual(p.netloc, "www.python.org")
 | |
|         self.assertEqual(p.path, "/javascript:alert('msg')/")
 | |
|         self.assertEqual(p.query, "query=something")
 | |
|         self.assertEqual(p.fragment, "fragment")
 | |
|         self.assertEqual(p.username, None)
 | |
|         self.assertEqual(p.password, None)
 | |
|         self.assertEqual(p.hostname, "www.python.org")
 | |
|         self.assertEqual(p.port, None)
 | |
|         self.assertEqual(p.geturl(), "http://www.python.org/javascript:alert('msg')/?query=something#fragment")
 | |
| 
 | |
|         # Remove ASCII tabs and newlines from input as bytes.
 | |
|         url = b"http\t://www.python\n.org\t/java\nscript:\talert('msg\r\n')/?query\n=\tsomething#frag\nment"
 | |
|         p = urllib.parse.urlsplit(url)
 | |
|         self.assertEqual(p.scheme, b"http")
 | |
|         self.assertEqual(p.netloc, b"www.python.org")
 | |
|         self.assertEqual(p.path, b"/javascript:alert('msg')/")
 | |
|         self.assertEqual(p.query, b"query=something")
 | |
|         self.assertEqual(p.fragment, b"fragment")
 | |
|         self.assertEqual(p.username, None)
 | |
|         self.assertEqual(p.password, None)
 | |
|         self.assertEqual(p.hostname, b"www.python.org")
 | |
|         self.assertEqual(p.port, None)
 | |
|         self.assertEqual(p.geturl(), b"http://www.python.org/javascript:alert('msg')/?query=something#fragment")
 | |
| 
 | |
|         # with scheme as cache-key
 | |
|         url = "http://www.python.org/java\nscript:\talert('msg\r\n')/?query\n=\tsomething#frag\nment"
 | |
|         scheme = "ht\ntp"
 | |
|         for _ in range(2):
 | |
|             p = urllib.parse.urlsplit(url, scheme=scheme)
 | |
|             self.assertEqual(p.scheme, "http")
 | |
|             self.assertEqual(p.geturl(), "http://www.python.org/javascript:alert('msg')/?query=something#fragment")
 | |
| 
 | |
|     def test_urlsplit_strip_url(self):
 | |
|         noise = bytes(range(0, 0x20 + 1))
 | |
|         base_url = "http://User:Pass@www.python.org:080/doc/?query=yes#frag"
 | |
| 
 | |
|         url = noise.decode("utf-8") + base_url
 | |
|         p = urllib.parse.urlsplit(url)
 | |
|         self.assertEqual(p.scheme, "http")
 | |
|         self.assertEqual(p.netloc, "User:Pass@www.python.org:080")
 | |
|         self.assertEqual(p.path, "/doc/")
 | |
|         self.assertEqual(p.query, "query=yes")
 | |
|         self.assertEqual(p.fragment, "frag")
 | |
|         self.assertEqual(p.username, "User")
 | |
|         self.assertEqual(p.password, "Pass")
 | |
|         self.assertEqual(p.hostname, "www.python.org")
 | |
|         self.assertEqual(p.port, 80)
 | |
|         self.assertEqual(p.geturl(), base_url)
 | |
| 
 | |
|         url = noise + base_url.encode("utf-8")
 | |
|         p = urllib.parse.urlsplit(url)
 | |
|         self.assertEqual(p.scheme, b"http")
 | |
|         self.assertEqual(p.netloc, b"User:Pass@www.python.org:080")
 | |
|         self.assertEqual(p.path, b"/doc/")
 | |
|         self.assertEqual(p.query, b"query=yes")
 | |
|         self.assertEqual(p.fragment, b"frag")
 | |
|         self.assertEqual(p.username, b"User")
 | |
|         self.assertEqual(p.password, b"Pass")
 | |
|         self.assertEqual(p.hostname, b"www.python.org")
 | |
|         self.assertEqual(p.port, 80)
 | |
|         self.assertEqual(p.geturl(), base_url.encode("utf-8"))
 | |
| 
 | |
|         # Test that trailing space is preserved as some applications rely on
 | |
|         # this within query strings.
 | |
|         query_spaces_url = "https://www.python.org:88/doc/?query=    "
 | |
|         p = urllib.parse.urlsplit(noise.decode("utf-8") + query_spaces_url)
 | |
|         self.assertEqual(p.scheme, "https")
 | |
|         self.assertEqual(p.netloc, "www.python.org:88")
 | |
|         self.assertEqual(p.path, "/doc/")
 | |
|         self.assertEqual(p.query, "query=    ")
 | |
|         self.assertEqual(p.port, 88)
 | |
|         self.assertEqual(p.geturl(), query_spaces_url)
 | |
| 
 | |
|         p = urllib.parse.urlsplit("www.pypi.org ")
 | |
|         # That "hostname" gets considered a "path" due to the
 | |
|         # trailing space and our existing logic...  YUCK...
 | |
|         # and re-assembles via geturl aka unurlsplit into the original.
 | |
|         # django.core.validators.URLValidator (at least through v3.2) relies on
 | |
|         # this, for better or worse, to catch it in a ValidationError via its
 | |
|         # regular expressions.
 | |
|         # Here we test the basic round trip concept of such a trailing space.
 | |
|         self.assertEqual(urllib.parse.urlunsplit(p), "www.pypi.org ")
 | |
| 
 | |
|         # with scheme as cache-key
 | |
|         url = "//www.python.org/"
 | |
|         scheme = noise.decode("utf-8") + "https" + noise.decode("utf-8")
 | |
|         for _ in range(2):
 | |
|             p = urllib.parse.urlsplit(url, scheme=scheme)
 | |
|             self.assertEqual(p.scheme, "https")
 | |
|             self.assertEqual(p.geturl(), "https://www.python.org/")
 | |
| 
 | |
|     def test_attributes_bad_port(self):
 | |
|         """Check handling of invalid ports."""
 | |
|         for bytes in (False, True):
 | |
|             for parse in (urllib.parse.urlsplit, urllib.parse.urlparse):
 | |
|                 for port in ("foo", "1.5", "-1", "0x10", "-0", "1_1", " 1", "1 ", "६"):
 | |
|                     with self.subTest(bytes=bytes, parse=parse, port=port):
 | |
|                         netloc = "www.example.net:" + port
 | |
|                         url = "http://" + netloc + "/"
 | |
|                         if bytes:
 | |
|                             if netloc.isascii() and port.isascii():
 | |
|                                 netloc = netloc.encode("ascii")
 | |
|                                 url = url.encode("ascii")
 | |
|                             else:
 | |
|                                 continue
 | |
|                         p = parse(url)
 | |
|                         self.assertEqual(p.netloc, netloc)
 | |
|                         with self.assertRaises(ValueError):
 | |
|                             p.port
 | |
| 
 | |
|     def test_attributes_bad_scheme(self):
 | |
|         """Check handling of invalid schemes."""
 | |
|         for bytes in (False, True):
 | |
|             for parse in (urllib.parse.urlsplit, urllib.parse.urlparse):
 | |
|                 for scheme in (".", "+", "-", "0", "http&", "६http"):
 | |
|                     with self.subTest(bytes=bytes, parse=parse, scheme=scheme):
 | |
|                         url = scheme + "://www.example.net"
 | |
|                         if bytes:
 | |
|                             if url.isascii():
 | |
|                                 url = url.encode("ascii")
 | |
|                             else:
 | |
|                                 continue
 | |
|                         p = parse(url)
 | |
|                         if bytes:
 | |
|                             self.assertEqual(p.scheme, b"")
 | |
|                         else:
 | |
|                             self.assertEqual(p.scheme, "")
 | |
| 
 | |
|     def test_attributes_without_netloc(self):
 | |
|         # This example is straight from RFC 3261.  It looks like it
 | |
|         # should allow the username, hostname, and port to be filled
 | |
|         # in, but doesn't.  Since it's a URI and doesn't use the
 | |
|         # scheme://netloc syntax, the netloc and related attributes
 | |
|         # should be left empty.
 | |
|         uri = "sip:alice@atlanta.com;maddr=239.255.255.1;ttl=15"
 | |
|         p = urllib.parse.urlsplit(uri)
 | |
|         self.assertEqual(p.netloc, "")
 | |
|         self.assertEqual(p.username, None)
 | |
|         self.assertEqual(p.password, None)
 | |
|         self.assertEqual(p.hostname, None)
 | |
|         self.assertEqual(p.port, None)
 | |
|         self.assertEqual(p.geturl(), uri)
 | |
| 
 | |
|         p = urllib.parse.urlparse(uri)
 | |
|         self.assertEqual(p.netloc, "")
 | |
|         self.assertEqual(p.username, None)
 | |
|         self.assertEqual(p.password, None)
 | |
|         self.assertEqual(p.hostname, None)
 | |
|         self.assertEqual(p.port, None)
 | |
|         self.assertEqual(p.geturl(), uri)
 | |
| 
 | |
|         # You guessed it, repeating the test with bytes input
 | |
|         uri = b"sip:alice@atlanta.com;maddr=239.255.255.1;ttl=15"
 | |
|         p = urllib.parse.urlsplit(uri)
 | |
|         self.assertEqual(p.netloc, b"")
 | |
|         self.assertEqual(p.username, None)
 | |
|         self.assertEqual(p.password, None)
 | |
|         self.assertEqual(p.hostname, None)
 | |
|         self.assertEqual(p.port, None)
 | |
|         self.assertEqual(p.geturl(), uri)
 | |
| 
 | |
|         p = urllib.parse.urlparse(uri)
 | |
|         self.assertEqual(p.netloc, b"")
 | |
|         self.assertEqual(p.username, None)
 | |
|         self.assertEqual(p.password, None)
 | |
|         self.assertEqual(p.hostname, None)
 | |
|         self.assertEqual(p.port, None)
 | |
|         self.assertEqual(p.geturl(), uri)
 | |
| 
 | |
|     def test_noslash(self):
 | |
|         # Issue 1637: http://foo.com?query is legal
 | |
|         self.assertEqual(urllib.parse.urlparse("http://example.com?blahblah=/foo"),
 | |
|                          ('http', 'example.com', '', '', 'blahblah=/foo', ''))
 | |
|         self.assertEqual(urllib.parse.urlparse(b"http://example.com?blahblah=/foo"),
 | |
|                          (b'http', b'example.com', b'', b'', b'blahblah=/foo', b''))
 | |
| 
 | |
|     def test_withoutscheme(self):
 | |
|         # Test urlparse without scheme
 | |
|         # Issue 754016: urlparse goes wrong with IP:port without scheme
 | |
|         # RFC 1808 specifies that netloc should start with //, urlparse expects
 | |
|         # the same, otherwise it classifies the portion of url as path.
 | |
|         self.assertEqual(urllib.parse.urlparse("path"),
 | |
|                 ('','','path','','',''))
 | |
|         self.assertEqual(urllib.parse.urlparse("//www.python.org:80"),
 | |
|                 ('','www.python.org:80','','','',''))
 | |
|         self.assertEqual(urllib.parse.urlparse("http://www.python.org:80"),
 | |
|                 ('http','www.python.org:80','','','',''))
 | |
|         # Repeat for bytes input
 | |
|         self.assertEqual(urllib.parse.urlparse(b"path"),
 | |
|                 (b'',b'',b'path',b'',b'',b''))
 | |
|         self.assertEqual(urllib.parse.urlparse(b"//www.python.org:80"),
 | |
|                 (b'',b'www.python.org:80',b'',b'',b'',b''))
 | |
|         self.assertEqual(urllib.parse.urlparse(b"http://www.python.org:80"),
 | |
|                 (b'http',b'www.python.org:80',b'',b'',b'',b''))
 | |
| 
 | |
|     def test_portseparator(self):
 | |
|         # Issue 754016 makes changes for port separator ':' from scheme separator
 | |
|         self.assertEqual(urllib.parse.urlparse("http:80"), ('http','','80','','',''))
 | |
|         self.assertEqual(urllib.parse.urlparse("https:80"), ('https','','80','','',''))
 | |
|         self.assertEqual(urllib.parse.urlparse("path:80"), ('path','','80','','',''))
 | |
|         self.assertEqual(urllib.parse.urlparse("http:"),('http','','','','',''))
 | |
|         self.assertEqual(urllib.parse.urlparse("https:"),('https','','','','',''))
 | |
|         self.assertEqual(urllib.parse.urlparse("http://www.python.org:80"),
 | |
|                 ('http','www.python.org:80','','','',''))
 | |
|         # As usual, need to check bytes input as well
 | |
|         self.assertEqual(urllib.parse.urlparse(b"http:80"), (b'http',b'',b'80',b'',b'',b''))
 | |
|         self.assertEqual(urllib.parse.urlparse(b"https:80"), (b'https',b'',b'80',b'',b'',b''))
 | |
|         self.assertEqual(urllib.parse.urlparse(b"path:80"), (b'path',b'',b'80',b'',b'',b''))
 | |
|         self.assertEqual(urllib.parse.urlparse(b"http:"),(b'http',b'',b'',b'',b'',b''))
 | |
|         self.assertEqual(urllib.parse.urlparse(b"https:"),(b'https',b'',b'',b'',b'',b''))
 | |
|         self.assertEqual(urllib.parse.urlparse(b"http://www.python.org:80"),
 | |
|                 (b'http',b'www.python.org:80',b'',b'',b'',b''))
 | |
| 
 | |
|     def test_usingsys(self):
 | |
|         # Issue 3314: sys module is used in the error
 | |
|         self.assertRaises(TypeError, urllib.parse.urlencode, "foo")
 | |
| 
 | |
|     def test_anyscheme(self):
 | |
|         # Issue 7904: s3://foo.com/stuff has netloc "foo.com".
 | |
|         self.assertEqual(urllib.parse.urlparse("s3://foo.com/stuff"),
 | |
|                          ('s3', 'foo.com', '/stuff', '', '', ''))
 | |
|         self.assertEqual(urllib.parse.urlparse("x-newscheme://foo.com/stuff"),
 | |
|                          ('x-newscheme', 'foo.com', '/stuff', '', '', ''))
 | |
|         self.assertEqual(urllib.parse.urlparse("x-newscheme://foo.com/stuff?query#fragment"),
 | |
|                          ('x-newscheme', 'foo.com', '/stuff', '', 'query', 'fragment'))
 | |
|         self.assertEqual(urllib.parse.urlparse("x-newscheme://foo.com/stuff?query"),
 | |
|                          ('x-newscheme', 'foo.com', '/stuff', '', 'query', ''))
 | |
| 
 | |
|         # And for bytes...
 | |
|         self.assertEqual(urllib.parse.urlparse(b"s3://foo.com/stuff"),
 | |
|                          (b's3', b'foo.com', b'/stuff', b'', b'', b''))
 | |
|         self.assertEqual(urllib.parse.urlparse(b"x-newscheme://foo.com/stuff"),
 | |
|                          (b'x-newscheme', b'foo.com', b'/stuff', b'', b'', b''))
 | |
|         self.assertEqual(urllib.parse.urlparse(b"x-newscheme://foo.com/stuff?query#fragment"),
 | |
|                          (b'x-newscheme', b'foo.com', b'/stuff', b'', b'query', b'fragment'))
 | |
|         self.assertEqual(urllib.parse.urlparse(b"x-newscheme://foo.com/stuff?query"),
 | |
|                          (b'x-newscheme', b'foo.com', b'/stuff', b'', b'query', b''))
 | |
| 
 | |
|     def test_default_scheme(self):
 | |
|         # Exercise the scheme parameter of urlparse() and urlsplit()
 | |
|         for func in (urllib.parse.urlparse, urllib.parse.urlsplit):
 | |
|             with self.subTest(function=func):
 | |
|                 result = func("http://example.net/", "ftp")
 | |
|                 self.assertEqual(result.scheme, "http")
 | |
|                 result = func(b"http://example.net/", b"ftp")
 | |
|                 self.assertEqual(result.scheme, b"http")
 | |
|                 self.assertEqual(func("path", "ftp").scheme, "ftp")
 | |
|                 self.assertEqual(func("path", scheme="ftp").scheme, "ftp")
 | |
|                 self.assertEqual(func(b"path", scheme=b"ftp").scheme, b"ftp")
 | |
|                 self.assertEqual(func("path").scheme, "")
 | |
|                 self.assertEqual(func(b"path").scheme, b"")
 | |
|                 self.assertEqual(func(b"path", "").scheme, b"")
 | |
| 
 | |
|     def test_parse_fragments(self):
 | |
|         # Exercise the allow_fragments parameter of urlparse() and urlsplit()
 | |
|         tests = (
 | |
|             ("http:#frag", "path", "frag"),
 | |
|             ("//example.net#frag", "path", "frag"),
 | |
|             ("index.html#frag", "path", "frag"),
 | |
|             (";a=b#frag", "params", "frag"),
 | |
|             ("?a=b#frag", "query", "frag"),
 | |
|             ("#frag", "path", "frag"),
 | |
|             ("abc#@frag", "path", "@frag"),
 | |
|             ("//abc#@frag", "path", "@frag"),
 | |
|             ("//abc:80#@frag", "path", "@frag"),
 | |
|             ("//abc#@frag:80", "path", "@frag:80"),
 | |
|         )
 | |
|         for url, attr, expected_frag in tests:
 | |
|             for func in (urllib.parse.urlparse, urllib.parse.urlsplit):
 | |
|                 if attr == "params" and func is urllib.parse.urlsplit:
 | |
|                     attr = "path"
 | |
|                 with self.subTest(url=url, function=func):
 | |
|                     result = func(url, allow_fragments=False)
 | |
|                     self.assertEqual(result.fragment, "")
 | |
|                     self.assertTrue(
 | |
|                             getattr(result, attr).endswith("#" + expected_frag))
 | |
|                     self.assertEqual(func(url, "", False).fragment, "")
 | |
| 
 | |
|                     result = func(url, allow_fragments=True)
 | |
|                     self.assertEqual(result.fragment, expected_frag)
 | |
|                     self.assertFalse(
 | |
|                             getattr(result, attr).endswith(expected_frag))
 | |
|                     self.assertEqual(func(url, "", True).fragment,
 | |
|                                      expected_frag)
 | |
|                     self.assertEqual(func(url).fragment, expected_frag)
 | |
| 
 | |
|     def test_mixed_types_rejected(self):
 | |
|         # Several functions that process either strings or ASCII encoded bytes
 | |
|         # accept multiple arguments. Check they reject mixed type input
 | |
|         with self.assertRaisesRegex(TypeError, "Cannot mix str"):
 | |
|             urllib.parse.urlparse("www.python.org", b"http")
 | |
|         with self.assertRaisesRegex(TypeError, "Cannot mix str"):
 | |
|             urllib.parse.urlparse(b"www.python.org", "http")
 | |
|         with self.assertRaisesRegex(TypeError, "Cannot mix str"):
 | |
|             urllib.parse.urlsplit("www.python.org", b"http")
 | |
|         with self.assertRaisesRegex(TypeError, "Cannot mix str"):
 | |
|             urllib.parse.urlsplit(b"www.python.org", "http")
 | |
|         with self.assertRaisesRegex(TypeError, "Cannot mix str"):
 | |
|             urllib.parse.urlunparse(( b"http", "www.python.org","","","",""))
 | |
|         with self.assertRaisesRegex(TypeError, "Cannot mix str"):
 | |
|             urllib.parse.urlunparse(("http", b"www.python.org","","","",""))
 | |
|         with self.assertRaisesRegex(TypeError, "Cannot mix str"):
 | |
|             urllib.parse.urlunsplit((b"http", "www.python.org","","",""))
 | |
|         with self.assertRaisesRegex(TypeError, "Cannot mix str"):
 | |
|             urllib.parse.urlunsplit(("http", b"www.python.org","","",""))
 | |
|         with self.assertRaisesRegex(TypeError, "Cannot mix str"):
 | |
|             urllib.parse.urljoin("http://python.org", b"http://python.org")
 | |
|         with self.assertRaisesRegex(TypeError, "Cannot mix str"):
 | |
|             urllib.parse.urljoin(b"http://python.org", "http://python.org")
 | |
| 
 | |
|     def _check_result_type(self, str_type):
 | |
|         num_args = len(str_type._fields)
 | |
|         bytes_type = str_type._encoded_counterpart
 | |
|         self.assertIs(bytes_type._decoded_counterpart, str_type)
 | |
|         str_args = ('',) * num_args
 | |
|         bytes_args = (b'',) * num_args
 | |
|         str_result = str_type(*str_args)
 | |
|         bytes_result = bytes_type(*bytes_args)
 | |
|         encoding = 'ascii'
 | |
|         errors = 'strict'
 | |
|         self.assertEqual(str_result, str_args)
 | |
|         self.assertEqual(bytes_result.decode(), str_args)
 | |
|         self.assertEqual(bytes_result.decode(), str_result)
 | |
|         self.assertEqual(bytes_result.decode(encoding), str_args)
 | |
|         self.assertEqual(bytes_result.decode(encoding), str_result)
 | |
|         self.assertEqual(bytes_result.decode(encoding, errors), str_args)
 | |
|         self.assertEqual(bytes_result.decode(encoding, errors), str_result)
 | |
|         self.assertEqual(bytes_result, bytes_args)
 | |
|         self.assertEqual(str_result.encode(), bytes_args)
 | |
|         self.assertEqual(str_result.encode(), bytes_result)
 | |
|         self.assertEqual(str_result.encode(encoding), bytes_args)
 | |
|         self.assertEqual(str_result.encode(encoding), bytes_result)
 | |
|         self.assertEqual(str_result.encode(encoding, errors), bytes_args)
 | |
|         self.assertEqual(str_result.encode(encoding, errors), bytes_result)
 | |
| 
 | |
|     def test_result_pairs(self):
 | |
|         # Check encoding and decoding between result pairs
 | |
|         result_types = [
 | |
|           urllib.parse.DefragResult,
 | |
|           urllib.parse.SplitResult,
 | |
|           urllib.parse.ParseResult,
 | |
|         ]
 | |
|         for result_type in result_types:
 | |
|             self._check_result_type(result_type)
 | |
| 
 | |
|     def test_parse_qs_encoding(self):
 | |
|         result = urllib.parse.parse_qs("key=\u0141%E9", encoding="latin-1")
 | |
|         self.assertEqual(result, {'key': ['\u0141\xE9']})
 | |
|         result = urllib.parse.parse_qs("key=\u0141%C3%A9", encoding="utf-8")
 | |
|         self.assertEqual(result, {'key': ['\u0141\xE9']})
 | |
|         result = urllib.parse.parse_qs("key=\u0141%C3%A9", encoding="ascii")
 | |
|         self.assertEqual(result, {'key': ['\u0141\ufffd\ufffd']})
 | |
|         result = urllib.parse.parse_qs("key=\u0141%E9-", encoding="ascii")
 | |
|         self.assertEqual(result, {'key': ['\u0141\ufffd-']})
 | |
|         result = urllib.parse.parse_qs("key=\u0141%E9-", encoding="ascii",
 | |
|                                                           errors="ignore")
 | |
|         self.assertEqual(result, {'key': ['\u0141-']})
 | |
| 
 | |
|     def test_parse_qsl_encoding(self):
 | |
|         result = urllib.parse.parse_qsl("key=\u0141%E9", encoding="latin-1")
 | |
|         self.assertEqual(result, [('key', '\u0141\xE9')])
 | |
|         result = urllib.parse.parse_qsl("key=\u0141%C3%A9", encoding="utf-8")
 | |
|         self.assertEqual(result, [('key', '\u0141\xE9')])
 | |
|         result = urllib.parse.parse_qsl("key=\u0141%C3%A9", encoding="ascii")
 | |
|         self.assertEqual(result, [('key', '\u0141\ufffd\ufffd')])
 | |
|         result = urllib.parse.parse_qsl("key=\u0141%E9-", encoding="ascii")
 | |
|         self.assertEqual(result, [('key', '\u0141\ufffd-')])
 | |
|         result = urllib.parse.parse_qsl("key=\u0141%E9-", encoding="ascii",
 | |
|                                                           errors="ignore")
 | |
|         self.assertEqual(result, [('key', '\u0141-')])
 | |
| 
 | |
|     def test_parse_qsl_max_num_fields(self):
 | |
|         with self.assertRaises(ValueError):
 | |
|             urllib.parse.parse_qsl('&'.join(['a=a']*11), max_num_fields=10)
 | |
|         urllib.parse.parse_qsl('&'.join(['a=a']*10), max_num_fields=10)
 | |
| 
 | |
|     def test_parse_qs_separator(self):
 | |
|         parse_qs_semicolon_cases = [
 | |
|             (";", {}),
 | |
|             (";;", {}),
 | |
|             (";a=b", {'a': ['b']}),
 | |
|             ("a=a+b;b=b+c", {'a': ['a b'], 'b': ['b c']}),
 | |
|             ("a=1;a=2", {'a': ['1', '2']}),
 | |
|             (b";", {}),
 | |
|             (b";;", {}),
 | |
|             (b";a=b", {b'a': [b'b']}),
 | |
|             (b"a=a+b;b=b+c", {b'a': [b'a b'], b'b': [b'b c']}),
 | |
|             (b"a=1;a=2", {b'a': [b'1', b'2']}),
 | |
|         ]
 | |
|         for orig, expect in parse_qs_semicolon_cases:
 | |
|             with self.subTest(f"Original: {orig!r}, Expected: {expect!r}"):
 | |
|                 result = urllib.parse.parse_qs(orig, separator=';')
 | |
|                 self.assertEqual(result, expect, "Error parsing %r" % orig)
 | |
|                 result_bytes = urllib.parse.parse_qs(orig, separator=b';')
 | |
|                 self.assertEqual(result_bytes, expect, "Error parsing %r" % orig)
 | |
| 
 | |
| 
 | |
|     def test_parse_qsl_separator(self):
 | |
|         parse_qsl_semicolon_cases = [
 | |
|             (";", []),
 | |
|             (";;", []),
 | |
|             (";a=b", [('a', 'b')]),
 | |
|             ("a=a+b;b=b+c", [('a', 'a b'), ('b', 'b c')]),
 | |
|             ("a=1;a=2", [('a', '1'), ('a', '2')]),
 | |
|             (b";", []),
 | |
|             (b";;", []),
 | |
|             (b";a=b", [(b'a', b'b')]),
 | |
|             (b"a=a+b;b=b+c", [(b'a', b'a b'), (b'b', b'b c')]),
 | |
|             (b"a=1;a=2", [(b'a', b'1'), (b'a', b'2')]),
 | |
|         ]
 | |
|         for orig, expect in parse_qsl_semicolon_cases:
 | |
|             with self.subTest(f"Original: {orig!r}, Expected: {expect!r}"):
 | |
|                 result = urllib.parse.parse_qsl(orig, separator=';')
 | |
|                 self.assertEqual(result, expect, "Error parsing %r" % orig)
 | |
|                 result_bytes = urllib.parse.parse_qsl(orig, separator=b';')
 | |
|                 self.assertEqual(result_bytes, expect, "Error parsing %r" % orig)
 | |
| 
 | |
|     def test_parse_qsl_bytes(self):
 | |
|         self.assertEqual(urllib.parse.parse_qsl(b'a=b'), [(b'a', b'b')])
 | |
|         self.assertEqual(urllib.parse.parse_qsl(bytearray(b'a=b')), [(b'a', b'b')])
 | |
|         self.assertEqual(urllib.parse.parse_qsl(memoryview(b'a=b')), [(b'a', b'b')])
 | |
| 
 | |
|     def test_parse_qsl_false_value(self):
 | |
|         kwargs = dict(keep_blank_values=True, strict_parsing=True)
 | |
|         for x in '', b'', None, 0, 0.0, [], {}, memoryview(b''):
 | |
|             self.assertEqual(urllib.parse.parse_qsl(x, **kwargs), [])
 | |
|             self.assertRaises(ValueError, urllib.parse.parse_qsl, x, separator=1)
 | |
| 
 | |
|     def test_parse_qsl_errors(self):
 | |
|         self.assertRaises(TypeError, urllib.parse.parse_qsl, list(b'a=b'))
 | |
|         self.assertRaises(TypeError, urllib.parse.parse_qsl, iter(b'a=b'))
 | |
|         self.assertRaises(TypeError, urllib.parse.parse_qsl, 1)
 | |
|         self.assertRaises(TypeError, urllib.parse.parse_qsl, object())
 | |
| 
 | |
|         for separator in '', b'', None, 0, 1, 0.0, 1.5:
 | |
|             with self.assertRaises(ValueError):
 | |
|                 urllib.parse.parse_qsl('a=b', separator=separator)
 | |
|         with self.assertRaises(UnicodeEncodeError):
 | |
|             urllib.parse.parse_qsl(b'a=b', separator='\xa6')
 | |
|         with self.assertRaises(UnicodeDecodeError):
 | |
|             urllib.parse.parse_qsl('a=b', separator=b'\xa6')
 | |
| 
 | |
|     def test_urlencode_sequences(self):
 | |
|         # Other tests incidentally urlencode things; test non-covered cases:
 | |
|         # Sequence and object values.
 | |
|         result = urllib.parse.urlencode({'a': [1, 2], 'b': (3, 4, 5)}, True)
 | |
|         # we cannot rely on ordering here
 | |
|         assert set(result.split('&')) == {'a=1', 'a=2', 'b=3', 'b=4', 'b=5'}
 | |
| 
 | |
|         class Trivial:
 | |
|             def __str__(self):
 | |
|                 return 'trivial'
 | |
| 
 | |
|         result = urllib.parse.urlencode({'a': Trivial()}, True)
 | |
|         self.assertEqual(result, 'a=trivial')
 | |
| 
 | |
|     def test_urlencode_quote_via(self):
 | |
|         result = urllib.parse.urlencode({'a': 'some value'})
 | |
|         self.assertEqual(result, "a=some+value")
 | |
|         result = urllib.parse.urlencode({'a': 'some value/another'},
 | |
|                                         quote_via=urllib.parse.quote)
 | |
|         self.assertEqual(result, "a=some%20value%2Fanother")
 | |
|         result = urllib.parse.urlencode({'a': 'some value/another'},
 | |
|                                         safe='/', quote_via=urllib.parse.quote)
 | |
|         self.assertEqual(result, "a=some%20value/another")
 | |
| 
 | |
|     def test_quote_from_bytes(self):
 | |
|         self.assertRaises(TypeError, urllib.parse.quote_from_bytes, 'foo')
 | |
|         result = urllib.parse.quote_from_bytes(b'archaeological arcana')
 | |
|         self.assertEqual(result, 'archaeological%20arcana')
 | |
|         result = urllib.parse.quote_from_bytes(b'')
 | |
|         self.assertEqual(result, '')
 | |
|         result = urllib.parse.quote_from_bytes(b'A'*10_000)
 | |
|         self.assertEqual(result, 'A'*10_000)
 | |
|         result = urllib.parse.quote_from_bytes(b'z\x01/ '*253_183)
 | |
|         self.assertEqual(result, 'z%01/%20'*253_183)
 | |
| 
 | |
|     def test_unquote_to_bytes(self):
 | |
|         result = urllib.parse.unquote_to_bytes('abc%20def')
 | |
|         self.assertEqual(result, b'abc def')
 | |
|         result = urllib.parse.unquote_to_bytes('')
 | |
|         self.assertEqual(result, b'')
 | |
| 
 | |
|     def test_quote_errors(self):
 | |
|         self.assertRaises(TypeError, urllib.parse.quote, b'foo',
 | |
|                           encoding='utf-8')
 | |
|         self.assertRaises(TypeError, urllib.parse.quote, b'foo', errors='strict')
 | |
| 
 | |
|     def test_issue14072(self):
 | |
|         p1 = urllib.parse.urlsplit('tel:+31-641044153')
 | |
|         self.assertEqual(p1.scheme, 'tel')
 | |
|         self.assertEqual(p1.path, '+31-641044153')
 | |
|         p2 = urllib.parse.urlsplit('tel:+31641044153')
 | |
|         self.assertEqual(p2.scheme, 'tel')
 | |
|         self.assertEqual(p2.path, '+31641044153')
 | |
|         # assert the behavior for urlparse
 | |
|         p1 = urllib.parse.urlparse('tel:+31-641044153')
 | |
|         self.assertEqual(p1.scheme, 'tel')
 | |
|         self.assertEqual(p1.path, '+31-641044153')
 | |
|         p2 = urllib.parse.urlparse('tel:+31641044153')
 | |
|         self.assertEqual(p2.scheme, 'tel')
 | |
|         self.assertEqual(p2.path, '+31641044153')
 | |
| 
 | |
|     def test_invalid_bracketed_hosts(self):
 | |
|         self.assertRaises(ValueError, urllib.parse.urlsplit, 'Scheme://user@[192.0.2.146]/Path?Query')
 | |
|         self.assertRaises(ValueError, urllib.parse.urlsplit, 'Scheme://user@[important.com:8000]/Path?Query')
 | |
|         self.assertRaises(ValueError, urllib.parse.urlsplit, 'Scheme://user@[v123r.IP]/Path?Query')
 | |
|         self.assertRaises(ValueError, urllib.parse.urlsplit, 'Scheme://user@[v12ae]/Path?Query')
 | |
|         self.assertRaises(ValueError, urllib.parse.urlsplit, 'Scheme://user@[v.IP]/Path?Query')
 | |
|         self.assertRaises(ValueError, urllib.parse.urlsplit, 'Scheme://user@[v123.]/Path?Query')
 | |
|         self.assertRaises(ValueError, urllib.parse.urlsplit, 'Scheme://user@[v]/Path?Query')
 | |
|         self.assertRaises(ValueError, urllib.parse.urlsplit, 'Scheme://user@[0439:23af::2309::fae7:1234]/Path?Query')
 | |
|         self.assertRaises(ValueError, urllib.parse.urlsplit, 'Scheme://user@[0439:23af:2309::fae7:1234:2342:438e:192.0.2.146]/Path?Query')
 | |
|         self.assertRaises(ValueError, urllib.parse.urlsplit, 'Scheme://user@]v6a.ip[/Path')
 | |
| 
 | |
|     def test_splitting_bracketed_hosts(self):
 | |
|         p1 = urllib.parse.urlsplit('scheme://user@[v6a.ip]/path?query')
 | |
|         self.assertEqual(p1.hostname, 'v6a.ip')
 | |
|         self.assertEqual(p1.username, 'user')
 | |
|         self.assertEqual(p1.path, '/path')
 | |
|         p2 = urllib.parse.urlsplit('scheme://user@[0439:23af:2309::fae7%test]/path?query')
 | |
|         self.assertEqual(p2.hostname, '0439:23af:2309::fae7%test')
 | |
|         self.assertEqual(p2.username, 'user')
 | |
|         self.assertEqual(p2.path, '/path')
 | |
|         p3 = urllib.parse.urlsplit('scheme://user@[0439:23af:2309::fae7:1234:192.0.2.146%test]/path?query')
 | |
|         self.assertEqual(p3.hostname, '0439:23af:2309::fae7:1234:192.0.2.146%test')
 | |
|         self.assertEqual(p3.username, 'user')
 | |
|         self.assertEqual(p3.path, '/path')
 | |
| 
 | |
|     def test_port_casting_failure_message(self):
 | |
|         message = "Port could not be cast to integer value as 'oracle'"
 | |
|         p1 = urllib.parse.urlparse('http://Server=sde; Service=sde:oracle')
 | |
|         with self.assertRaisesRegex(ValueError, message):
 | |
|             p1.port
 | |
| 
 | |
|         p2 = urllib.parse.urlsplit('http://Server=sde; Service=sde:oracle')
 | |
|         with self.assertRaisesRegex(ValueError, message):
 | |
|             p2.port
 | |
| 
 | |
|     def test_telurl_params(self):
 | |
|         p1 = urllib.parse.urlparse('tel:123-4;phone-context=+1-650-516')
 | |
|         self.assertEqual(p1.scheme, 'tel')
 | |
|         self.assertEqual(p1.path, '123-4')
 | |
|         self.assertEqual(p1.params, 'phone-context=+1-650-516')
 | |
| 
 | |
|         p1 = urllib.parse.urlparse('tel:+1-201-555-0123')
 | |
|         self.assertEqual(p1.scheme, 'tel')
 | |
|         self.assertEqual(p1.path, '+1-201-555-0123')
 | |
|         self.assertEqual(p1.params, '')
 | |
| 
 | |
|         p1 = urllib.parse.urlparse('tel:7042;phone-context=example.com')
 | |
|         self.assertEqual(p1.scheme, 'tel')
 | |
|         self.assertEqual(p1.path, '7042')
 | |
|         self.assertEqual(p1.params, 'phone-context=example.com')
 | |
| 
 | |
|         p1 = urllib.parse.urlparse('tel:863-1234;phone-context=+1-914-555')
 | |
|         self.assertEqual(p1.scheme, 'tel')
 | |
|         self.assertEqual(p1.path, '863-1234')
 | |
|         self.assertEqual(p1.params, 'phone-context=+1-914-555')
 | |
| 
 | |
|     def test_Quoter_repr(self):
 | |
|         quoter = urllib.parse._Quoter(urllib.parse._ALWAYS_SAFE)
 | |
|         self.assertIn('Quoter', repr(quoter))
 | |
| 
 | |
|     def test_clear_cache_for_code_coverage(self):
 | |
|         urllib.parse.clear_cache()
 | |
| 
 | |
|     def test_urllib_parse_getattr_failure(self):
 | |
|         """Test that urllib.parse.__getattr__() fails correctly."""
 | |
|         with self.assertRaises(AttributeError):
 | |
|             unused = urllib.parse.this_does_not_exist
 | |
| 
 | |
|     def test_all(self):
 | |
|         expected = []
 | |
|         undocumented = {
 | |
|             'splitattr', 'splithost', 'splitnport', 'splitpasswd',
 | |
|             'splitport', 'splitquery', 'splittag', 'splittype', 'splituser',
 | |
|             'splitvalue',
 | |
|             'ResultBase', 'clear_cache', 'to_bytes', 'unwrap',
 | |
|         }
 | |
|         for name in dir(urllib.parse):
 | |
|             if name.startswith('_') or name in undocumented:
 | |
|                 continue
 | |
|             object = getattr(urllib.parse, name)
 | |
|             if getattr(object, '__module__', None) == 'urllib.parse':
 | |
|                 expected.append(name)
 | |
|         self.assertCountEqual(urllib.parse.__all__, expected)
 | |
| 
 | |
|     def test_urlsplit_normalization(self):
 | |
|         # Certain characters should never occur in the netloc,
 | |
|         # including under normalization.
 | |
|         # Ensure that ALL of them are detected and cause an error
 | |
|         illegal_chars = '/:#?@'
 | |
|         hex_chars = {'{:04X}'.format(ord(c)) for c in illegal_chars}
 | |
|         denorm_chars = [
 | |
|             c for c in map(chr, range(128, sys.maxunicode))
 | |
|             if unicodedata.decomposition(c)
 | |
|             and (hex_chars & set(unicodedata.decomposition(c).split()))
 | |
|             and c not in illegal_chars
 | |
|         ]
 | |
|         # Sanity check that we found at least one such character
 | |
|         self.assertIn('\u2100', denorm_chars)
 | |
|         self.assertIn('\uFF03', denorm_chars)
 | |
| 
 | |
|         # bpo-36742: Verify port separators are ignored when they
 | |
|         # existed prior to decomposition
 | |
|         urllib.parse.urlsplit('http://\u30d5\u309a:80')
 | |
|         with self.assertRaises(ValueError):
 | |
|             urllib.parse.urlsplit('http://\u30d5\u309a\ufe1380')
 | |
| 
 | |
|         for scheme in ["http", "https", "ftp"]:
 | |
|             for netloc in ["netloc{}false.netloc", "n{}user@netloc"]:
 | |
|                 for c in denorm_chars:
 | |
|                     url = "{}://{}/path".format(scheme, netloc.format(c))
 | |
|                     with self.subTest(url=url, char='{:04X}'.format(ord(c))):
 | |
|                         with self.assertRaises(ValueError):
 | |
|                             urllib.parse.urlsplit(url)
 | |
| 
 | |
| class Utility_Tests(unittest.TestCase):
 | |
|     """Testcase to test the various utility functions in the urllib."""
 | |
|     # In Python 2 this test class was in test_urllib.
 | |
| 
 | |
|     def test_splittype(self):
 | |
|         splittype = urllib.parse._splittype
 | |
|         self.assertEqual(splittype('type:opaquestring'), ('type', 'opaquestring'))
 | |
|         self.assertEqual(splittype('opaquestring'), (None, 'opaquestring'))
 | |
|         self.assertEqual(splittype(':opaquestring'), (None, ':opaquestring'))
 | |
|         self.assertEqual(splittype('type:'), ('type', ''))
 | |
|         self.assertEqual(splittype('type:opaque:string'), ('type', 'opaque:string'))
 | |
| 
 | |
|     def test_splithost(self):
 | |
|         splithost = urllib.parse._splithost
 | |
|         self.assertEqual(splithost('//www.example.org:80/foo/bar/baz.html'),
 | |
|                          ('www.example.org:80', '/foo/bar/baz.html'))
 | |
|         self.assertEqual(splithost('//www.example.org:80'),
 | |
|                          ('www.example.org:80', ''))
 | |
|         self.assertEqual(splithost('/foo/bar/baz.html'),
 | |
|                          (None, '/foo/bar/baz.html'))
 | |
| 
 | |
|         # bpo-30500: # starts a fragment.
 | |
|         self.assertEqual(splithost('//127.0.0.1#@host.com'),
 | |
|                          ('127.0.0.1', '/#@host.com'))
 | |
|         self.assertEqual(splithost('//127.0.0.1#@host.com:80'),
 | |
|                          ('127.0.0.1', '/#@host.com:80'))
 | |
|         self.assertEqual(splithost('//127.0.0.1:80#@host.com'),
 | |
|                          ('127.0.0.1:80', '/#@host.com'))
 | |
| 
 | |
|         # Empty host is returned as empty string.
 | |
|         self.assertEqual(splithost("///file"),
 | |
|                          ('', '/file'))
 | |
| 
 | |
|         # Trailing semicolon, question mark and hash symbol are kept.
 | |
|         self.assertEqual(splithost("//example.net/file;"),
 | |
|                          ('example.net', '/file;'))
 | |
|         self.assertEqual(splithost("//example.net/file?"),
 | |
|                          ('example.net', '/file?'))
 | |
|         self.assertEqual(splithost("//example.net/file#"),
 | |
|                          ('example.net', '/file#'))
 | |
| 
 | |
|     def test_splituser(self):
 | |
|         splituser = urllib.parse._splituser
 | |
|         self.assertEqual(splituser('User:Pass@www.python.org:080'),
 | |
|                          ('User:Pass', 'www.python.org:080'))
 | |
|         self.assertEqual(splituser('@www.python.org:080'),
 | |
|                          ('', 'www.python.org:080'))
 | |
|         self.assertEqual(splituser('www.python.org:080'),
 | |
|                          (None, 'www.python.org:080'))
 | |
|         self.assertEqual(splituser('User:Pass@'),
 | |
|                          ('User:Pass', ''))
 | |
|         self.assertEqual(splituser('User@example.com:Pass@www.python.org:080'),
 | |
|                          ('User@example.com:Pass', 'www.python.org:080'))
 | |
| 
 | |
|     def test_splitpasswd(self):
 | |
|         # Some of the password examples are not sensible, but it is added to
 | |
|         # confirming to RFC2617 and addressing issue4675.
 | |
|         splitpasswd = urllib.parse._splitpasswd
 | |
|         self.assertEqual(splitpasswd('user:ab'), ('user', 'ab'))
 | |
|         self.assertEqual(splitpasswd('user:a\nb'), ('user', 'a\nb'))
 | |
|         self.assertEqual(splitpasswd('user:a\tb'), ('user', 'a\tb'))
 | |
|         self.assertEqual(splitpasswd('user:a\rb'), ('user', 'a\rb'))
 | |
|         self.assertEqual(splitpasswd('user:a\fb'), ('user', 'a\fb'))
 | |
|         self.assertEqual(splitpasswd('user:a\vb'), ('user', 'a\vb'))
 | |
|         self.assertEqual(splitpasswd('user:a:b'), ('user', 'a:b'))
 | |
|         self.assertEqual(splitpasswd('user:a b'), ('user', 'a b'))
 | |
|         self.assertEqual(splitpasswd('user 2:ab'), ('user 2', 'ab'))
 | |
|         self.assertEqual(splitpasswd('user+1:a+b'), ('user+1', 'a+b'))
 | |
|         self.assertEqual(splitpasswd('user:'), ('user', ''))
 | |
|         self.assertEqual(splitpasswd('user'), ('user', None))
 | |
|         self.assertEqual(splitpasswd(':ab'), ('', 'ab'))
 | |
| 
 | |
|     def test_splitport(self):
 | |
|         splitport = urllib.parse._splitport
 | |
|         self.assertEqual(splitport('parrot:88'), ('parrot', '88'))
 | |
|         self.assertEqual(splitport('parrot'), ('parrot', None))
 | |
|         self.assertEqual(splitport('parrot:'), ('parrot', None))
 | |
|         self.assertEqual(splitport('127.0.0.1'), ('127.0.0.1', None))
 | |
|         self.assertEqual(splitport('parrot:cheese'), ('parrot:cheese', None))
 | |
|         self.assertEqual(splitport('[::1]:88'), ('[::1]', '88'))
 | |
|         self.assertEqual(splitport('[::1]'), ('[::1]', None))
 | |
|         self.assertEqual(splitport(':88'), ('', '88'))
 | |
| 
 | |
|     def test_splitnport(self):
 | |
|         splitnport = urllib.parse._splitnport
 | |
|         self.assertEqual(splitnport('parrot:88'), ('parrot', 88))
 | |
|         self.assertEqual(splitnport('parrot'), ('parrot', -1))
 | |
|         self.assertEqual(splitnport('parrot', 55), ('parrot', 55))
 | |
|         self.assertEqual(splitnport('parrot:'), ('parrot', -1))
 | |
|         self.assertEqual(splitnport('parrot:', 55), ('parrot', 55))
 | |
|         self.assertEqual(splitnport('127.0.0.1'), ('127.0.0.1', -1))
 | |
|         self.assertEqual(splitnport('127.0.0.1', 55), ('127.0.0.1', 55))
 | |
|         self.assertEqual(splitnport('parrot:cheese'), ('parrot', None))
 | |
|         self.assertEqual(splitnport('parrot:cheese', 55), ('parrot', None))
 | |
|         self.assertEqual(splitnport('parrot: +1_0 '), ('parrot', None))
 | |
| 
 | |
|     def test_splitquery(self):
 | |
|         # Normal cases are exercised by other tests; ensure that we also
 | |
|         # catch cases with no port specified (testcase ensuring coverage)
 | |
|         splitquery = urllib.parse._splitquery
 | |
|         self.assertEqual(splitquery('http://python.org/fake?foo=bar'),
 | |
|                          ('http://python.org/fake', 'foo=bar'))
 | |
|         self.assertEqual(splitquery('http://python.org/fake?foo=bar?'),
 | |
|                          ('http://python.org/fake?foo=bar', ''))
 | |
|         self.assertEqual(splitquery('http://python.org/fake'),
 | |
|                          ('http://python.org/fake', None))
 | |
|         self.assertEqual(splitquery('?foo=bar'), ('', 'foo=bar'))
 | |
| 
 | |
|     def test_splittag(self):
 | |
|         splittag = urllib.parse._splittag
 | |
|         self.assertEqual(splittag('http://example.com?foo=bar#baz'),
 | |
|                          ('http://example.com?foo=bar', 'baz'))
 | |
|         self.assertEqual(splittag('http://example.com?foo=bar#'),
 | |
|                          ('http://example.com?foo=bar', ''))
 | |
|         self.assertEqual(splittag('#baz'), ('', 'baz'))
 | |
|         self.assertEqual(splittag('http://example.com?foo=bar'),
 | |
|                          ('http://example.com?foo=bar', None))
 | |
|         self.assertEqual(splittag('http://example.com?foo=bar#baz#boo'),
 | |
|                          ('http://example.com?foo=bar#baz', 'boo'))
 | |
| 
 | |
|     def test_splitattr(self):
 | |
|         splitattr = urllib.parse._splitattr
 | |
|         self.assertEqual(splitattr('/path;attr1=value1;attr2=value2'),
 | |
|                          ('/path', ['attr1=value1', 'attr2=value2']))
 | |
|         self.assertEqual(splitattr('/path;'), ('/path', ['']))
 | |
|         self.assertEqual(splitattr(';attr1=value1;attr2=value2'),
 | |
|                          ('', ['attr1=value1', 'attr2=value2']))
 | |
|         self.assertEqual(splitattr('/path'), ('/path', []))
 | |
| 
 | |
|     def test_splitvalue(self):
 | |
|         # Normal cases are exercised by other tests; test pathological cases
 | |
|         # with no key/value pairs. (testcase ensuring coverage)
 | |
|         splitvalue = urllib.parse._splitvalue
 | |
|         self.assertEqual(splitvalue('foo=bar'), ('foo', 'bar'))
 | |
|         self.assertEqual(splitvalue('foo='), ('foo', ''))
 | |
|         self.assertEqual(splitvalue('=bar'), ('', 'bar'))
 | |
|         self.assertEqual(splitvalue('foobar'), ('foobar', None))
 | |
|         self.assertEqual(splitvalue('foo=bar=baz'), ('foo', 'bar=baz'))
 | |
| 
 | |
|     def test_to_bytes(self):
 | |
|         result = urllib.parse._to_bytes('http://www.python.org')
 | |
|         self.assertEqual(result, 'http://www.python.org')
 | |
|         self.assertRaises(UnicodeError, urllib.parse._to_bytes,
 | |
|                           'http://www.python.org/medi\u00e6val')
 | |
| 
 | |
|     def test_unwrap(self):
 | |
|         for wrapped_url in ('<URL:scheme://host/path>', '<scheme://host/path>',
 | |
|                             'URL:scheme://host/path', 'scheme://host/path'):
 | |
|             url = urllib.parse.unwrap(wrapped_url)
 | |
|             self.assertEqual(url, 'scheme://host/path')
 | |
| 
 | |
| 
 | |
| class DeprecationTest(unittest.TestCase):
 | |
| 
 | |
|     def test_Quoter_deprecation(self):
 | |
|         with self.assertWarns(DeprecationWarning) as cm:
 | |
|             old_class = urllib.parse.Quoter
 | |
|             self.assertIs(old_class, urllib.parse._Quoter)
 | |
|         self.assertIn('Quoter will be removed', str(cm.warning))
 | |
| 
 | |
|     def test_splittype_deprecation(self):
 | |
|         with self.assertWarns(DeprecationWarning) as cm:
 | |
|             urllib.parse.splittype('')
 | |
|         self.assertEqual(str(cm.warning),
 | |
|                          'urllib.parse.splittype() is deprecated as of 3.8, '
 | |
|                          'use urllib.parse.urlparse() instead')
 | |
| 
 | |
|     def test_splithost_deprecation(self):
 | |
|         with self.assertWarns(DeprecationWarning) as cm:
 | |
|             urllib.parse.splithost('')
 | |
|         self.assertEqual(str(cm.warning),
 | |
|                          'urllib.parse.splithost() is deprecated as of 3.8, '
 | |
|                          'use urllib.parse.urlparse() instead')
 | |
| 
 | |
|     def test_splituser_deprecation(self):
 | |
|         with self.assertWarns(DeprecationWarning) as cm:
 | |
|             urllib.parse.splituser('')
 | |
|         self.assertEqual(str(cm.warning),
 | |
|                          'urllib.parse.splituser() is deprecated as of 3.8, '
 | |
|                          'use urllib.parse.urlparse() instead')
 | |
| 
 | |
|     def test_splitpasswd_deprecation(self):
 | |
|         with self.assertWarns(DeprecationWarning) as cm:
 | |
|             urllib.parse.splitpasswd('')
 | |
|         self.assertEqual(str(cm.warning),
 | |
|                          'urllib.parse.splitpasswd() is deprecated as of 3.8, '
 | |
|                          'use urllib.parse.urlparse() instead')
 | |
| 
 | |
|     def test_splitport_deprecation(self):
 | |
|         with self.assertWarns(DeprecationWarning) as cm:
 | |
|             urllib.parse.splitport('')
 | |
|         self.assertEqual(str(cm.warning),
 | |
|                          'urllib.parse.splitport() is deprecated as of 3.8, '
 | |
|                          'use urllib.parse.urlparse() instead')
 | |
| 
 | |
|     def test_splitnport_deprecation(self):
 | |
|         with self.assertWarns(DeprecationWarning) as cm:
 | |
|             urllib.parse.splitnport('')
 | |
|         self.assertEqual(str(cm.warning),
 | |
|                          'urllib.parse.splitnport() is deprecated as of 3.8, '
 | |
|                          'use urllib.parse.urlparse() instead')
 | |
| 
 | |
|     def test_splitquery_deprecation(self):
 | |
|         with self.assertWarns(DeprecationWarning) as cm:
 | |
|             urllib.parse.splitquery('')
 | |
|         self.assertEqual(str(cm.warning),
 | |
|                          'urllib.parse.splitquery() is deprecated as of 3.8, '
 | |
|                          'use urllib.parse.urlparse() instead')
 | |
| 
 | |
|     def test_splittag_deprecation(self):
 | |
|         with self.assertWarns(DeprecationWarning) as cm:
 | |
|             urllib.parse.splittag('')
 | |
|         self.assertEqual(str(cm.warning),
 | |
|                          'urllib.parse.splittag() is deprecated as of 3.8, '
 | |
|                          'use urllib.parse.urlparse() instead')
 | |
| 
 | |
|     def test_splitattr_deprecation(self):
 | |
|         with self.assertWarns(DeprecationWarning) as cm:
 | |
|             urllib.parse.splitattr('')
 | |
|         self.assertEqual(str(cm.warning),
 | |
|                          'urllib.parse.splitattr() is deprecated as of 3.8, '
 | |
|                          'use urllib.parse.urlparse() instead')
 | |
| 
 | |
|     def test_splitvalue_deprecation(self):
 | |
|         with self.assertWarns(DeprecationWarning) as cm:
 | |
|             urllib.parse.splitvalue('')
 | |
|         self.assertEqual(str(cm.warning),
 | |
|                          'urllib.parse.splitvalue() is deprecated as of 3.8, '
 | |
|                          'use urllib.parse.parse_qsl() instead')
 | |
| 
 | |
|     def test_to_bytes_deprecation(self):
 | |
|         with self.assertWarns(DeprecationWarning) as cm:
 | |
|             urllib.parse.to_bytes('')
 | |
|         self.assertEqual(str(cm.warning),
 | |
|                          'urllib.parse.to_bytes() is deprecated as of 3.8')
 | |
| 
 | |
| 
 | |
| if __name__ == "__main__":
 | |
|     unittest.main()
 |