Home | History | Annotate | Download | only in test

Lines Matching refs:http

22 import http.client
96 url = "http://example.com"
108 backwards-compatibility. (In the HTTP case, normalization to
110 http.client).
119 url = "http://example.com"
145 add("Some Realm", "http://example.com/", "joe", "password")
146 add("Some Realm", "http://example.com/ni", "ni", "ni")
147 add("Some Realm", "http://c.example.com:3128", "3", "c")
154 self.assertEqual(find_user_pass("Some Realm", "http://example.com/ni"),
156 self.assertEqual(find_user_pass("Some Realm", "http://example.com"),
158 self.assertEqual(find_user_pass("Some Realm", "http://example.com/"),
161 "http://example.com/spam"),
165 "http://example.com/spam/spam"),
170 add("c", "http://example.com/foo", "foo", "ni")
171 add("c", "http://example.com/bar", "bar", "nini")
173 self.assertEqual(find_user_pass("c", "http://example.com/foo"),
176 self.assertEqual(find_user_pass("c", "http://example.com/bar"),
181 add("b", "http://example.com/", "first", "blah")
182 add("b", "http://example.com/", "second", "spam")
184 self.assertEqual(find_user_pass("b", "http://example.com/"),
189 add("a", "http://example.com", "1", "a")
190 self.assertEqual(find_user_pass("a", "http://example.com/"),
193 self.assertEqual(find_user_pass("a", "http://a.example.com/"),
203 find_user_pass("Some Realm", "http://c.example.com:3128"),
219 add("f", "http://g.example.com:80", "10", "j")
220 add("g", "http://h.example.com", "11", "k")
229 self.assertEqual(find_user_pass("f", "http://g.example.com:100"),
231 self.assertEqual(find_user_pass("f", "http://g.example.com:80"),
233 self.assertEqual(find_user_pass("f", "http://g.example.com"),
237 self.assertEqual(find_user_pass("g", "http://h.example.com:80"),
241 self.assertEqual(find_user_pass("h", "http://i.example.com:80"),
246 self.assertEqual(find_user_pass("i", "http://j.example.com"),
248 self.assertEqual(find_user_pass("i", "http://j.example.com:80"),
398 return Request("http://blah/")
406 return self.parent.error("http", args[0], res, code, "", {})
484 name = http.client.responses[self.code]
487 "http", req, MockFile(), self.code, name, msg)
520 name = http.client.responses[self.code]
578 req = Request("http://example.com/")
606 o.open("http://example.com/")
620 req = Request("http://example.com/")
626 # http errors are a special case
641 req = Request("http://example.com/")
663 req = Request("http://example.com/")
863 url = "http://example.com/"
868 http = MockHTTPClass()
869 r = h.do_open(http, req)
879 self.assertEqual(http.host, "example.com")
880 self.assertEqual(http.level, 0)
881 self.assertEqual(http.method, method)
882 self.assertEqual(http.selector, "/")
883 self.assertEqual(http.req_headers,
886 self.assertEqual(http.data, data)
889 http.raise_on_endheaders = True
890 self.assertRaises(urllib.error.URLError, h.do_open, http, req)
893 req = Request("http://example.com/","badpost")
899 req = Request("http://example.com/", data)
937 req = Request("http://example.com/", f, {})
944 req = Request("http://example.com/", f, {"Content-Length": 30})
959 req = Request("http://example.com/", file_obj, {})
965 req = Request("http://example.com/", file_obj, headers)
984 req = Request("http://example.com/", proc.stdout, headers)
1005 req = Request("http://example.com/", iterable_body(), headers)
1018 req = h.do_request_(Request("http://example.com/", ()))
1031 req = Request("http://example.com/", iterable_array, headers)
1051 "http://example.com/foo/bar/baz.html",
1052 "http://example.com//foo/bar/baz.html",
1053 "http://example.com/foo//bar/baz.html",
1054 "http://example.com/foo/bar//baz.html"
1074 'http://example.com?foo=bar#baz',
1075 'http://example.com?foo=bar&spam=eggs#bash',
1076 'http://example.com',
1081 r = Request('http://example.com')
1093 r = Request('http://www.example.com')
1106 weird_url = 'http://www.python.org?getspam'
1112 url_without_path = 'http://www.python.org'
1122 url = "http://example.com/"
1140 self.assertEqual(o.proto, "http") # o.error called
1148 req = Request("http://example.com/")
1161 from_url = "http://example.com/a.html"
1162 to_url = "http://example.com/b.html"
1216 redirect(h, req, "http://example.com/")
1228 redirect(h, req, "http://example.com/%d" % count)
1235 from_url = "http://example.com/a.html"
1236 valid_schemes = ['http','https','ftp']
1257 from_url = "http://example.com/a.html"
1271 from http.cookiejar import CookieJar
1275 interact_netscape(cj, "http://www.example.com/", "spam=eggs")
1276 hh = MockHTTPHandler(302, "Location: http://www.cracker.com/\r\n\r\n")
1281 o.open("http://www.example.com/")
1285 redirected_url = 'http://www.example.com/index.html#OK\r\n\r\n'
1290 fp = o.open('http://www.example.com')
1295 real_class = http.client.HTTPConnection
1296 response1 = b"HTTP/1.1 302 Found\r\nLocation: ?query\r\n\r\n"
1297 http.client.HTTPConnection = test_urllib.fakehttp(response1)
1298 self.addCleanup(setattr, http.client, "HTTPConnection", real_class)
1304 conn.__class__.fakedata = b"HTTP/1.1 200 OK\r\n\r\nHello!"
1305 http.client.HTTPConnection.request = request
1306 fp = urllib.request.urlopen("http://python.org/path")
1307 self.assertEqual(fp.geturl(), "http://python.org/path?query")
1318 b'HTTP/1.1 200 OK\r\n'
1335 b'HTTP/1.1 302 Redirect\r\n'
1339 response = opener.open('http://example.com/')
1346 ph = urllib.request.ProxyHandler(dict(http="proxy.example.com:3128"))
1353 req = Request("http://acme.example.com/")
1364 ph = urllib.request.ProxyHandler(dict(http="proxy.example.com"))
1366 req = Request("http://www.perl.org/")
1370 req = Request("http://www.python.org")
1379 ph = urllib.request.ProxyHandler(dict(http="proxy.example.com"))
1381 req = Request("http://www.python.org")
1460 "http://acme.example.com/protected",
1461 "http://acme.example.com/protected",
1479 "http://acme.example.com/protected",
1480 "http://acme.example.com/protected",
1485 ph = urllib.request.ProxyHandler(dict(http="proxy.example.com:3128"))
1496 "http://acme.example.com:3128/protected",
1502 # response (http://python.org/sf/1479302), where it should instead
1506 # Also (http://python.org/sf/14797027, RFC 2617 section 1.2), we must
1543 "http://acme.example.com/protected",
1544 "http://acme.example.com/protected",
1558 self.assertRaises(ValueError, opener.open, "http://www.example.com")
1568 self.assertRaises(ValueError, opener.open, "http://www.example.com")
1600 # handle the HTTP auth error
1612 request_url = "http://acme.example.com/protected"
1637 request_url = 'http://acme.example.com/protected'
1680 header = "HTTP/1.1 200 OK\r\n{}\r\n\r\n".format(transfer)
1683 req = Request("http://dummy/")
1694 req = Request("http://dummy/")
1696 with self.assertRaises(http.client.BadStatusLine):
1754 with support.transient_internet("http://www.example.com/"):
1756 request = urllib.request.Request("http://www.example.com/")
1780 expected_errmsg = 'HTTP Error %s: %s' % (err.code, err.msg)
1801 ('http://proxy.example.com/',
1802 ('http', None, None, 'proxy.example.com')),
1803 ('http://proxy.example.com:3128/',
1804 ('http', None, None, 'proxy.example.com:3128')),
1805 ('http://joe:password@proxy.example.com/',
1806 ('http', 'joe', 'password', 'proxy.example.com')),
1807 ('http://joe:password@proxy.example.com:3128',
1808 ('http', 'joe', 'password', 'proxy.example.com:3128')),
1813 ('http://joe:password@proxy.example.com',
1814 ('http', 'joe', 'password', 'proxy.example.com'))
1837 self.get = Request("http://www.python.org/~jeremy/")
1838 self.post = Request("http://www.python.org/~jeremy/",
1841 self.head = Request("http://www.python.org/~jeremy/", method='HEAD')
1842 self.put = self.PutRequest("http://www.python.org/~jeremy/")
1843 self.force_post = self.PutRequest("http://www.python.org/~jeremy/",
1880 self.assertEqual("http://www.python.org/~jeremy/",
1885 req = Request("http://www.python.org/")
1889 self.assertEqual("http", self.get.type)
1895 req = Request("http://www.%70ython.org/")
1900 self.get.set_proxy("www.perl.org", "http")
1906 req = Request("<URL:http://www.python.org>")
1910 req = Request("http://www.python.org/?qs=query#fragment=true")
1912 req = Request("http://www.python.org/#fun=true")
1916 url = 'http://docs.python.org/library/urllib2.html#OK'
1921 urls = ['http://docs.python.org',
1922 'http://docs.python.org/library/urllib2.html#OK',
1923 'http://www.python.org/?qs=query#fragment=true']