Jellyfin(8096), OrbStack(8097) 포트 충돌으로 변경. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
1131 lines
42 KiB
Python
1131 lines
42 KiB
Python
from __future__ import annotations
|
|
|
|
import logging
|
|
import os
|
|
import re
|
|
import socket
|
|
import time
|
|
import typing
|
|
import warnings
|
|
from datetime import datetime, timedelta
|
|
from socket import timeout as SocketTimeout
|
|
|
|
|
|
from .util import ssl_
|
|
|
|
if typing.TYPE_CHECKING:
|
|
from typing_extensions import Literal
|
|
|
|
from .util.ssltransport import SSLTransport
|
|
from ._typing import (
|
|
_TYPE_BODY,
|
|
_TYPE_PEER_CERT_RET_DICT,
|
|
_TYPE_SOCKET_OPTIONS,
|
|
_TYPE_TIMEOUT_INTERNAL,
|
|
ProxyConfig,
|
|
)
|
|
from .util.traffic_police import TrafficPolice
|
|
from .backend._base import LowLevelResponse
|
|
|
|
from ._constant import DEFAULT_BLOCKSIZE, DEFAULT_KEEPALIVE_DELAY
|
|
from .response import HTTPResponse
|
|
from .util.timeout import _DEFAULT_TIMEOUT, Timeout
|
|
from .util.util import to_str
|
|
|
|
try: # Compiled with SSL?
|
|
import ssl
|
|
|
|
except (ImportError, AttributeError):
|
|
ssl = None # type: ignore[assignment]
|
|
|
|
from ._version import __version__
|
|
from .backend import HfaceBackend, HttpVersion, QuicPreemptiveCacheType, ResponsePromise
|
|
from .contrib.resolver import BaseResolver, ResolverDescription
|
|
from .exceptions import BaseSSLError # noqa
|
|
from .exceptions import ConnectTimeoutError, EarlyResponse
|
|
from .exceptions import HTTPError as HTTPException # noqa
|
|
from .exceptions import (
|
|
NameResolutionError,
|
|
NewConnectionError,
|
|
ProxyError,
|
|
ResponseNotReady,
|
|
)
|
|
from .util import SKIP_HEADER, SKIPPABLE_HEADERS
|
|
from .util.request import body_to_chunks
|
|
from .util.ssl_ import assert_fingerprint as _assert_fingerprint
|
|
from .util.ssl_ import (
|
|
is_capable_for_quic,
|
|
is_ipaddress,
|
|
resolve_cert_reqs,
|
|
resolve_ssl_version,
|
|
ssl_wrap_socket,
|
|
)
|
|
from .util.ssl_match_hostname import CertificateError, match_hostname
|
|
from .util.url import Url
|
|
from .util.socket_state import is_established
|
|
|
|
# Not a no-op, we're adding this to the namespace so it can be imported.
|
|
ConnectionError = ConnectionError
|
|
BrokenPipeError = BrokenPipeError
|
|
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
port_by_scheme = {"http": 80, "https": 443}
|
|
|
|
_CONTAINS_CONTROL_CHAR_RE = re.compile(r"[^-!#$%&'*+.^_`|~0-9a-zA-Z]")
|
|
|
|
|
|
class _ResponseOptions(typing.NamedTuple):
|
|
# TODO: Remove this in favor of a better
|
|
# HTTP request/response lifecycle tracking.
|
|
request_method: str
|
|
request_url: str
|
|
preload_content: bool
|
|
decode_content: bool
|
|
enforce_content_length: bool
|
|
|
|
|
|
class HTTPConnection(HfaceBackend):
|
|
"""
|
|
Based on :class:`urllib3.backend.BaseBackend` but provides an extra constructor
|
|
backwards-compatibility layer between older and newer Pythons.
|
|
|
|
Additional keyword parameters are used to configure attributes of the connection.
|
|
Accepted parameters include:
|
|
|
|
- ``source_address``: Set the source address for the current connection.
|
|
- ``socket_options``: Set specific options on the underlying socket. If not specified, then
|
|
defaults are loaded from ``HTTPConnection.default_socket_options`` which includes disabling
|
|
Nagle's algorithm (sets TCP_NODELAY to 1) unless the connection is behind a proxy.
|
|
|
|
For example, if you wish to enable TCP Keep Alive in addition to the defaults,
|
|
you might pass:
|
|
|
|
.. code-block:: python
|
|
|
|
HTTPConnection.default_socket_options + [
|
|
(socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1),
|
|
]
|
|
|
|
Or you may want to disable the defaults by passing an empty list (e.g., ``[]``).
|
|
"""
|
|
|
|
scheme = "http"
|
|
default_port: typing.ClassVar[int] = port_by_scheme[scheme]
|
|
|
|
blocksize: int
|
|
source_address: tuple[str, int] | None
|
|
socket_options: _TYPE_SOCKET_OPTIONS | None
|
|
|
|
_has_connected_to_proxy: bool
|
|
_response_options: _ResponseOptions | None
|
|
_tunnel_host: str | None
|
|
_tunnel_port: int | None
|
|
_tunnel_scheme: str | None
|
|
|
|
def __init__(
|
|
self,
|
|
host: str,
|
|
port: int | None = None,
|
|
*,
|
|
timeout: _TYPE_TIMEOUT_INTERNAL = _DEFAULT_TIMEOUT,
|
|
source_address: tuple[str, int] | None = None,
|
|
blocksize: int = DEFAULT_BLOCKSIZE,
|
|
socket_options: _TYPE_SOCKET_OPTIONS
|
|
| None = HfaceBackend.default_socket_options,
|
|
proxy: Url | None = None,
|
|
proxy_config: ProxyConfig | None = None,
|
|
disabled_svn: set[HttpVersion] | None = None,
|
|
preemptive_quic_cache: QuicPreemptiveCacheType | None = None,
|
|
resolver: BaseResolver | None = None,
|
|
socket_family: socket.AddressFamily = socket.AF_UNSPEC,
|
|
keepalive_delay: float | int | None = DEFAULT_KEEPALIVE_DELAY,
|
|
) -> None:
|
|
super().__init__(
|
|
host=host,
|
|
port=port,
|
|
timeout=Timeout.resolve_default_timeout(timeout),
|
|
source_address=source_address,
|
|
blocksize=blocksize,
|
|
socket_options=socket_options,
|
|
disabled_svn=disabled_svn,
|
|
preemptive_quic_cache=preemptive_quic_cache,
|
|
keepalive_delay=keepalive_delay,
|
|
)
|
|
self.proxy = proxy
|
|
self.proxy_config = proxy_config
|
|
|
|
self._has_connected_to_proxy = False
|
|
|
|
if resolver is None:
|
|
resolver = ResolverDescription.from_url("system://").new()
|
|
|
|
self._resolver: BaseResolver = resolver
|
|
|
|
#: This struct hold: resolution delay, established delay, and after established datetime.
|
|
self._connect_timings: tuple[timedelta, timedelta, datetime] | None = None
|
|
|
|
if socket_family not in [socket.AF_UNSPEC, socket.AF_INET, socket.AF_INET6]:
|
|
raise ValueError(
|
|
"Unsupported socket_family argument value. Supported values are: socket.AF_UNSPEC, socket.AF_INET, socket.AF_INET6"
|
|
)
|
|
|
|
#: Restrict/Scope IP family per connection.
|
|
self._socket_family = socket_family
|
|
|
|
@property
|
|
def host(self) -> str:
|
|
"""
|
|
Getter method to remove any trailing dots that indicate the hostname is an FQDN.
|
|
|
|
In general, SSL certificates don't include the trailing dot indicating a
|
|
fully-qualified domain name, and thus, they don't validate properly when
|
|
checked against a domain name that includes the dot. In addition, some
|
|
servers may not expect to receive the trailing dot when provided.
|
|
|
|
However, the hostname with trailing dot is critical to DNS resolution; doing a
|
|
lookup with the trailing dot will properly only resolve the appropriate FQDN,
|
|
whereas a lookup without a trailing dot will search the system's search domain
|
|
list. Thus, it's important to keep the original host around for use only in
|
|
those cases where it's appropriate (i.e., when doing DNS lookup to establish the
|
|
actual TCP connection across which we're going to send HTTP requests).
|
|
"""
|
|
return self._dns_host.rstrip(".")
|
|
|
|
@host.setter
|
|
def host(self, value: str) -> None:
|
|
"""
|
|
Setter for the `host` property.
|
|
|
|
We assume that only urllib3 uses the _dns_host attribute; httplib itself
|
|
only uses `host`, and it seems reasonable that other libraries follow suit.
|
|
"""
|
|
self._dns_host = value
|
|
|
|
def _new_conn(self) -> socket.socket:
|
|
"""Establish a socket connection and set nodelay settings on it.
|
|
|
|
:return: New socket connection.
|
|
"""
|
|
super()._new_conn()
|
|
|
|
backup_timeout: float | None = -1.0
|
|
|
|
# we want to purposely mitigate the following scenario:
|
|
# "A server yield its support for HTTP/2 or HTTP/3 through Alt-Svc, but
|
|
# it cannot connect to the alt-svc, thus confusing the end-user on why it
|
|
# waits forever for the 2nd request."
|
|
if self._max_tolerable_delay_for_upgrade is not None:
|
|
backup_timeout = self.timeout
|
|
self.timeout = self._max_tolerable_delay_for_upgrade
|
|
|
|
# it may be required to send some initial data, aka. magic header (PRI * HTTP/2..)
|
|
try:
|
|
sock = self._resolver.create_connection(
|
|
(self._dns_host, self.port or self.default_port),
|
|
self.timeout,
|
|
source_address=self.source_address,
|
|
socket_options=self.socket_options,
|
|
socket_kind=self.socket_kind,
|
|
quic_upgrade_via_dns_rr=self.scheme == "https"
|
|
and HttpVersion.h3 not in self._disabled_svn
|
|
and self.socket_kind != socket.SOCK_DGRAM,
|
|
timing_hook=lambda _: setattr(self, "_connect_timings", _),
|
|
default_socket_family=self._socket_family,
|
|
)
|
|
except socket.gaierror as e:
|
|
raise NameResolutionError(self.host, self, e) from e
|
|
except SocketTimeout as e:
|
|
raise ConnectTimeoutError(
|
|
self,
|
|
f"Connection to {self.host} timed out. (connect timeout={self.timeout})",
|
|
) from e
|
|
except OSError as e:
|
|
raise NewConnectionError(
|
|
self, f"Failed to establish a new connection: {e}"
|
|
) from e
|
|
finally:
|
|
if backup_timeout != -1:
|
|
self.timeout = backup_timeout
|
|
|
|
# We can, migrate to a DGRAM socket if DNS HTTPS/RR record exist and yield HTTP/3+QUIC support.
|
|
if sock.type == socket.SOCK_DGRAM and self.socket_kind == socket.SOCK_STREAM:
|
|
self.socket_kind = socket.SOCK_DGRAM
|
|
self._svn = HttpVersion.h3
|
|
|
|
return sock
|
|
|
|
def set_tunnel(
|
|
self,
|
|
host: str,
|
|
port: int | None = None,
|
|
headers: typing.Mapping[str, str] | None = None,
|
|
scheme: str = "http",
|
|
) -> None:
|
|
if scheme not in ("http", "https"):
|
|
raise ValueError(
|
|
f"Invalid proxy scheme for tunneling: {scheme!r}, must be either 'http' or 'https'"
|
|
)
|
|
super().set_tunnel(host, port=port, headers=headers)
|
|
self._tunnel_scheme = scheme
|
|
|
|
def connect(self) -> None:
|
|
self.sock = self._new_conn()
|
|
if self._tunnel_host:
|
|
self._post_conn()
|
|
# If we're tunneling it means we're connected to our proxy.
|
|
self._has_connected_to_proxy = True
|
|
|
|
# TODO: Fix tunnel so it doesn't depend on self.sock state.
|
|
self._tunnel()
|
|
|
|
self._post_conn()
|
|
# If there's a proxy to be connected to we are fully connected.
|
|
# This is set twice (once above and here) due to forwarding proxies
|
|
# not using tunnelling.
|
|
self._has_connected_to_proxy = bool(self.proxy)
|
|
|
|
if self._has_connected_to_proxy:
|
|
self.proxy_is_verified = False
|
|
|
|
@property
|
|
def is_closed(self) -> bool:
|
|
return self.sock is None
|
|
|
|
@property
|
|
def is_connected(self) -> bool:
|
|
if self.sock is None:
|
|
return False
|
|
|
|
if self.sock.fileno() == -1 or self._protocol is None:
|
|
return False
|
|
|
|
# has_expired can be True when the connection isn't dead
|
|
# with GoAway for example.
|
|
if self._promises or self._pending_responses:
|
|
return True
|
|
|
|
# consider the conn dead after our keep alive delay passed.
|
|
if (
|
|
self._keepalive_delay is not None
|
|
and self.connected_at is not None
|
|
and time.monotonic() - self.connected_at >= self._keepalive_delay
|
|
):
|
|
return False
|
|
|
|
if self._protocol.has_expired():
|
|
return False
|
|
|
|
return is_established(self.sock)
|
|
|
|
@property
|
|
def has_connected_to_proxy(self) -> bool:
|
|
return self._has_connected_to_proxy
|
|
|
|
@property
|
|
def proxy_is_forwarding(self) -> bool:
|
|
"""
|
|
Return True if a forwarding proxy is configured, else return False
|
|
"""
|
|
return bool(self.proxy) and self._tunnel_host is None
|
|
|
|
@property
|
|
def proxy_is_tunneling(self) -> bool:
|
|
"""
|
|
Return True if a tunneling proxy is configured, else return False
|
|
"""
|
|
return self._tunnel_host is not None
|
|
|
|
def close(self) -> None:
|
|
try:
|
|
super().close()
|
|
finally:
|
|
# Reset all stateful properties so connection
|
|
# can be re-used without leaking prior configs.
|
|
self.sock = None
|
|
self.is_verified = False
|
|
self.proxy_is_verified = None
|
|
self._has_connected_to_proxy = False
|
|
self._response_options = None
|
|
self._tunnel_host = None
|
|
self._tunnel_port = None
|
|
self._tunnel_scheme = None
|
|
|
|
def putrequest(
|
|
self,
|
|
method: str,
|
|
url: str,
|
|
skip_host: bool = False,
|
|
skip_accept_encoding: bool = False,
|
|
) -> None:
|
|
""""""
|
|
return super().putrequest(
|
|
method, url, skip_host=skip_host, skip_accept_encoding=skip_accept_encoding
|
|
)
|
|
|
|
def putheader(self, header: str, *values: str) -> None:
|
|
""""""
|
|
if not any(isinstance(v, str) and v == SKIP_HEADER for v in values):
|
|
super().putheader(header, *values)
|
|
elif to_str(header.lower()) not in SKIPPABLE_HEADERS:
|
|
skippable_headers = "', '".join(
|
|
[str.title(header) for header in sorted(SKIPPABLE_HEADERS)]
|
|
)
|
|
raise ValueError(
|
|
f"urllib3.util.SKIP_HEADER only supports '{skippable_headers}'"
|
|
)
|
|
|
|
def request(
|
|
self,
|
|
method: str,
|
|
url: str,
|
|
body: _TYPE_BODY | None = None,
|
|
headers: typing.Mapping[str, str] | None = None,
|
|
*,
|
|
chunked: bool = False,
|
|
preload_content: bool = True,
|
|
decode_content: bool = True,
|
|
enforce_content_length: bool = True,
|
|
on_upload_body: typing.Callable[[int, int | None, bool, bool], None]
|
|
| None = None,
|
|
) -> ResponsePromise:
|
|
# Update the inner socket's timeout value to send the request.
|
|
# This only triggers if the connection is re-used.
|
|
if self.sock is not None:
|
|
self.sock.settimeout(self.timeout)
|
|
|
|
# Store these values to be fed into the HTTPResponse
|
|
# object later. TODO: Remove this in favor of a real
|
|
# HTTP lifecycle mechanism.
|
|
|
|
# We have to store these before we call .request()
|
|
# because sometimes we can still salvage a response
|
|
# off the wire even if we aren't able to completely
|
|
# send the request body.
|
|
response_options = _ResponseOptions(
|
|
request_method=method,
|
|
request_url=url,
|
|
preload_content=preload_content,
|
|
decode_content=decode_content,
|
|
enforce_content_length=enforce_content_length,
|
|
)
|
|
|
|
if headers is None:
|
|
headers = {}
|
|
header_keys = frozenset(to_str(k.lower()) for k in headers)
|
|
|
|
self.putrequest(
|
|
method,
|
|
url,
|
|
skip_accept_encoding="accept-encoding" in header_keys,
|
|
skip_host="host" in header_keys,
|
|
)
|
|
|
|
# Transform the body into an iterable of sendall()-able chunks
|
|
# and detect if an explicit Content-Length is doable.
|
|
chunks_and_cl = body_to_chunks(
|
|
body,
|
|
method=method,
|
|
blocksize=self.max_frame_size,
|
|
force=self._svn != HttpVersion.h11,
|
|
)
|
|
is_sending_string = chunks_and_cl.is_string
|
|
chunks = chunks_and_cl.chunks
|
|
content_length = chunks_and_cl.content_length
|
|
|
|
overrule_content_length: bool = False
|
|
enforce_charset_transparency: bool = False
|
|
|
|
# users may send plain 'str' and assign a Content-Length that will
|
|
# disagree with the actual amount of data to send (encoded, aka. bytes)
|
|
if (
|
|
isinstance(body, str)
|
|
and "content-length" in header_keys
|
|
and len(body) != content_length
|
|
):
|
|
overrule_content_length = True
|
|
|
|
# We shall make our intent clear as we are sending a string.
|
|
# Not being explicit is like doing the same mistake as the early 2k years.
|
|
# No more guessing game based on "Our time make X prevalent, no need to say it! It will never change!" ><'
|
|
if is_sending_string:
|
|
if "content-type" in header_keys:
|
|
enforce_charset_transparency = True
|
|
else:
|
|
self.putheader("Content-Type", "text/plain; charset=utf-8")
|
|
|
|
# When chunked is explicit set to 'True' we respect that.
|
|
if chunked:
|
|
if "transfer-encoding" not in header_keys:
|
|
self.putheader("Transfer-Encoding", "chunked")
|
|
else:
|
|
# Otherwise we go off the recommendation of 'body_to_chunks()'.
|
|
if (
|
|
"content-length" not in header_keys
|
|
and "transfer-encoding" not in header_keys
|
|
):
|
|
if content_length is None:
|
|
if chunks is not None:
|
|
self.putheader("Transfer-Encoding", "chunked")
|
|
else:
|
|
self.putheader("Content-Length", str(content_length))
|
|
|
|
# Now that framing headers are out of the way we send all the other headers.
|
|
if "user-agent" not in header_keys:
|
|
self.putheader("User-Agent", _get_default_user_agent())
|
|
for header, value in headers.items():
|
|
if overrule_content_length and header.lower() == "content-length":
|
|
value = str(content_length)
|
|
if enforce_charset_transparency and header.lower() == "content-type":
|
|
value_lower = value.lower()
|
|
# even if not "officially" supported
|
|
# some may send values as bytes, and we have to
|
|
# cast "temporarily" the value
|
|
# this case is already covered in the parent class.
|
|
if isinstance(value_lower, bytes):
|
|
value_lower = value_lower.decode()
|
|
value = value.decode()
|
|
if "charset" in value_lower:
|
|
if (
|
|
"utf-8" not in value_lower
|
|
and "utf_8" not in value_lower
|
|
and "utf8" not in value_lower
|
|
):
|
|
warnings.warn(
|
|
"A conflicting charset has been set in Content-Type while sending a 'string' as the body. "
|
|
"Beware that urllib3.future always encode a string to unicode. "
|
|
f"Expected 'charset=utf-8', got: {value} "
|
|
"Either encode your string to bytes or open your file in bytes mode.",
|
|
UserWarning,
|
|
stacklevel=2,
|
|
)
|
|
self.putheader(header, value)
|
|
|
|
try:
|
|
rp = self.endheaders(expect_body_afterward=chunks is not None)
|
|
except BrokenPipeError as e:
|
|
rp = e.promise # type: ignore[attr-defined]
|
|
assert rp is not None
|
|
rp.set_parameter("response_options", response_options)
|
|
raise e
|
|
|
|
if rp:
|
|
rp.set_parameter("response_options", response_options)
|
|
return rp
|
|
|
|
total_sent = 0
|
|
|
|
try:
|
|
# If we're given a body we start sending that in chunks.
|
|
if chunks is not None:
|
|
if hasattr(chunks, "__aiter__"):
|
|
raise RuntimeError(
|
|
"Unable to send an async iterable through a synchronous connection"
|
|
)
|
|
for chunk in chunks:
|
|
# Sending empty chunks isn't allowed for TE: chunked
|
|
# as it indicates the end of the body.
|
|
if not chunk:
|
|
continue
|
|
if isinstance(chunk, str):
|
|
chunk = chunk.encode("utf-8")
|
|
self.send(chunk)
|
|
total_sent += len(chunk)
|
|
if on_upload_body is not None:
|
|
on_upload_body(total_sent, content_length, False, False)
|
|
try:
|
|
rp = self.send(b"", eot=True)
|
|
except TypeError:
|
|
# AWSConnection override the send() method
|
|
# thus preventing us to add an additional kwarg in send(...)
|
|
# reason (AWS side): urllib3 2.0 chunks and calls send potentially thousands of
|
|
# times inside `request` unlike the standard library[...]
|
|
# response (urllib3-future): not concerned by this. bypass the bypass.
|
|
rp = super().send(b"", eot=True)
|
|
if on_upload_body is not None:
|
|
on_upload_body(total_sent, content_length, True, False)
|
|
except EarlyResponse as e:
|
|
rp = e.promise
|
|
if on_upload_body is not None:
|
|
on_upload_body(total_sent, content_length, False, True)
|
|
except BrokenPipeError as e:
|
|
if on_upload_body is not None:
|
|
on_upload_body(
|
|
total_sent,
|
|
content_length,
|
|
total_sent == content_length,
|
|
total_sent != content_length,
|
|
)
|
|
rp = e.promise # type: ignore[attr-defined]
|
|
assert rp is not None
|
|
rp.set_parameter("response_options", response_options)
|
|
raise e
|
|
|
|
assert rp is not None
|
|
rp.set_parameter("response_options", response_options)
|
|
return rp
|
|
|
|
def getresponse( # type: ignore[override]
|
|
self,
|
|
*,
|
|
promise: ResponsePromise | None = None,
|
|
police_officer: TrafficPolice[HTTPConnection] | None = None,
|
|
early_response_callback: typing.Callable[[HTTPResponse], None] | None = None,
|
|
) -> HTTPResponse:
|
|
"""
|
|
Get the response from the server.
|
|
|
|
If the HTTPConnection is in the correct state, returns an instance of HTTPResponse or of whatever object is returned by the response_class variable.
|
|
|
|
If a request has not been sent or if a previous response has not be handled, ResponseNotReady is raised. If the HTTP response indicates that the connection should be closed, then it will be closed before the response is returned. When the connection is closed, the underlying socket is closed.
|
|
"""
|
|
# Raise the same error as http.client.HTTPConnection
|
|
if self.sock is None:
|
|
raise ResponseNotReady()
|
|
|
|
# Since the connection's timeout value may have been updated
|
|
# we need to set the timeout on the socket.
|
|
self.sock.settimeout(self.timeout)
|
|
|
|
def early_response_handler(early_low_response: LowLevelResponse) -> None:
|
|
"""Handle unexpected early response. Notify the upper stack!"""
|
|
nonlocal promise, early_response_callback
|
|
|
|
_promise = None
|
|
|
|
if promise is None:
|
|
_promise = early_low_response.from_promise
|
|
else:
|
|
_promise = promise
|
|
|
|
if _promise is None:
|
|
raise OSError
|
|
|
|
if early_response_callback is None:
|
|
early_response_callback = _promise.get_parameter("on_early_response")
|
|
|
|
if early_response_callback is None:
|
|
return
|
|
|
|
early_resp_options: _ResponseOptions = _promise.get_parameter(
|
|
"response_options"
|
|
) # type: ignore[assignment]
|
|
|
|
early_response = HTTPResponse(
|
|
body=early_low_response,
|
|
headers=early_low_response.msg,
|
|
status=early_low_response.status,
|
|
version=early_low_response.version,
|
|
reason=early_low_response.reason,
|
|
preload_content=False,
|
|
decode_content=early_resp_options.decode_content,
|
|
original_response=early_low_response,
|
|
enforce_content_length=False,
|
|
request_method=early_resp_options.request_method,
|
|
request_url=early_resp_options.request_url,
|
|
connection=None,
|
|
police_officer=None,
|
|
)
|
|
|
|
early_response_callback(early_response)
|
|
|
|
# Get the response from backend._base.BaseBackend
|
|
low_response = super().getresponse(
|
|
promise=promise,
|
|
early_response_callback=early_response_handler,
|
|
)
|
|
|
|
if promise is None:
|
|
promise = low_response.from_promise
|
|
|
|
if promise is None:
|
|
raise OSError
|
|
|
|
resp_options: _ResponseOptions = promise.get_parameter("response_options") # type: ignore[assignment]
|
|
|
|
response = HTTPResponse(
|
|
body=low_response,
|
|
headers=low_response.msg,
|
|
status=low_response.status,
|
|
version=low_response.version,
|
|
reason=low_response.reason,
|
|
preload_content=resp_options.preload_content,
|
|
decode_content=resp_options.decode_content,
|
|
original_response=low_response,
|
|
enforce_content_length=resp_options.enforce_content_length,
|
|
request_method=resp_options.request_method,
|
|
request_url=resp_options.request_url,
|
|
connection=self,
|
|
police_officer=police_officer,
|
|
)
|
|
return response
|
|
|
|
|
|
class HTTPSConnection(HTTPConnection):
|
|
"""
|
|
Many of the parameters to this constructor are passed to the underlying SSL
|
|
socket by means of :py:func:`urllib3.util.ssl_wrap_socket`.
|
|
"""
|
|
|
|
scheme = "https"
|
|
default_port = port_by_scheme[scheme]
|
|
|
|
cert_reqs: int | str | None = None
|
|
ca_certs: str | None = None
|
|
ca_cert_dir: str | None = None
|
|
ca_cert_data: None | str | bytes = None
|
|
ssl_version: int | str | None = None
|
|
ssl_minimum_version: int | None = None
|
|
ssl_maximum_version: int | None = None
|
|
assert_fingerprint: str | None = None
|
|
cert_file: str | None = None
|
|
key_file: str | None = None
|
|
key_password: str | None = None
|
|
cert_data: str | bytes | None = None
|
|
key_data: str | bytes | None = None
|
|
ciphers: str | None = None
|
|
|
|
def __init__(
|
|
self,
|
|
host: str,
|
|
port: int | None = None,
|
|
*,
|
|
timeout: _TYPE_TIMEOUT_INTERNAL = _DEFAULT_TIMEOUT,
|
|
source_address: tuple[str, int] | None = None,
|
|
blocksize: int = DEFAULT_BLOCKSIZE,
|
|
socket_options: _TYPE_SOCKET_OPTIONS
|
|
| None = HTTPConnection.default_socket_options,
|
|
disabled_svn: set[HttpVersion] | None = None,
|
|
preemptive_quic_cache: QuicPreemptiveCacheType | None = None,
|
|
resolver: BaseResolver | None = None,
|
|
socket_family: socket.AddressFamily = socket.AF_UNSPEC,
|
|
keepalive_delay: float | int | None = DEFAULT_KEEPALIVE_DELAY,
|
|
proxy: Url | None = None,
|
|
proxy_config: ProxyConfig | None = None,
|
|
cert_reqs: int | str | None = None,
|
|
assert_hostname: None | str | Literal[False] = None,
|
|
assert_fingerprint: str | None = None,
|
|
server_hostname: str | None = None,
|
|
ssl_context: ssl.SSLContext | None = None,
|
|
ca_certs: str | None = None,
|
|
ca_cert_dir: str | None = None,
|
|
ca_cert_data: None | str | bytes = None,
|
|
ssl_minimum_version: int | None = None,
|
|
ssl_maximum_version: int | None = None,
|
|
ssl_version: int | str | None = None,
|
|
cert_file: str | None = None,
|
|
key_file: str | None = None,
|
|
key_password: str | None = None,
|
|
cert_data: str | bytes | None = None,
|
|
key_data: str | bytes | None = None,
|
|
ciphers: str | None = None,
|
|
) -> None:
|
|
if not is_capable_for_quic(ssl_context, ssl_maximum_version):
|
|
if disabled_svn is None:
|
|
disabled_svn = set()
|
|
|
|
disabled_svn.add(HttpVersion.h3)
|
|
|
|
super().__init__(
|
|
host,
|
|
port=port,
|
|
timeout=timeout,
|
|
source_address=source_address,
|
|
blocksize=blocksize,
|
|
socket_options=socket_options,
|
|
proxy=proxy,
|
|
proxy_config=proxy_config,
|
|
disabled_svn=disabled_svn,
|
|
preemptive_quic_cache=preemptive_quic_cache,
|
|
resolver=resolver,
|
|
socket_family=socket_family,
|
|
keepalive_delay=keepalive_delay,
|
|
)
|
|
|
|
self.key_file = key_file
|
|
self.cert_file = cert_file
|
|
self.cert_data = cert_data
|
|
self.key_data = key_data
|
|
self.key_password = key_password
|
|
self.ssl_context = ssl_context
|
|
self.server_hostname = server_hostname
|
|
self.assert_hostname = assert_hostname
|
|
self.assert_fingerprint = assert_fingerprint
|
|
self.ssl_version = ssl_version
|
|
self.ssl_minimum_version = ssl_minimum_version
|
|
self.ssl_maximum_version = ssl_maximum_version
|
|
self.ca_certs = ca_certs and os.path.expanduser(ca_certs)
|
|
self.ca_cert_dir = ca_cert_dir and os.path.expanduser(ca_cert_dir)
|
|
self.ca_cert_data = ca_cert_data
|
|
self.ciphers = ciphers
|
|
|
|
# cert_reqs depends on ssl_context so calculate last.
|
|
if cert_reqs is None:
|
|
if self.ssl_context is not None:
|
|
cert_reqs = self.ssl_context.verify_mode
|
|
else:
|
|
cert_reqs = resolve_cert_reqs(None)
|
|
self.cert_reqs = cert_reqs
|
|
|
|
#: used to store the last used/working ssl context
|
|
self._upgrade_ctx: ssl.SSLContext | None = None
|
|
|
|
def connect(self) -> None:
|
|
sock: socket.socket | ssl.SSLSocket
|
|
self.sock = sock = self._new_conn()
|
|
|
|
# the protocol/state-machine may also ship with an external TLS Engine.
|
|
if (
|
|
self._custom_tls(
|
|
self.ssl_context or self._upgrade_ctx,
|
|
self.ca_certs,
|
|
self.ca_cert_dir,
|
|
self.ca_cert_data,
|
|
self.ssl_minimum_version,
|
|
self.ssl_maximum_version,
|
|
self.cert_file or self.cert_data,
|
|
self.key_file or self.key_data,
|
|
self.key_password,
|
|
self.assert_fingerprint,
|
|
self.assert_hostname,
|
|
self.cert_reqs,
|
|
)
|
|
is False
|
|
):
|
|
server_hostname: str = self.host
|
|
tls_in_tls = False
|
|
|
|
alpn_protocols: list[str] = []
|
|
|
|
# we explicitly skip h3 while still over TCP
|
|
for svn in reversed(HTTPSConnection.supported_svn):
|
|
if svn in self.disabled_svn:
|
|
continue
|
|
if svn == HttpVersion.h11:
|
|
alpn_protocols.append("http/1.1")
|
|
elif svn == HttpVersion.h2:
|
|
alpn_protocols.append("h2")
|
|
|
|
# Do we need to establish a tunnel?
|
|
if self.proxy_is_tunneling:
|
|
# We're tunneling to an HTTPS origin so need to do TLS-in-TLS.
|
|
if self._tunnel_scheme == "https":
|
|
self.sock = sock = self._connect_tls_proxy(
|
|
self.host, sock, ["http/1.1"]
|
|
)
|
|
tls_in_tls = True
|
|
elif self._tunnel_scheme == "http":
|
|
self.proxy_is_verified = False
|
|
|
|
self._post_conn()
|
|
|
|
# If we're tunneling it means we're connected to our proxy.
|
|
self._has_connected_to_proxy = True
|
|
|
|
self._tunnel()
|
|
# Override the host with the one we're requesting data from.
|
|
server_hostname = self._tunnel_host # type: ignore[assignment]
|
|
|
|
if self.server_hostname is not None:
|
|
server_hostname = self.server_hostname
|
|
|
|
sock_and_verified = _ssl_wrap_socket_and_match_hostname(
|
|
sock=sock,
|
|
cert_reqs=self.cert_reqs,
|
|
ssl_version=self.ssl_version,
|
|
ssl_minimum_version=self.ssl_minimum_version,
|
|
ssl_maximum_version=self.ssl_maximum_version,
|
|
ca_certs=self.ca_certs,
|
|
ca_cert_dir=self.ca_cert_dir,
|
|
ca_cert_data=self.ca_cert_data,
|
|
cert_file=self.cert_file,
|
|
key_file=self.key_file,
|
|
key_password=self.key_password,
|
|
server_hostname=server_hostname,
|
|
ssl_context=self.ssl_context,
|
|
tls_in_tls=tls_in_tls,
|
|
assert_hostname=self.assert_hostname,
|
|
assert_fingerprint=self.assert_fingerprint,
|
|
alpn_protocols=alpn_protocols or None,
|
|
cert_data=self.cert_data,
|
|
key_data=self.key_data,
|
|
ciphers=self.ciphers,
|
|
)
|
|
|
|
# we want the http3 upgrade to behave
|
|
# exactly as http1/http2 ssl handshake
|
|
# configuration CAstore wise for example
|
|
# only if not using tls in tls
|
|
if hasattr(sock_and_verified.socket, "context"):
|
|
self._upgrade_ctx = sock_and_verified.socket.context
|
|
|
|
self.sock = sock_and_verified.socket # type: ignore[assignment]
|
|
|
|
# Forwarding proxies can never have a verified target since
|
|
# the proxy is the one doing the verification. Should instead
|
|
# use a CONNECT tunnel in order to verify the target.
|
|
# See: https://github.com/urllib3/urllib3/issues/3267.
|
|
if self.proxy_is_forwarding:
|
|
self.is_verified = False
|
|
else:
|
|
self.is_verified = sock_and_verified.is_verified
|
|
|
|
# If there's a proxy to be connected to we are fully connected.
|
|
# This is set twice (once above and here) due to forwarding proxies
|
|
# not using tunnelling.
|
|
self._has_connected_to_proxy = bool(self.proxy)
|
|
|
|
# Set `self.proxy_is_verified` unless it's already set while
|
|
# establishing a tunnel.
|
|
if self._has_connected_to_proxy and self.proxy_is_verified is None:
|
|
self.proxy_is_verified = sock_and_verified.is_verified
|
|
|
|
self._post_conn()
|
|
|
|
def _connect_tls_proxy(
|
|
self,
|
|
hostname: str,
|
|
sock: socket.socket,
|
|
alpn_protocols: list[str] | None = None,
|
|
) -> ssl.SSLSocket:
|
|
"""
|
|
Establish a TLS connection to the proxy using the provided SSL context.
|
|
"""
|
|
# `_connect_tls_proxy` is called when self._tunnel_host is truthy.
|
|
assert self.proxy_config is not None
|
|
proxy_config = self.proxy_config
|
|
ssl_context = proxy_config.ssl_context
|
|
sock_and_verified = _ssl_wrap_socket_and_match_hostname(
|
|
sock,
|
|
cert_reqs=self.cert_reqs,
|
|
ssl_version=self.ssl_version,
|
|
ssl_minimum_version=self.ssl_minimum_version,
|
|
ssl_maximum_version=self.ssl_maximum_version,
|
|
ca_certs=self.ca_certs,
|
|
ca_cert_dir=self.ca_cert_dir,
|
|
ca_cert_data=self.ca_cert_data,
|
|
server_hostname=hostname,
|
|
ssl_context=ssl_context,
|
|
assert_hostname=proxy_config.assert_hostname,
|
|
assert_fingerprint=proxy_config.assert_fingerprint,
|
|
ciphers=self.ciphers,
|
|
# Features that aren't implemented for proxies yet:
|
|
cert_file=None,
|
|
key_file=None,
|
|
key_password=None,
|
|
tls_in_tls=False,
|
|
alpn_protocols=alpn_protocols,
|
|
cert_data=None,
|
|
key_data=None,
|
|
)
|
|
self.proxy_is_verified = sock_and_verified.is_verified
|
|
return sock_and_verified.socket # type: ignore[return-value]
|
|
|
|
|
|
class _WrappedAndVerifiedSocket(typing.NamedTuple):
|
|
"""
|
|
Wrapped socket and whether the connection is
|
|
verified after the TLS handshake
|
|
"""
|
|
|
|
socket: ssl.SSLSocket | SSLTransport
|
|
is_verified: bool
|
|
|
|
|
|
def _ssl_wrap_socket_and_match_hostname(
|
|
sock: socket.socket,
|
|
*,
|
|
cert_reqs: None | str | int,
|
|
ssl_version: None | str | int,
|
|
ssl_minimum_version: int | None,
|
|
ssl_maximum_version: int | None,
|
|
cert_file: str | None,
|
|
key_file: str | None,
|
|
key_password: str | None,
|
|
ca_certs: str | None,
|
|
ca_cert_dir: str | None,
|
|
ca_cert_data: None | str | bytes,
|
|
assert_hostname: None | str | Literal[False],
|
|
assert_fingerprint: str | None,
|
|
server_hostname: str | None,
|
|
ssl_context: ssl.SSLContext | None,
|
|
tls_in_tls: bool = False,
|
|
alpn_protocols: list[str] | None = None,
|
|
cert_data: str | bytes | None = None,
|
|
key_data: str | bytes | None = None,
|
|
ciphers: str | None = None,
|
|
) -> _WrappedAndVerifiedSocket:
|
|
"""Logic for constructing an SSLContext from all TLS parameters, passing
|
|
that down into ssl_wrap_socket, and then doing certificate verification
|
|
either via hostname or fingerprint. This function exists to guarantee
|
|
that both proxies and targets have the same behavior when connecting via TLS.
|
|
"""
|
|
default_ssl_context = False
|
|
|
|
if ssl_context is None:
|
|
default_ssl_context = True
|
|
context = None
|
|
else:
|
|
context = ssl_context
|
|
|
|
check_hostname: bool | None = None
|
|
|
|
# In some cases, we want to verify hostnames ourselves
|
|
if (
|
|
# `ssl` can't verify fingerprints or alternate hostnames
|
|
assert_fingerprint
|
|
or assert_hostname
|
|
# assert_hostname can be set to False to disable hostname checking
|
|
or assert_hostname is False
|
|
or not ssl_.HAS_NEVER_CHECK_COMMON_NAME
|
|
):
|
|
check_hostname = False
|
|
|
|
# Ensure that IPv6 addresses are in the proper format and don't have a
|
|
# scope ID. Python's SSL module fails to recognize scoped IPv6 addresses
|
|
# and interprets them as DNS hostnames.
|
|
if server_hostname is not None:
|
|
normalized = server_hostname.strip("[]")
|
|
if "%" in normalized:
|
|
normalized = normalized[: normalized.rfind("%")]
|
|
if is_ipaddress(normalized):
|
|
server_hostname = normalized
|
|
|
|
ssl_sock = ssl_wrap_socket(
|
|
sock=sock,
|
|
keyfile=key_file,
|
|
certfile=cert_file,
|
|
key_password=key_password,
|
|
ca_certs=ca_certs,
|
|
ca_cert_dir=ca_cert_dir,
|
|
ca_cert_data=ca_cert_data,
|
|
server_hostname=server_hostname,
|
|
ssl_context=context,
|
|
tls_in_tls=tls_in_tls,
|
|
alpn_protocols=alpn_protocols,
|
|
certdata=cert_data,
|
|
keydata=key_data,
|
|
ciphers=ciphers,
|
|
cert_reqs=resolve_cert_reqs(cert_reqs),
|
|
check_hostname=check_hostname,
|
|
ssl_version=resolve_ssl_version(ssl_version, mitigate_tls_version=True),
|
|
ssl_minimum_version=ssl_minimum_version,
|
|
ssl_maximum_version=ssl_maximum_version,
|
|
)
|
|
|
|
context = ssl_sock.context
|
|
|
|
try:
|
|
if assert_fingerprint:
|
|
_assert_fingerprint(
|
|
ssl_sock.getpeercert(binary_form=True), assert_fingerprint
|
|
)
|
|
elif (
|
|
context.verify_mode != ssl.CERT_NONE
|
|
and not context.check_hostname
|
|
and assert_hostname is not False
|
|
):
|
|
cert: _TYPE_PEER_CERT_RET_DICT = ssl_sock.getpeercert() # type: ignore[assignment]
|
|
|
|
# Need to signal to our match_hostname whether to use 'commonName' or not.
|
|
# If we're using our own constructed SSLContext we explicitly set 'False'
|
|
# because PyPy hard-codes 'True' from SSLContext.hostname_checks_common_name.
|
|
if default_ssl_context:
|
|
hostname_checks_common_name = False
|
|
else:
|
|
hostname_checks_common_name = (
|
|
getattr(context, "hostname_checks_common_name", False) or False
|
|
)
|
|
|
|
_match_hostname(
|
|
cert,
|
|
assert_hostname or server_hostname, # type: ignore[arg-type]
|
|
hostname_checks_common_name,
|
|
)
|
|
|
|
return _WrappedAndVerifiedSocket(
|
|
socket=ssl_sock,
|
|
is_verified=context.verify_mode == ssl.CERT_REQUIRED
|
|
or bool(assert_fingerprint),
|
|
)
|
|
except BaseException:
|
|
ssl_sock.close()
|
|
raise
|
|
|
|
|
|
def _match_hostname(
|
|
cert: _TYPE_PEER_CERT_RET_DICT | None,
|
|
asserted_hostname: str,
|
|
hostname_checks_common_name: bool = False,
|
|
) -> None:
|
|
# Our upstream implementation of ssl.match_hostname()
|
|
# only applies this normalization to IP addresses so it doesn't
|
|
# match DNS SANs so we do the same thing!
|
|
stripped_hostname = asserted_hostname.strip("[]")
|
|
if is_ipaddress(stripped_hostname):
|
|
asserted_hostname = stripped_hostname
|
|
|
|
try:
|
|
match_hostname(cert, asserted_hostname, hostname_checks_common_name)
|
|
except CertificateError as e:
|
|
log.warning(
|
|
"Certificate did not match expected hostname: %s. Certificate: %s",
|
|
asserted_hostname,
|
|
cert,
|
|
)
|
|
# Add cert to exception and reraise so client code can inspect
|
|
# the cert when catching the exception, if they want to
|
|
e._peer_cert = cert # type: ignore[attr-defined]
|
|
raise
|
|
|
|
|
|
def _wrap_proxy_error(err: Exception, proxy_scheme: str | None) -> ProxyError:
|
|
# Look for the phrase 'wrong version number', if found
|
|
# then we should warn the user that we're very sure that
|
|
# this proxy is HTTP-only and they have a configuration issue.
|
|
error_normalized = " ".join(re.split("[^a-z]", str(err).lower()))
|
|
is_likely_http_proxy = (
|
|
"wrong version number" in error_normalized
|
|
or "unknown protocol" in error_normalized
|
|
or "record layer failure" in error_normalized
|
|
)
|
|
http_proxy_warning = (
|
|
". Your proxy appears to only use HTTP and not HTTPS, "
|
|
"try changing your proxy URL to be HTTP. See: "
|
|
"https://urllib3future.readthedocs.io/en/latest/advanced-usage.html"
|
|
"#https-proxy-error-http-proxy"
|
|
)
|
|
new_err = ProxyError(
|
|
f"Unable to connect to proxy"
|
|
f"{http_proxy_warning if is_likely_http_proxy and proxy_scheme == 'https' else ''}",
|
|
err,
|
|
)
|
|
new_err.__cause__ = err
|
|
return new_err
|
|
|
|
|
|
def _get_default_user_agent() -> str:
|
|
return f"urllib3.future/{__version__}"
|
|
|
|
|
|
class DummyConnection:
|
|
"""Used to detect a failed ConnectionCls import."""
|
|
|
|
|
|
if not ssl:
|
|
HTTPSConnection = DummyConnection # type: ignore[misc, assignment] # noqa: F811
|
|
|
|
|
|
VerifiedHTTPSConnection = HTTPSConnection
|
|
|
|
|
|
def _url_from_connection(
|
|
conn: HTTPConnection | HTTPSConnection, path: str | None = None
|
|
) -> str:
|
|
"""Returns the URL from a given connection. This is mainly used for testing and logging."""
|
|
|
|
scheme = "https" if isinstance(conn, HTTPSConnection) else "http"
|
|
|
|
return Url(scheme=scheme, host=conn.host, port=conn.port, path=path).url
|