forked from pool/python-urllib3
Compare commits
79 Commits
| Author | SHA256 | Date | |
|---|---|---|---|
| 5596e5dd87 | |||
|
|
036b503cc3 | ||
| c7e3c17e34 | |||
| 75573f8736 | |||
| b90c81c378 | |||
| fa15163672 | |||
| af9a86ac19 | |||
| 2e3a3af491 | |||
| c9bda474fd | |||
| 9860f9689b | |||
| ef1a31c0b1 | |||
| a5d1101265 | |||
| e84edfbd9b | |||
| a6661b64be | |||
| 259b7264d6 | |||
| 020c67f782 | |||
| 6533ff8336 | |||
| 91110ae749 | |||
| 027dfe1969 | |||
| b0cb844771 | |||
| 0aa6795e5a | |||
| 053c01be78 | |||
| 2682ddb5ff | |||
| 2ff9f6959c | |||
| e8a8c2a6d9 | |||
| 0f2030b111 | |||
| eff7ddf428 | |||
| 59f03c6819 | |||
| c394dddeeb | |||
| aff513686e | |||
| e2b98a98d7 | |||
| b0f3a26d76 | |||
| c87c2bf46d | |||
| ce2ab856af | |||
| 9835754219 | |||
|
|
febcba0e07 | ||
| 65aa1cce40 | |||
| 7b9ba5e57a | |||
| 009a47f5e0 | |||
| 7f246b4425 | |||
| 048c180cf3 | |||
| ddc0ac103e | |||
| 66fc397097 | |||
| 16f20fe873 | |||
|
|
feb9b404d6 | ||
| 4e31bbb477 | |||
| 2681f3eb9e | |||
| 5a3f9776b8 | |||
| 18fb3073a5 | |||
| 2ef6d5ab08 | |||
| bbb00acf42 | |||
| d0c1256114 | |||
| b0f8bd2d46 | |||
| d23bce13ee | |||
| b2ede91d4c | |||
| ff60925403 | |||
| a446ee64df | |||
| b5f5150e3f | |||
| d013f9dceb | |||
| a8fbd10bf2 | |||
| a4bd82b802 | |||
| 111bf3a794 | |||
| 7ff5e21031 | |||
| 7038a55ead | |||
| cf481d062a | |||
| a65e3ac77e | |||
| 16a7a9807b | |||
| 6543261589 | |||
| ea31972c50 | |||
| 4f0e4bb597 | |||
| 2477a01c1d | |||
|
|
99632586bf | ||
| f86bcb8d83 | |||
| c968421ea5 | |||
| 00592aae4d | |||
|
|
8af9d6b346 | ||
| 8685af2a7f | |||
|
|
ad9cfbca7e | ||
|
|
d130b9b182 |
@@ -1,154 +0,0 @@
|
||||
From accff72ecc2f6cf5a76d9570198a93ac7c90270e Mon Sep 17 00:00:00 2001
|
||||
From: Quentin Pradet <quentin.pradet@gmail.com>
|
||||
Date: Mon, 17 Jun 2024 11:09:06 +0400
|
||||
Subject: [PATCH] Merge pull request from GHSA-34jh-p97f-mpxf
|
||||
|
||||
* Strip Proxy-Authorization header on redirects
|
||||
|
||||
* Fix test_retry_default_remove_headers_on_redirect
|
||||
|
||||
* Set release date
|
||||
---
|
||||
CHANGES.rst | 5 +++++
|
||||
src/urllib3/util/retry.py | 4 +++-
|
||||
test/test_retry.py | 6 ++++-
|
||||
test/with_dummyserver/test_poolmanager.py | 27 ++++++++++++++++++++---
|
||||
4 files changed, 37 insertions(+), 5 deletions(-)
|
||||
|
||||
|
||||
diff --git a/src/urllib3/util/retry.py b/src/urllib3/util/retry.py
|
||||
index 7a76a4a6ad..0456cceba4 100644
|
||||
--- a/src/urllib3/util/retry.py
|
||||
+++ b/src/urllib3/util/retry.py
|
||||
@@ -189,7 +189,9 @@ class Retry:
|
||||
RETRY_AFTER_STATUS_CODES = frozenset([413, 429, 503])
|
||||
|
||||
#: Default headers to be used for ``remove_headers_on_redirect``
|
||||
- DEFAULT_REMOVE_HEADERS_ON_REDIRECT = frozenset(["Cookie", "Authorization"])
|
||||
+ DEFAULT_REMOVE_HEADERS_ON_REDIRECT = frozenset(
|
||||
+ ["Cookie", "Authorization", "Proxy-Authorization"]
|
||||
+ )
|
||||
|
||||
#: Default maximum backoff time.
|
||||
DEFAULT_BACKOFF_MAX = 120
|
||||
diff --git a/test/test_retry.py b/test/test_retry.py
|
||||
index f71e7acc9e..ac3ce4ca73 100644
|
||||
--- a/test/test_retry.py
|
||||
+++ b/test/test_retry.py
|
||||
@@ -334,7 +334,11 @@ def test_retry_method_not_allowed(self) -> None:
|
||||
def test_retry_default_remove_headers_on_redirect(self) -> None:
|
||||
retry = Retry()
|
||||
|
||||
- assert retry.remove_headers_on_redirect == {"authorization", "cookie"}
|
||||
+ assert retry.remove_headers_on_redirect == {
|
||||
+ "authorization",
|
||||
+ "proxy-authorization",
|
||||
+ "cookie",
|
||||
+ }
|
||||
|
||||
def test_retry_set_remove_headers_on_redirect(self) -> None:
|
||||
retry = Retry(remove_headers_on_redirect=["X-API-Secret"])
|
||||
diff --git a/test/with_dummyserver/test_poolmanager.py b/test/with_dummyserver/test_poolmanager.py
|
||||
index 4fa9ec850a..af77241d6c 100644
|
||||
--- a/test/with_dummyserver/test_poolmanager.py
|
||||
+++ b/test/with_dummyserver/test_poolmanager.py
|
||||
@@ -144,7 +144,11 @@ def test_redirect_cross_host_remove_headers(self) -> None:
|
||||
"GET",
|
||||
f"{self.base_url}/redirect",
|
||||
fields={"target": f"{self.base_url_alt}/headers"},
|
||||
- headers={"Authorization": "foo", "Cookie": "foo=bar"},
|
||||
+ headers={
|
||||
+ "Authorization": "foo",
|
||||
+ "Proxy-Authorization": "bar",
|
||||
+ "Cookie": "foo=bar",
|
||||
+ },
|
||||
)
|
||||
|
||||
assert r.status == 200
|
||||
@@ -152,13 +156,18 @@ def test_redirect_cross_host_remove_headers(self) -> None:
|
||||
data = r.json()
|
||||
|
||||
assert "Authorization" not in data
|
||||
+ assert "Proxy-Authorization" not in data
|
||||
assert "Cookie" not in data
|
||||
|
||||
r = http.request(
|
||||
"GET",
|
||||
f"{self.base_url}/redirect",
|
||||
fields={"target": f"{self.base_url_alt}/headers"},
|
||||
- headers={"authorization": "foo", "cookie": "foo=bar"},
|
||||
+ headers={
|
||||
+ "authorization": "foo",
|
||||
+ "proxy-authorization": "baz",
|
||||
+ "cookie": "foo=bar",
|
||||
+ },
|
||||
)
|
||||
|
||||
assert r.status == 200
|
||||
@@ -167,6 +176,8 @@ def test_redirect_cross_host_remove_headers(self) -> None:
|
||||
|
||||
assert "authorization" not in data
|
||||
assert "Authorization" not in data
|
||||
+ assert "proxy-authorization" not in data
|
||||
+ assert "Proxy-Authorization" not in data
|
||||
assert "cookie" not in data
|
||||
assert "Cookie" not in data
|
||||
|
||||
@@ -176,7 +187,11 @@ def test_redirect_cross_host_no_remove_headers(self) -> None:
|
||||
"GET",
|
||||
f"{self.base_url}/redirect",
|
||||
fields={"target": f"{self.base_url_alt}/headers"},
|
||||
- headers={"Authorization": "foo", "Cookie": "foo=bar"},
|
||||
+ headers={
|
||||
+ "Authorization": "foo",
|
||||
+ "Proxy-Authorization": "bar",
|
||||
+ "Cookie": "foo=bar",
|
||||
+ },
|
||||
retries=Retry(remove_headers_on_redirect=[]),
|
||||
)
|
||||
|
||||
@@ -185,6 +200,7 @@ def test_redirect_cross_host_no_remove_headers(self) -> None:
|
||||
data = r.json()
|
||||
|
||||
assert data["Authorization"] == "foo"
|
||||
+ assert data["Proxy-Authorization"] == "bar"
|
||||
assert data["Cookie"] == "foo=bar"
|
||||
|
||||
def test_redirect_cross_host_set_removed_headers(self) -> None:
|
||||
@@ -196,6 +212,7 @@ def test_redirect_cross_host_set_removed_headers(self) -> None:
|
||||
headers={
|
||||
"X-API-Secret": "foo",
|
||||
"Authorization": "bar",
|
||||
+ "Proxy-Authorization": "baz",
|
||||
"Cookie": "foo=bar",
|
||||
},
|
||||
retries=Retry(remove_headers_on_redirect=["X-API-Secret"]),
|
||||
@@ -207,11 +224,13 @@ def test_redirect_cross_host_set_removed_headers(self) -> None:
|
||||
|
||||
assert "X-API-Secret" not in data
|
||||
assert data["Authorization"] == "bar"
|
||||
+ assert data["Proxy-Authorization"] == "baz"
|
||||
assert data["Cookie"] == "foo=bar"
|
||||
|
||||
headers = {
|
||||
"x-api-secret": "foo",
|
||||
"authorization": "bar",
|
||||
+ "proxy-authorization": "baz",
|
||||
"cookie": "foo=bar",
|
||||
}
|
||||
r = http.request(
|
||||
@@ -229,12 +248,14 @@ def test_redirect_cross_host_set_removed_headers(self) -> None:
|
||||
assert "x-api-secret" not in data
|
||||
assert "X-API-Secret" not in data
|
||||
assert data["Authorization"] == "bar"
|
||||
+ assert data["Proxy-Authorization"] == "baz"
|
||||
assert data["Cookie"] == "foo=bar"
|
||||
|
||||
# Ensure the header argument itself is not modified in-place.
|
||||
assert headers == {
|
||||
"x-api-secret": "foo",
|
||||
"authorization": "bar",
|
||||
+ "proxy-authorization": "baz",
|
||||
"cookie": "foo=bar",
|
||||
}
|
||||
|
||||
60
CVE-2025-66418.patch
Normal file
60
CVE-2025-66418.patch
Normal file
@@ -0,0 +1,60 @@
|
||||
From 24d7b67eac89f94e11003424bcf0d8f7b72222a8 Mon Sep 17 00:00:00 2001
|
||||
From: Illia Volochii <illia.volochii@gmail.com>
|
||||
Date: Fri, 5 Dec 2025 16:41:33 +0200
|
||||
Subject: [PATCH] Merge commit from fork
|
||||
|
||||
* Add a hard-coded limit for the decompression chain
|
||||
|
||||
* Reuse new list
|
||||
---
|
||||
changelog/GHSA-gm62-xv2j-4w53.security.rst | 4 ++++
|
||||
src/urllib3/response.py | 12 +++++++++++-
|
||||
test/test_response.py | 10 ++++++++++
|
||||
3 files changed, 25 insertions(+), 1 deletion(-)
|
||||
create mode 100644 changelog/GHSA-gm62-xv2j-4w53.security.rst
|
||||
|
||||
Index: urllib3-2.5.0/src/urllib3/response.py
|
||||
===================================================================
|
||||
--- urllib3-2.5.0.orig/src/urllib3/response.py
|
||||
+++ urllib3-2.5.0/src/urllib3/response.py
|
||||
@@ -342,8 +342,18 @@ class MultiDecoder(ContentDecoder):
|
||||
they were applied.
|
||||
"""
|
||||
|
||||
+ # Maximum allowed number of chained HTTP encodings in the
|
||||
+ # Content-Encoding header.
|
||||
+ max_decode_links = 5
|
||||
+
|
||||
def __init__(self, modes: str) -> None:
|
||||
- self._decoders = [_get_decoder(m.strip()) for m in modes.split(",")]
|
||||
+ encodings = [m.strip() for m in modes.split(",")]
|
||||
+ if len(encodings) > self.max_decode_links:
|
||||
+ raise DecodeError(
|
||||
+ "Too many content encodings in the chain: "
|
||||
+ f"{len(encodings)} > {self.max_decode_links}"
|
||||
+ )
|
||||
+ self._decoders = [_get_decoder(e) for e in encodings]
|
||||
|
||||
def flush(self) -> bytes:
|
||||
return self._decoders[0].flush()
|
||||
Index: urllib3-2.5.0/test/test_response.py
|
||||
===================================================================
|
||||
--- urllib3-2.5.0.orig/test/test_response.py
|
||||
+++ urllib3-2.5.0/test/test_response.py
|
||||
@@ -843,6 +843,16 @@ class TestResponse:
|
||||
assert r.read(9 * 37) == b"foobarbaz" * 37
|
||||
assert r.read() == b""
|
||||
|
||||
+ def test_read_multi_decoding_too_many_links(self) -> None:
|
||||
+ fp = BytesIO(b"foo")
|
||||
+ with pytest.raises(
|
||||
+ DecodeError, match="Too many content encodings in the chain: 6 > 5"
|
||||
+ ):
|
||||
+ HTTPResponse(
|
||||
+ fp,
|
||||
+ headers={"content-encoding": "gzip, deflate, br, zstd, gzip, deflate"},
|
||||
+ )
|
||||
+
|
||||
def test_body_blob(self) -> None:
|
||||
resp = HTTPResponse(b"foo")
|
||||
assert resp.data == b"foo"
|
||||
898
CVE-2025-66471.patch
Normal file
898
CVE-2025-66471.patch
Normal file
@@ -0,0 +1,898 @@
|
||||
From c19571de34c47de3a766541b041637ba5f716ed7 Mon Sep 17 00:00:00 2001
|
||||
From: Illia Volochii <illia.volochii@gmail.com>
|
||||
Date: Fri, 5 Dec 2025 16:40:41 +0200
|
||||
Subject: [PATCH] Merge commit from fork
|
||||
|
||||
* Prevent decompression bomb for zstd in Python 3.14
|
||||
|
||||
* Add experimental `decompress_iter` for Brotli
|
||||
|
||||
* Update changes for Brotli
|
||||
|
||||
* Add `GzipDecoder.decompress_iter`
|
||||
|
||||
* Test https://github.com/python-hyper/brotlicffi/pull/207
|
||||
|
||||
* Pin Brotli
|
||||
|
||||
* Add `decompress_iter` to all decoders and make tests pass
|
||||
|
||||
* Pin brotlicffi to an official release
|
||||
|
||||
* Revert changes to response.py
|
||||
|
||||
* Add `max_length` parameter to all `decompress` methods
|
||||
|
||||
* Fix the `test_brotlipy` session
|
||||
|
||||
* Unset `_data` on gzip error
|
||||
|
||||
* Add a test for memory usage
|
||||
|
||||
* Test more methods
|
||||
|
||||
* Fix the test for `stream`
|
||||
|
||||
* Cover more lines with tests
|
||||
|
||||
* Add more coverage
|
||||
|
||||
* Make `read1` a bit more efficient
|
||||
|
||||
* Fix PyPy tests for Brotli
|
||||
|
||||
* Revert an unnecessarily moved check
|
||||
|
||||
* Add some comments
|
||||
|
||||
* Leave just one `self._obj.decompress` call in `GzipDecoder`
|
||||
|
||||
* Refactor test params
|
||||
|
||||
* Test reads with all data already in the decompressor
|
||||
|
||||
* Prevent needless copying of data decoded with `max_length`
|
||||
|
||||
* Rename the changed test
|
||||
|
||||
* Note that responses of unknown length should be streamed too
|
||||
|
||||
* Add a changelog entry
|
||||
|
||||
* Avoid returning a memory view from `BytesQueueBuffer`
|
||||
|
||||
* Add one more note to the changelog entry
|
||||
---
|
||||
CHANGES.rst | 22 ++++
|
||||
docs/advanced-usage.rst | 3 +-
|
||||
docs/user-guide.rst | 4 +-
|
||||
noxfile.py | 16 ++-
|
||||
pyproject.toml | 5 +-
|
||||
src/urllib3/response.py | 279 ++++++++++++++++++++++++++++++++++------
|
||||
test/test_response.py | 269 +++++++++++++++++++++++++++++++++++++-
|
||||
uv.lock | 177 +++++++++++--------------
|
||||
8 files changed, 621 insertions(+), 154 deletions(-)
|
||||
|
||||
Index: urllib3-2.5.0/docs/advanced-usage.rst
|
||||
===================================================================
|
||||
--- urllib3-2.5.0.orig/docs/advanced-usage.rst
|
||||
+++ urllib3-2.5.0/docs/advanced-usage.rst
|
||||
@@ -66,7 +66,8 @@ When using ``preload_content=True`` (the
|
||||
response body will be read immediately into memory and the HTTP connection
|
||||
will be released back into the pool without manual intervention.
|
||||
|
||||
-However, when dealing with large responses it's often better to stream the response
|
||||
+However, when dealing with responses of large or unknown length,
|
||||
+it's often better to stream the response
|
||||
content using ``preload_content=False``. Setting ``preload_content`` to ``False`` means
|
||||
that urllib3 will only read from the socket when data is requested.
|
||||
|
||||
Index: urllib3-2.5.0/docs/user-guide.rst
|
||||
===================================================================
|
||||
--- urllib3-2.5.0.orig/docs/user-guide.rst
|
||||
+++ urllib3-2.5.0/docs/user-guide.rst
|
||||
@@ -145,8 +145,8 @@ to a byte string representing the respon
|
||||
print(resp.data)
|
||||
# b"\xaa\xa5H?\x95\xe9\x9b\x11"
|
||||
|
||||
-.. note:: For larger responses, it's sometimes better to :ref:`stream <stream>`
|
||||
- the response.
|
||||
+.. note:: For responses of large or unknown length, it's sometimes better to
|
||||
+ :ref:`stream <stream>` the response.
|
||||
|
||||
Using io Wrappers with Response Content
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
Index: urllib3-2.5.0/pyproject.toml
|
||||
===================================================================
|
||||
--- urllib3-2.5.0.orig/pyproject.toml
|
||||
+++ urllib3-2.5.0/pyproject.toml
|
||||
@@ -41,8 +41,8 @@ dynamic = ["version"]
|
||||
|
||||
[project.optional-dependencies]
|
||||
brotli = [
|
||||
- "brotli>=1.0.9; platform_python_implementation == 'CPython'",
|
||||
- "brotlicffi>=0.8.0; platform_python_implementation != 'CPython'"
|
||||
+ "brotli>=1.2.0; platform_python_implementation == 'CPython'",
|
||||
+ "brotlicffi>=1.2.0.0; platform_python_implementation != 'CPython'"
|
||||
]
|
||||
# Once we drop support for Python 3.13 this extra can be removed.
|
||||
# We'll need a deprecation period for the 'zstandard' module support
|
||||
@@ -160,6 +160,7 @@ filterwarnings = [
|
||||
'''default:ssl\.PROTOCOL_TLSv1_1 is deprecated:DeprecationWarning''',
|
||||
'''default:ssl\.PROTOCOL_TLSv1_2 is deprecated:DeprecationWarning''',
|
||||
'''default:ssl NPN is deprecated, use ALPN instead:DeprecationWarning''',
|
||||
+ '''default:Brotli >= 1.2.0 is required to prevent decompression bombs\.:urllib3.exceptions.DependencyWarning''',
|
||||
# https://github.com/SeleniumHQ/selenium/issues/13328
|
||||
'''default:unclosed file <_io\.BufferedWriter name='/dev/null'>:ResourceWarning''',
|
||||
# https://github.com/SeleniumHQ/selenium/issues/14686
|
||||
Index: urllib3-2.5.0/src/urllib3/response.py
|
||||
===================================================================
|
||||
--- urllib3-2.5.0.orig/src/urllib3/response.py
|
||||
+++ urllib3-2.5.0/src/urllib3/response.py
|
||||
@@ -33,6 +33,7 @@ from .connection import BaseSSLError, HT
|
||||
from .exceptions import (
|
||||
BodyNotHttplibCompatible,
|
||||
DecodeError,
|
||||
+ DependencyWarning,
|
||||
HTTPError,
|
||||
IncompleteRead,
|
||||
InvalidChunkLength,
|
||||
@@ -52,7 +53,11 @@ log = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class ContentDecoder:
|
||||
- def decompress(self, data: bytes) -> bytes:
|
||||
+ def decompress(self, data: bytes, max_length: int = -1) -> bytes:
|
||||
+ raise NotImplementedError()
|
||||
+
|
||||
+ @property
|
||||
+ def has_unconsumed_tail(self) -> bool:
|
||||
raise NotImplementedError()
|
||||
|
||||
def flush(self) -> bytes:
|
||||
@@ -62,30 +67,57 @@ class ContentDecoder:
|
||||
class DeflateDecoder(ContentDecoder):
|
||||
def __init__(self) -> None:
|
||||
self._first_try = True
|
||||
- self._data = b""
|
||||
+ self._first_try_data = b""
|
||||
+ self._unfed_data = b""
|
||||
self._obj = zlib.decompressobj()
|
||||
|
||||
- def decompress(self, data: bytes) -> bytes:
|
||||
- if not data:
|
||||
+ def decompress(self, data: bytes, max_length: int = -1) -> bytes:
|
||||
+ data = self._unfed_data + data
|
||||
+ self._unfed_data = b""
|
||||
+ if not data and not self._obj.unconsumed_tail:
|
||||
return data
|
||||
+ original_max_length = max_length
|
||||
+ if original_max_length < 0:
|
||||
+ max_length = 0
|
||||
+ elif original_max_length == 0:
|
||||
+ # We should not pass 0 to the zlib decompressor because 0 is
|
||||
+ # the default value that will make zlib decompress without a
|
||||
+ # length limit.
|
||||
+ # Data should be stored for subsequent calls.
|
||||
+ self._unfed_data = data
|
||||
+ return b""
|
||||
|
||||
+ # Subsequent calls always reuse `self._obj`. zlib requires
|
||||
+ # passing the unconsumed tail if decompression is to continue.
|
||||
if not self._first_try:
|
||||
- return self._obj.decompress(data)
|
||||
+ return self._obj.decompress(
|
||||
+ self._obj.unconsumed_tail + data, max_length=max_length
|
||||
+ )
|
||||
|
||||
- self._data += data
|
||||
+ # First call tries with RFC 1950 ZLIB format.
|
||||
+ self._first_try_data += data
|
||||
try:
|
||||
- decompressed = self._obj.decompress(data)
|
||||
+ decompressed = self._obj.decompress(data, max_length=max_length)
|
||||
if decompressed:
|
||||
self._first_try = False
|
||||
- self._data = None # type: ignore[assignment]
|
||||
+ self._first_try_data = b""
|
||||
return decompressed
|
||||
+ # On failure, it falls back to RFC 1951 DEFLATE format.
|
||||
except zlib.error:
|
||||
self._first_try = False
|
||||
self._obj = zlib.decompressobj(-zlib.MAX_WBITS)
|
||||
try:
|
||||
- return self.decompress(self._data)
|
||||
+ return self.decompress(
|
||||
+ self._first_try_data, max_length=original_max_length
|
||||
+ )
|
||||
finally:
|
||||
- self._data = None # type: ignore[assignment]
|
||||
+ self._first_try_data = b""
|
||||
+
|
||||
+ @property
|
||||
+ def has_unconsumed_tail(self) -> bool:
|
||||
+ return bool(self._unfed_data) or (
|
||||
+ bool(self._obj.unconsumed_tail) and not self._first_try
|
||||
+ )
|
||||
|
||||
def flush(self) -> bytes:
|
||||
return self._obj.flush()
|
||||
@@ -101,27 +133,61 @@ class GzipDecoder(ContentDecoder):
|
||||
def __init__(self) -> None:
|
||||
self._obj = zlib.decompressobj(16 + zlib.MAX_WBITS)
|
||||
self._state = GzipDecoderState.FIRST_MEMBER
|
||||
+ self._unconsumed_tail = b""
|
||||
|
||||
- def decompress(self, data: bytes) -> bytes:
|
||||
+ def decompress(self, data: bytes, max_length: int = -1) -> bytes:
|
||||
ret = bytearray()
|
||||
- if self._state == GzipDecoderState.SWALLOW_DATA or not data:
|
||||
+ if self._state == GzipDecoderState.SWALLOW_DATA:
|
||||
+ return bytes(ret)
|
||||
+
|
||||
+ if max_length == 0:
|
||||
+ # We should not pass 0 to the zlib decompressor because 0 is
|
||||
+ # the default value that will make zlib decompress without a
|
||||
+ # length limit.
|
||||
+ # Data should be stored for subsequent calls.
|
||||
+ self._unconsumed_tail += data
|
||||
+ return b""
|
||||
+
|
||||
+ # zlib requires passing the unconsumed tail to the subsequent
|
||||
+ # call if decompression is to continue.
|
||||
+ data = self._unconsumed_tail + data
|
||||
+ if not data and self._obj.eof:
|
||||
return bytes(ret)
|
||||
+
|
||||
while True:
|
||||
try:
|
||||
- ret += self._obj.decompress(data)
|
||||
+ ret += self._obj.decompress(
|
||||
+ data, max_length=max(max_length - len(ret), 0)
|
||||
+ )
|
||||
except zlib.error:
|
||||
previous_state = self._state
|
||||
# Ignore data after the first error
|
||||
self._state = GzipDecoderState.SWALLOW_DATA
|
||||
+ self._unconsumed_tail = b""
|
||||
if previous_state == GzipDecoderState.OTHER_MEMBERS:
|
||||
# Allow trailing garbage acceptable in other gzip clients
|
||||
return bytes(ret)
|
||||
raise
|
||||
- data = self._obj.unused_data
|
||||
+
|
||||
+ self._unconsumed_tail = data = (
|
||||
+ self._obj.unconsumed_tail or self._obj.unused_data
|
||||
+ )
|
||||
+ if max_length > 0 and len(ret) >= max_length:
|
||||
+ break
|
||||
+
|
||||
if not data:
|
||||
return bytes(ret)
|
||||
- self._state = GzipDecoderState.OTHER_MEMBERS
|
||||
- self._obj = zlib.decompressobj(16 + zlib.MAX_WBITS)
|
||||
+ # When the end of a gzip member is reached, a new decompressor
|
||||
+ # must be created for unused (possibly future) data.
|
||||
+ if self._obj.eof:
|
||||
+ self._state = GzipDecoderState.OTHER_MEMBERS
|
||||
+ self._obj = zlib.decompressobj(16 + zlib.MAX_WBITS)
|
||||
+
|
||||
+ return bytes(ret)
|
||||
+
|
||||
+ @property
|
||||
+ def has_unconsumed_tail(self) -> bool:
|
||||
+ return bool(self._unconsumed_tail)
|
||||
|
||||
def flush(self) -> bytes:
|
||||
return self._obj.flush()
|
||||
@@ -136,9 +202,35 @@ if brotli is not None:
|
||||
def __init__(self) -> None:
|
||||
self._obj = brotli.Decompressor()
|
||||
if hasattr(self._obj, "decompress"):
|
||||
- setattr(self, "decompress", self._obj.decompress)
|
||||
+ setattr(self, "_decompress", self._obj.decompress)
|
||||
else:
|
||||
- setattr(self, "decompress", self._obj.process)
|
||||
+ setattr(self, "_decompress", self._obj.process)
|
||||
+
|
||||
+ # Requires Brotli >= 1.2.0 for `output_buffer_limit`.
|
||||
+ def _decompress(self, data: bytes, output_buffer_limit: int = -1) -> bytes:
|
||||
+ raise NotImplementedError()
|
||||
+
|
||||
+ def decompress(self, data: bytes, max_length: int = -1) -> bytes:
|
||||
+ try:
|
||||
+ if max_length > 0:
|
||||
+ return self._decompress(data, output_buffer_limit=max_length)
|
||||
+ else:
|
||||
+ return self._decompress(data)
|
||||
+ except TypeError:
|
||||
+ # Fallback for Brotli/brotlicffi/brotlipy versions without
|
||||
+ # the `output_buffer_limit` parameter.
|
||||
+ warnings.warn(
|
||||
+ "Brotli >= 1.2.0 is required to prevent decompression bombs.",
|
||||
+ DependencyWarning,
|
||||
+ )
|
||||
+ return self._decompress(data)
|
||||
+
|
||||
+ @property
|
||||
+ def has_unconsumed_tail(self) -> bool:
|
||||
+ try:
|
||||
+ return not self._obj.can_accept_more_data()
|
||||
+ except AttributeError:
|
||||
+ return False
|
||||
|
||||
def flush(self) -> bytes:
|
||||
if hasattr(self._obj, "flush"):
|
||||
@@ -156,16 +248,46 @@ try:
|
||||
def __init__(self) -> None:
|
||||
self._obj = zstd.ZstdDecompressor()
|
||||
|
||||
- def decompress(self, data: bytes) -> bytes:
|
||||
- if not data:
|
||||
+ def decompress(self, data: bytes, max_length: int = -1) -> bytes:
|
||||
+ if not data and not self.has_unconsumed_tail:
|
||||
return b""
|
||||
- data_parts = [self._obj.decompress(data)]
|
||||
- while self._obj.eof and self._obj.unused_data:
|
||||
- unused_data = self._obj.unused_data
|
||||
+ if self._obj.eof:
|
||||
+ data = self._obj.unused_data + data
|
||||
self._obj = zstd.ZstdDecompressor()
|
||||
- data_parts.append(self._obj.decompress(unused_data))
|
||||
+ part = self._obj.decompress(data, max_length=max_length)
|
||||
+ length = len(part)
|
||||
+ data_parts = [part]
|
||||
+ # Every loop iteration is supposed to read data from a separate frame.
|
||||
+ # The loop breaks when:
|
||||
+ # - enough data is read;
|
||||
+ # - no more unused data is available;
|
||||
+ # - end of the last read frame has not been reached (i.e.,
|
||||
+ # more data has to be fed).
|
||||
+ while (
|
||||
+ self._obj.eof
|
||||
+ and self._obj.unused_data
|
||||
+ and (max_length < 0 or length < max_length)
|
||||
+ ):
|
||||
+ unused_data = self._obj.unused_data
|
||||
+ if not self._obj.needs_input:
|
||||
+ self._obj = zstd.ZstdDecompressor()
|
||||
+ part = self._obj.decompress(
|
||||
+ unused_data,
|
||||
+ max_length=(max_length - length) if max_length > 0 else -1,
|
||||
+ )
|
||||
+ if part_length := len(part):
|
||||
+ data_parts.append(part)
|
||||
+ length += part_length
|
||||
+ elif self._obj.needs_input:
|
||||
+ break
|
||||
return b"".join(data_parts)
|
||||
|
||||
+ @property
|
||||
+ def has_unconsumed_tail(self) -> bool:
|
||||
+ return not (self._obj.needs_input or self._obj.eof) or bool(
|
||||
+ self._obj.unused_data
|
||||
+ )
|
||||
+
|
||||
def flush(self) -> bytes:
|
||||
if not self._obj.eof:
|
||||
raise DecodeError("Zstandard data is incomplete")
|
||||
@@ -226,10 +348,35 @@ class MultiDecoder(ContentDecoder):
|
||||
def flush(self) -> bytes:
|
||||
return self._decoders[0].flush()
|
||||
|
||||
- def decompress(self, data: bytes) -> bytes:
|
||||
- for d in reversed(self._decoders):
|
||||
- data = d.decompress(data)
|
||||
- return data
|
||||
+ def decompress(self, data: bytes, max_length: int = -1) -> bytes:
|
||||
+ if max_length <= 0:
|
||||
+ for d in reversed(self._decoders):
|
||||
+ data = d.decompress(data)
|
||||
+ return data
|
||||
+
|
||||
+ ret = bytearray()
|
||||
+ # Every while loop iteration goes through all decoders once.
|
||||
+ # It exits when enough data is read or no more data can be read.
|
||||
+ # It is possible that the while loop iteration does not produce
|
||||
+ # any data because we retrieve up to `max_length` from every
|
||||
+ # decoder, and the amount of bytes may be insufficient for the
|
||||
+ # next decoder to produce enough/any output.
|
||||
+ while True:
|
||||
+ any_data = False
|
||||
+ for d in reversed(self._decoders):
|
||||
+ data = d.decompress(data, max_length=max_length - len(ret))
|
||||
+ if data:
|
||||
+ any_data = True
|
||||
+ # We should not break when no data is returned because
|
||||
+ # next decoders may produce data even with empty input.
|
||||
+ ret += data
|
||||
+ if not any_data or len(ret) >= max_length:
|
||||
+ return bytes(ret)
|
||||
+ data = b""
|
||||
+
|
||||
+ @property
|
||||
+ def has_unconsumed_tail(self) -> bool:
|
||||
+ return any(d.has_unconsumed_tail for d in self._decoders)
|
||||
|
||||
|
||||
def _get_decoder(mode: str) -> ContentDecoder:
|
||||
@@ -262,9 +409,6 @@ class BytesQueueBuffer:
|
||||
|
||||
* self.buffer, which contains the full data
|
||||
* the largest chunk that we will copy in get()
|
||||
-
|
||||
- The worst case scenario is a single chunk, in which case we'll make a full copy of
|
||||
- the data inside get().
|
||||
"""
|
||||
|
||||
def __init__(self) -> None:
|
||||
@@ -286,6 +430,10 @@ class BytesQueueBuffer:
|
||||
elif n < 0:
|
||||
raise ValueError("n should be > 0")
|
||||
|
||||
+ if len(self.buffer[0]) == n and isinstance(self.buffer[0], bytes):
|
||||
+ self._size -= n
|
||||
+ return self.buffer.popleft()
|
||||
+
|
||||
fetched = 0
|
||||
ret = io.BytesIO()
|
||||
while fetched < n:
|
||||
@@ -492,7 +640,11 @@ class BaseHTTPResponse(io.IOBase):
|
||||
self._decoder = _get_decoder(content_encoding)
|
||||
|
||||
def _decode(
|
||||
- self, data: bytes, decode_content: bool | None, flush_decoder: bool
|
||||
+ self,
|
||||
+ data: bytes,
|
||||
+ decode_content: bool | None,
|
||||
+ flush_decoder: bool,
|
||||
+ max_length: int | None = None,
|
||||
) -> bytes:
|
||||
"""
|
||||
Decode the data passed in and potentially flush the decoder.
|
||||
@@ -505,9 +657,12 @@ class BaseHTTPResponse(io.IOBase):
|
||||
)
|
||||
return data
|
||||
|
||||
+ if max_length is None or flush_decoder:
|
||||
+ max_length = -1
|
||||
+
|
||||
try:
|
||||
if self._decoder:
|
||||
- data = self._decoder.decompress(data)
|
||||
+ data = self._decoder.decompress(data, max_length=max_length)
|
||||
self._has_decoded_content = True
|
||||
except self.DECODER_ERROR_CLASSES as e:
|
||||
content_encoding = self.headers.get("content-encoding", "").lower()
|
||||
@@ -978,6 +1133,14 @@ class HTTPResponse(BaseHTTPResponse):
|
||||
elif amt is not None:
|
||||
cache_content = False
|
||||
|
||||
+ if self._decoder and self._decoder.has_unconsumed_tail:
|
||||
+ decoded_data = self._decode(
|
||||
+ b"",
|
||||
+ decode_content,
|
||||
+ flush_decoder=False,
|
||||
+ max_length=amt - len(self._decoded_buffer),
|
||||
+ )
|
||||
+ self._decoded_buffer.put(decoded_data)
|
||||
if len(self._decoded_buffer) >= amt:
|
||||
return self._decoded_buffer.get(amt)
|
||||
|
||||
@@ -985,7 +1148,11 @@ class HTTPResponse(BaseHTTPResponse):
|
||||
|
||||
flush_decoder = amt is None or (amt != 0 and not data)
|
||||
|
||||
- if not data and len(self._decoded_buffer) == 0:
|
||||
+ if (
|
||||
+ not data
|
||||
+ and len(self._decoded_buffer) == 0
|
||||
+ and not (self._decoder and self._decoder.has_unconsumed_tail)
|
||||
+ ):
|
||||
return data
|
||||
|
||||
if amt is None:
|
||||
@@ -1002,7 +1169,12 @@ class HTTPResponse(BaseHTTPResponse):
|
||||
)
|
||||
return data
|
||||
|
||||
- decoded_data = self._decode(data, decode_content, flush_decoder)
|
||||
+ decoded_data = self._decode(
|
||||
+ data,
|
||||
+ decode_content,
|
||||
+ flush_decoder,
|
||||
+ max_length=amt - len(self._decoded_buffer),
|
||||
+ )
|
||||
self._decoded_buffer.put(decoded_data)
|
||||
|
||||
while len(self._decoded_buffer) < amt and data:
|
||||
@@ -1010,7 +1182,12 @@ class HTTPResponse(BaseHTTPResponse):
|
||||
# For example, the GZ file header takes 10 bytes, we don't want to read
|
||||
# it one byte at a time
|
||||
data = self._raw_read(amt)
|
||||
- decoded_data = self._decode(data, decode_content, flush_decoder)
|
||||
+ decoded_data = self._decode(
|
||||
+ data,
|
||||
+ decode_content,
|
||||
+ flush_decoder,
|
||||
+ max_length=amt - len(self._decoded_buffer),
|
||||
+ )
|
||||
self._decoded_buffer.put(decoded_data)
|
||||
data = self._decoded_buffer.get(amt)
|
||||
|
||||
@@ -1045,6 +1222,20 @@ class HTTPResponse(BaseHTTPResponse):
|
||||
"Calling read1(decode_content=False) is not supported after "
|
||||
"read1(decode_content=True) was called."
|
||||
)
|
||||
+ if (
|
||||
+ self._decoder
|
||||
+ and self._decoder.has_unconsumed_tail
|
||||
+ and (amt is None or len(self._decoded_buffer) < amt)
|
||||
+ ):
|
||||
+ decoded_data = self._decode(
|
||||
+ b"",
|
||||
+ decode_content,
|
||||
+ flush_decoder=False,
|
||||
+ max_length=(
|
||||
+ amt - len(self._decoded_buffer) if amt is not None else None
|
||||
+ ),
|
||||
+ )
|
||||
+ self._decoded_buffer.put(decoded_data)
|
||||
if len(self._decoded_buffer) > 0:
|
||||
if amt is None:
|
||||
return self._decoded_buffer.get_all()
|
||||
@@ -1060,7 +1251,9 @@ class HTTPResponse(BaseHTTPResponse):
|
||||
self._init_decoder()
|
||||
while True:
|
||||
flush_decoder = not data
|
||||
- decoded_data = self._decode(data, decode_content, flush_decoder)
|
||||
+ decoded_data = self._decode(
|
||||
+ data, decode_content, flush_decoder, max_length=amt
|
||||
+ )
|
||||
self._decoded_buffer.put(decoded_data)
|
||||
if decoded_data or flush_decoder:
|
||||
break
|
||||
@@ -1091,7 +1284,11 @@ class HTTPResponse(BaseHTTPResponse):
|
||||
if self.chunked and self.supports_chunked_reads():
|
||||
yield from self.read_chunked(amt, decode_content=decode_content)
|
||||
else:
|
||||
- while not is_fp_closed(self._fp) or len(self._decoded_buffer) > 0:
|
||||
+ while (
|
||||
+ not is_fp_closed(self._fp)
|
||||
+ or len(self._decoded_buffer) > 0
|
||||
+ or (self._decoder and self._decoder.has_unconsumed_tail)
|
||||
+ ):
|
||||
data = self.read(amt=amt, decode_content=decode_content)
|
||||
|
||||
if data:
|
||||
@@ -1254,7 +1451,10 @@ class HTTPResponse(BaseHTTPResponse):
|
||||
break
|
||||
chunk = self._handle_chunk(amt)
|
||||
decoded = self._decode(
|
||||
- chunk, decode_content=decode_content, flush_decoder=False
|
||||
+ chunk,
|
||||
+ decode_content=decode_content,
|
||||
+ flush_decoder=False,
|
||||
+ max_length=amt,
|
||||
)
|
||||
if decoded:
|
||||
yield decoded
|
||||
Index: urllib3-2.5.0/test/test_response.py
|
||||
===================================================================
|
||||
--- urllib3-2.5.0.orig/test/test_response.py
|
||||
+++ urllib3-2.5.0/test/test_response.py
|
||||
@@ -1,6 +1,7 @@
|
||||
from __future__ import annotations
|
||||
|
||||
import contextlib
|
||||
+import gzip
|
||||
import http.client as httplib
|
||||
import socket
|
||||
import ssl
|
||||
@@ -43,6 +44,26 @@ def zstd_compress(data: bytes) -> bytes:
|
||||
return zstd.compress(data) # type: ignore[no-any-return]
|
||||
|
||||
|
||||
+def deflate2_compress(data: bytes) -> bytes:
|
||||
+ compressor = zlib.compressobj(6, zlib.DEFLATED, -zlib.MAX_WBITS)
|
||||
+ return compressor.compress(data) + compressor.flush()
|
||||
+
|
||||
+
|
||||
+if brotli:
|
||||
+ try:
|
||||
+ brotli.Decompressor().process(b"", output_buffer_limit=1024)
|
||||
+ _brotli_gte_1_2_0_available = True
|
||||
+ except (AttributeError, TypeError):
|
||||
+ _brotli_gte_1_2_0_available = False
|
||||
+else:
|
||||
+ _brotli_gte_1_2_0_available = False
|
||||
+try:
|
||||
+ zstd_compress(b"")
|
||||
+ _zstd_available = True
|
||||
+except ModuleNotFoundError:
|
||||
+ _zstd_available = False
|
||||
+
|
||||
+
|
||||
class TestBytesQueueBuffer:
|
||||
def test_single_chunk(self) -> None:
|
||||
buffer = BytesQueueBuffer()
|
||||
@@ -118,12 +139,19 @@ class TestBytesQueueBuffer:
|
||||
|
||||
assert len(get_func(buffer)) == 10 * 2**20
|
||||
|
||||
+ @pytest.mark.parametrize(
|
||||
+ "get_func",
|
||||
+ (lambda b: b.get(len(b)), lambda b: b.get_all()),
|
||||
+ ids=("get", "get_all"),
|
||||
+ )
|
||||
@pytest.mark.limit_memory("10.01 MB", current_thread_only=True)
|
||||
- def test_get_all_memory_usage_single_chunk(self) -> None:
|
||||
+ def test_memory_usage_single_chunk(
|
||||
+ self, get_func: typing.Callable[[BytesQueueBuffer], bytes]
|
||||
+ ) -> None:
|
||||
buffer = BytesQueueBuffer()
|
||||
chunk = bytes(10 * 2**20) # 10 MiB
|
||||
buffer.put(chunk)
|
||||
- assert buffer.get_all() is chunk
|
||||
+ assert get_func(buffer) is chunk
|
||||
|
||||
|
||||
# A known random (i.e, not-too-compressible) payload generated with:
|
||||
@@ -426,7 +454,26 @@ class TestResponse:
|
||||
assert r.data == b"foo"
|
||||
|
||||
@onlyZstd()
|
||||
- def test_decode_multiframe_zstd(self) -> None:
|
||||
+ @pytest.mark.parametrize(
|
||||
+ "read_amt",
|
||||
+ (
|
||||
+ # Read all data at once.
|
||||
+ None,
|
||||
+ # Read one byte at a time, data of frames will be returned
|
||||
+ # separately.
|
||||
+ 1,
|
||||
+ # Read two bytes at a time, the second read should return
|
||||
+ # data from both frames.
|
||||
+ 2,
|
||||
+ # Read three bytes at a time, the whole frames will be
|
||||
+ # returned separately in two calls.
|
||||
+ 3,
|
||||
+ # Read four bytes at a time, the first read should return
|
||||
+ # data from the first frame and a part of the second frame.
|
||||
+ 4,
|
||||
+ ),
|
||||
+ )
|
||||
+ def test_decode_multiframe_zstd(self, read_amt: int | None) -> None:
|
||||
data = (
|
||||
# Zstandard frame
|
||||
zstd_compress(b"foo")
|
||||
@@ -441,8 +488,57 @@ class TestResponse:
|
||||
)
|
||||
|
||||
fp = BytesIO(data)
|
||||
- r = HTTPResponse(fp, headers={"content-encoding": "zstd"})
|
||||
- assert r.data == b"foobar"
|
||||
+ result = bytearray()
|
||||
+ r = HTTPResponse(
|
||||
+ fp, headers={"content-encoding": "zstd"}, preload_content=False
|
||||
+ )
|
||||
+ total_length = 6
|
||||
+ while len(result) < total_length:
|
||||
+ chunk = r.read(read_amt, decode_content=True)
|
||||
+ if read_amt is None:
|
||||
+ assert len(chunk) == total_length
|
||||
+ else:
|
||||
+ assert len(chunk) == min(read_amt, total_length - len(result))
|
||||
+ result += chunk
|
||||
+ assert bytes(result) == b"foobar"
|
||||
+
|
||||
+ @onlyZstd()
|
||||
+ def test_decode_multiframe_zstd_with_max_length_close_to_compressed_data_size(
|
||||
+ self,
|
||||
+ ) -> None:
|
||||
+ """
|
||||
+ Test decoding when the first read from the socket returns all
|
||||
+ the compressed frames, but then it has to be decompressed in a
|
||||
+ couple of read calls.
|
||||
+ """
|
||||
+ data = (
|
||||
+ # Zstandard frame
|
||||
+ zstd_compress(b"x" * 1024)
|
||||
+ # skippable frame (must be ignored)
|
||||
+ + bytes.fromhex(
|
||||
+ "50 2A 4D 18" # Magic_Number (little-endian)
|
||||
+ "07 00 00 00" # Frame_Size (little-endian)
|
||||
+ "00 00 00 00 00 00 00" # User_Data
|
||||
+ )
|
||||
+ # Zstandard frame
|
||||
+ + zstd_compress(b"y" * 1024)
|
||||
+ )
|
||||
+
|
||||
+ fp = BytesIO(data)
|
||||
+ r = HTTPResponse(
|
||||
+ fp, headers={"content-encoding": "zstd"}, preload_content=False
|
||||
+ )
|
||||
+ # Read the whole first frame.
|
||||
+ assert r.read(1024) == b"x" * 1024
|
||||
+ assert len(r._decoded_buffer) == 0
|
||||
+ # Read the whole second frame in two reads.
|
||||
+ assert r.read(512) == b"y" * 512
|
||||
+ assert len(r._decoded_buffer) == 0
|
||||
+ assert r.read(512) == b"y" * 512
|
||||
+ assert len(r._decoded_buffer) == 0
|
||||
+ # Ensure no more data is left.
|
||||
+ assert r.read() == b""
|
||||
+ assert len(r._decoded_buffer) == 0
|
||||
|
||||
@onlyZstd()
|
||||
def test_chunked_decoding_zstd(self) -> None:
|
||||
@@ -535,6 +631,169 @@ class TestResponse:
|
||||
decoded_data += part
|
||||
assert decoded_data == data
|
||||
|
||||
+ _test_compressor_params: list[
|
||||
+ tuple[str, tuple[str, typing.Callable[[bytes], bytes]] | None]
|
||||
+ ] = [
|
||||
+ ("deflate1", ("deflate", zlib.compress)),
|
||||
+ ("deflate2", ("deflate", deflate2_compress)),
|
||||
+ ("gzip", ("gzip", gzip.compress)),
|
||||
+ ]
|
||||
+ if _brotli_gte_1_2_0_available:
|
||||
+ _test_compressor_params.append(("brotli", ("br", brotli.compress)))
|
||||
+ else:
|
||||
+ _test_compressor_params.append(("brotli", None))
|
||||
+ if _zstd_available:
|
||||
+ _test_compressor_params.append(("zstd", ("zstd", zstd_compress)))
|
||||
+ else:
|
||||
+ _test_compressor_params.append(("zstd", None))
|
||||
+
|
||||
+ @pytest.mark.parametrize("read_method", ("read", "read1"))
|
||||
+ @pytest.mark.parametrize(
|
||||
+ "data",
|
||||
+ [d[1] for d in _test_compressor_params],
|
||||
+ ids=[d[0] for d in _test_compressor_params],
|
||||
+ )
|
||||
+ def test_read_with_all_data_already_in_decompressor(
|
||||
+ self,
|
||||
+ request: pytest.FixtureRequest,
|
||||
+ read_method: str,
|
||||
+ data: tuple[str, typing.Callable[[bytes], bytes]] | None,
|
||||
+ ) -> None:
|
||||
+ if data is None:
|
||||
+ pytest.skip(f"Proper {request.node.callspec.id} decoder is not available")
|
||||
+ original_data = b"bar" * 1000
|
||||
+ name, compress_func = data
|
||||
+ compressed_data = compress_func(original_data)
|
||||
+ fp = mock.Mock(read=mock.Mock(return_value=b""))
|
||||
+ r = HTTPResponse(fp, headers={"content-encoding": name}, preload_content=False)
|
||||
+ # Put all data in the decompressor's buffer.
|
||||
+ r._init_decoder()
|
||||
+ assert r._decoder is not None # for mypy
|
||||
+ decoded = r._decoder.decompress(compressed_data, max_length=0)
|
||||
+ if name == "br":
|
||||
+ # It's known that some Brotli libraries do not respect
|
||||
+ # `max_length`.
|
||||
+ r._decoded_buffer.put(decoded)
|
||||
+ else:
|
||||
+ assert decoded == b""
|
||||
+ # Read the data via `HTTPResponse`.
|
||||
+ read = getattr(r, read_method)
|
||||
+ assert read(0) == b""
|
||||
+ assert read(2500) == original_data[:2500]
|
||||
+ assert read(500) == original_data[2500:]
|
||||
+ assert read(0) == b""
|
||||
+ assert read() == b""
|
||||
+
|
||||
+ @pytest.mark.parametrize(
|
||||
+ "delta",
|
||||
+ (
|
||||
+ 0, # First read from socket returns all compressed data.
|
||||
+ -1, # First read from socket returns all but one byte of compressed data.
|
||||
+ ),
|
||||
+ )
|
||||
+ @pytest.mark.parametrize("read_method", ("read", "read1"))
|
||||
+ @pytest.mark.parametrize(
|
||||
+ "data",
|
||||
+ [d[1] for d in _test_compressor_params],
|
||||
+ ids=[d[0] for d in _test_compressor_params],
|
||||
+ )
|
||||
+ def test_decode_with_max_length_close_to_compressed_data_size(
|
||||
+ self,
|
||||
+ request: pytest.FixtureRequest,
|
||||
+ delta: int,
|
||||
+ read_method: str,
|
||||
+ data: tuple[str, typing.Callable[[bytes], bytes]] | None,
|
||||
+ ) -> None:
|
||||
+ """
|
||||
+ Test decoding when the first read from the socket returns all or
|
||||
+ almost all the compressed data, but then it has to be
|
||||
+ decompressed in a couple of read calls.
|
||||
+ """
|
||||
+ if data is None:
|
||||
+ pytest.skip(f"Proper {request.node.callspec.id} decoder is not available")
|
||||
+
|
||||
+ original_data = b"foo" * 1000
|
||||
+ name, compress_func = data
|
||||
+ compressed_data = compress_func(original_data)
|
||||
+ fp = BytesIO(compressed_data)
|
||||
+ r = HTTPResponse(fp, headers={"content-encoding": name}, preload_content=False)
|
||||
+ initial_limit = len(compressed_data) + delta
|
||||
+ read = getattr(r, read_method)
|
||||
+ initial_chunk = read(amt=initial_limit, decode_content=True)
|
||||
+ assert len(initial_chunk) == initial_limit
|
||||
+ assert (
|
||||
+ len(read(amt=len(original_data), decode_content=True))
|
||||
+ == len(original_data) - initial_limit
|
||||
+ )
|
||||
+
|
||||
+ # Prepare 50 MB of compressed data outside of the test measuring
|
||||
+ # memory usage.
|
||||
+ _test_memory_usage_decode_with_max_length_params: list[
|
||||
+ tuple[str, tuple[str, bytes] | None]
|
||||
+ ] = [
|
||||
+ (
|
||||
+ params[0],
|
||||
+ (params[1][0], params[1][1](b"A" * (50 * 2**20))) if params[1] else None,
|
||||
+ )
|
||||
+ for params in _test_compressor_params
|
||||
+ ]
|
||||
+
|
||||
+ @pytest.mark.parametrize(
|
||||
+ "data",
|
||||
+ [d[1] for d in _test_memory_usage_decode_with_max_length_params],
|
||||
+ ids=[d[0] for d in _test_memory_usage_decode_with_max_length_params],
|
||||
+ )
|
||||
+ @pytest.mark.parametrize("read_method", ("read", "read1", "read_chunked", "stream"))
|
||||
+ # Decoders consume different amounts of memory during decompression.
|
||||
+ # We set the 10 MB limit to ensure that the whole decompressed data
|
||||
+ # is not stored unnecessarily.
|
||||
+ #
|
||||
+ # FYI, the following consumption was observed for the test with
|
||||
+ # `read` on CPython 3.14.0:
|
||||
+ # - deflate: 2.3 MiB
|
||||
+ # - deflate2: 2.1 MiB
|
||||
+ # - gzip: 2.1 MiB
|
||||
+ # - brotli:
|
||||
+ # - brotli v1.2.0: 9 MiB
|
||||
+ # - brotlicffi v1.2.0.0: 6 MiB
|
||||
+ # - brotlipy v0.7.0: 105.8 MiB
|
||||
+ # - zstd: 4.5 MiB
|
||||
+ @pytest.mark.limit_memory("10 MB", current_thread_only=True)
|
||||
+ def test_memory_usage_decode_with_max_length(
|
||||
+ self,
|
||||
+ request: pytest.FixtureRequest,
|
||||
+ read_method: str,
|
||||
+ data: tuple[str, bytes] | None,
|
||||
+ ) -> None:
|
||||
+ if data is None:
|
||||
+ pytest.skip(f"Proper {request.node.callspec.id} decoder is not available")
|
||||
+
|
||||
+ name, compressed_data = data
|
||||
+ limit = 1024 * 1024 # 1 MiB
|
||||
+ if read_method in ("read_chunked", "stream"):
|
||||
+ httplib_r = httplib.HTTPResponse(MockSock) # type: ignore[arg-type]
|
||||
+ httplib_r.fp = MockChunkedEncodingResponse([compressed_data]) # type: ignore[assignment]
|
||||
+ r = HTTPResponse(
|
||||
+ httplib_r,
|
||||
+ preload_content=False,
|
||||
+ headers={"transfer-encoding": "chunked", "content-encoding": name},
|
||||
+ )
|
||||
+ next(getattr(r, read_method)(amt=limit, decode_content=True))
|
||||
+ else:
|
||||
+ fp = BytesIO(compressed_data)
|
||||
+ r = HTTPResponse(
|
||||
+ fp, headers={"content-encoding": name}, preload_content=False
|
||||
+ )
|
||||
+ getattr(r, read_method)(amt=limit, decode_content=True)
|
||||
+
|
||||
+ # Check that the internal decoded buffer is empty unless brotli
|
||||
+ # is used.
|
||||
+ # Google's brotli library does not fully respect the output
|
||||
+ # buffer limit: https://github.com/google/brotli/issues/1396
|
||||
+ # And unmaintained brotlipy cannot limit the output buffer size.
|
||||
+ if name != "br" or brotli.__name__ == "brotlicffi":
|
||||
+ assert len(r._decoded_buffer) == 0
|
||||
+
|
||||
def test_multi_decoding_deflate_deflate(self) -> None:
|
||||
data = zlib.compress(zlib.compress(b"foo"))
|
||||
|
||||
88
CVE-2026-21441.patch
Normal file
88
CVE-2026-21441.patch
Normal file
@@ -0,0 +1,88 @@
|
||||
From f3ef966d0c717099d2f4a1697bd661b48c703efd Mon Sep 17 00:00:00 2001
|
||||
From: Illia Volochii <illia.volochii@gmail.com>
|
||||
Date: Wed, 7 Jan 2026 18:07:30 +0200
|
||||
Subject: [PATCH] Merge commit from fork
|
||||
|
||||
* Stop decoding response content during redirects needlessly
|
||||
|
||||
* Rename the new query parameter
|
||||
|
||||
* Add a changelog entry
|
||||
---
|
||||
dummyserver/app.py | 8 +++++++-
|
||||
src/urllib3/response.py | 6 +++++-
|
||||
test/with_dummyserver/test_connectionpool.py | 19 +++++++++++++++++++
|
||||
3 files changed, 31 insertions(+), 2 deletions(-)
|
||||
|
||||
diff --git a/dummyserver/app.py b/dummyserver/app.py
|
||||
index 0eeb93f7..5b82e932 100644
|
||||
--- a/dummyserver/app.py
|
||||
+++ b/dummyserver/app.py
|
||||
@@ -233,10 +233,16 @@ async def redirect() -> ResponseReturnValue:
|
||||
values = await request.values
|
||||
target = values.get("target", "/")
|
||||
status = values.get("status", "303 See Other")
|
||||
+ compressed = values.get("compressed") == "true"
|
||||
status_code = status.split(" ")[0]
|
||||
|
||||
headers = [("Location", target)]
|
||||
- return await make_response("", status_code, headers)
|
||||
+ if compressed:
|
||||
+ headers.append(("Content-Encoding", "gzip"))
|
||||
+ data = gzip.compress(b"foo")
|
||||
+ else:
|
||||
+ data = b""
|
||||
+ return await make_response(data, status_code, headers)
|
||||
|
||||
|
||||
@hypercorn_app.route("/redirect_after")
|
||||
diff --git a/src/urllib3/response.py b/src/urllib3/response.py
|
||||
index 5632dab3..720fbf26 100644
|
||||
--- a/src/urllib3/response.py
|
||||
+++ b/src/urllib3/response.py
|
||||
@@ -677,7 +677,11 @@ class HTTPResponse(BaseHTTPResponse):
|
||||
Unread data in the HTTPResponse connection blocks the connection from being released back to the pool.
|
||||
"""
|
||||
try:
|
||||
- self.read()
|
||||
+ self.read(
|
||||
+ # Do not spend resources decoding the content unless
|
||||
+ # decoding has already been initiated.
|
||||
+ decode_content=self._has_decoded_content,
|
||||
+ )
|
||||
except (HTTPError, OSError, BaseSSLError, HTTPException):
|
||||
pass
|
||||
|
||||
diff --git a/test/with_dummyserver/test_connectionpool.py b/test/with_dummyserver/test_connectionpool.py
|
||||
index ce165e24..8d6107ae 100644
|
||||
--- a/test/with_dummyserver/test_connectionpool.py
|
||||
+++ b/test/with_dummyserver/test_connectionpool.py
|
||||
@@ -508,6 +508,25 @@ class TestConnectionPool(HypercornDummyServerTestCase):
|
||||
assert r.status == 200
|
||||
assert r.data == b"Dummy server!"
|
||||
|
||||
+ @mock.patch("urllib3.response.GzipDecoder.decompress")
|
||||
+ def test_no_decoding_with_redirect_when_preload_disabled(
|
||||
+ self, gzip_decompress: mock.MagicMock
|
||||
+ ) -> None:
|
||||
+ """
|
||||
+ Test that urllib3 does not attempt to decode a gzipped redirect
|
||||
+ response when `preload_content` is set to `False`.
|
||||
+ """
|
||||
+ with HTTPConnectionPool(self.host, self.port) as pool:
|
||||
+ # Three requests are expected: two redirects and one final / 200 OK.
|
||||
+ response = pool.request(
|
||||
+ "GET",
|
||||
+ "/redirect",
|
||||
+ fields={"target": "/redirect?compressed=true", "compressed": "true"},
|
||||
+ preload_content=False,
|
||||
+ )
|
||||
+ assert response.status == 200
|
||||
+ gzip_decompress.assert_not_called()
|
||||
+
|
||||
def test_303_redirect_makes_request_lose_body(self) -> None:
|
||||
with HTTPConnectionPool(self.host, self.port) as pool:
|
||||
response = pool.request(
|
||||
--
|
||||
2.52.0
|
||||
|
||||
BIN
hypercorn-d1719f8c1570cbd8e6a3719ffdb14a4d72880abb.tar.gz
LFS
Normal file
BIN
hypercorn-d1719f8c1570cbd8e6a3719ffdb14a4d72880abb.tar.gz
LFS
Normal file
Binary file not shown.
@@ -1,32 +0,0 @@
|
||||
Index: urllib3-2.1.0/changelog/3268.bugfix.rst
|
||||
===================================================================
|
||||
--- /dev/null
|
||||
+++ urllib3-2.1.0/changelog/3268.bugfix.rst
|
||||
@@ -0,0 +1 @@
|
||||
+Fixed handling of OpenSSL 3.2.0 new error message for misconfiguring an HTTP proxy as HTTPS.
|
||||
Index: urllib3-2.1.0/src/urllib3/connection.py
|
||||
===================================================================
|
||||
--- urllib3-2.1.0.orig/src/urllib3/connection.py
|
||||
+++ urllib3-2.1.0/src/urllib3/connection.py
|
||||
@@ -864,6 +864,7 @@ def _wrap_proxy_error(err: Exception, pr
|
||||
is_likely_http_proxy = (
|
||||
"wrong version number" in error_normalized
|
||||
or "unknown protocol" in error_normalized
|
||||
+ or "record layer failure" in error_normalized
|
||||
)
|
||||
http_proxy_warning = (
|
||||
". Your proxy appears to only use HTTP and not HTTPS, "
|
||||
Index: urllib3-2.1.0/test/with_dummyserver/test_socketlevel.py
|
||||
===================================================================
|
||||
--- urllib3-2.1.0.orig/test/with_dummyserver/test_socketlevel.py
|
||||
+++ urllib3-2.1.0/test/with_dummyserver/test_socketlevel.py
|
||||
@@ -1297,7 +1297,8 @@ class TestSSL(SocketDummyServerTestCase)
|
||||
self._start_server(socket_handler)
|
||||
with HTTPSConnectionPool(self.host, self.port, ca_certs=DEFAULT_CA) as pool:
|
||||
with pytest.raises(
|
||||
- SSLError, match=r"(wrong version number|record overflow)"
|
||||
+ SSLError,
|
||||
+ match=r"(wrong version number|record overflow|record layer failure)",
|
||||
):
|
||||
pool.request("GET", "/", retries=False)
|
||||
|
||||
@@ -1,3 +1,141 @@
|
||||
-------------------------------------------------------------------
|
||||
Tue Jan 20 12:36:47 UTC 2026 - Nico Krapp <nico.krapp@suse.com>
|
||||
|
||||
- Add security patch:
|
||||
* CVE-2025-66471.patch (bsc#1254867)
|
||||
* CVE-2025-66418.patch (bsc#1254866)
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Tue Jan 13 09:58:43 UTC 2026 - John Paul Adrian Glaubitz <adrian.glaubitz@suse.com>
|
||||
|
||||
- Add CVE-2026-21441.patch to fix excessive resource consumption
|
||||
during decompression of data in HTTP redirect responses
|
||||
(bsc#1256331, CVE-2026-21441)
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Mon Jun 23 02:03:12 UTC 2025 - Steve Kowalik <steven.kowalik@suse.com>
|
||||
|
||||
- Update to 2.5.0:
|
||||
* Security issues
|
||||
Pool managers now properly control redirects when retries is passed
|
||||
(CVE-2025-50181, GHSA-pq67-6m6q-mj2v, bsc#1244925)
|
||||
Redirects are now controlled by urllib3 in the Node.js runtime
|
||||
(CVE-2025-50182, GHSA-48p4-8xcf-vxj5, bsc#1244924)
|
||||
* Features
|
||||
Added support for the compression.zstd module that is new in Python 3.14.
|
||||
Added support for version 0.5 of hatch-vcs
|
||||
* Bugfixes
|
||||
Raised exception for HTTPResponse.shutdown on a connection already
|
||||
released to the pool.
|
||||
Fixed incorrect CONNECT statement when using an IPv6 proxy with
|
||||
connection_from_host. Previously would not be wrapped in [].
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Tue May 27 08:56:43 UTC 2025 - Daniel Garcia <daniel.garcia@suse.com>
|
||||
|
||||
- Update to 2.4.0
|
||||
* Applied PEP 639 by specifying the license fields in
|
||||
pyproject.toml. (#3522)
|
||||
* Updated exceptions to save and restore more properties during the
|
||||
pickle/serialization process. (#3567)
|
||||
* Added verify_flags option to create_urllib3_context with a default
|
||||
of VERIFY_X509_PARTIAL_CHAIN and VERIFY_X509_STRICT for Python
|
||||
3.13+. (#3571)
|
||||
* Fixed a bug with partial reads of streaming data in Emscripten.
|
||||
(#3555)
|
||||
* Switched to uv for installing development dependecies. (#3550)
|
||||
* Removed the multiple.intoto.jsonl asset from GitHub releases.
|
||||
Attestation of release files since v2.3.0 can be found on PyPI.
|
||||
(#3566)
|
||||
- 2.3.0:
|
||||
* Added HTTPResponse.shutdown() to stop any ongoing or future reads
|
||||
for a specific response. It calls shutdown(SHUT_RD) on the
|
||||
underlying socket. This feature was sponsored by LaunchDarkly.
|
||||
(#2868)
|
||||
* Added support for JavaScript Promise Integration on Emscripten.
|
||||
This enables more efficient WebAssembly requests and streaming,
|
||||
and makes it possible to use in Node.js if you launch it as node
|
||||
--experimental-wasm-stack-switching. (#3400)
|
||||
* Added the proxy_is_tunneling property to HTTPConnection and
|
||||
HTTPSConnection. (#3285)
|
||||
* Added pickling support to NewConnectionError and
|
||||
NameResolutionError. (#3480)
|
||||
* Fixed an issue in debug logs where the HTTP version was rendering
|
||||
as "HTTP/11" instead of "HTTP/1.1". (#3489)
|
||||
* Removed support for Python 3.8. (#3492)
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Tue May 27 08:51:09 UTC 2025 - Daniel Garcia <daniel.garcia@suse.com>
|
||||
|
||||
- Skip test_close_after_handshake flaky test, it fails sometimes in
|
||||
ppc64le and s390x architectures, bsc#1243583
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Thu Dec 19 07:20:32 UTC 2024 - Daniel Garcia <daniel.garcia@suse.com>
|
||||
|
||||
- Skip some flaky tests that fail sometimes in OBS (bsc#1234681)
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Wed Dec 18 08:41:22 UTC 2024 - Daniel Garcia <daniel.garcia@suse.com>
|
||||
|
||||
- Ignore DeprecationWarning in tests (bsc#1234681)
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Thu Oct 3 05:10:09 UTC 2024 - Steve Kowalik <steven.kowalik@suse.com>
|
||||
|
||||
- Update to 2.2.3:
|
||||
* Features
|
||||
+ Added support for Python 3.13.
|
||||
* Bugfixes
|
||||
+ Fixed the default encoding of chunked request bodies to be UTF-8
|
||||
instead of ISO-8859-1. All other methods of supplying a request body
|
||||
already use UTF-8 starting in urllib3 v2.0.
|
||||
+ Fixed ResourceWarning on CONNECT with Python < 3.11.4 by backporting
|
||||
python/cpython#103472.
|
||||
+ Fixed a crash where certain standard library hash functions were absent
|
||||
in restricted environments.
|
||||
+ Added the Proxy-Authorization header to the list of headers to strip
|
||||
from requests when redirecting to a different host. As before,
|
||||
different headers can be set via Retry.remove_headers_on_redirect.
|
||||
+ Allowed passing negative integers as amt to read methods of
|
||||
http.client.HTTPResponse as an alternative to None.
|
||||
+ Fixed issue where InsecureRequestWarning was emitted for HTTPS
|
||||
connections when using Emscripten.
|
||||
+ Fixed HTTPConnectionPool.urlopen to stop automatically casting
|
||||
non-proxy headers to HTTPHeaderDict. This change was premature as it
|
||||
did not apply to proxy headers and HTTPHeaderDict does not handle byte
|
||||
header values correctly yet.
|
||||
+ Changed InvalidChunkLength to ProtocolError when response terminates
|
||||
before the chunk length is sent.
|
||||
+ Changed ProtocolError to be more verbose on incomplete reads with
|
||||
excess content.
|
||||
+ Added support for HTTPResponse.read1() method.
|
||||
+ Fixed issue where requests against urls with trailing dots were
|
||||
failing due to SSL errors when using proxy.
|
||||
+ Fixed HTTPConnection.proxy_is_verified and
|
||||
HTTPSConnection.proxy_is_verified to be always set to a boolean after
|
||||
connecting to a proxy. It could be None in some cases previously.
|
||||
+ Fixed an issue where headers passed in a request with json= would be
|
||||
mutated
|
||||
+ Fixed HTTPSConnection.is_verified to be set to False when connecting
|
||||
from a HTTPS proxy to an HTTP target. It was set to True previously.
|
||||
+ Fixed handling of new error message from OpenSSL 3.2.0 when configuring
|
||||
an HTTP proxy as HTTPS
|
||||
+ Fixed TLS 1.3 post-handshake auth when the server certificate
|
||||
validation is disabled
|
||||
* HTTP/2 (experimental)
|
||||
+ Excluded Transfer-Encoding: chunked from HTTP/2 request body
|
||||
+ Added a probing mechanism for determining whether a given target
|
||||
origin supports HTTP/2 via ALPN.
|
||||
+ Add support for sending a request body with HTTP/2
|
||||
* Removals
|
||||
+ Drop support for end-of-life PyPy3.8 and PyPy3.9.
|
||||
- Drop patches, they are now included upstream:
|
||||
* CVE-2024-37891.patch
|
||||
* openssl-3.2.patch
|
||||
- Included patched hypercorn, which is only unpacked and used for the test
|
||||
suite.
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Tue Jun 18 09:46:57 UTC 2024 - Markéta Machová <mmachova@suse.com>
|
||||
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
#
|
||||
# spec file for package python-urllib3
|
||||
#
|
||||
# Copyright (c) 2024 SUSE LLC
|
||||
# Copyright (c) 2025 SUSE LLC
|
||||
#
|
||||
# All modifications and additions to the file contributed by third parties
|
||||
# remain the property of their copyright owners, unless otherwise agreed
|
||||
@@ -26,42 +26,52 @@
|
||||
%endif
|
||||
%{?sle15_python_module_pythons}
|
||||
Name: python-urllib3%{psuffix}
|
||||
Version: 2.1.0
|
||||
Version: 2.5.0
|
||||
Release: 0
|
||||
Summary: HTTP library with thread-safe connection pooling, file post, and more
|
||||
License: MIT
|
||||
URL: https://urllib3.readthedocs.org/
|
||||
Source: https://files.pythonhosted.org/packages/source/u/urllib3/urllib3-%{version}.tar.gz
|
||||
# PATCH-FIX-OPENSUSE openssl-3.2.patch gh#urllib3/urllib3#3271
|
||||
Patch1: openssl-3.2.patch
|
||||
# PATCH-FIX-UPSTREAM https://github.com/urllib3/urllib3/commit/accff72ecc2f6cf5a76d9570198a93ac7c90270e Strip Proxy-Authorization header on redirects
|
||||
Patch2: CVE-2024-37891.patch
|
||||
BuildRequires: %{python_module base >= 3.7}
|
||||
# https://github.com/urllib3/urllib3/issues/3334
|
||||
%define hypercorn_commit d1719f8c1570cbd8e6a3719ffdb14a4d72880abb
|
||||
Source1: https://github.com/urllib3/hypercorn/archive/%{hypercorn_commit}/hypercorn-%{hypercorn_commit}.tar.gz
|
||||
# PATCH-FIX-UPSTREAM https://github.com/urllib3/urllib3/commit/8864ac407bba8607950025e0979c4c69bc7abc7b
|
||||
# Stop decoding response content during redirects needlessly (CVE-2026-21441)
|
||||
Patch1: CVE-2026-21441.patch
|
||||
# PATCH-FIX-UPSTREAM CVE-2025-66471.patch bsc#1254867 gh#urllib3/urllib3#c19571d
|
||||
Patch2: CVE-2025-66471.patch
|
||||
# PATCH-FIX-UPSTREAM CVE-2025-66418.patch bsc#1254866 gh#urllib3/urllib3#24d7b67
|
||||
Patch3: CVE-2025-66418.patch
|
||||
BuildRequires: %{python_module base >= 3.8}
|
||||
BuildRequires: %{python_module hatch-vcs}
|
||||
BuildRequires: %{python_module hatchling}
|
||||
BuildRequires: %{python_module pip}
|
||||
BuildRequires: fdupes
|
||||
BuildRequires: python-rpm-macros
|
||||
#!BuildIgnore: python-requests
|
||||
Requires: ca-certificates-mozilla
|
||||
Requires: python-certifi
|
||||
Requires: python-cryptography >= 1.9
|
||||
Requires: python-idna >= 3.4
|
||||
Requires: python-pyOpenSSL >= 23.2.0
|
||||
Recommends: python-Brotli >= 1.0.9
|
||||
Recommends: python-PySocks >= 1.7.1
|
||||
Recommends: python-h2 >= 4
|
||||
Recommends: python-zstandard >= 0.18
|
||||
BuildArch: noarch
|
||||
%if %{with test}
|
||||
BuildRequires: %{python_module Brotli >= 1.0.9}
|
||||
BuildRequires: %{python_module PySocks >= 1.7.1}
|
||||
BuildRequires: %{python_module certifi}
|
||||
BuildRequires: %{python_module cryptography >= 1.9}
|
||||
BuildRequires: %{python_module Quart >= 0.19}
|
||||
BuildRequires: %{python_module cryptography >= 43}
|
||||
BuildRequires: %{python_module flaky}
|
||||
BuildRequires: %{python_module idna >= 3.4}
|
||||
BuildRequires: %{python_module h2 >= 4.1}
|
||||
BuildRequires: %{python_module httpx >= 0.25}
|
||||
BuildRequires: %{python_module idna >= 3.7}
|
||||
BuildRequires: %{python_module psutil}
|
||||
BuildRequires: %{python_module pyOpenSSL >= 24.2}
|
||||
BuildRequires: %{python_module pytest >= 7.4.0}
|
||||
BuildRequires: %{python_module pytest-socket >= 0.7}
|
||||
BuildRequires: %{python_module pytest-timeout >= 2.1.0}
|
||||
BuildRequires: %{python_module pytest-xdist}
|
||||
BuildRequires: %{python_module tornado >= 6.2}
|
||||
BuildRequires: %{python_module quart-trio >= 0.11}
|
||||
BuildRequires: %{python_module trio >= 0.26}
|
||||
BuildRequires: %{python_module trustme >= 0.9.0}
|
||||
BuildRequires: %{python_module urllib3 >= %{version}}
|
||||
BuildRequires: timezone
|
||||
@@ -88,6 +98,11 @@ Highlights
|
||||
|
||||
%prep
|
||||
%autosetup -p1 -n urllib3-%{version}
|
||||
# https://github.com/urllib3/urllib3/issues/3334
|
||||
%if %{with test}
|
||||
mkdir ../patched-hypercorn
|
||||
tar -C ../patched-hypercorn -zxf %{SOURCE1}
|
||||
%endif
|
||||
|
||||
find . -type f -exec chmod a-x '{}' \;
|
||||
find . -name __pycache__ -type d -exec rm -fr {} +
|
||||
@@ -104,10 +119,12 @@ find . -name __pycache__ -type d -exec rm -fr {} +
|
||||
|
||||
%if %{with test}
|
||||
%check
|
||||
# https://github.com/urllib3/urllib3/issues/3334
|
||||
export PYTHONPATH="$PWD/../patched-hypercorn/hypercorn-%{hypercorn_commit}/src"
|
||||
# gh#urllib3/urllib3#2109
|
||||
export CI="true"
|
||||
# skip some randomly failing tests (mostly on i586, but sometimes they fail on other architectures)
|
||||
skiplist="test_ssl_read_timeout or test_ssl_failed_fingerprint_verification or test_ssl_custom_validation_failure_terminates"
|
||||
skiplist="test_ssl_read_timeout or test_ssl_failed_fingerprint_verification or test_ssl_custom_validation_failure_terminates or test_close_after_handshake"
|
||||
# gh#urllib3/urllib3#1752 and others: upstream's way of checking that the build
|
||||
# system has a correct system time breaks (re-)building the package after too
|
||||
# many months have passed since the last release.
|
||||
@@ -116,7 +133,12 @@ skiplist+=" or test_recent_date"
|
||||
skiplist+=" or test_requesting_large_resources_via_ssl"
|
||||
# Try to access external evil.com
|
||||
skiplist+=" or test_deprecated_no_scheme"
|
||||
%pytest %{?jobs:-n %jobs} -k "not (${skiplist})" --ignore test/with_dummyserver/test_socketlevel.py
|
||||
# weird threading issues on OBS runners
|
||||
skiplist+=" or test_http2_probe_blocked_per_thread"
|
||||
# flaky test, works locally but fails in OBS with
|
||||
# TypeError: _wrap_bio() argument 'incoming' must be _ssl.MemoryBIO, not _ssl.MemoryBIO
|
||||
skiplist+=" or test_https_proxy_forwarding_for_https or test_https_headers_forwarding_for_https"
|
||||
%pytest -W ignore::DeprecationWarning %{?jobs:-n %jobs} -k "not (${skiplist})" --ignore test/with_dummyserver/test_socketlevel.py
|
||||
%endif
|
||||
|
||||
%if ! %{with test}
|
||||
@@ -124,7 +146,7 @@ skiplist+=" or test_deprecated_no_scheme"
|
||||
%license LICENSE.txt
|
||||
%doc CHANGES.rst README.md
|
||||
%{python_sitelib}/urllib3
|
||||
%{python_sitelib}/urllib3-%{version}*-info
|
||||
%{python_sitelib}/urllib3-%{version}.dist-info
|
||||
%endif
|
||||
|
||||
%changelog
|
||||
|
||||
BIN
urllib3-2.1.0.tar.gz
LFS
BIN
urllib3-2.1.0.tar.gz
LFS
Binary file not shown.
BIN
urllib3-2.5.0.tar.gz
LFS
Normal file
BIN
urllib3-2.5.0.tar.gz
LFS
Normal file
Binary file not shown.
Reference in New Issue
Block a user