Compare commits
2 Commits
| Author | SHA256 | Date | |
|---|---|---|---|
| a232380790 | |||
| 9dcc206ed6 |
219
CVE-2025-54121.patch
Normal file
219
CVE-2025-54121.patch
Normal file
@@ -0,0 +1,219 @@
|
||||
From 9f7ec2eb512fcc3fe90b43cb9dd9e1d08696bec1 Mon Sep 17 00:00:00 2001
|
||||
From: Michael Honaker <37811263+HonakerM@users.noreply.github.com>
|
||||
Date: Mon, 21 Jul 2025 02:24:02 +0900
|
||||
Subject: [PATCH] Make UploadFile check for future rollover (#2962)
|
||||
|
||||
Co-authored-by: Marcelo Trylesinski <marcelotryle@gmail.com>
|
||||
---
|
||||
starlette/datastructures.py | 22 ++++++++++---
|
||||
tests/test_formparsers.py | 66 +++++++++++++++++++++++++++++++++++--
|
||||
2 files changed, 82 insertions(+), 6 deletions(-)
|
||||
|
||||
Index: starlette-0.41.3/starlette/datastructures.py
|
||||
===================================================================
|
||||
--- starlette-0.41.3.orig/starlette/datastructures.py
|
||||
+++ starlette-0.41.3/starlette/datastructures.py
|
||||
@@ -424,6 +424,10 @@ class UploadFile:
|
||||
self.size = size
|
||||
self.headers = headers or Headers()
|
||||
|
||||
+ # Capture max size from SpooledTemporaryFile if one is provided. This slightly speeds up future checks.
|
||||
+ # Note 0 means unlimited mirroring SpooledTemporaryFile's __init__
|
||||
+ self._max_mem_size = getattr(self.file, "_max_size", 0)
|
||||
+
|
||||
@property
|
||||
def content_type(self) -> str | None:
|
||||
return self.headers.get("content-type", None)
|
||||
@@ -434,14 +438,24 @@ class UploadFile:
|
||||
rolled_to_disk = getattr(self.file, "_rolled", True)
|
||||
return not rolled_to_disk
|
||||
|
||||
+ def _will_roll(self, size_to_add: int) -> bool:
|
||||
+ # If we're not in_memory then we will always roll
|
||||
+ if not self._in_memory:
|
||||
+ return True
|
||||
+
|
||||
+ # Check for SpooledTemporaryFile._max_size
|
||||
+ future_size = self.file.tell() + size_to_add
|
||||
+ return bool(future_size > self._max_mem_size) if self._max_mem_size else False
|
||||
+
|
||||
async def write(self, data: bytes) -> None:
|
||||
+ new_data_len = len(data)
|
||||
if self.size is not None:
|
||||
- self.size += len(data)
|
||||
+ self.size += new_data_len
|
||||
|
||||
- if self._in_memory:
|
||||
- self.file.write(data)
|
||||
- else:
|
||||
+ if self._will_roll(new_data_len):
|
||||
await run_in_threadpool(self.file.write, data)
|
||||
+ else:
|
||||
+ self.file.write(data)
|
||||
|
||||
async def read(self, size: int = -1) -> bytes:
|
||||
if self._in_memory:
|
||||
Index: starlette-0.41.3/tests/test_formparsers.py
|
||||
===================================================================
|
||||
--- starlette-0.41.3.orig/tests/test_formparsers.py
|
||||
+++ starlette-0.41.3/tests/test_formparsers.py
|
||||
@@ -1,15 +1,21 @@
|
||||
from __future__ import annotations
|
||||
|
||||
import os
|
||||
+import threading
|
||||
import typing
|
||||
+from collections.abc import Generator
|
||||
from contextlib import nullcontext as does_not_raise
|
||||
+from io import BytesIO
|
||||
from pathlib import Path
|
||||
+from tempfile import SpooledTemporaryFile
|
||||
+from typing import Any, ClassVar
|
||||
+from unittest import mock
|
||||
|
||||
import pytest
|
||||
|
||||
from starlette.applications import Starlette
|
||||
from starlette.datastructures import UploadFile
|
||||
-from starlette.formparsers import MultiPartException, _user_safe_decode
|
||||
+from starlette.formparsers import MultiPartException, MultiPartParser, _user_safe_decode
|
||||
from starlette.requests import Request
|
||||
from starlette.responses import JSONResponse
|
||||
from starlette.routing import Mount
|
||||
@@ -104,6 +110,22 @@ async def app_read_body(scope: Scope, re
|
||||
await response(scope, receive, send)
|
||||
|
||||
|
||||
+async def app_monitor_thread(scope: Scope, receive: Receive, send: Send) -> None:
|
||||
+ """Helper app to monitor what thread the app was called on.
|
||||
+
|
||||
+ This can later be used to validate thread/event loop operations.
|
||||
+ """
|
||||
+ request = Request(scope, receive)
|
||||
+
|
||||
+ # Make sure we parse the form
|
||||
+ await request.form()
|
||||
+ await request.close()
|
||||
+
|
||||
+ # Send back the current thread id
|
||||
+ response = JSONResponse({"thread_ident": threading.current_thread().ident})
|
||||
+ await response(scope, receive, send)
|
||||
+
|
||||
+
|
||||
def make_app_max_parts(max_files: int = 1000, max_fields: int = 1000) -> ASGIApp:
|
||||
async def app(scope: Scope, receive: Receive, send: Send) -> None:
|
||||
request = Request(scope, receive)
|
||||
@@ -303,6 +325,88 @@ def test_multipart_request_mixed_files_a
|
||||
}
|
||||
|
||||
|
||||
+class ThreadTrackingSpooledTemporaryFile(SpooledTemporaryFile[bytes]):
|
||||
+ """Helper class to track which threads performed the rollover operation.
|
||||
+
|
||||
+ This is not threadsafe/multi-test safe.
|
||||
+ """
|
||||
+
|
||||
+ rollover_threads: ClassVar[set[int | None]] = set()
|
||||
+
|
||||
+ def rollover(self) -> None:
|
||||
+ ThreadTrackingSpooledTemporaryFile.rollover_threads.add(threading.current_thread().ident)
|
||||
+ super().rollover()
|
||||
+
|
||||
+
|
||||
+@pytest.fixture
|
||||
+def mock_spooled_temporary_file() -> Generator[None]:
|
||||
+ try:
|
||||
+ with mock.patch("starlette.formparsers.SpooledTemporaryFile", ThreadTrackingSpooledTemporaryFile):
|
||||
+ yield
|
||||
+ finally:
|
||||
+ ThreadTrackingSpooledTemporaryFile.rollover_threads.clear()
|
||||
+
|
||||
+
|
||||
+def test_multipart_request_large_file_rollover_in_background_thread(
|
||||
+ mock_spooled_temporary_file: None, test_client_factory: TestClientFactory
|
||||
+) -> None:
|
||||
+ """Test that Spooled file rollovers happen in background threads."""
|
||||
+ data = BytesIO(b" " * (MultiPartParser.spool_max_size + 1))
|
||||
+
|
||||
+ client = test_client_factory(app_monitor_thread)
|
||||
+ response = client.post("/", files=[("test_large", data)])
|
||||
+ assert response.status_code == 200
|
||||
+
|
||||
+ # Parse the event thread id from the API response and ensure we have one
|
||||
+ app_thread_ident = response.json().get("thread_ident")
|
||||
+ assert app_thread_ident is not None
|
||||
+
|
||||
+ # Ensure the app thread was not the same as the rollover one and that a rollover thread exists
|
||||
+ assert app_thread_ident not in ThreadTrackingSpooledTemporaryFile.rollover_threads
|
||||
+ assert len(ThreadTrackingSpooledTemporaryFile.rollover_threads) == 1
|
||||
+
|
||||
+
|
||||
+class ThreadTrackingSpooledTemporaryFile(SpooledTemporaryFile[bytes]):
|
||||
+ """Helper class to track which threads performed the rollover operation.
|
||||
+
|
||||
+ This is not threadsafe/multi-test safe.
|
||||
+ """
|
||||
+
|
||||
+ rollover_threads: ClassVar[set[int | None]] = set()
|
||||
+
|
||||
+ def rollover(self) -> None:
|
||||
+ ThreadTrackingSpooledTemporaryFile.rollover_threads.add(threading.current_thread().ident)
|
||||
+ super().rollover()
|
||||
+
|
||||
+
|
||||
+@pytest.fixture
|
||||
+def mock_spooled_temporary_file() -> Generator[None]:
|
||||
+ try:
|
||||
+ with mock.patch("starlette.formparsers.SpooledTemporaryFile", ThreadTrackingSpooledTemporaryFile):
|
||||
+ yield
|
||||
+ finally:
|
||||
+ ThreadTrackingSpooledTemporaryFile.rollover_threads.clear()
|
||||
+
|
||||
+
|
||||
+def test_multipart_request_large_file_rollover_in_background_thread(
|
||||
+ mock_spooled_temporary_file: None, test_client_factory: TestClientFactory
|
||||
+) -> None:
|
||||
+ """Test that Spooled file rollovers happen in background threads."""
|
||||
+ data = BytesIO(b" " * (MultiPartParser.spool_max_size + 1))
|
||||
+
|
||||
+ client = test_client_factory(app_monitor_thread)
|
||||
+ response = client.post("/", files=[("test_large", data)])
|
||||
+ assert response.status_code == 200
|
||||
+
|
||||
+ # Parse the event thread id from the API response and ensure we have one
|
||||
+ app_thread_ident = response.json().get("thread_ident")
|
||||
+ assert app_thread_ident is not None
|
||||
+
|
||||
+ # Ensure the app thread was not the same as the rollover one and that a rollover thread exists
|
||||
+ assert app_thread_ident not in ThreadTrackingSpooledTemporaryFile.rollover_threads
|
||||
+ assert len(ThreadTrackingSpooledTemporaryFile.rollover_threads) == 1
|
||||
+
|
||||
+
|
||||
def test_multipart_request_with_charset_for_filename(tmpdir: Path, test_client_factory: TestClientFactory) -> None:
|
||||
client = test_client_factory(app)
|
||||
response = client.post(
|
||||
Index: starlette-0.41.3/starlette/formparsers.py
|
||||
===================================================================
|
||||
--- starlette-0.41.3.orig/starlette/formparsers.py
|
||||
+++ starlette-0.41.3/starlette/formparsers.py
|
||||
@@ -122,7 +122,10 @@ class FormParser:
|
||||
|
||||
|
||||
class MultiPartParser:
|
||||
- max_file_size = 1024 * 1024 # 1MB
|
||||
+ spool_max_size = 1024 * 1024 # 1MB
|
||||
+ """The maximum size of the spooled temporary file used to store file data."""
|
||||
+ max_part_size = 1024 * 1024 # 1MB
|
||||
+ """The maximum size of a part in the multipart request."""
|
||||
max_part_size = 1024 * 1024 # 1MB
|
||||
|
||||
def __init__(
|
||||
@@ -201,7 +204,7 @@ class MultiPartParser:
|
||||
if self._current_files > self.max_files:
|
||||
raise MultiPartException(f"Too many files. Maximum number of files is {self.max_files}.")
|
||||
filename = _user_safe_decode(options[b"filename"], self._charset)
|
||||
- tempfile = SpooledTemporaryFile(max_size=self.max_file_size)
|
||||
+ tempfile = SpooledTemporaryFile(max_size=self.spool_max_size)
|
||||
self._files_to_close_on_error.append(tempfile)
|
||||
self._current_part.file = UploadFile(
|
||||
file=tempfile, # type: ignore[arg-type]
|
||||
134
CVE-2025-62727.patch
Normal file
134
CVE-2025-62727.patch
Normal file
@@ -0,0 +1,134 @@
|
||||
From 4ea6e22b489ec388d6004cfbca52dd5b147127c5 Mon Sep 17 00:00:00 2001
|
||||
From: Marcelo Trylesinski <marcelotryle@gmail.com>
|
||||
Date: Tue, 28 Oct 2025 18:14:01 +0100
|
||||
Subject: [PATCH] Merge commit from fork
|
||||
|
||||
---
|
||||
starlette/responses.py | 46 ++++++++++++++++++++++++++++-------------
|
||||
tests/test_responses.py | 28 +++++++++++++++++++++++++
|
||||
2 files changed, 60 insertions(+), 14 deletions(-)
|
||||
|
||||
Index: starlette-0.41.3/starlette/responses.py
|
||||
===================================================================
|
||||
--- starlette-0.41.3.orig/starlette/responses.py
|
||||
+++ starlette-0.41.3/starlette/responses.py
|
||||
@@ -3,7 +3,6 @@ from __future__ import annotations
|
||||
import http.cookies
|
||||
import json
|
||||
import os
|
||||
-import re
|
||||
import stat
|
||||
import typing
|
||||
import warnings
|
||||
@@ -272,9 +271,6 @@ class RangeNotSatisfiable(Exception):
|
||||
self.max_size = max_size
|
||||
|
||||
|
||||
-_RANGE_PATTERN = re.compile(r"(\d*)-(\d*)")
|
||||
-
|
||||
-
|
||||
class FileResponse(Response):
|
||||
chunk_size = 64 * 1024
|
||||
|
||||
@@ -435,8 +431,8 @@ class FileResponse(Response):
|
||||
etag = f'"{md5_hexdigest(etag_base.encode(), usedforsecurity=False)}"'
|
||||
return http_if_range == formatdate(stat_result.st_mtime, usegmt=True) or http_if_range == etag
|
||||
|
||||
- @staticmethod
|
||||
- def _parse_range_header(http_range: str, file_size: int) -> list[tuple[int, int]]:
|
||||
+ @classmethod
|
||||
+ def _parse_range_header(cls, http_range: str, file_size: int) -> list[tuple[int, int]]:
|
||||
ranges: list[tuple[int, int]] = []
|
||||
try:
|
||||
units, range_ = http_range.split("=", 1)
|
||||
@@ -448,14 +444,7 @@ class FileResponse(Response):
|
||||
if units != "bytes":
|
||||
raise MalformedRangeHeader("Only support bytes range")
|
||||
|
||||
- ranges = [
|
||||
- (
|
||||
- int(_[0]) if _[0] else file_size - int(_[1]),
|
||||
- int(_[1]) + 1 if _[0] and _[1] and int(_[1]) < file_size else file_size,
|
||||
- )
|
||||
- for _ in _RANGE_PATTERN.findall(range_)
|
||||
- if _ != ("", "")
|
||||
- ]
|
||||
+ ranges = cls._parse_ranges(range_, file_size)
|
||||
|
||||
if len(ranges) == 0:
|
||||
raise MalformedRangeHeader("Range header: range must be requested")
|
||||
@@ -487,6 +476,35 @@ class FileResponse(Response):
|
||||
|
||||
return result
|
||||
|
||||
+ @classmethod
|
||||
+ def _parse_ranges(cls, range_: str, file_size: int) -> list[tuple[int, int]]:
|
||||
+ ranges: list[tuple[int, int]] = []
|
||||
+
|
||||
+ for part in range_.split(","):
|
||||
+ part = part.strip()
|
||||
+
|
||||
+ # If the range is empty or a single dash, we ignore it.
|
||||
+ if not part or part == "-":
|
||||
+ continue
|
||||
+
|
||||
+ # If the range is not in the format "start-end", we ignore it.
|
||||
+ if "-" not in part:
|
||||
+ continue
|
||||
+
|
||||
+ start_str, end_str = part.split("-", 1)
|
||||
+ start_str = start_str.strip()
|
||||
+ end_str = end_str.strip()
|
||||
+
|
||||
+ try:
|
||||
+ start = int(start_str) if start_str else file_size - int(end_str)
|
||||
+ end = int(end_str) + 1 if start_str and end_str and int(end_str) < file_size else file_size
|
||||
+ ranges.append((start, end))
|
||||
+ except ValueError:
|
||||
+ # If the range is not numeric, we ignore it.
|
||||
+ continue
|
||||
+
|
||||
+ return ranges
|
||||
+
|
||||
def generate_multipart(
|
||||
self,
|
||||
ranges: typing.Sequence[tuple[int, int]],
|
||||
Index: starlette-0.41.3/tests/test_responses.py
|
||||
===================================================================
|
||||
--- starlette-0.41.3.orig/tests/test_responses.py
|
||||
+++ starlette-0.41.3/tests/test_responses.py
|
||||
@@ -684,6 +684,34 @@ def test_file_response_insert_ranges(fil
|
||||
]
|
||||
|
||||
|
||||
+def test_file_response_range_without_dash(file_response_client: TestClient) -> None:
|
||||
+ response = file_response_client.get("/", headers={"Range": "bytes=100, 0-50"})
|
||||
+ assert response.status_code == 206
|
||||
+ assert response.headers["content-range"] == f"bytes 0-50/{len(README.encode('utf8'))}"
|
||||
+
|
||||
+
|
||||
+def test_file_response_range_empty_start_and_end(file_response_client: TestClient) -> None:
|
||||
+ response = file_response_client.get("/", headers={"Range": "bytes= - , 0-50"})
|
||||
+ assert response.status_code == 206
|
||||
+ assert response.headers["content-range"] == f"bytes 0-50/{len(README.encode('utf8'))}"
|
||||
+
|
||||
+
|
||||
+def test_file_response_range_ignore_non_numeric(file_response_client: TestClient) -> None:
|
||||
+ response = file_response_client.get("/", headers={"Range": "bytes=abc-def, 0-50"})
|
||||
+ assert response.status_code == 206
|
||||
+ assert response.headers["content-range"] == f"bytes 0-50/{len(README.encode('utf8'))}"
|
||||
+
|
||||
+
|
||||
+def test_file_response_suffix_range(file_response_client: TestClient) -> None:
|
||||
+ # Test suffix range (last N bytes) - line 523 with empty start_str
|
||||
+ response = file_response_client.get("/", headers={"Range": "bytes=-100"})
|
||||
+ assert response.status_code == 206
|
||||
+ file_size = len(README.encode("utf8"))
|
||||
+ assert response.headers["content-range"] == f"bytes {file_size - 100}-{file_size - 1}/{file_size}"
|
||||
+ assert response.headers["content-length"] == "100"
|
||||
+ assert response.content == README.encode("utf8")[-100:]
|
||||
+
|
||||
+
|
||||
@pytest.mark.anyio
|
||||
async def test_file_response_multi_small_chunk_size(readme_file: Path) -> None:
|
||||
class SmallChunkSizeFileResponse(FileResponse):
|
||||
@@ -1,3 +1,13 @@
|
||||
-------------------------------------------------------------------
|
||||
Wed Oct 29 14:29:58 UTC 2025 - Nico Krapp <nico.krapp@suse.com>
|
||||
|
||||
- Add CVE-2025-62727.patch to fix CVE-2025-62727 (bsc#1252805)
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Thu Jul 31 09:20:02 UTC 2025 - Nico Krapp <nico.krapp@suse.com>
|
||||
|
||||
- Add CVE-2025-54121.patch to fix CVE-2025-54121 (bsc#1246855)
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Thu Dec 12 16:09:18 UTC 2024 - Ben Greiner <code@bnavigator.de>
|
||||
|
||||
|
||||
@@ -35,6 +35,12 @@ URL: https://github.com/encode/starlette
|
||||
Source: https://github.com/encode/starlette/archive/refs/tags/%{version}.tar.gz#/starlette-%{version}.tar.gz
|
||||
# PATCH-FIX-UPSTREAM starlette-pr2773-httpx0.28.patch gh#encode/starlette#2773
|
||||
Patch0: https://github.com/encode/starlette/pull/2773.patch#/starlette-pr2773-httpx0.28.patch
|
||||
# PATCH-FIX-UPSTREAM CVE-2025-54121.patch bsc#1246855
|
||||
# taken from https://github.com/encode/starlette/commit/9f7ec2eb512fcc3fe90b43cb9dd9e1d08696bec1
|
||||
# and https://github.com/encode/starlette/commit/4ae3213ca557a25aff91f1d43d530ca3962c729d
|
||||
Patch1: CVE-2025-54121.patch
|
||||
# PATCH-FIX-UPSTREAM CVE-2025-62727.patch https://github.com/Kludex/starlette/commit/4ea6e22b489ec388d6004cfbca52dd5b147127c5
|
||||
Patch2: CVE-2025-62727.patch
|
||||
BuildRequires: %{python_module base >= 3.8}
|
||||
BuildRequires: %{python_module hatchling}
|
||||
BuildRequires: %{python_module pip}
|
||||
|
||||
Reference in New Issue
Block a user