Compare commits
7 Commits
| Author | SHA256 | Date | |
|---|---|---|---|
| 61ebda221a | |||
| 4f2796afa6 | |||
| 8d3a5f7bb5 | |||
| c6a0d90815 | |||
| 1c5571fbe0 | |||
| 52207b5f15 | |||
| 92830197a0 |
3
beautifulsoup4-4.13.4.tar.gz
Normal file
3
beautifulsoup4-4.13.4.tar.gz
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:dbb3c4e1ceae6aefebdaf2423247260cd062430a410e38c66f2baa50a8437195
|
||||||
|
size 621067
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
version https://git-lfs.github.com/spec/v1
|
|
||||||
oid sha256:6292b1c5186d356bba669ef9f7f051757099565ad9ada5dd630bd9de5fa7fb86
|
|
||||||
size 627737
|
|
||||||
148
htmlparser.patch
148
htmlparser.patch
@@ -1,148 +0,0 @@
|
|||||||
From 55f655ffb7ef03bdd1df0f013743831fe54e3c7a Mon Sep 17 00:00:00 2001
|
|
||||||
From: Leonard Richardson <leonardr@segfault.org>
|
|
||||||
Date: Mon, 8 Dec 2025 19:34:16 -0500
|
|
||||||
Subject: * Change the html.parser tree builder's code for handling numeric
|
|
||||||
character references, to avoid a crash when using Python versions that
|
|
||||||
include the fix to Python issue https://bugs.python.org/issue13633 (e.g.
|
|
||||||
Python 3.11.13). [bug=2134393]
|
|
||||||
|
|
||||||
---
|
|
||||||
CHANGELOG | 5 +++
|
|
||||||
bs4/builder/_htmlparser.py | 78 ++++++++++++++++++++++++++++++++++++--------
|
|
||||||
bs4/tests/test_htmlparser.py | 17 ++++++++++
|
|
||||||
3 files changed, 86 insertions(+), 14 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/bs4/builder/_htmlparser.py b/bs4/builder/_htmlparser.py
|
|
||||||
index 165a3d8..ead800f 100644
|
|
||||||
--- a/bs4/builder/_htmlparser.py
|
|
||||||
+++ b/bs4/builder/_htmlparser.py
|
|
||||||
@@ -10,6 +10,7 @@ __all__ = [
|
|
||||||
]
|
|
||||||
|
|
||||||
from html.parser import HTMLParser
|
|
||||||
+import re
|
|
||||||
|
|
||||||
from typing import (
|
|
||||||
Any,
|
|
||||||
@@ -223,6 +224,64 @@ class BeautifulSoupHTMLParser(HTMLParser, DetectsXMLParsedAsHTML):
|
|
||||||
"""Handle some textual data that shows up between tags."""
|
|
||||||
self.soup.handle_data(data)
|
|
||||||
|
|
||||||
+ _DECIMAL_REFERENCE_WITH_FOLLOWING_DATA = re.compile("^([0-9]+)(.*)")
|
|
||||||
+ _HEX_REFERENCE_WITH_FOLLOWING_DATA = re.compile("^([0-9a-f]+)(.*)")
|
|
||||||
+
|
|
||||||
+ @classmethod
|
|
||||||
+ def _dereference_numeric_character_reference(cls, name:str) -> Tuple[str, bool, str]:
|
|
||||||
+ """Convert a numeric character reference into an actual character.
|
|
||||||
+
|
|
||||||
+ :param name: The number of the character reference, as
|
|
||||||
+ obtained by html.parser
|
|
||||||
+
|
|
||||||
+ :return: A 3-tuple (dereferenced, replacement_added,
|
|
||||||
+ extra_data). `dereferenced` is the dereferenced character
|
|
||||||
+ reference, or the empty string if there was no
|
|
||||||
+ reference. `replacement_added` is True if the reference
|
|
||||||
+ could only be dereferenced by replacing content with U+FFFD
|
|
||||||
+ REPLACEMENT CHARACTER. `extra_data` is a portion of data
|
|
||||||
+ following the character reference, which was deemed to be
|
|
||||||
+ normal data and not part of the reference at all.
|
|
||||||
+ """
|
|
||||||
+ dereferenced:str = ""
|
|
||||||
+ replacement_added:bool = False
|
|
||||||
+ extra_data:str = ""
|
|
||||||
+
|
|
||||||
+ base:int = 10
|
|
||||||
+ reg = cls._DECIMAL_REFERENCE_WITH_FOLLOWING_DATA
|
|
||||||
+ if name.startswith("x") or name.startswith("X"):
|
|
||||||
+ # Hex reference
|
|
||||||
+ name = name[1:]
|
|
||||||
+ base = 16
|
|
||||||
+ reg = cls._HEX_REFERENCE_WITH_FOLLOWING_DATA
|
|
||||||
+
|
|
||||||
+ real_name:Optional[int] = None
|
|
||||||
+ try:
|
|
||||||
+ real_name = int(name, base)
|
|
||||||
+ except ValueError:
|
|
||||||
+ # This is either bad data that starts with what looks like
|
|
||||||
+ # a numeric character reference, or a real numeric
|
|
||||||
+ # reference that wasn't terminated by a semicolon.
|
|
||||||
+ #
|
|
||||||
+ # The fix to https://bugs.python.org/issue13633 made it
|
|
||||||
+ # our responsibility to handle the extra data.
|
|
||||||
+ #
|
|
||||||
+ # To preserve the old behavior, we extract the numeric
|
|
||||||
+ # portion of the incoming "reference" and treat that as a
|
|
||||||
+ # numeric reference. All subsequent data will be processed
|
|
||||||
+ # as string data.
|
|
||||||
+ match = reg.search(name)
|
|
||||||
+ if match is not None:
|
|
||||||
+ real_name = int(match.groups()[0], base)
|
|
||||||
+ extra_data = match.groups()[1]
|
|
||||||
+
|
|
||||||
+ if real_name is None:
|
|
||||||
+ dereferenced = ""
|
|
||||||
+ extra_data = name
|
|
||||||
+ else:
|
|
||||||
+ dereferenced, replacement_added = UnicodeDammit.numeric_character_reference(real_name)
|
|
||||||
+ return dereferenced, replacement_added, extra_data
|
|
||||||
+
|
|
||||||
def handle_charref(self, name: str) -> None:
|
|
||||||
"""Handle a numeric character reference by converting it to the
|
|
||||||
corresponding Unicode character and treating it as textual
|
|
||||||
@@ -230,22 +289,13 @@ class BeautifulSoupHTMLParser(HTMLParser, DetectsXMLParsedAsHTML):
|
|
||||||
|
|
||||||
:param name: Character number, possibly in hexadecimal.
|
|
||||||
"""
|
|
||||||
- # TODO: This was originally a workaround for a bug in
|
|
||||||
- # HTMLParser. (http://bugs.python.org/issue13633) The bug has
|
|
||||||
- # been fixed, but removing this code still makes some
|
|
||||||
- # Beautiful Soup tests fail. This needs investigation.
|
|
||||||
- real_name:int
|
|
||||||
- if name.startswith("x"):
|
|
||||||
- real_name = int(name.lstrip("x"), 16)
|
|
||||||
- elif name.startswith("X"):
|
|
||||||
- real_name = int(name.lstrip("X"), 16)
|
|
||||||
- else:
|
|
||||||
- real_name = int(name)
|
|
||||||
-
|
|
||||||
- data, replacement_added = UnicodeDammit.numeric_character_reference(real_name)
|
|
||||||
+ dereferenced, replacement_added, extra_data = self._dereference_numeric_character_reference(name)
|
|
||||||
if replacement_added:
|
|
||||||
self.soup.contains_replacement_characters = True
|
|
||||||
- self.handle_data(data)
|
|
||||||
+ if dereferenced is not None:
|
|
||||||
+ self.handle_data(dereferenced)
|
|
||||||
+ if extra_data is not None:
|
|
||||||
+ self.handle_data(extra_data)
|
|
||||||
|
|
||||||
def handle_entityref(self, name: str) -> None:
|
|
||||||
"""Handle a named entity reference by converting it to the
|
|
||||||
diff --git a/bs4/tests/test_htmlparser.py b/bs4/tests/test_htmlparser.py
|
|
||||||
index 0086a9d..cb85b53 100644
|
|
||||||
--- a/bs4/tests/test_htmlparser.py
|
|
||||||
+++ b/bs4/tests/test_htmlparser.py
|
|
||||||
@@ -162,3 +162,20 @@ class TestHTMLParserTreeBuilder(HTMLTreeBuilderSmokeTest):
|
|
||||||
# Since we do the replacement ourselves, we can set contains_replacement_characters appropriately.
|
|
||||||
# lxml and html5lib do the replacement so all we ever see is REPLACEMENT CHARACTER.
|
|
||||||
assert soup.contains_replacement_characters == True
|
|
||||||
+
|
|
||||||
+class TestBeautifulSoupHTMLParser:
|
|
||||||
+ def test_dereference_numeric_character_reference(self):
|
|
||||||
+ m = BeautifulSoupHTMLParser._dereference_numeric_character_reference
|
|
||||||
+ assert m("64") == ("@", False, "")
|
|
||||||
+ assert m("x64") == ("d", False, "")
|
|
||||||
+ assert m("X64") == ("d", False, "")
|
|
||||||
+ assert m("64andsomeextra") == ("@", False, "andsomeextra")
|
|
||||||
+ assert m("") == ("", False, "")
|
|
||||||
+ assert m("00whee") == ("<22>", True, "whee")
|
|
||||||
+ assert m("xfffdthatsit") == ("<22>", False, "thatsit")
|
|
||||||
+ assert m("xabcdplussomeextra") == ("ꯍ", False, "plussomeextra")
|
|
||||||
+ assert m("obviouslynotnumeric") == ("", False, "obviouslynotnumeric")
|
|
||||||
+
|
|
||||||
+ # These are almost certainly wrong but at least it doesn't crash.
|
|
||||||
+ assert m("xabcdandsomeextra") == ("\U000abcda", False, "ndsomeextra")
|
|
||||||
+ assert m("xffffffffffffffffffffffbeep") == ("<22>", True, "p")
|
|
||||||
--
|
|
||||||
cgit v1.2.3
|
|
||||||
|
|
||||||
|
|
||||||
@@ -1,84 +1,3 @@
|
|||||||
-------------------------------------------------------------------
|
|
||||||
Mon Dec 29 09:58:48 UTC 2025 - Markéta Machová <mmachova@suse.com>
|
|
||||||
|
|
||||||
- update to 4.14.3
|
|
||||||
* When using one of the lxml tree builders, you can pass in
|
|
||||||
huge_tree=True to disable lxml's security restrictions and process
|
|
||||||
files that include huge text nodes. ("huge" means more than
|
|
||||||
10,000,000 bytes of text in a single node). Without this, lxml may
|
|
||||||
silently stop processing the file after encountering a huge text
|
|
||||||
node.
|
|
||||||
* The html.parser tree builder processes numeric character entities
|
|
||||||
using the algorithm described in the HTML spec. If this means
|
|
||||||
replacing some other character with REPLACEMENT CHARACTER, it will
|
|
||||||
set BeautifulSoup.contains_replacement_characters.
|
|
||||||
* Added a general test of the html.parser tree builder's ability to
|
|
||||||
turn any parsing exception from html.parser into a
|
|
||||||
ParserRejectedMarkup exception. This makes it possible to remove
|
|
||||||
version-dependent tests that depended on the existence of specific
|
|
||||||
bugs in html.parser.
|
|
||||||
- Add htmlparser.patch to fix behaviour with cpython interpreters
|
|
||||||
|
|
||||||
-------------------------------------------------------------------
|
|
||||||
Mon Oct 13 09:11:52 UTC 2025 - Dirk Müller <dmueller@suse.com>
|
|
||||||
|
|
||||||
- update to 4.14.2:
|
|
||||||
* Making ResultSet inherit from MutableSequence still resulted
|
|
||||||
in too many breaking changes in users of the library,
|
|
||||||
so I reverted the ResultSet code back to where it was in 4.13.5
|
|
||||||
and added tests of all known breaking behavior. [bug=2125906]
|
|
||||||
* Made ResultSet inherit from MutableSequence instead of
|
|
||||||
Sequence, since lots of existing code treats ResultSet as a
|
|
||||||
mutable list.
|
|
||||||
* This version adds function overloading to the find_* methods
|
|
||||||
to make it easier to write type-safe Python.
|
|
||||||
* The typing for find_parent() and find_parents() was improved
|
|
||||||
without any overloading. Casts should never be necessary,
|
|
||||||
since those methods only ever return Tag and ResultSet[Tag],
|
|
||||||
respectively.
|
|
||||||
* ResultSet now inherits from Sequence. This should make it
|
|
||||||
easier to incorporate ResultSet objects into your type system
|
|
||||||
without needing to handle ResultSet specially.
|
|
||||||
* Fixed an unhandled exception when creating the string
|
|
||||||
representation of a decomposed element.
|
|
||||||
* The default value for the 'attrs' attribute in find* methods
|
|
||||||
is now None, not the empty dictionary. This should have no visible
|
|
||||||
effect on anything.
|
|
||||||
|
|
||||||
-------------------------------------------------------------------
|
|
||||||
Wed Sep 10 07:16:20 UTC 2025 - John Paul Adrian Glaubitz <adrian.glaubitz@suse.com>
|
|
||||||
|
|
||||||
- Update to 4.13.5
|
|
||||||
* Fixed an unhandled exception when parsing invalid markup that contains the { character
|
|
||||||
when using lxml==6.0.0. [bug=2116306]
|
|
||||||
* Fixed a regression when matching a multi-valued attribute against the
|
|
||||||
empty string. [bug=2115352]
|
|
||||||
* Unit tests and test case data are no longer packaged with the wheel. [bug=2107495]
|
|
||||||
* Fixed a bug that gave the wrong result when parsing the empty bytestring. [bug=2110492]
|
|
||||||
* Brought the Spanish translation of the documentation up to date with
|
|
||||||
4.13.4. Courtesy of Carlos Romero.
|
|
||||||
* For Python 3.13 and above, disabled tests that verify Beautiful Soup's handling of htmlparser's
|
|
||||||
exceptions when given very bad markup. The bug in htmlparser that caused
|
|
||||||
this behavior has been fixed. Patch courtesy of Stefano Rivera.
|
|
||||||
* Used overloading to improve type hints for prettify().
|
|
||||||
* Updated the SoupStrainer documentation to clarify that during initial
|
|
||||||
parsing, attribute values are always passed into the SoupStrainer as raw strings. [bug=2111651]
|
|
||||||
* Fixed all type checking errors issued by pyright. (Previously only mypy
|
|
||||||
was used for type checking.)
|
|
||||||
* Improved the type hints for PageElement.replace_with. [bug=2114746]
|
|
||||||
* Improved the type hint for the arguments of the lambda function that can
|
|
||||||
be used to match a tag's attribute. [bug=2110401]
|
|
||||||
* Modified some of the lxml tests to accommodate behavioral changes in libxml2
|
|
||||||
2.14.3. Specifically:
|
|
||||||
|
|
||||||
1. XML declarations and processing instructions in HTML documents
|
|
||||||
are rewritten as comments. Note that this means XHTML documents will
|
|
||||||
now turn into regular HTML documents if run through the 'lxml'
|
|
||||||
parser. The 'xml' parser is unaffected.
|
|
||||||
|
|
||||||
2. Out-of-range numeric entities are replaced with REPLACEMENT
|
|
||||||
CHARACTER rather than omitted entirely. [bug=2112242]
|
|
||||||
|
|
||||||
-------------------------------------------------------------------
|
-------------------------------------------------------------------
|
||||||
Sun Jul 13 14:04:39 UTC 2025 - Ben Greiner <code@bnavigator.de>
|
Sun Jul 13 14:04:39 UTC 2025 - Ben Greiner <code@bnavigator.de>
|
||||||
|
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
#
|
#
|
||||||
# spec file for package python-beautifulsoup4
|
# spec file for package python-beautifulsoup4
|
||||||
#
|
#
|
||||||
# Copyright (c) 2025 SUSE LLC and contributors
|
# Copyright (c) 2025 SUSE LLC
|
||||||
#
|
#
|
||||||
# All modifications and additions to the file contributed by third parties
|
# All modifications and additions to the file contributed by third parties
|
||||||
# remain the property of their copyright owners, unless otherwise agreed
|
# remain the property of their copyright owners, unless otherwise agreed
|
||||||
@@ -18,14 +18,12 @@
|
|||||||
|
|
||||||
%{?sle15_python_module_pythons}
|
%{?sle15_python_module_pythons}
|
||||||
Name: python-beautifulsoup4
|
Name: python-beautifulsoup4
|
||||||
Version: 4.14.3
|
Version: 4.13.4
|
||||||
Release: 0
|
Release: 0
|
||||||
Summary: HTML/XML Parser for Quick-Turnaround Applications Like Screen-Scraping
|
Summary: HTML/XML Parser for Quick-Turnaround Applications Like Screen-Scraping
|
||||||
License: MIT
|
License: MIT
|
||||||
URL: https://www.crummy.com/software/BeautifulSoup/
|
URL: https://www.crummy.com/software/BeautifulSoup/
|
||||||
Source: https://files.pythonhosted.org/packages/source/b/beautifulsoup4/beautifulsoup4-%{version}.tar.gz
|
Source: https://files.pythonhosted.org/packages/source/b/beautifulsoup4/beautifulsoup4-%{version}.tar.gz
|
||||||
# PATCH-FIX-UPSTREAM 55f655ffb7ef03bdd1df0f013743831fe54e3c7a Change the html.parser tree builder's code for handling numeric character references
|
|
||||||
Patch0: htmlparser.patch
|
|
||||||
BuildRequires: %{python_module base >= 3.7}
|
BuildRequires: %{python_module base >= 3.7}
|
||||||
BuildRequires: %{python_module hatchling}
|
BuildRequires: %{python_module hatchling}
|
||||||
BuildRequires: %{python_module pip}
|
BuildRequires: %{python_module pip}
|
||||||
|
|||||||
Reference in New Issue
Block a user