From 82d122dfa526e379d7e259f0061c12909fe2e098d64cdd7fd63edc28798bd2e5 Mon Sep 17 00:00:00 2001 From: Guang Yee Date: Wed, 18 Dec 2024 17:58:42 +0000 Subject: [PATCH] Accepting request 1231910 from home:uncomfyhalomacro:branches:science:machinelearning - Update to version 0.21.0: * More cache options. * Disable caching for long strings. * Testing ABI3 wheels to reduce number of wheels * Adding an API for decode streaming. * Decode stream python * Fix encode_batch and encode_batch_fast to accept ndarrays again OBS-URL: https://build.opensuse.org/request/show/1231910 OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/python-tokenizers?expand=0&rev=8 --- _service | 2 +- python-tokenizers.changes | 11 +++++++++++ python-tokenizers.spec | 2 +- registry.tar.zst | 4 ++-- tokenizers-0.20.3.tar.gz | 3 --- tokenizers-0.21.0.tar.gz | 3 +++ 6 files changed, 18 insertions(+), 7 deletions(-) delete mode 100644 tokenizers-0.20.3.tar.gz create mode 100644 tokenizers-0.21.0.tar.gz diff --git a/_service b/_service index 468871a..48f9076 100644 --- a/_service +++ b/_service @@ -2,7 +2,7 @@ - tokenizers*.tar.gz + tokenizers-*.tar.gz registry true zst diff --git a/python-tokenizers.changes b/python-tokenizers.changes index c3d3203..b9f61d0 100644 --- a/python-tokenizers.changes +++ b/python-tokenizers.changes @@ -1,3 +1,14 @@ +------------------------------------------------------------------- +Wed Dec 18 14:20:07 UTC 2024 - Soc Virnyl Estela + +- Update to version 0.21.0: + * More cache options. + * Disable caching for long strings. + * Testing ABI3 wheels to reduce number of wheels + * Adding an API for decode streaming. + * Decode stream python + * Fix encode_batch and encode_batch_fast to accept ndarrays again + ------------------------------------------------------------------- Thu Nov 7 11:30:50 UTC 2024 - Soc Virnyl Estela diff --git a/python-tokenizers.spec b/python-tokenizers.spec index adf2c45..728c8cb 100644 --- a/python-tokenizers.spec +++ b/python-tokenizers.spec @@ -22,7 +22,7 @@ %{?sle15_python_module_pythons} Name: python-tokenizers -Version: 0.20.3 +Version: 0.21.0 Release: 0 Summary: Provides an implementation of today's most used tokenizers License: Apache-2.0 diff --git a/registry.tar.zst b/registry.tar.zst index a52b9c1..4d1963b 100644 --- a/registry.tar.zst +++ b/registry.tar.zst @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:24cd5d9d9fa151e42867be3979aaa00ab3ea4ac1fa1bb7650c7b0803986d7cad -size 46442951 +oid sha256:c6e2a7b1f52a7523f92f36544bd7088ee7a96622f493a7c1a4f353c25078bb74 +size 46535229 diff --git a/tokenizers-0.20.3.tar.gz b/tokenizers-0.20.3.tar.gz deleted file mode 100644 index 60b37c9..0000000 --- a/tokenizers-0.20.3.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:21e9a235c72e49cafb7ed29829650f6a49f0a951194e1f0168b3b2f547362569 -size 1539739 diff --git a/tokenizers-0.21.0.tar.gz b/tokenizers-0.21.0.tar.gz new file mode 100644 index 0000000..dbfb422 --- /dev/null +++ b/tokenizers-0.21.0.tar.gz @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:841279ad797d575ed3cf31fc4f30e09e37acbd35028d30c51fc0879ef7ed4094 +size 1544853