Accepting request 1222485 from science:machinelearning
OBS-URL: https://build.opensuse.org/request/show/1222485 OBS-URL: https://build.opensuse.org/package/show/openSUSE:Factory/ollama?expand=0&rev=24
This commit is contained in:
commit
d62d34d2ac
2
_service
2
_service
@ -3,7 +3,7 @@
|
|||||||
<service name="obs_scm" mode="manual">
|
<service name="obs_scm" mode="manual">
|
||||||
<param name="url">https://github.com/ollama/ollama.git</param>
|
<param name="url">https://github.com/ollama/ollama.git</param>
|
||||||
<param name="scm">git</param>
|
<param name="scm">git</param>
|
||||||
<param name="revision">v0.3.14</param>
|
<param name="revision">v0.4.0</param>
|
||||||
<param name="versionformat">@PARENT_TAG@</param>
|
<param name="versionformat">@PARENT_TAG@</param>
|
||||||
<param name="versionrewrite-pattern">v(.*)</param>
|
<param name="versionrewrite-pattern">v(.*)</param>
|
||||||
<param name="changesgenerate">enable</param>
|
<param name="changesgenerate">enable</param>
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
<servicedata>
|
<servicedata>
|
||||||
<service name="tar_scm">
|
<service name="tar_scm">
|
||||||
<param name="url">https://github.com/ollama/ollama.git</param>
|
<param name="url">https://github.com/ollama/ollama.git</param>
|
||||||
<param name="changesrevision">f2890a4494f9fb3722ee7a4c506252362d1eab65</param></service></servicedata>
|
<param name="changesrevision">9d71bcc3e2a97c8e62d758450f43aa212346410e</param></service></servicedata>
|
@ -1,20 +0,0 @@
|
|||||||
--- a/llm/generate/gen_linux.sh.orig 2024-09-17 12:52:41.511508050 +0200
|
|
||||||
+++ b/llm/generate/gen_linux.sh 2024-09-17 13:01:55.316347171 +0200
|
|
||||||
@@ -52,7 +52,7 @@
|
|
||||||
export CUDACXX=$(command -v nvcc)
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
-COMMON_CMAKE_DEFS="-DCMAKE_SKIP_RPATH=on -DBUILD_SHARED_LIBS=on -DCMAKE_POSITION_INDEPENDENT_CODE=on -DGGML_NATIVE=off -DGGML_AVX=on -DGGML_AVX2=off -DGGML_AVX512=off -DGGML_FMA=off -DGGML_F16C=off -DGGML_OPENMP=off"
|
|
||||||
+COMMON_CMAKE_DEFS="-DCMAKE_SKIP_RPATH=on -DBUILD_SHARED_LIBS=on -DCMAKE_POSITION_INDEPENDENT_CODE=on -DGGML_NATIVE=off -DGGML_AVX=on -DGGML_AVX2=off -DGGML_AVX512=off -DGGML_FMA=off -DGGML_F16C=off -DGGML_OPENMP=off -DGGML_LTO=on"
|
|
||||||
source $(dirname $0)/gen_common.sh
|
|
||||||
init_vars
|
|
||||||
git_module_setup
|
|
||||||
@@ -95,7 +95,7 @@
|
|
||||||
# -DGGML_AVX512_VBMI -- 2018 Intel Cannon Lake
|
|
||||||
# -DGGML_AVX512_VNNI -- 2021 Intel Alder Lake
|
|
||||||
|
|
||||||
- COMMON_CPU_DEFS="-DBUILD_SHARED_LIBS=on -DCMAKE_POSITION_INDEPENDENT_CODE=on -DGGML_NATIVE=off -DGGML_OPENMP=off"
|
|
||||||
+ COMMON_CPU_DEFS="-DBUILD_SHARED_LIBS=on -DCMAKE_POSITION_INDEPENDENT_CODE=on -DGGML_NATIVE=off -DGGML_OPENMP=off -DGGML_LTO=on"
|
|
||||||
if [ -z "${OLLAMA_CPU_TARGET}" -o "${OLLAMA_CPU_TARGET}" = "cpu" ]; then
|
|
||||||
#
|
|
||||||
# CPU first for the default library, set up as lowest common denominator for maximum compatibility (including Rosetta)
|
|
BIN
ollama-0.3.14.obscpio
(Stored with Git LFS)
BIN
ollama-0.3.14.obscpio
(Stored with Git LFS)
Binary file not shown.
BIN
ollama-0.4.0.obscpio
(Stored with Git LFS)
Normal file
BIN
ollama-0.4.0.obscpio
(Stored with Git LFS)
Normal file
Binary file not shown.
@ -1,3 +1,65 @@
|
|||||||
|
-------------------------------------------------------------------
|
||||||
|
Thu Nov 07 12:06:09 UTC 2024 - adrian@suse.de
|
||||||
|
|
||||||
|
- Update to version 0.4.0:
|
||||||
|
* Update README.md (#7516)
|
||||||
|
* One corrupt manifest should not wedge model operations (#7515)
|
||||||
|
* prompt: Use a single token when estimating mllama context size
|
||||||
|
* readme: add Hexabot to the list of community integrations
|
||||||
|
* Quiet down debug log of image payload (#7454)
|
||||||
|
|
||||||
|
-------------------------------------------------------------------
|
||||||
|
Wed Nov 06 12:31:53 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
|
||||||
|
|
||||||
|
- Update to version 0.4.0-rc8:
|
||||||
|
* CI: Switch to v13 macos runner (#7498)
|
||||||
|
* CI: matrix strategy fix (#7496)
|
||||||
|
* Sign windows arm64 official binaries (#7493)
|
||||||
|
* readme: add TextCraft to community integrations (#7377)
|
||||||
|
* nvidia libs have inconsistent ordering (#7473)
|
||||||
|
* CI: omit unused tools for faster release builds (#7432)
|
||||||
|
* llama: Improve error handling
|
||||||
|
* runner.go: Only allocate 1 element embedding batches for mllama
|
||||||
|
* refactor kv estimation
|
||||||
|
* mllama cross attention
|
||||||
|
* Add basic mllama integration tests (#7455)
|
||||||
|
* runner.go: Don't set cross attention before sending embeddings
|
||||||
|
* Give unicode test more time to run (#7437)
|
||||||
|
|
||||||
|
-------------------------------------------------------------------
|
||||||
|
Fri Nov 01 02:18:50 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
|
||||||
|
|
||||||
|
- Remove enable-lto.patch
|
||||||
|
|
||||||
|
- Update to version 0.4.0-rc6:
|
||||||
|
* Refine default thread selection for NUMA systems (#7322)
|
||||||
|
* runner.go: Better abstract vision model integration
|
||||||
|
* Soften windows clang requirement (#7428)
|
||||||
|
* Remove submodule and shift to Go server - 0.4.0 (#7157)
|
||||||
|
* Move windows app out of preview (#7347)
|
||||||
|
* windows: Support alt install paths, fit and finish (#6967)
|
||||||
|
* add more tests for getting the optimal tiled canvas (#7411)
|
||||||
|
* Switch windows to clang (#7407)
|
||||||
|
* tests: Add test for Unicode processing
|
||||||
|
* runner.go: Better handle return NULL values from llama.cpp
|
||||||
|
* add mllama image processing to the generate handler (#7384)
|
||||||
|
* Bump to latest Go 1.22 patch (#7379)
|
||||||
|
* Fix deepseek deseret regex (#7369)
|
||||||
|
* Better support for AMD multi-GPU on linux (#7212)
|
||||||
|
* Fix unicode output on windows with redirect to file (#7358)
|
||||||
|
* Fix incremental build file deps (#7361)
|
||||||
|
* Improve dependency gathering logic (#7345)
|
||||||
|
* fix #7247 - invalid image input (#7249)
|
||||||
|
* integration: harden embedding test (#7306)
|
||||||
|
* default to "FROM ." if a Modelfile isn't present (#7250)
|
||||||
|
* Fix rocm windows build and clean up dependency gathering (#7305)
|
||||||
|
* runner.go: Merge partial unicode characters before sending
|
||||||
|
* readme: add Ollama for Swift to the community integrations (#7295)
|
||||||
|
* server: allow vscode-webview origin (#7273)
|
||||||
|
* image processing for llama3.2 (#6963)
|
||||||
|
* llama: Decouple patching script from submodule (#7139)
|
||||||
|
* llama: add compiler tags for cpu features (#7137)
|
||||||
|
|
||||||
-------------------------------------------------------------------
|
-------------------------------------------------------------------
|
||||||
Wed Oct 30 01:47:37 UTC 2024 - Alessandro de Oliveira Faria <cabelo@opensuse.org>
|
Wed Oct 30 01:47:37 UTC 2024 - Alessandro de Oliveira Faria <cabelo@opensuse.org>
|
||||||
|
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
name: ollama
|
name: ollama
|
||||||
version: 0.3.14
|
version: 0.4.0
|
||||||
mtime: 1729191592
|
mtime: 1730848045
|
||||||
commit: f2890a4494f9fb3722ee7a4c506252362d1eab65
|
commit: 9d71bcc3e2a97c8e62d758450f43aa212346410e
|
||||||
|
@ -17,7 +17,7 @@
|
|||||||
|
|
||||||
|
|
||||||
Name: ollama
|
Name: ollama
|
||||||
Version: 0.3.14
|
Version: 0.4.0
|
||||||
Release: 0
|
Release: 0
|
||||||
Summary: Tool for running AI models on-premise
|
Summary: Tool for running AI models on-premise
|
||||||
License: MIT
|
License: MIT
|
||||||
@ -26,7 +26,6 @@ Source: %{name}-%{version}.tar
|
|||||||
Source1: vendor.tar.zstd
|
Source1: vendor.tar.zstd
|
||||||
Source2: ollama.service
|
Source2: ollama.service
|
||||||
Source3: %{name}-user.conf
|
Source3: %{name}-user.conf
|
||||||
Patch0: enable-lto.patch
|
|
||||||
BuildRequires: cmake >= 3.24
|
BuildRequires: cmake >= 3.24
|
||||||
BuildRequires: git
|
BuildRequires: git
|
||||||
BuildRequires: sysuser-tools
|
BuildRequires: sysuser-tools
|
||||||
|
BIN
vendor.tar.zstd
(Stored with Git LFS)
BIN
vendor.tar.zstd
(Stored with Git LFS)
Binary file not shown.
Loading…
x
Reference in New Issue
Block a user