From 8ef2b26afe5567048056b677cf93594a4fda8694914a1ecf54738fe69392c3f3 Mon Sep 17 00:00:00 2001 From: Loren Burkholder Date: Wed, 17 Apr 2024 00:53:52 +0000 Subject: [PATCH] Accepting request 1168020 from home:bmwiedemann:branches:science:machinelearning Update to version 0.1.31: * Backport MacOS SDK fix from main * Apply 01-cache.diff * fix: workflows * stub stub * mangle arch * only generate on changes to llm subdirectory * only generate cuda/rocm when changes to llm detected * Detect arrow keys on windows (#3363) * add license in file header for vendored llama.cpp code (#3351) * remove need for `$VSINSTALLDIR` since build will fail if `ninja` cannot be found (#3350) * change `github.com/jmorganca/ollama` to `github.com/ollama/ollama` (#3347) * malformed markdown link (#3358) * Switch runner for final release job * Use Rocky Linux Vault to get GCC 10.2 installed * Revert "Switch arm cuda base image to centos 7" * Switch arm cuda base image to centos 7 * Bump llama.cpp to b2527 * Fix ROCm link in `development.md` * adds ooo to community integrations (#1623) * Add cliobot to ollama supported list (#1873) * Add Dify.AI to community integrations (#1944) * enh: add ollero.nvim to community applications (#1905) * Add typechat-cli to Terminal apps (#2428) * add new Web & Desktop link in readme for alpaca webui (#2881) * Add LibreChat to Web & Desktop Apps (#2918) * Add Community Integration: OllamaGUI (#2927) * Add Community Integration: OpenAOE (#2946) * Add Saddle (#3178) * tlm added to README.md terminal section. (#3274) ... OBS-URL: https://build.opensuse.org/request/show/1168020 OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=7 --- _service | 2 +- _servicedata | 4 ++ ollama-0.1.27.tar.gz | 3 -- ollama-0.1.28.tar.gz | 3 -- ollama-0.1.31.tar.gz | 3 ++ ollama.changes | 126 +++++++++++++++++++++++++++++++++++++++++++ ollama.spec | 4 +- vendor.tar.xz | 4 +- 8 files changed, 138 insertions(+), 11 deletions(-) create mode 100644 _servicedata delete mode 100644 ollama-0.1.27.tar.gz delete mode 100644 ollama-0.1.28.tar.gz create mode 100644 ollama-0.1.31.tar.gz diff --git a/_service b/_service index 35c0b4e..3b6781c 100644 --- a/_service +++ b/_service @@ -4,7 +4,7 @@ https://github.com/ollama/ollama.git git - v0.1.28 + v0.1.31 @PARENT_TAG@ v(.*) enable diff --git a/_servicedata b/_servicedata new file mode 100644 index 0000000..a2bd40b --- /dev/null +++ b/_servicedata @@ -0,0 +1,4 @@ + + + https://github.com/ollama/ollama.git + dc011d16b9ff160c0be3829fc39a43054f0315d0 \ No newline at end of file diff --git a/ollama-0.1.27.tar.gz b/ollama-0.1.27.tar.gz deleted file mode 100644 index 7d96a3e..0000000 --- a/ollama-0.1.27.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c9b7005256616e8161cc2800cec78b0f43ab4c05ae78b18a7337756dacf5b97a -size 63206855 diff --git a/ollama-0.1.28.tar.gz b/ollama-0.1.28.tar.gz deleted file mode 100644 index 28355a7..0000000 --- a/ollama-0.1.28.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:30225f7d1a96b8a573e82810584950ed2f9c95dcd2157d794c278ca44a43861b -size 75624882 diff --git a/ollama-0.1.31.tar.gz b/ollama-0.1.31.tar.gz new file mode 100644 index 0000000..9f437dd --- /dev/null +++ b/ollama-0.1.31.tar.gz @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e55ac3bbb965f56c6920b793f254e814f4bf5fea77c81e8d8d867e850b15394 +size 80992137 diff --git a/ollama.changes b/ollama.changes index 0d124e7..9a54f83 100644 --- a/ollama.changes +++ b/ollama.changes @@ -1,3 +1,129 @@ +------------------------------------------------------------------- +Tue Apr 16 10:52:25 UTC 2024 - bwiedemann@suse.com + +- Update to version 0.1.31: + * Backport MacOS SDK fix from main + * Apply 01-cache.diff + * fix: workflows + * stub stub + * mangle arch + * only generate on changes to llm subdirectory + * only generate cuda/rocm when changes to llm detected + * Detect arrow keys on windows (#3363) + * add license in file header for vendored llama.cpp code (#3351) + * remove need for `$VSINSTALLDIR` since build will fail if `ninja` cannot be found (#3350) + * change `github.com/jmorganca/ollama` to `github.com/ollama/ollama` (#3347) + * malformed markdown link (#3358) + * Switch runner for final release job + * Use Rocky Linux Vault to get GCC 10.2 installed + * Revert "Switch arm cuda base image to centos 7" + * Switch arm cuda base image to centos 7 + * Bump llama.cpp to b2527 + * Fix ROCm link in `development.md` + * adds ooo to community integrations (#1623) + * Add cliobot to ollama supported list (#1873) + * Add Dify.AI to community integrations (#1944) + * enh: add ollero.nvim to community applications (#1905) + * Add typechat-cli to Terminal apps (#2428) + * add new Web & Desktop link in readme for alpaca webui (#2881) + * Add LibreChat to Web & Desktop Apps (#2918) + * Add Community Integration: OllamaGUI (#2927) + * Add Community Integration: OpenAOE (#2946) + * Add Saddle (#3178) + * tlm added to README.md terminal section. (#3274) + * Update README.md (#3288) + * Update README.md (#3338) + * Integration tests conditionally pull + * add support for libcudart.so for CUDA devices (adds Jetson support) + * llm: prevent race appending to slice (#3320) + * Bump llama.cpp to b2510 + * Add Testcontainers into Libraries section (#3291) + * Revamp go based integration tests + * rename `.gitattributes` + * Bump llama.cpp to b2474 + * Add docs for GPU selection and nvidia uvm workaround + * doc: faq gpu compatibility (#3142) + * Update faq.md + * Better tmpdir cleanup + * Update faq.md + * update `faq.md` + * dyn global + * llama: remove server static assets (#3174) + * add `llm/ext_server` directory to `linguist-vendored` (#3173) + * Add Radeon gfx940-942 GPU support + * Wire up more complete CI for releases + * llm,readline: use errors.Is instead of simple == check (#3161) + * server: replace blob prefix separator from ':' to '-' (#3146) + * Add ROCm support to linux install script (#2966) + * .github: fix model and feature request yml (#3155) + * .github: add issue templates (#3143) + * fix: clip memory leak + * Update README.md + * add `OLLAMA_KEEP_ALIVE` to environment variable docs for `ollama serve` (#3127) + * Default Keep Alive environment variable (#3094) + * Use stdin for term discovery on windows + * Update ollama.iss + * restore locale patch (#3091) + * token repeat limit for prediction requests (#3080) + * Fix iGPU detection for linux + * add more docs on for the modelfile message command (#3087) + * warn when json format is expected but not mentioned in prompt (#3081) + * Adapt our build for imported server.cpp + * Import server.cpp as of b2356 + * refactor readseeker + * Add docs explaining GPU selection env vars + * chore: fix typo (#3073) + * fix gpu_info_cuda.c compile warning (#3077) + * use `-trimpath` when building releases (#3069) + * relay load model errors to the client (#3065) + * Update troubleshooting.md + * update llama.cpp submodule to `ceca1ae` (#3064) + * convert: fix shape + * Avoid rocm runner and dependency clash + * fix `03-locale.diff` + * Harden for deps file being empty (or short) + * Add ollama executable peer dir for rocm + * patch: use default locale in wpm tokenizer (#3034) + * only copy deps for `amd64` in `build_linux.sh` + * Rename ROCm deps file to avoid confusion (#3025) + * add `macapp` to `.dockerignore` + * add `bundle_metal` and `cleanup_metal` funtions to `gen_darwin.sh` + * tidy cleanup logs + * update llama.cpp submodule to `77d1ac7` (#3030) + * disable gpu for certain model architectures and fix divide-by-zero on memory estimation + * Doc how to set up ROCm builds on windows + * Finish unwinding idempotent payload logic + * update llama.cpp submodule to `c2101a2` (#3020) + * separate out `isLocalIP` + * simplify host checks + * add additional allowed hosts + * Update docs `README.md` and table of contents + * add allowed host middleware and remove `workDir` middleware (#3018) + * decode ggla + * convert: fix default shape + * fix: allow importing a model from name reference (#3005) + * update llama.cpp submodule to `6cdabe6` (#2999) + * Update api.md + * Revert "adjust download and upload concurrency based on available bandwidth" (#2995) + * cmd: tighten up env var usage sections (#2962) + * default terminal width, height + * Refined ROCm troubleshooting docs + * Revamp ROCm support + * update go to 1.22 in other places (#2975) + * docs: Add LLM-X to Web Integration section (#2759) + * fix some typos (#2973) + * Convert Safetensors to an Ollama model (#2824) + * Allow setting max vram for workarounds + * cmd: document environment variables for serve command + * Add Odin Runes, a Feature-Rich Java UI for Ollama, to README (#2440) + * Update api.md + * Add NotesOllama to Community Integrations (#2909) + * Added community link for Ollama Copilot (#2582) + * use LimitGroup for uploads + * adjust group limit based on download speed + * add new LimitGroup for dynamic concurrency + * refactor download run + ------------------------------------------------------------------- Wed Mar 06 23:51:28 UTC 2024 - computersemiexpert@outlook.com diff --git a/ollama.spec b/ollama.spec index 0860883..0fbbc20 100644 --- a/ollama.spec +++ b/ollama.spec @@ -17,7 +17,7 @@ Name: ollama -Version: 0.1.28 +Version: 0.1.31 Release: 0 Summary: Tool for running AI models on-premise License: MIT @@ -31,7 +31,7 @@ BuildRequires: cmake >= 3.24 BuildRequires: gcc-c++ >= 11.4.0 BuildRequires: git BuildRequires: sysuser-tools -BuildRequires: golang(API) >= 1.21 +BuildRequires: golang(API) >= 1.22 %{sysusers_requires} diff --git a/vendor.tar.xz b/vendor.tar.xz index 363d08a..e3679a7 100644 --- a/vendor.tar.xz +++ b/vendor.tar.xz @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:548f8d5870f6b0b2881f5f68ef3f45c8b77bba282e10dc1aecafe14396213327 -size 2993296 +oid sha256:924b704ac695115d54330397b6d02d737de88d5f67fa760fd1065521357193fd +size 3656224