Accepting request 1189591 from science:machinelearning
OBS-URL: https://build.opensuse.org/request/show/1189591 OBS-URL: https://build.opensuse.org/package/show/openSUSE:Factory/ollama?expand=0&rev=14
This commit is contained in:
commit
973749ceec
2
_service
2
_service
@ -3,7 +3,7 @@
|
||||
<service name="obs_scm" mode="manual">
|
||||
<param name="url">https://github.com/ollama/ollama.git</param>
|
||||
<param name="scm">git</param>
|
||||
<param name="revision">v0.2.6</param>
|
||||
<param name="revision">v0.2.8</param>
|
||||
<param name="versionformat">@PARENT_TAG@</param>
|
||||
<param name="versionrewrite-pattern">v(.*)</param>
|
||||
<param name="changesgenerate">enable</param>
|
||||
|
@ -1,4 +1,4 @@
|
||||
<servicedata>
|
||||
<service name="tar_scm">
|
||||
<param name="url">https://github.com/ollama/ollama.git</param>
|
||||
<param name="changesrevision">b2554455572b28c0e18423d6fe6896cf7137dbd6</param></service></servicedata>
|
||||
<param name="changesrevision">c0648233f2236f82f6830d2aaed552ae0f72379b</param></service></servicedata>
|
@ -1,3 +0,0 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:391fad97bacee37e8fab00273fd5d5a0a20912fd47c51907131ee1f274c7d2bf
|
||||
size 161902606
|
3
ollama-0.2.8.obscpio
Normal file
3
ollama-0.2.8.obscpio
Normal file
@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:f1dfa7d3fc6d8dc35af4bd9a458a9f22ab613d07c1e5e48db2b2803ff7f77214
|
||||
size 151425038
|
@ -1,3 +1,29 @@
|
||||
-------------------------------------------------------------------
|
||||
Wed Jul 24 14:28:08 UTC 2024 - adrian@suse.de
|
||||
|
||||
- Update to version 0.2.8:
|
||||
* api embed docs (#5282)
|
||||
* convert: capture `head_dim` for mistral (#5818)
|
||||
* Update llama.cpp submodule commit to `d94c6e0c` (#5805)
|
||||
* server: collect nested tool call objects when parsing (#5824)
|
||||
* Remove no longer supported max vram var
|
||||
* Refine error reporting for subprocess crash
|
||||
* Remove out of space test temporarily (#5825)
|
||||
* llm: consider `head_dim` in llama arch (#5817)
|
||||
* Adjust windows ROCm discovery
|
||||
* add patch for tekken (#5807)
|
||||
* preserve last assistant message (#5802)
|
||||
* Fix generate test flakyness (#5804)
|
||||
* server: validate template (#5734)
|
||||
* OpenAI: Function Based Testing (#5752)
|
||||
* adjust openai chat msg processing (#5729)
|
||||
* fix parsing tool calls
|
||||
* server: check for empty tools array too (#5779)
|
||||
* always provide content even if empty (#5778)
|
||||
* server: only parse tool calls if tools are provided (#5771)
|
||||
* Fix context exhaustion integration test for small gpus
|
||||
* Refine scheduler unit tests for reliability
|
||||
|
||||
-------------------------------------------------------------------
|
||||
Thu Jul 18 13:09:10 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
|
||||
|
||||
|
@ -1,4 +1,4 @@
|
||||
name: ollama
|
||||
version: 0.2.6
|
||||
mtime: 1721255711
|
||||
commit: b2554455572b28c0e18423d6fe6896cf7137dbd6
|
||||
version: 0.2.8
|
||||
mtime: 1721680628
|
||||
commit: c0648233f2236f82f6830d2aaed552ae0f72379b
|
||||
|
@ -17,7 +17,7 @@
|
||||
|
||||
|
||||
Name: ollama
|
||||
Version: 0.2.6
|
||||
Version: 0.2.8
|
||||
Release: 0
|
||||
Summary: Tool for running AI models on-premise
|
||||
License: MIT
|
||||
|
@ -1,3 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:b9dabb1b28321cce2672e5b37eb792e904715539dad5ecabc0eee92d6b0b10e1
|
||||
size 5355343
|
||||
oid sha256:e4b96cf9ccbb2b5ac6750dd67375daafbdcda9e2db58d0673a2566478b776878
|
||||
size 5355002
|
||||
|
Loading…
Reference in New Issue
Block a user