Accepting request 1191409 from science:machinelearning

- Update to version 0.3.3:
  * The /api/embed endpoint now returns statistics: total_duration,
    load_duration, and prompt_eval_count
  * Added usage metrics to the /v1/embeddings OpenAI compatibility
    API
  * Fixed issue where /api/generate would respond with an empty 
    string if provided a context
  * Fixed issue where /api/generate would return an incorrect 
    value for context
  * /show modefile will now render MESSAGE commands correctly
- Update to version 0.3.2:
  * Fixed issue where ollama pull would not resume download 
    progress
  * Fixed issue where phi3 would report an error on older versions

OBS-URL: https://build.opensuse.org/request/show/1191409
OBS-URL: https://build.opensuse.org/package/show/openSUSE:Factory/ollama?expand=0&rev=17
This commit is contained in:
Dominique Leuenberger 2024-08-03 18:06:55 +00:00 committed by Git OBS Bridge
commit 20accf24c0
8 changed files with 30 additions and 11 deletions

View File

@ -3,7 +3,7 @@
<service name="obs_scm" mode="manual">
<param name="url">https://github.com/ollama/ollama.git</param>
<param name="scm">git</param>
<param name="revision">v0.3.1</param>
<param name="revision">v0.3.3</param>
<param name="versionformat">@PARENT_TAG@</param>
<param name="versionrewrite-pattern">v(.*)</param>
<param name="changesgenerate">enable</param>

View File

@ -1,4 +1,4 @@
<servicedata>
<service name="tar_scm">
<param name="url">https://github.com/ollama/ollama.git</param>
<param name="changesrevision">1a83581a8e1063418f5f1fec14638409d0681b68</param></service></servicedata>
<param name="changesrevision">ce1fb4447efc9958dcf279f7eb2ae6941bec1220</param></service></servicedata>

BIN
ollama-0.3.1.obscpio (Stored with Git LFS)

Binary file not shown.

BIN
ollama-0.3.3.obscpio (Stored with Git LFS) Normal file

Binary file not shown.

View File

@ -1,3 +1,22 @@
-------------------------------------------------------------------
Sat Aug 03 09:41:56 UTC 2024 - eyadlorenzo@gmail.com
- Update to version 0.3.3:
* The /api/embed endpoint now returns statistics: total_duration,
load_duration, and prompt_eval_count
* Added usage metrics to the /v1/embeddings OpenAI compatibility
API
* Fixed issue where /api/generate would respond with an empty
string if provided a context
* Fixed issue where /api/generate would return an incorrect
value for context
* /show modefile will now render MESSAGE commands correctly
- Update to version 0.3.2:
* Fixed issue where ollama pull would not resume download
progress
* Fixed issue where phi3 would report an error on older versions
-------------------------------------------------------------------
Tue Jul 30 07:08:37 UTC 2024 - Adrian Schröter <adrian@suse.de>

View File

@ -1,4 +1,4 @@
name: ollama
version: 0.3.1
mtime: 1722288341
commit: 1a83581a8e1063418f5f1fec14638409d0681b68
version: 0.3.3
mtime: 1722555107
commit: ce1fb4447efc9958dcf279f7eb2ae6941bec1220

View File

@ -17,7 +17,7 @@
Name: ollama
Version: 0.3.1
Version: 0.3.3
Release: 0
Summary: Tool for running AI models on-premise
License: MIT

BIN
vendor.tar.zstd (Stored with Git LFS)

Binary file not shown.