- Update to version 0.4.0-rc8:

OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=62
This commit is contained in:
Eyad Issa 2024-11-06 12:33:58 +00:00 committed by Git OBS Bridge
parent d5f830b976
commit 7cc332278e
6 changed files with 99 additions and 81 deletions

View File

@ -3,9 +3,9 @@
<service name="obs_scm" mode="manual"> <service name="obs_scm" mode="manual">
<param name="url">https://github.com/ollama/ollama.git</param> <param name="url">https://github.com/ollama/ollama.git</param>
<param name="scm">git</param> <param name="scm">git</param>
<param name="revision">v0.4.0-rc6</param> <param name="revision">v0.4.0-rc8</param>
<param name="versionformat">@PARENT_TAG@</param> <param name="versionformat">@PARENT_TAG@</param>
<param name="versionrewrite-pattern">v(.*)-rc6</param> <param name="versionrewrite-pattern">v(.*)-rc8</param>
<param name="changesgenerate">enable</param> <param name="changesgenerate">enable</param>
<param name="submodules">enable</param> <param name="submodules">enable</param>
<param name="exclude">macapp</param> <param name="exclude">macapp</param>

View File

@ -1,4 +1,4 @@
<servicedata> <servicedata>
<service name="tar_scm"> <service name="tar_scm">
<param name="url">https://github.com/ollama/ollama.git</param> <param name="url">https://github.com/ollama/ollama.git</param>
<param name="changesrevision">16f4eabe2d409b2b8a6e50fa08c8ce3a2a3b18d1</param></service></servicedata> <param name="changesrevision">046054fa3bba6d6511bcf46ca53f3ee8bc972df6</param></service></servicedata>

View File

@ -1,3 +1,3 @@
version https://git-lfs.github.com/spec/v1 version https://git-lfs.github.com/spec/v1
oid sha256:a170a1b1dad3a0414739389095d73562aa9e2038357f4fc1a4c5db344f836547 oid sha256:bcbcbb9aa1cdde96a51a5bbf25f1b6a3cb97d71ed7589a17e68f9a30287bd450
size 16461325 size 16564237

View File

@ -1,5 +1,23 @@
------------------------------------------------------------------- -------------------------------------------------------------------
Fri Nov 01 02:18:50 UTC 2024 - eyadlorenzo@gmail.com Wed Nov 06 12:31:53 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
- Update to version 0.4.0-rc8:
* CI: Switch to v13 macos runner (#7498)
* CI: matrix strategy fix (#7496)
* Sign windows arm64 official binaries (#7493)
* readme: add TextCraft to community integrations (#7377)
* nvidia libs have inconsistent ordering (#7473)
* CI: omit unused tools for faster release builds (#7432)
* llama: Improve error handling
* runner.go: Only allocate 1 element embedding batches for mllama
* refactor kv estimation
* mllama cross attention
* Add basic mllama integration tests (#7455)
* runner.go: Don't set cross attention before sending embeddings
* Give unicode test more time to run (#7437)
-------------------------------------------------------------------
Fri Nov 01 02:18:50 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
- Remove enable-lto.patch - Remove enable-lto.patch
@ -37,12 +55,12 @@ Wed Oct 30 01:47:37 UTC 2024 - Alessandro de Oliveira Faria <cabelo@opensuse.org
- Update to version 0.3.14: - Update to version 0.3.14:
* New Models * New Models
+ Granite 3 MoE: The IBM Granite 1B and 3B models are the + Granite 3 MoE: The IBM Granite 1B and 3B models are the
first mixture of experts (MoE) Granite models from IBM first mixture of experts (MoE) Granite models from IBM
designed for low latency usage. designed for low latency usage.
+ Granite 3 Dense: The IBM Granite 2B and 8B models are + Granite 3 Dense: The IBM Granite 2B and 8B models are
designed to support tool-based use cases and support for designed to support tool-based use cases and support for
retrieval augmented generation (RAG), streamlining code retrieval augmented generation (RAG), streamlining code
generation, translation and bug fixing. generation, translation and bug fixing.
------------------------------------------------------------------- -------------------------------------------------------------------
@ -53,21 +71,21 @@ Sat Oct 12 20:55:18 UTC 2024 - eyadlorenzo@gmail.com
~ Llama Guard 3: a series of models by Meta, fine-tuned for ~ Llama Guard 3: a series of models by Meta, fine-tuned for
content safety classification of LLM inputs and responses. content safety classification of LLM inputs and responses.
~ ShieldGemma: ShieldGemma is set of instruction tuned models ~ ShieldGemma: ShieldGemma is set of instruction tuned models
from Google DeepMind for evaluating the safety of text from Google DeepMind for evaluating the safety of text
prompt input and text output responses against a set of prompt input and text output responses against a set of
defined safety policies. defined safety policies.
* Fixed issue where ollama pull would leave connections when * Fixed issue where ollama pull would leave connections when
encountering an error encountering an error
* ollama rm will now stop a model if it is running prior to * ollama rm will now stop a model if it is running prior to
deleting it deleting it
------------------------------------------------------------------- -------------------------------------------------------------------
Sat Sep 28 03:53:10 UTC 2024 - Alessandro de Oliveira Faria <cabelo@opensuse.org> Sat Sep 28 03:53:10 UTC 2024 - Alessandro de Oliveira Faria <cabelo@opensuse.org>
- Update to version 0.3.12: - Update to version 0.3.12:
* Llama 3.2: Meta's Llama 3.2 goes small with 1B and 3B * Llama 3.2: Meta's Llama 3.2 goes small with 1B and 3B
models. models.
* Qwen 2.5 Coder: The latest series of Code-Specific Qwen * Qwen 2.5 Coder: The latest series of Code-Specific Qwen
models, with significant improvements in code generation, models, with significant improvements in code generation,
code reasoning, and code fixing. code reasoning, and code fixing.
* Ollama now supports ARM Windows machines * Ollama now supports ARM Windows machines
* Fixed rare issue where Ollama would report a missing .dll * Fixed rare issue where Ollama would report a missing .dll
@ -241,23 +259,23 @@ Sun Aug 11 02:40:06 UTC 2024 - Alessandro de Oliveira Faria <cabelo@opensuse.org
- Update to version 0.3.4: - Update to version 0.3.4:
* New embedding models * New embedding models
- BGE-M3: a large embedding model from BAAI distinguished for - BGE-M3: a large embedding model from BAAI distinguished for
its versatility in Multi-Functionality, Multi-Linguality, and its versatility in Multi-Functionality, Multi-Linguality, and
Multi-Granularity. Multi-Granularity.
- BGE-Large: a large embedding model trained in english. - BGE-Large: a large embedding model trained in english.
- Paraphrase-Multilingual: A multilingual embedding model - Paraphrase-Multilingual: A multilingual embedding model
trained on parallel data for 50+ languages. trained on parallel data for 50+ languages.
* New embedding API with batch support * New embedding API with batch support
- Ollama now supports a new API endpoint /api/embed for - Ollama now supports a new API endpoint /api/embed for
embedding generation: embedding generation:
* This API endpoint supports new features: * This API endpoint supports new features:
- Batches: generate embeddings for several documents in - Batches: generate embeddings for several documents in
one request one request
- Normalized embeddings: embeddings are now normalized, - Normalized embeddings: embeddings are now normalized,
improving similarity results improving similarity results
- Truncation: a new truncate parameter that will error if - Truncation: a new truncate parameter that will error if
set to false set to false
- Metrics: responses include load_duration, total_duration and - Metrics: responses include load_duration, total_duration and
prompt_eval_count metrics prompt_eval_count metrics
------------------------------------------------------------------- -------------------------------------------------------------------
@ -268,17 +286,17 @@ Sat Aug 03 09:41:56 UTC 2024 - eyadlorenzo@gmail.com
load_duration, and prompt_eval_count load_duration, and prompt_eval_count
* Added usage metrics to the /v1/embeddings OpenAI compatibility * Added usage metrics to the /v1/embeddings OpenAI compatibility
API API
* Fixed issue where /api/generate would respond with an empty * Fixed issue where /api/generate would respond with an empty
string if provided a context string if provided a context
* Fixed issue where /api/generate would return an incorrect * Fixed issue where /api/generate would return an incorrect
value for context value for context
* /show modefile will now render MESSAGE commands correctly * /show modefile will now render MESSAGE commands correctly
- Update to version 0.3.2: - Update to version 0.3.2:
* Fixed issue where ollama pull would not resume download * Fixed issue where ollama pull would not resume download
progress progress
* Fixed issue where phi3 would report an error on older versions * Fixed issue where phi3 would report an error on older versions
------------------------------------------------------------------- -------------------------------------------------------------------
Tue Jul 30 07:08:37 UTC 2024 - Adrian Schröter <adrian@suse.de> Tue Jul 30 07:08:37 UTC 2024 - Adrian Schröter <adrian@suse.de>
@ -343,16 +361,16 @@ Wed Jul 24 14:28:08 UTC 2024 - adrian@suse.de
------------------------------------------------------------------- -------------------------------------------------------------------
Thu Jul 18 13:09:10 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com> Thu Jul 18 13:09:10 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
- Fixed issue with shared libraries - Fixed issue with shared libraries
------------------------------------------------------------------- -------------------------------------------------------------------
Thu Jul 18 12:27:54 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com> Thu Jul 18 12:27:54 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
- Added %check section - Added %check section
- Use -v when building - Use -v when building
- Update to version 0.2.6: - Update to version 0.2.6:
* New models: MathΣtral is a 7B model designed for math * New models: MathΣtral is a 7B model designed for math
reasoning and scientific discovery by Mistral AI. reasoning and scientific discovery by Mistral AI.
* Fixed issue where uppercase roles such as USER would no longer * Fixed issue where uppercase roles such as USER would no longer
work in the chat endpoints work in the chat endpoints
@ -366,62 +384,62 @@ Sun Jul 14 17:48:36 UTC 2024 - eyadlorenzo@gmail.com
* Fixed issue where a model's SYSTEM message not be applied * Fixed issue where a model's SYSTEM message not be applied
- Update to version 0.2.4: - Update to version 0.2.4:
* Fixed issue where context, load_duration and total_duration * Fixed issue where context, load_duration and total_duration
fields would not be set in the /api/generate endpoint. fields would not be set in the /api/generate endpoint.
* Ollama will no longer error if loading models larger than * Ollama will no longer error if loading models larger than
system memory if disk space is available system memory if disk space is available
- Update to version 0.2.3: - Update to version 0.2.3:
* Fix issue where system prompt would not be applied * Fix issue where system prompt would not be applied
- Update to version 0.2.2: - Update to version 0.2.2:
* Fixed errors that occurred when using Ollama with Nvidia V100 * Fixed errors that occurred when using Ollama with Nvidia V100
GPUs GPUs
* glm4 models will no longer fail to load from out of memory * glm4 models will no longer fail to load from out of memory
errors errors
* Fixed error that would occur when running deepseek-v2 and * Fixed error that would occur when running deepseek-v2 and
deepseek-coder-v2 models deepseek-coder-v2 models
* Fixed a series of out of memory issues when using Nvidia * Fixed a series of out of memory issues when using Nvidia
GPUs GPUs
* Fixed a series of errors that would occur when using multiple * Fixed a series of errors that would occur when using multiple
Radeon GPUs Radeon GPUs
- Update to version 0.2.1: - Update to version 0.2.1:
* Fixed issue where setting OLLAMA_NUM_PARALLEL would cause * Fixed issue where setting OLLAMA_NUM_PARALLEL would cause
models to be reloaded after each request models to be reloaded after each request
- Update to version 0.2.0: - Update to version 0.2.0:
* Ollama 0.2.0 is now available with concurrency support. * Ollama 0.2.0 is now available with concurrency support.
This unlocks 2 specific features: This unlocks 2 specific features:
~ Ollama can now serve multiple requests at the same time ~ Ollama can now serve multiple requests at the same time
~ Ollama now supports loading different models at the same time ~ Ollama now supports loading different models at the same time
* New models: GLM-4: A strong multi-lingual general language * New models: GLM-4: A strong multi-lingual general language
model with competitive performance to Llama 3. model with competitive performance to Llama 3.
* New models: CodeGeeX4: A versatile model for AI software * New models: CodeGeeX4: A versatile model for AI software
development scenarios, including code completion. development scenarios, including code completion.
* New models: Gemma 2: Improved output quality and base text * New models: Gemma 2: Improved output quality and base text
generation models now available generation models now available
* Ollama will now show a better error if a model architecture * Ollama will now show a better error if a model architecture
isn't supported isn't supported
* Improved handling of quotes and spaces in Modelfile FROM lines * Improved handling of quotes and spaces in Modelfile FROM lines
* Ollama will now return an error if the system does not have * Ollama will now return an error if the system does not have
enough memory to run a model on Linux enough memory to run a model on Linux
------------------------------------------------------------------- -------------------------------------------------------------------
Sun Jul 07 19:18:11 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com> Sun Jul 07 19:18:11 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
- Update to version 0.1.48: - Update to version 0.1.48:
* Fixed issue where Gemma 2 would continuously output when * Fixed issue where Gemma 2 would continuously output when
reaching context limits reaching context limits
* Fixed out of memory and core dump errors when running Gemma 2 * Fixed out of memory and core dump errors when running Gemma 2
* /show info will now show additional model information in * /show info will now show additional model information in
ollama run ollama run
* Fixed issue where ollama show would result in an error on * Fixed issue where ollama show would result in an error on
certain vision models certain vision models
- Update to version 0.1.48: - Update to version 0.1.48:
* Added support for Google Gemma 2 models (9B and 27B) * Added support for Google Gemma 2 models (9B and 27B)
* Fixed issues with ollama create when importing from Safetensors * Fixed issues with ollama create when importing from Safetensors
------------------------------------------------------------------- -------------------------------------------------------------------
Mon Jun 24 10:11:17 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com> Mon Jun 24 10:11:17 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
@ -456,44 +474,44 @@ Sat Jun 22 10:08:00 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
------------------------------------------------------------------- -------------------------------------------------------------------
Tue Jun 18 12:12:41 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com> Tue Jun 18 12:12:41 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
- Added documentation files to .spec - Added documentation files to .spec
- Update to version 0.1.44: - Update to version 0.1.44:
* Fixed issue where unicode characters such as emojis would not * Fixed issue where unicode characters such as emojis would not
be loaded correctly when running ollama create be loaded correctly when running ollama create
* Fixed certain cases where Nvidia GPUs would not be detected and * Fixed certain cases where Nvidia GPUs would not be detected and
reported as compute capability 1.0 devices reported as compute capability 1.0 devices
- Update to version 0.1.43: - Update to version 0.1.43:
* New import.md guide for converting and importing models to * New import.md guide for converting and importing models to
Ollama Ollama
* Fixed issue where embedding vectors resulting from * Fixed issue where embedding vectors resulting from
/api/embeddings would not be accurate /api/embeddings would not be accurate
* JSON mode responses will no longer include invalid escape * JSON mode responses will no longer include invalid escape
characters characters
* Removing a model will no longer show incorrect File not found * Removing a model will no longer show incorrect File not found
errors errors
* Fixed issue where running ollama create would result in an * Fixed issue where running ollama create would result in an
error on Windows with certain file formatting error on Windows with certain file formatting
- Update to version 0.1.42: - Update to version 0.1.42:
* New models: Qwen 2: a new series of large language models * New models: Qwen 2: a new series of large language models
from Alibaba group from Alibaba group
* Qwen 2: a new series of large language models from Alibaba * Qwen 2: a new series of large language models from Alibaba
group group
* ollama pull is now faster if it detects a model is already * ollama pull is now faster if it detects a model is already
downloaded downloaded
* ollama create will now automatically detect prompt templates * ollama create will now automatically detect prompt templates
for popular model architectures such as Llama, Gemma, Phi and for popular model architectures such as Llama, Gemma, Phi and
more. more.
* Ollama can now be accessed from local apps built with Electron * Ollama can now be accessed from local apps built with Electron
and Tauri, as well as in developing apps in local html files and Tauri, as well as in developing apps in local html files
* Update welcome prompt in Windows to llama3 * Update welcome prompt in Windows to llama3
* Fixed issues where /api/ps and /api/tags would show invalid * Fixed issues where /api/ps and /api/tags would show invalid
timestamps in responses timestamps in responses
- Update to version 0.1.41: - Update to version 0.1.41:
* Fixed issue on Windows 10 and 11 with Intel CPUs with * Fixed issue on Windows 10 and 11 with Intel CPUs with
integrated GPUs where Ollama would encounter an error integrated GPUs where Ollama would encounter an error
------------------------------------------------------------------- -------------------------------------------------------------------
@ -503,12 +521,12 @@ Sat Jun 01 21:12:20 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
* New model: Codestral: Codestral is Mistral AIs first-ever code * New model: Codestral: Codestral is Mistral AIs first-ever code
model designed for code generation tasks. model designed for code generation tasks.
* New model: IBM Granite Code: now in 3B and 8B parameter sizes. * New model: IBM Granite Code: now in 3B and 8B parameter sizes.
* New model: Deepseek V2: A Strong, Economical, and Efficient * New model: Deepseek V2: A Strong, Economical, and Efficient
Mixture-of-Experts Language Model Mixture-of-Experts Language Model
* Fixed out of memory and incorrect token issues when running * Fixed out of memory and incorrect token issues when running
Codestral on 16GB Macs Codestral on 16GB Macs
* Fixed issue where full-width characters (e.g. Japanese, * Fixed issue where full-width characters (e.g. Japanese,
Chinese, Russian) were deleted at end of the line when using Chinese, Russian) were deleted at end of the line when using
ollama run ollama run
------------------------------------------------------------------- -------------------------------------------------------------------
@ -517,9 +535,9 @@ Wed May 29 11:38:26 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
- Update to version 0.1.39: - Update to version 0.1.39:
* New model: Cohere Aya 23: A new state-of-the-art, multilingual * New model: Cohere Aya 23: A new state-of-the-art, multilingual
LLM covering 23 different languages. LLM covering 23 different languages.
* New model: Mistral 7B 0.3: A new version of Mistral 7B with * New model: Mistral 7B 0.3: A new version of Mistral 7B with
initial support for function calling. initial support for function calling.
* New model: Phi-3 Medium: a 14B parameters, lightweight, * New model: Phi-3 Medium: a 14B parameters, lightweight,
state-of-the-art open model by Microsoft. state-of-the-art open model by Microsoft.
* New model: Phi-3 Mini 128K and Phi-3 Medium 128K: versions of * New model: Phi-3 Mini 128K and Phi-3 Medium 128K: versions of
the Phi-3 models that support a context window size of 128K the Phi-3 models that support a context window size of 128K
@ -527,7 +545,7 @@ Wed May 29 11:38:26 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
IBM for Code Intelligence IBM for Code Intelligence
* It is now possible to import and quantize Llama 3 and its * It is now possible to import and quantize Llama 3 and its
finetunes from Safetensors format to Ollama. finetunes from Safetensors format to Ollama.
* Full changelog at * Full changelog at
https://github.com/ollama/ollama/releases/tag/v0.1.39 https://github.com/ollama/ollama/releases/tag/v0.1.39
------------------------------------------------------------------- -------------------------------------------------------------------
@ -541,7 +559,7 @@ Thu May 16 19:55:51 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
- Update to version 0.1.38: - Update to version 0.1.38:
* New model: Falcon 2: A new 11B parameters causal decoder-only * New model: Falcon 2: A new 11B parameters causal decoder-only
model built by TII and trained over 5T tokens. model built by TII and trained over 5T tokens.
* New model: Yi 1.5: A new high-performing version of Yi, now * New model: Yi 1.5: A new high-performing version of Yi, now
licensed as Apache 2.0. Available in 6B, 9B and 34B sizes. licensed as Apache 2.0. Available in 6B, 9B and 34B sizes.
* Added ollama ps command * Added ollama ps command
* Added /clear command * Added /clear command
@ -566,7 +584,7 @@ Sun May 12 19:05:53 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
Sun May 12 15:20:28 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com> Sun May 12 15:20:28 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
- Use obs_scm service instead of the deprecated tar_scm - Use obs_scm service instead of the deprecated tar_scm
- Use zstd for vendor tarball compression - Use zstd for vendor tarball compression
------------------------------------------------------------------- -------------------------------------------------------------------
Sun May 12 01:39:26 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com> Sun May 12 01:39:26 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
@ -604,11 +622,11 @@ Sun May 12 01:39:26 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
* New model: CodeGemma 1.1 * New model: CodeGemma 1.1
* New model: StableLM2 12B * New model: StableLM2 12B
* New model: Moondream 2 * New model: Moondream 2
* Fixed issues with LLaVa models where they would respond * Fixed issues with LLaVa models where they would respond
incorrectly after the first request incorrectly after the first request
* Fixed out of memory errors when running large models such as * Fixed out of memory errors when running large models such as
Llama 3 70B Llama 3 70B
* Fixed various issues with Nvidia GPU discovery on Linux and * Fixed various issues with Nvidia GPU discovery on Linux and
Windows Windows
* Fixed a series of Modelfile errors when running ollama create * Fixed a series of Modelfile errors when running ollama create
* Fixed no slots available error that occurred when cancelling a * Fixed no slots available error that occurred when cancelling a
@ -626,13 +644,13 @@ Sun May 12 01:39:26 UTC 2024 - Eyad Issa <eyadlorenzo@gmail.com>
* New model: Llama 3 Gradient 1048K * New model: Llama 3 Gradient 1048K
* New model: Dolphin Llama 3 * New model: Dolphin Llama 3
* New model: Qwen 110B * New model: Qwen 110B
* Fixed issues where the model would not terminate, causing the * Fixed issues where the model would not terminate, causing the
API to hang. API to hang.
* Fixed a series of out of memory errors on Apple Silicon Macs * Fixed a series of out of memory errors on Apple Silicon Macs
* Fixed out of memory errors when running Mixtral architecture * Fixed out of memory errors when running Mixtral architecture
models models
* Aded experimental concurrency features: * Aded experimental concurrency features:
~ OLLAMA_NUM_PARALLEL: Handle multiple requests simultaneously ~ OLLAMA_NUM_PARALLEL: Handle multiple requests simultaneously
for a single model for a single model
~ OLLAMA_MAX_LOADED_MODELS: Load multiple models simultaneously ~ OLLAMA_MAX_LOADED_MODELS: Load multiple models simultaneously

View File

@ -1,4 +1,4 @@
name: ollama name: ollama
version: 0.4.0 version: 0.4.0
mtime: 1730325945 mtime: 1730754127
commit: 16f4eabe2d409b2b8a6e50fa08c8ce3a2a3b18d1 commit: 046054fa3bba6d6511bcf46ca53f3ee8bc972df6

View File

@ -1,3 +1,3 @@
version https://git-lfs.github.com/spec/v1 version https://git-lfs.github.com/spec/v1
oid sha256:f66046626b5f525abb5373c347640f617eb7bd368d2a7f7f5b298a1db6e2b7b7 oid sha256:a465edc1e925c1c066e9a5923c6dd3b0534f3cef0ee0d32646b84d1347a126e6
size 5367921 size 5367830