to go build
- Update to version 0.4.1:
* runner.go: Check for zero length images
* docs: update langchainpy.md with proper model name (#7527)
* Set macos min version for all architectures (#7579)
* win: remove preview title from installer (#7529)
* Workaround buggy P2P ROCm copy on windows (#7466)
* Debug logging for nvcuda init (#7532)
* Align rocm compiler flags (#7467)
* Be explicit for gpu library link dir (#7560)
* docs: OLLAMA_NEW_RUNNERS no longer exists
* runner.go: Remove unused arguments
* sched: Lift parallel restriction for multimodal models except mllama
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=65
* Refine default thread selection for NUMA systems (#7322)
* runner.go: Better abstract vision model integration
* Soften windows clang requirement (#7428)
* Remove submodule and shift to Go server - 0.4.0 (#7157)
* Move windows app out of preview (#7347)
* windows: Support alt install paths, fit and finish (#6967)
* add more tests for getting the optimal tiled canvas (#7411)
* Switch windows to clang (#7407)
* tests: Add test for Unicode processing
* runner.go: Better handle return NULL values from llama.cpp
* add mllama image processing to the generate handler (#7384)
* Bump to latest Go 1.22 patch (#7379)
* Fix deepseek deseret regex (#7369)
* Better support for AMD multi-GPU on linux (#7212)
* Fix unicode output on windows with redirect to file (#7358)
* Fix incremental build file deps (#7361)
* Improve dependency gathering logic (#7345)
* fix#7247 - invalid image input (#7249)
* integration: harden embedding test (#7306)
* default to "FROM ." if a Modelfile isn't present (#7250)
* Fix rocm windows build and clean up dependency gathering (#7305)
* runner.go: Merge partial unicode characters before sending
* readme: add Ollama for Swift to the community integrations (#7295)
* server: allow vscode-webview origin (#7273)
* image processing for llama3.2 (#6963)
* llama: Decouple patching script from submodule (#7139)
* llama: add compiler tags for cpu features (#7137)
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=59
* New Models
+ Granite 3 MoE: The IBM Granite 1B and 3B models are the
first mixture of experts (MoE) Granite models from IBM
designed for low latency usage.
+ Granite 3 Dense: The IBM Granite 2B and 8B models are
designed to support tool-based use cases and support for
retrieval augmented generation (RAG), streamlining code
generation, translation and bug fixing.
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=57
* New safety models:
~ Llama Guard 3: a series of models by Meta, fine-tuned for
content safety classification of LLM inputs and responses.
~ ShieldGemma: ShieldGemma is set of instruction tuned models
from Google DeepMind for evaluating the safety of text
prompt input and text output responses against a set of
defined safety policies.
* Fixed issue where ollama pull would leave connections when
encountering an error
* ollama rm will now stop a model if it is running prior to
deleting it
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=55
- Update to version 0.3.12:
* Llama 3.2: Meta's Llama 3.2 goes small with 1B and 3B
models.
* Qwen 2.5 Coder: The latest series of Code-Specific Qwen
models, with significant improvements in code generation,
code reasoning, and code fixing.
* Ollama now supports ARM Windows machines
* Fixed rare issue where Ollama would report a missing .dll
file on Windows
* Fixed performance issue for Windows without GPUs (forwarded request 1204394 from cabelo)
OBS-URL: https://build.opensuse.org/request/show/1204591
OBS-URL: https://build.opensuse.org/package/show/openSUSE:Factory/ollama?expand=0&rev=21
* Llama 3.2: Meta's Llama 3.2 goes small with 1B and 3B
models.
* Qwen 2.5 Coder: The latest series of Code-Specific Qwen
models, with significant improvements in code generation,
code reasoning, and code fixing.
* Ollama now supports ARM Windows machines
* Fixed rare issue where Ollama would report a missing .dll
file on Windows
* Fixed performance issue for Windows without GPUs
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=53
* llm: add solar pro (preview) (#6846)
* server: add tool parsing support for nemotron-mini (#6849)
* make patches git am-able
* CI: dist directories no longer present (#6834)
* CI: clean up naming, fix tagging latest (#6832)
* CI: set platform build build_linux script to keep buildx happy (#6829)
* readme: add Agents-Flex to community integrations (#6788)
* fix typo in import docs (#6828)
* readme: add vim-intelligence-bridge to Terminal section (#6818)
* readme: add Obsidian Quiz Generator plugin to community integrations (#6789)
* Fix incremental builds on linux (#6780)
* Use GOARCH for build dirs (#6779)
* Optimize container images for startup (#6547)
* examples: updated requirements.txt for privategpt example
* examples: polish loganalyzer example (#6744)
* readme: add ollama_moe to community integrations (#6752)
* runner: Flush pending responses before returning
* add "stop" command (#6739)
* refactor show ouput
* readme: add QodeAssist to community integrations (#6754)
* Verify permissions for AMD GPU (#6736)
* add *_proxy for debugging
* docs: update examples to use llama3.1 (#6718)
* Quiet down dockers new lint warnings (#6716)
* catch when model vocab size is set correctly (#6714)
* readme: add crewAI to community integrations (#6699)
* readme: add crewAI with mesop to community integrations
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=51
* openai: align chat temperature and frequency_penalty options with completion (#6688)
* docs: improve linux install documentation (#6683)
* openai: don't scale temperature or frequency_penalty (#6514)
* readme: add Archyve to community integrations (#6680)
* readme: add Plasmoid Ollama Control to community integrations (#6681)
* Improve logging on GPU too small (#6666)
* openai: fix "presence_penalty" typo and add test (#6665)
* Fix gemma2 2b conversion (#6645)
* Document uninstall on windows (#6663)
* Revert "Detect running in a container (#6495)" (#6662)
* llm: make load time stall duration configurable via OLLAMA_LOAD_TIMEOUT
* Introduce GPU Overhead env var (#5922)
* Detect running in a container (#6495)
* readme: add AiLama to the list of community integrations (#4957)
* Update gpu.md: Add RTX 3050 Ti and RTX 3050 Ti (#5888)
* server: fix blob download when receiving a 200 response (#6656)
* readme: add Gentoo package manager entry to community integrations (#5714)
* Update install.sh:Replace "command -v" with encapsulated functionality (#6035)
* readme: include Enchanted for Apple Vision Pro (#4949)
* readme: add lsp-ai to community integrations (#5063)
* readme: add ollama-php library to community integrations (#6361)
* readme: add vnc-lm discord bot community integration (#6644)
* llm: use json.hpp from common (#6642)
* readme: add confichat to community integrations (#6378)
* docs: add group to manual Linux isntructions and verify service is running (#6430)
* readme: add gollm to the list of community libraries (#6099)
* readme: add Cherry Studio to community integrations (#6633)
* readme: add Go fun package (#6421)
* docs: fix spelling error (#6391)
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=49
- Update to version 0.3.6:
* Fixed issue where /api/embed would return an error instead of
loading the model when the input field was not provided.
* ollama create can now import Phi-3 models from Safetensors
* Added progress information to ollama create when importing GGUF
files
* Ollama will now import GGUF files faster by minimizing file
copies
- Update to version 0.3.6:
* Fixed issue where temporary files would not be cleaned up
* Fix rare error when Ollama would start up due to invalid model
data
- Update to version 0.3.4:
* New embedding models
- BGE-M3: a large embedding model from BAAI distinguished for
its versatility in Multi-Functionality, Multi-Linguality, and
Multi-Granularity.
- BGE-Large: a large embedding model trained in english.
- Paraphrase-Multilingual: A multilingual embedding model
trained on parallel data for 50+ languages.
* New embedding API with batch support
- Ollama now supports a new API endpoint /api/embed for
embedding generation:
* This API endpoint supports new features:
- Batches: generate embeddings for several documents in
one request
- Normalized embeddings: embeddings are now normalized,
improving similarity results
- Truncation: a new truncate parameter that will error if
set to false
- Metrics: responses include load_duration, total_duration and
prompt_eval_count metrics
OBS-URL: https://build.opensuse.org/request/show/1194354
OBS-URL: https://build.opensuse.org/package/show/openSUSE:Factory/ollama?expand=0&rev=18
* Fixed issue where /api/embed would return an error instead of
loading the model when the input field was not provided.
* ollama create can now import Phi-3 models from Safetensors
* Added progress information to ollama create when importing GGUF
files
* Ollama will now import GGUF files faster by minimizing file
copies
- Update to version 0.3.6:
* Fixed issue where temporary files would not be cleaned up
* Fix rare error when Ollama would start up due to invalid model
data
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=47
* New embedding models
- BGE-M3: a large embedding model from BAAI distinguished for
its versatility in Multi-Functionality, Multi-Linguality, and
Multi-Granularity.
- BGE-Large: a large embedding model trained in english.
- Paraphrase-Multilingual: A multilingual embedding model
trained on parallel data for 50+ languages.
* New embedding API with batch support
- Ollama now supports a new API endpoint /api/embed for
embedding generation:
* This API endpoint supports new features:
- Batches: generate embeddings for several documents in
one request
- Normalized embeddings: embeddings are now normalized,
improving similarity results
- Truncation: a new truncate parameter that will error if
set to false
- Metrics: responses include load_duration, total_duration and
prompt_eval_count metrics
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=46
- Update to version 0.3.3:
* The /api/embed endpoint now returns statistics: total_duration,
load_duration, and prompt_eval_count
* Added usage metrics to the /v1/embeddings OpenAI compatibility
API
* Fixed issue where /api/generate would respond with an empty
string if provided a context
* Fixed issue where /api/generate would return an incorrect
value for context
* /show modefile will now render MESSAGE commands correctly
- Update to version 0.3.2:
* Fixed issue where ollama pull would not resume download
progress
* Fixed issue where phi3 would report an error on older versions
OBS-URL: https://build.opensuse.org/request/show/1191409
OBS-URL: https://build.opensuse.org/package/show/openSUSE:Factory/ollama?expand=0&rev=17
- Update to version 0.3.1:
* Added support for min_p sampling option
* Lowered number of requests required when downloading models
with ollama pull
* ollama create will now autodetect required stop parameters
when importing certain models
* Fixed issue where /save would cause parameters to be saved
incorrectly.
* OpenAI-compatible API will now return a finish_reason of
tool_calls if a tool call occured.
- fix build on leap 15.6
- exclude builds on 32bit due to build failures
OBS-URL: https://build.opensuse.org/request/show/1190611
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=41
- Update to version 0.3.0:
* Ollama now supports tool calling with popular models such
as Llama 3.1. This enables a model to answer a given prompt
using tool(s) it knows about, making it possible for models to
perform more complex tasks or interact with the outside world.
* New models:
~ Llama 3.1
~ Mistral Large 2
~ Firefunction v2
~ Llama-3-Groq-Tool-Use
* Fixed duplicate error message when running ollama create
OBS-URL: https://build.opensuse.org/request/show/1189982
OBS-URL: https://build.opensuse.org/package/show/openSUSE:Factory/ollama?expand=0&rev=15
* Ollama now supports tool calling with popular models such
as Llama 3.1. This enables a model to answer a given prompt
using tool(s) it knows about, making it possible for models to
perform more complex tasks or interact with the outside world.
* New models:
~ Llama 3.1
~ Mistral Large 2
~ Firefunction v2
~ Llama-3-Groq-Tool-Use
* Fixed duplicate error message when running ollama create
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=39
* api embed docs (#5282)
* convert: capture `head_dim` for mistral (#5818)
* Update llama.cpp submodule commit to `d94c6e0c` (#5805)
* server: collect nested tool call objects when parsing (#5824)
* Remove no longer supported max vram var
* Refine error reporting for subprocess crash
* Remove out of space test temporarily (#5825)
* llm: consider `head_dim` in llama arch (#5817)
* Adjust windows ROCm discovery
* add patch for tekken (#5807)
* preserve last assistant message (#5802)
* Fix generate test flakyness (#5804)
* server: validate template (#5734)
* OpenAI: Function Based Testing (#5752)
* adjust openai chat msg processing (#5729)
* fix parsing tool calls
* server: check for empty tools array too (#5779)
* always provide content even if empty (#5778)
* server: only parse tool calls if tools are provided (#5771)
* Fix context exhaustion integration test for small gpus
* Refine scheduler unit tests for reliability
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=37
- Fixed issue with shared libraries
- Added %check section
- Use -v when building
- Update to version 0.2.6:
* New models: MathΣtral is a 7B model designed for math
reasoning and scientific discovery by Mistral AI.
* Fixed issue where uppercase roles such as USER would no longer
work in the chat endpoints
* Fixed issue where empty system message would be included in the
prompt
OBS-URL: https://build.opensuse.org/request/show/1188404
OBS-URL: https://build.opensuse.org/package/show/openSUSE:Factory/ollama?expand=0&rev=13
* New models: MathΣtral is a 7B model designed for math
reasoning and scientific discovery by Mistral AI.
* Fixed issue where uppercase roles such as USER would no longer
work in the chat endpoints
* Fixed issue where empty system message would be included in the
prompt
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=33
- Update to version 0.1.48:
* Fixed issue where Gemma 2 would continuously output when
reaching context limits
* Fixed out of memory and core dump errors when running Gemma 2
* /show info will now show additional model information in
ollama run
* Fixed issue where ollama show would result in an error on
certain vision models
- Update to version 0.1.48:
* Added support for Google Gemma 2 models (9B and 27B)
* Fixed issues with ollama create when importing from Safetensors
- Update to version 0.1.46:
* Docs (#5149)
* fix: quantization with template
* Fix use_mmap parsing for modelfiles
* Refine mmap default logic on linux
* Bump latest fedora cuda repo to 39
OBS-URL: https://build.opensuse.org/request/show/1186033
OBS-URL: https://build.opensuse.org/package/show/openSUSE:Factory/ollama?expand=0&rev=11
* Fixed issue where Gemma 2 would continuously output when
reaching context limits
* Fixed out of memory and core dump errors when running Gemma 2
* /show info will now show additional model information in
ollama run
* Fixed issue where ollama show would result in an error on
certain vision models
- Update to version 0.1.48:
* Added support for Google Gemma 2 models (9B and 27B)
* Fixed issues with ollama create when importing from Safetensors
- Update to version 0.1.46:
* Docs (#5149)
* fix: quantization with template
* Fix use_mmap parsing for modelfiles
* Refine mmap default logic on linux
* Bump latest fedora cuda repo to 39
OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=29