a6be67967f
* New ollama launch clawdbot command for launching Clawdbot Image generation:
Eyad Issa2026-01-28 23:21:33 +00:00
aca5577791
Fix version and remove duplicated copyright comment in .spec
Eyad Issa2026-01-28 23:20:58 +00:00
96c17fa94a
Accepting request 1329648 from home:mslacken:ml
Eyad Issa2026-01-28 23:19:26 +00:00
d1fcc05458
Accepting request 1328568 from science:machinelearning
Ana Guerrero2026-01-22 14:15:44 +00:00
26551ac3a3
- Update to version 0.14.2: * New models: TranslateGemma * Shift + Enter will now enter a newline in Ollama's CLI * Improve /v1/responses API to better confirm to OpenResponses specification - Update to version 0.14.1: * Experimental image generation models are available Linux (CUDA) ollama run x/z-image-turboEyad Issa2026-01-21 18:48:33 +00:00
a1aa13bec5
- Update to version 0.14.0: * ollama run --experimental CLI will now open a new Ollama CLI that includes an agent loop and the bash tool * Anthropic API compatibility: support for the /v1/messages API * A new REQUIRES command for the Modelfile allows declaring which version of Ollama is required for the model * For older models, Ollama will avoid an integer underflow on low VRAM systems during memory estimation * More accurate VRAM measurements for AMD iGPUs * An error will now return when embeddings return NaN or -Inf
Eyad Issa2026-01-14 18:41:04 +00:00
98424f4344
Accepting request 1312121 from home:mimosius:science:machinelearning
Eyad Issa2025-10-18 14:10:00 +00:00
944f54f239
- Update vendored golang.org/x/net/html to v0.45.0 [boo#1251413] [CVE-2025-47911] [boo#1241757] [CVE-2025-22872] - Update to version 0.12.5: * Fixed issue where "think": false would show an error instead of being silently ignored * Fixed deepseek-r1 output issues - Update to version 0.12.4: * Flash attention is now enabled by default for Qwen 3 and Qwen 3 Coder * Fixed an issue where keep_alive in the API would accept different values for the /api/chat and /api/generate endpoints * Fixed tool calling rendering with qwen3-coder * More reliable and accurate VRAM detection * OLLAMA_FLASH_ATTENTION can now be overridden to 0 for models that have flash attention enabled by default * Fixed crash where templates were not correctly defined * openai: always provide reasoning * Bug fixes * No notable changes. * Fixed issue when quantizing models with the Gemma 3n * Ollama will now limit context length to what the model was * Fixed issue where tool calls without parameters would not be returned correctly * Fixed issue where some special tokens would not be tokenized - Allow to build for Package Hub for SLE-15-SP7
Eyad Issa2025-10-11 15:15:50 +00:00
65512c06ec
- Update to version 0.12.3: * New models: DeepSeek-V3.1-Terminus, Kimi K2-Instruct-0905 * Fixed issue where tool calls provided as stringified JSON would not be parsed correctly * ollama push will now provide a URL to follow to sign in * Fixed issues where qwen3-coder would output unicode characters incorrectly * Fix issue where loading a model with /load would crash - Update to version 0.12.2: * A new web search API is now available in Ollama * Models with Qwen3's architecture including MoE now run in Ollama's new engine * Fixed issue where built-in tools for gpt-oss were not being rendered correctly * Support multi-regex pretokenizers in Ollama's new engine * Ollama's new engine can now load tensors by matching a prefix or suffix - Update to version 0.12.1: * New model: Qwen3 Embedding: state of the art open embedding model by the Qwen team * Qwen3-Coder now supports tool calling * Fixed issue where Gemma3 QAT models would not output correct tokens * Fix issue where & characters in Qwen3-Coder would not be parsed correctly when function calling * Fixed issues where ollama signin would not work properly - Update to version 0.12.0: * Cloud models are now available in preview * Models with the Bert architecture now run on Ollama's engine * Models with the Qwen 3 architecture now run on Ollama's engine
Eyad Issa2025-10-04 21:24:33 +00:00
aaf1294c95
- Update to version 0.11.4: * openai: allow for content and tool calls in the same message * openai: when converting role=tool messages, propagate the tool name * openai: always provide reasoning * Bug fixes
Eyad Issa2025-08-07 23:21:08 +00:00
172a9eec78
- Update to version 0.10.1: * No notable changes. - Update to version 0.10.0: * ollama ps will now show the context length of loaded models * Improved performance in gemma3n models by 2-3x * Parallel request processing now defaults to 1 * Fixed issue where tool calling would not work correctly with granite3.3 and mistral-nemo models * Fixed issue where Ollama's tool calling would not work correctly if a tool's name was part of of another one, such as add and get_address * Improved performance when using multiple GPUs by 10-30% * Ollama's OpenAI-compatible API will now support WebP images * Fixed issue where ollama show would report an error * ollama run will more gracefully display errors
Eyad Issa2025-08-05 00:09:03 +00:00
7624762a9a
Accepting request 1290234 from science:machinelearning
Ana Guerrero2025-07-06 15:07:50 +00:00
9b2f052e10
- Update to version 0.9.5: * No notable changes. - Update to version 0.9.4: * The directory in which models are stored can now be modified. * Tool calling with empty parameters will now work correctly * Fixed issue when quantizing models with the Gemma 3n architecture - Update to version 0.9.3: * Ollama now supports Gemma 3n * Ollama will now limit context length to what the model was trained against to avoid strange overflow behavior - Update to version 0.9.2: * Fixed issue where tool calls without parameters would not be returned correctly * Fixed does not support generate errors * Fixed issue where some special tokens would not be tokenized properly for some model architectures
Eyad Issa2025-07-03 00:15:58 +00:00
b72896bbb7
Accepting request 1288227 from science:machinelearning
Ana Guerrero2025-06-24 18:50:15 +00:00
73961852fd
- Update to version 0.9.1: * Tool calling reliability and performance has been improved for the following models: Magistral Llama 4 Mistral DeepSeek-R1-2508 * Magistral now supports disabling thinking mode * Error messages that previously showed POST predict will now be more informative
Eyad Issa2025-06-17 10:54:45 +00:00
d9f77ab949
Accepting request 1283893 from science:machinelearning
Ana Guerrero2025-06-10 07:05:27 +00:00
9b760ab447
- Update to version 0.9.0: * Ollama now has the ability to enable or disable thinking. This gives users the flexibility to choose the model’s thinking behavior for different applications and use cases. - Update to version 0.8.0: * Ollama will now stream responses with tool calls * Logs will now include better memory estimate debug information when running models in Ollama's engine. - Update to version 0.7.1: * Improved model memory management to allocate sufficient memory to prevent crashes when running multimodal models in certain situations * Enhanced memory estimation for models to prevent unintended memory offloading * ollama show will now show ... when data is truncated * Fixed crash that would occur with qwen2.5vl * Fixed crash on Nvidia's CUDA for llama3.2-vision * Support for Alibaba's Qwen 3 and Qwen 2 architectures in Ollama's new multimodal engine
Eyad Issa2025-06-01 00:00:21 +00:00
b7464d2582
Accepting request 1279778 from science:machinelearning
Ana Guerrero2025-05-26 16:32:37 +00:00
c4d3594049
Accepting request 1279534 from home:bigironman:branches:science:machinelearning
Eyad Issa2025-05-23 12:19:15 +00:00
329d0b3b07
Accepting request 1279105 from home:bigironman:branches:science:machinelearning
Eyad Issa2025-05-21 21:23:32 +00:00
cf4576f29f
Accepting request 1278142 from science:machinelearning
Ana Guerrero2025-05-20 07:36:41 +00:00
4bf7dbc507
- Update to version 0.7.0: * Ollama now supports multimodal models via Ollama’s new engine, starting with new vision multimodal models: ~ Meta Llama 4 ~ Google Gemma 3 ~ Qwen 2.5 VL ~ Qwen 2.5 VL * Ollama now supports providing WebP images as input to multimodal models * Improved performance of importing safetensors models via ollama create * Various bug fixes and performance enhancements
Eyad Issa2025-05-17 14:49:57 +00:00
46e1c2bc3f
Accepting request 1277233 from science:machinelearning
Ana Guerrero2025-05-14 15:01:10 +00:00
e550128c4d
- Update to version 0.6.8: * Performance improvements for Qwen 3 MoE models on NVIDIA and AMD GPUs * Fixed a memory leak that occurred when providing images as input * ollama show will now correctly label older vision models such as llava * Reduced out of memory errors by improving worst-case memory estimations * Fix issue that resulted in a context canceled error - Update to version 0.6.7: * New model: Qwen 3 * New model: Phi 4 reasoning and Phi 4 mini reasoning * New model: llama 4 * Increased default context window to 4096 tokens * Fixed issue where image paths would not be recognized with ~ when being provided to ollama run * Improved output quality when using JSON mode in certain scenarios * Fixed issue where model would be stuck in the Stopping... state - Use source url (https://en.opensuse.org/SourceUrls)
Eyad Issa2025-05-06 10:13:41 +00:00
0bf309e93f
Accepting request 1272498 from science:machinelearning
Ana Guerrero2025-04-25 20:19:03 +00:00
ff0b24ba61
- Update to version 0.6.6: * New model: IBM Granite 3.3 * New model: DeepCoder * New, faster model downloading: OLLAMA_EXPERIMENT=client2 ollama serve will run Ollama using a new downloader with improved performance and reliability when running ollama pull * Fixed memory leak issues when running Gemma 3, Mistral Small 3.1 and other models on Ollama * Improved performance of ollama create when importing models from Safetensors * Ollama will now allow tool function parameters with either a single type or an array of types * Fixed certain out-of-memory issues caused by not reserving enough memory at startup * Fixed nondeterministic model unload order * Included the items and $defs fields to properly handle array types in the API * OpenAI-Beta headers are now included in the CORS safelist * Fixed issue where model tensor data would be corrupted when importing models from Safetensors
Eyad Issa2025-04-24 16:37:52 +00:00
b2c17bb47d
- Add ollama to the video group - Update to version 0.6.5: * Add support for mistral-small * Fix issues with spm tokenizer for Gemma 3 models * Add checks for values falling out of sliding window cache * Improve file descriptor management for tensors and Pull operations * Add gfx1200 & gfx1201 GPU support on Linux * Optimize sliding window attention and KV cache implementations * Implement loading tensors in 32KiB chunks for better performance * Add autotemplate for gemma3 models * Add benchmarking for ollama server performance * Fix file handling in /proc/cpuinfo discovery * Support heterogeneous KV cache layer sizes in memory estimation * Fix debug logging for memory estimates * Improve error handling for empty logits and tensor data reading * Return model capabilities from the show endpoint - Update BuildRequires to go1.24
Eyad Issa2025-04-19 22:15:43 +00:00
8bad7a10ef
Accepting request 1256309 from science:machinelearning
Ana Guerrero2025-03-27 21:31:53 +00:00
3af7d3df29
Accepting request 1256012 from home:Levitating:branches:science:machinelearning
Eyad Issa2025-03-26 20:03:43 +00:00
a83e47ea03
Accepting request 1254230 from science:machinelearning
Ana Guerrero2025-03-19 21:33:26 +00:00
800abc486e
Accepting request 1254170 from home:bmwiedemann:branches:science:machinelearning
Eyad Issa2025-03-18 20:06:03 +00:00
fb36d5b698
Accepting request 1252927 from science:machinelearning
Ana Guerrero2025-03-14 22:52:08 +00:00
b50fb8cb36
- Update BuildRequires to go1.24
Eyad Issa2025-03-14 01:22:52 +00:00
08a952a4a0
- Update to version 0.6.0: * New model: Gemma 3 * Fixed error that would occur when running snowflake-arctic-embed and snowflake-arctic-embed2 models * Various performance improvements and bug fixes
Eyad Issa2025-03-14 01:19:14 +00:00
2ab2860f07
- Use Ninja instead of Make and update the build script to match the new version
Eyad Issa2025-02-15 02:49:21 +00:00
b8329e61c4
- Update to version 0.5.11: * No notable changes for Linux - Update to version 0.5.10: * Fixed issue on multi-GPU Windows and Linux machines where memory estimations would be incorrect - Update to version 0.5.9: * New model: DeepScaleR * New model: OpenThinker - Update to version 0.5.8: * Ollama will now use AVX-512 instructions where available for additional CPU acceleration * Fixed indexing error that would occur when downloading a model with ollama run or ollama pull * Fixes cases where download progress would reverse
Eyad Issa2025-02-15 01:36:40 +00:00
30357c31e8
Accepting request 1240594 from science:machinelearning
Ana Guerrero2025-01-29 15:10:09 +00:00
5e1bd9f77a
Accepting request 1240545 from home:adrianSuSE
Eyad Issa2025-01-27 16:15:11 +00:00
fe39e419e6
- Removed 01-build-verbose.patch: embedded GOFLAG into .spec file - Disabled reproducible.patch: should be not needed, as .gz is not produced anymore - Update to version 0.5.7:
Eyad Issa2025-01-17 00:02:27 +00:00
7a79526c46
Accepting request 1230609 from science:machinelearning
Ana Guerrero2024-12-12 20:18:15 +00:00
399e882fde
Accepting request 1230608 from home:bmwiedemann:branches:science:machinelearning
Eyad Issa2024-12-12 14:52:49 +00:00
3bc79498e9
- Update to version 0.5.1: - Update to version 0.5.0: - Update to version 0.4.7:
Eyad Issa2024-12-07 18:30:08 +00:00
e1749d13d5
- Update to version 0.4.6: - Update to version 0.4.5: - Update to version 0.4.4: - Update to version 0.4.3:
Eyad Issa2024-11-30 20:05:29 +00:00
c01fcf4bf9
Accepting request 1225993 from science:machinelearning
Ana Guerrero2024-11-24 10:04:51 +00:00
523c01be8b
- Add patch 01-build-verbose.patch to add the -v option to go build - Update to version 0.4.1: * runner.go: Check for zero length images * docs: update langchainpy.md with proper model name (#7527) * Set macos min version for all architectures (#7579) * win: remove preview title from installer (#7529) * Workaround buggy P2P ROCm copy on windows (#7466) * Debug logging for nvcuda init (#7532) * Align rocm compiler flags (#7467) * Be explicit for gpu library link dir (#7560) * docs: OLLAMA_NEW_RUNNERS no longer exists * runner.go: Remove unused arguments * sched: Lift parallel restriction for multimodal models except mllama
Eyad Issa2024-11-11 14:46:13 +00:00
924fe15781
- Update to version 0.4.0-rc6: * Refine default thread selection for NUMA systems (#7322) * runner.go: Better abstract vision model integration * Soften windows clang requirement (#7428) * Remove submodule and shift to Go server - 0.4.0 (#7157) * Move windows app out of preview (#7347) * windows: Support alt install paths, fit and finish (#6967) * add more tests for getting the optimal tiled canvas (#7411) * Switch windows to clang (#7407) * tests: Add test for Unicode processing * runner.go: Better handle return NULL values from llama.cpp * add mllama image processing to the generate handler (#7384) * Bump to latest Go 1.22 patch (#7379) * Fix deepseek deseret regex (#7369) * Better support for AMD multi-GPU on linux (#7212) * Fix unicode output on windows with redirect to file (#7358) * Fix incremental build file deps (#7361) * Improve dependency gathering logic (#7345) * fix#7247 - invalid image input (#7249) * integration: harden embedding test (#7306) * default to "FROM ." if a Modelfile isn't present (#7250) * Fix rocm windows build and clean up dependency gathering (#7305) * runner.go: Merge partial unicode characters before sending * readme: add Ollama for Swift to the community integrations (#7295) * server: allow vscode-webview origin (#7273) * image processing for llama3.2 (#6963) * llama: Decouple patching script from submodule (#7139) * llama: add compiler tags for cpu features (#7137)
Eyad Issa2024-11-01 02:20:51 +00:00
332862e2b8
- Update to version 0.3.13: * New safety models: ~ Llama Guard 3: a series of models by Meta, fine-tuned for content safety classification of LLM inputs and responses. ~ ShieldGemma: ShieldGemma is set of instruction tuned models from Google DeepMind for evaluating the safety of text prompt input and text output responses against a set of defined safety policies. * Fixed issue where ollama pull would leave connections when encountering an error * ollama rm will now stop a model if it is running prior to deleting it
Guillaume GARDET2024-10-14 07:28:18 +00:00
374f7878c3
Accepting request 1207484 from home:VaiTon:branches:science:machinelearning
Guillaume GARDET2024-10-14 07:28:18 +00:00
bd7fc28fe4
Accepting request 1204591 from science:machinelearning
Ana Guerrero2024-09-30 13:40:27 +00:00
56b72bf587
Accepting request 1204591 from science:machinelearning
Ana Guerrero2024-09-30 13:40:27 +00:00
2808304cf4
- Update to version 0.3.12: * Llama 3.2: Meta's Llama 3.2 goes small with 1B and 3B models. * Qwen 2.5 Coder: The latest series of Code-Specific Qwen models, with significant improvements in code generation, code reasoning, and code fixing. * Ollama now supports ARM Windows machines * Fixed rare issue where Ollama would report a missing .dll file on Windows * Fixed performance issue for Windows without GPUs
Eyad Issa2024-09-29 21:30:54 +00:00
e29af59180
Accepting request 1204394 from home:cabelo:branches:science:machinelearning
Eyad Issa2024-09-29 21:30:54 +00:00
f7aaf9b2af
Accepting request 1202264 from science:machinelearning
Ana Guerrero2024-09-22 09:06:09 +00:00
edcd06fbe6
Accepting request 1202264 from science:machinelearning
Ana Guerrero2024-09-22 09:06:09 +00:00
5bb20bbdee
- Update to version 0.3.11: * llm: add solar pro (preview) (#6846) * server: add tool parsing support for nemotron-mini (#6849) * make patches git am-able * CI: dist directories no longer present (#6834) * CI: clean up naming, fix tagging latest (#6832) * CI: set platform build build_linux script to keep buildx happy (#6829) * readme: add Agents-Flex to community integrations (#6788) * fix typo in import docs (#6828) * readme: add vim-intelligence-bridge to Terminal section (#6818) * readme: add Obsidian Quiz Generator plugin to community integrations (#6789) * Fix incremental builds on linux (#6780) * Use GOARCH for build dirs (#6779) * Optimize container images for startup (#6547) * examples: updated requirements.txt for privategpt example * examples: polish loganalyzer example (#6744) * readme: add ollama_moe to community integrations (#6752) * runner: Flush pending responses before returning * add "stop" command (#6739) * refactor show ouput * readme: add QodeAssist to community integrations (#6754) * Verify permissions for AMD GPU (#6736) * add *_proxy for debugging * docs: update examples to use llama3.1 (#6718) * Quiet down dockers new lint warnings (#6716) * catch when model vocab size is set correctly (#6714) * readme: add crewAI to community integrations (#6699) * readme: add crewAI with mesop to community integrations
Eyad Issa2024-09-20 20:29:36 +00:00
b79ecf65ca
Accepting request 1202133 from home:adrianSuSE:branches:science:machinelearning
Eyad Issa2024-09-20 20:29:36 +00:00
e5b1fec77c
Accepting request 1201962 from science:machinelearning
Ana Guerrero2024-09-19 19:17:44 +00:00