Commit Graph

  • 8f6a537dea Accepting request 1329736 from science:machinelearning factory Ana Guerrero 2026-01-29 16:46:03 +00:00
  • a6be67967f * New ollama launch clawdbot command for launching Clawdbot Image generation: Eyad Issa 2026-01-28 23:21:33 +00:00
  • aca5577791 Fix version and remove duplicated copyright comment in .spec Eyad Issa 2026-01-28 23:20:58 +00:00
  • 96c17fa94a Accepting request 1329648 from home:mslacken:ml Eyad Issa 2026-01-28 23:19:26 +00:00
  • d1fcc05458 Accepting request 1328568 from science:machinelearning Ana Guerrero 2026-01-22 14:15:44 +00:00
  • 26551ac3a3 - Update to version 0.14.2: * New models: TranslateGemma * Shift + Enter will now enter a newline in Ollama's CLI * Improve /v1/responses API to better confirm to OpenResponses specification - Update to version 0.14.1: * Experimental image generation models are available Linux (CUDA) ollama run x/z-image-turbo Eyad Issa 2026-01-21 18:48:33 +00:00
  • a1aa13bec5 - Update to version 0.14.0: * ollama run --experimental CLI will now open a new Ollama CLI that includes an agent loop and the bash tool * Anthropic API compatibility: support for the /v1/messages API * A new REQUIRES command for the Modelfile allows declaring which version of Ollama is required for the model * For older models, Ollama will avoid an integer underflow on low VRAM systems during memory estimation * More accurate VRAM measurements for AMD iGPUs * An error will now return when embeddings return NaN or -Inf Eyad Issa 2026-01-14 18:41:04 +00:00
  • c5264cdec3 Accepting request 1324368 from science:machinelearning Dominique Leuenberger 2025-12-25 18:57:30 +00:00
  • 5a22398f23 Accepting request 1323716 from home:mimosius:science:machinelearning Eyad Issa 2025-12-24 12:55:19 +00:00
  • d70048e75c Accepting request 1323403 from science:machinelearning Ana Guerrero 2025-12-18 17:32:23 +00:00
  • c0713fb29e OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=133 Eyad Issa 2025-12-17 18:26:26 +00:00
  • 6bb25745d3 Accepting request 1323340 from home:mimosius:science:machinelearning Eyad Issa 2025-12-17 18:18:59 +00:00
  • 6ab999dfe8 Accepting request 1321364 from science:machinelearning Ana Guerrero 2025-12-08 10:54:38 +00:00
  • 8fce087f1b Accepting request 1321193 from home:VaiTon:branches:science:machinelearning Eyad Issa 2025-12-06 14:54:55 +00:00
  • 81f96995e6 Accepting request 1319255 from science:machinelearning Ana Guerrero 2025-11-24 13:10:47 +00:00
  • bc0f28f53d Accepting request 1319192 from home:mimosius:science:machinelearning Eyad Issa 2025-11-22 16:05:11 +00:00
  • 225f77e0f6 Accepting request 1317118 from science:machinelearning Dominique Leuenberger 2025-11-11 18:22:26 +00:00
  • e5457d3d02 Accepting request 1317097 from home:eeich:branches:science:machinelearning Christian Goll 2025-11-11 14:19:36 +00:00
  • 2dd2e711ba Accepting request 1316476 from science:machinelearning Ana Guerrero 2025-11-08 15:36:34 +00:00
  • ef07c70fef Accepting request 1316467 from home:mimosius:science:machinelearning Eyad Issa 2025-11-07 16:45:55 +00:00
  • 625a28aed6 Accepting request 1315207 from science:machinelearning Ana Guerrero 2025-11-03 17:54:58 +00:00
  • ffa132f9b9 Accepting request 1315028 from home:mimosius:science:machinelearning Eyad Issa 2025-11-03 01:56:17 +00:00
  • 68cb516a9e Accepting request 1313741 from science:machinelearning Ana Guerrero 2025-10-27 13:41:10 +00:00
  • dda32dab82 Accepting request 1313740 from home:Yoshio_Sato:branches:science:machinelearning Eyad Issa 2025-10-26 19:51:04 +00:00
  • 3c45ab8003 Accepting request 1312160 from science:machinelearning Dominique Leuenberger 2025-10-18 16:36:26 +00:00
  • 98424f4344 Accepting request 1312121 from home:mimosius:science:machinelearning Eyad Issa 2025-10-18 14:10:00 +00:00
  • 944f54f239 - Update vendored golang.org/x/net/html to v0.45.0 [boo#1251413] [CVE-2025-47911] [boo#1241757] [CVE-2025-22872] - Update to version 0.12.5: * Fixed issue where "think": false would show an error instead of being silently ignored * Fixed deepseek-r1 output issues - Update to version 0.12.4: * Flash attention is now enabled by default for Qwen 3 and Qwen 3 Coder * Fixed an issue where keep_alive in the API would accept different values for the /api/chat and /api/generate endpoints * Fixed tool calling rendering with qwen3-coder * More reliable and accurate VRAM detection * OLLAMA_FLASH_ATTENTION can now be overridden to 0 for models that have flash attention enabled by default * Fixed crash where templates were not correctly defined * openai: always provide reasoning * Bug fixes * No notable changes. * Fixed issue when quantizing models with the Gemma 3n * Ollama will now limit context length to what the model was * Fixed issue where tool calls without parameters would not be returned correctly * Fixed issue where some special tokens would not be tokenized - Allow to build for Package Hub for SLE-15-SP7 Eyad Issa 2025-10-11 15:15:50 +00:00
  • bf8ed313e6 Accepting request 1309021 from science:machinelearning Dominique Leuenberger 2025-10-05 15:51:13 +00:00
  • 65512c06ec - Update to version 0.12.3: * New models: DeepSeek-V3.1-Terminus, Kimi K2-Instruct-0905 * Fixed issue where tool calls provided as stringified JSON would not be parsed correctly * ollama push will now provide a URL to follow to sign in * Fixed issues where qwen3-coder would output unicode characters incorrectly * Fix issue where loading a model with /load would crash - Update to version 0.12.2: * A new web search API is now available in Ollama * Models with Qwen3's architecture including MoE now run in Ollama's new engine * Fixed issue where built-in tools for gpt-oss were not being rendered correctly * Support multi-regex pretokenizers in Ollama's new engine * Ollama's new engine can now load tensors by matching a prefix or suffix - Update to version 0.12.1: * New model: Qwen3 Embedding: state of the art open embedding model by the Qwen team * Qwen3-Coder now supports tool calling * Fixed issue where Gemma3 QAT models would not output correct tokens * Fix issue where & characters in Qwen3-Coder would not be parsed correctly when function calling * Fixed issues where ollama signin would not work properly - Update to version 0.12.0: * Cloud models are now available in preview * Models with the Bert architecture now run on Ollama's engine * Models with the Qwen 3 architecture now run on Ollama's engine Eyad Issa 2025-10-04 21:24:33 +00:00
  • 44b7a8907c Accepting request 1298233 from science:machinelearning Dominique Leuenberger 2025-08-08 13:12:48 +00:00
  • aaf1294c95 - Update to version 0.11.4: * openai: allow for content and tool calls in the same message * openai: when converting role=tool messages, propagate the tool name * openai: always provide reasoning * Bug fixes Eyad Issa 2025-08-07 23:21:08 +00:00
  • a33f11b76e Accepting request 1298006 from science:machinelearning Dominique Leuenberger 2025-08-07 14:48:46 +00:00
  • 349d4de67b - Update to version 0.11.0: * New model: OpenAI gpt-oss 20B and 120B * Quantization - MXFP4 format Eyad Issa 2025-08-06 12:51:20 +00:00
  • 9bbc684da6 Accepting request 1297591 from science:machinelearning Dominique Leuenberger 2025-08-05 12:21:43 +00:00
  • 172a9eec78 - Update to version 0.10.1: * No notable changes. - Update to version 0.10.0: * ollama ps will now show the context length of loaded models * Improved performance in gemma3n models by 2-3x * Parallel request processing now defaults to 1 * Fixed issue where tool calling would not work correctly with granite3.3 and mistral-nemo models * Fixed issue where Ollama's tool calling would not work correctly if a tool's name was part of of another one, such as add and get_address * Improved performance when using multiple GPUs by 10-30% * Ollama's OpenAI-compatible API will now support WebP images * Fixed issue where ollama show would report an error * ollama run will more gracefully display errors Eyad Issa 2025-08-05 00:09:03 +00:00
  • 7624762a9a Accepting request 1290234 from science:machinelearning Ana Guerrero 2025-07-06 15:07:50 +00:00
  • 9b2f052e10 - Update to version 0.9.5: * No notable changes. - Update to version 0.9.4: * The directory in which models are stored can now be modified. * Tool calling with empty parameters will now work correctly * Fixed issue when quantizing models with the Gemma 3n architecture - Update to version 0.9.3: * Ollama now supports Gemma 3n * Ollama will now limit context length to what the model was trained against to avoid strange overflow behavior - Update to version 0.9.2: * Fixed issue where tool calls without parameters would not be returned correctly * Fixed does not support generate errors * Fixed issue where some special tokens would not be tokenized properly for some model architectures Eyad Issa 2025-07-03 00:15:58 +00:00
  • b72896bbb7 Accepting request 1288227 from science:machinelearning Ana Guerrero 2025-06-24 18:50:15 +00:00
  • 73961852fd - Update to version 0.9.1: * Tool calling reliability and performance has been improved for the following models: Magistral Llama 4 Mistral DeepSeek-R1-2508 * Magistral now supports disabling thinking mode * Error messages that previously showed POST predict will now be more informative Eyad Issa 2025-06-17 10:54:45 +00:00
  • d9f77ab949 Accepting request 1283893 from science:machinelearning Ana Guerrero 2025-06-10 07:05:27 +00:00
  • 9b760ab447 - Update to version 0.9.0: * Ollama now has the ability to enable or disable thinking. This gives users the flexibility to choose the model’s thinking behavior for different applications and use cases. - Update to version 0.8.0: * Ollama will now stream responses with tool calls * Logs will now include better memory estimate debug information when running models in Ollama's engine. - Update to version 0.7.1: * Improved model memory management to allocate sufficient memory to prevent crashes when running multimodal models in certain situations * Enhanced memory estimation for models to prevent unintended memory offloading * ollama show will now show ... when data is truncated * Fixed crash that would occur with qwen2.5vl * Fixed crash on Nvidia's CUDA for llama3.2-vision * Support for Alibaba's Qwen 3 and Qwen 2 architectures in Ollama's new multimodal engine Eyad Issa 2025-06-01 00:00:21 +00:00
  • b7464d2582 Accepting request 1279778 from science:machinelearning Ana Guerrero 2025-05-26 16:32:37 +00:00
  • fa54a05c54 OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=101 Eyad Issa 2025-05-24 13:40:07 +00:00
  • c4d3594049 Accepting request 1279534 from home:bigironman:branches:science:machinelearning Eyad Issa 2025-05-23 12:19:15 +00:00
  • 329d0b3b07 Accepting request 1279105 from home:bigironman:branches:science:machinelearning Eyad Issa 2025-05-21 21:23:32 +00:00
  • cf4576f29f Accepting request 1278142 from science:machinelearning Ana Guerrero 2025-05-20 07:36:41 +00:00
  • 4bf7dbc507 - Update to version 0.7.0: * Ollama now supports multimodal models via Ollama’s new engine, starting with new vision multimodal models: ~ Meta Llama 4 ~ Google Gemma 3 ~ Qwen 2.5 VL ~ Qwen 2.5 VL * Ollama now supports providing WebP images as input to multimodal models * Improved performance of importing safetensors models via ollama create * Various bug fixes and performance enhancements Eyad Issa 2025-05-17 14:49:57 +00:00
  • 46e1c2bc3f Accepting request 1277233 from science:machinelearning Ana Guerrero 2025-05-14 15:01:10 +00:00
  • f968f1aa72 - Use source url (https://en.opensuse.org/SourceUrls) Eyad Issa 2025-05-13 17:03:33 +00:00
  • 3f196feccd OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=94 Eyad Issa 2025-05-13 17:02:11 +00:00
  • e550128c4d - Update to version 0.6.8: * Performance improvements for Qwen 3 MoE models on NVIDIA and AMD GPUs * Fixed a memory leak that occurred when providing images as input * ollama show will now correctly label older vision models such as llava * Reduced out of memory errors by improving worst-case memory estimations * Fix issue that resulted in a context canceled error - Update to version 0.6.7: * New model: Qwen 3 * New model: Phi 4 reasoning and Phi 4 mini reasoning * New model: llama 4 * Increased default context window to 4096 tokens * Fixed issue where image paths would not be recognized with ~ when being provided to ollama run * Improved output quality when using JSON mode in certain scenarios * Fixed issue where model would be stuck in the Stopping... state - Use source url (https://en.opensuse.org/SourceUrls) Eyad Issa 2025-05-06 10:13:41 +00:00
  • 0bf309e93f Accepting request 1272498 from science:machinelearning Ana Guerrero 2025-04-25 20:19:03 +00:00
  • ff0b24ba61 - Update to version 0.6.6: * New model: IBM Granite 3.3 * New model: DeepCoder * New, faster model downloading: OLLAMA_EXPERIMENT=client2 ollama serve will run Ollama using a new downloader with improved performance and reliability when running ollama pull * Fixed memory leak issues when running Gemma 3, Mistral Small 3.1 and other models on Ollama * Improved performance of ollama create when importing models from Safetensors * Ollama will now allow tool function parameters with either a single type or an array of types * Fixed certain out-of-memory issues caused by not reserving enough memory at startup * Fixed nondeterministic model unload order * Included the items and $defs fields to properly handle array types in the API * OpenAI-Beta headers are now included in the CORS safelist * Fixed issue where model tensor data would be corrupted when importing models from Safetensors Eyad Issa 2025-04-24 16:37:52 +00:00
  • f805b78847 Accepting request 1271013 from science:machinelearning Dominique Leuenberger 2025-04-20 07:36:28 +00:00
  • 794cdfc5f1 OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=89 Eyad Issa 2025-04-19 22:30:16 +00:00
  • b2c17bb47d - Add ollama to the video group - Update to version 0.6.5: * Add support for mistral-small * Fix issues with spm tokenizer for Gemma 3 models * Add checks for values falling out of sliding window cache * Improve file descriptor management for tensors and Pull operations * Add gfx1200 & gfx1201 GPU support on Linux * Optimize sliding window attention and KV cache implementations * Implement loading tensors in 32KiB chunks for better performance * Add autotemplate for gemma3 models * Add benchmarking for ollama server performance * Fix file handling in /proc/cpuinfo discovery * Support heterogeneous KV cache layer sizes in memory estimation * Fix debug logging for memory estimates * Improve error handling for empty logits and tensor data reading * Return model capabilities from the show endpoint - Update BuildRequires to go1.24 Eyad Issa 2025-04-19 22:15:43 +00:00
  • 8bad7a10ef Accepting request 1256309 from science:machinelearning Ana Guerrero 2025-03-27 21:31:53 +00:00
  • 3f63c81429 Re-wrap .changes file to 67 chars Eyad Issa 2025-03-26 20:06:49 +00:00
  • 3af7d3df29 Accepting request 1256012 from home:Levitating:branches:science:machinelearning Eyad Issa 2025-03-26 20:03:43 +00:00
  • a83e47ea03 Accepting request 1254230 from science:machinelearning Ana Guerrero 2025-03-19 21:33:26 +00:00
  • 800abc486e Accepting request 1254170 from home:bmwiedemann:branches:science:machinelearning Eyad Issa 2025-03-18 20:06:03 +00:00
  • fb36d5b698 Accepting request 1252927 from science:machinelearning Ana Guerrero 2025-03-14 22:52:08 +00:00
  • b50fb8cb36 - Update BuildRequires to go1.24 Eyad Issa 2025-03-14 01:22:52 +00:00
  • 08a952a4a0 - Update to version 0.6.0: * New model: Gemma 3 * Fixed error that would occur when running snowflake-arctic-embed and snowflake-arctic-embed2 models * Various performance improvements and bug fixes Eyad Issa 2025-03-14 01:19:14 +00:00
  • 3286c88fc9 Accepting request 1249021 from science:machinelearning Dominique Leuenberger 2025-02-28 16:38:42 +00:00
  • bd42b42b28 Accepting request 1249020 from home:VaiTon:branches:science:machinelearning Eyad Issa 2025-02-27 14:02:42 +00:00
  • cd41c1501e Accepting request 1247774 from science:machinelearning Dominique Leuenberger 2025-02-22 18:05:12 +00:00
  • 2ab2860f07 - Use Ninja instead of Make and update the build script to match the new version Eyad Issa 2025-02-15 02:49:21 +00:00
  • b8329e61c4 - Update to version 0.5.11: * No notable changes for Linux - Update to version 0.5.10: * Fixed issue on multi-GPU Windows and Linux machines where memory estimations would be incorrect - Update to version 0.5.9: * New model: DeepScaleR * New model: OpenThinker - Update to version 0.5.8: * Ollama will now use AVX-512 instructions where available for additional CPU acceleration * Fixed indexing error that would occur when downloading a model with ollama run or ollama pull * Fixes cases where download progress would reverse Eyad Issa 2025-02-15 01:36:40 +00:00
  • 30357c31e8 Accepting request 1240594 from science:machinelearning Ana Guerrero 2025-01-29 15:10:09 +00:00
  • 5e1bd9f77a Accepting request 1240545 from home:adrianSuSE Eyad Issa 2025-01-27 16:15:11 +00:00
  • fe39e419e6 - Removed 01-build-verbose.patch: embedded GOFLAG into .spec file - Disabled reproducible.patch: should be not needed, as .gz is not produced anymore - Update to version 0.5.7: Eyad Issa 2025-01-17 00:02:27 +00:00
  • 7a79526c46 Accepting request 1230609 from science:machinelearning Ana Guerrero 2024-12-12 20:18:15 +00:00
  • 399e882fde Accepting request 1230608 from home:bmwiedemann:branches:science:machinelearning Eyad Issa 2024-12-12 14:52:49 +00:00
  • 3bc79498e9 - Update to version 0.5.1: - Update to version 0.5.0: - Update to version 0.4.7: Eyad Issa 2024-12-07 18:30:08 +00:00
  • e1749d13d5 - Update to version 0.4.6: - Update to version 0.4.5: - Update to version 0.4.4: - Update to version 0.4.3: Eyad Issa 2024-11-30 20:05:29 +00:00
  • c01fcf4bf9 Accepting request 1225993 from science:machinelearning Ana Guerrero 2024-11-24 10:04:51 +00:00
  • 7617b8c945 - Update to version 0.4.2: Eyad Issa 2024-11-16 16:22:34 +00:00
  • 523c01be8b - Add patch 01-build-verbose.patch to add the -v option to go build - Update to version 0.4.1: * runner.go: Check for zero length images * docs: update langchainpy.md with proper model name (#7527) * Set macos min version for all architectures (#7579) * win: remove preview title from installer (#7529) * Workaround buggy P2P ROCm copy on windows (#7466) * Debug logging for nvcuda init (#7532) * Align rocm compiler flags (#7467) * Be explicit for gpu library link dir (#7560) * docs: OLLAMA_NEW_RUNNERS no longer exists * runner.go: Remove unused arguments * sched: Lift parallel restriction for multimodal models except mllama Eyad Issa 2024-11-11 14:46:13 +00:00
  • 133893abde Accepting request 1222485 from science:machinelearning Dominique Leuenberger 2024-11-07 17:17:05 +00:00
  • 7c33e09d61 Accepting request 1222177 from home:adrianSuSE Guillaume GARDET 2024-11-07 15:09:03 +00:00
  • bd7edb56ee - Update to version 0.4.0-rc8: Eyad Issa 2024-11-06 12:33:58 +00:00
  • 00757f2949 - Remove enable-lto.patch Eyad Issa 2024-11-01 02:38:53 +00:00
  • eda01dcf91 OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=60 Eyad Issa 2024-11-01 02:23:50 +00:00
  • 924fe15781 - Update to version 0.4.0-rc6: * Refine default thread selection for NUMA systems (#7322) * runner.go: Better abstract vision model integration * Soften windows clang requirement (#7428) * Remove submodule and shift to Go server - 0.4.0 (#7157) * Move windows app out of preview (#7347) * windows: Support alt install paths, fit and finish (#6967) * add more tests for getting the optimal tiled canvas (#7411) * Switch windows to clang (#7407) * tests: Add test for Unicode processing * runner.go: Better handle return NULL values from llama.cpp * add mllama image processing to the generate handler (#7384) * Bump to latest Go 1.22 patch (#7379) * Fix deepseek deseret regex (#7369) * Better support for AMD multi-GPU on linux (#7212) * Fix unicode output on windows with redirect to file (#7358) * Fix incremental build file deps (#7361) * Improve dependency gathering logic (#7345) * fix #7247 - invalid image input (#7249) * integration: harden embedding test (#7306) * default to "FROM ." if a Modelfile isn't present (#7250) * Fix rocm windows build and clean up dependency gathering (#7305) * runner.go: Merge partial unicode characters before sending * readme: add Ollama for Swift to the community integrations (#7295) * server: allow vscode-webview origin (#7273) * image processing for llama3.2 (#6963) * llama: Decouple patching script from submodule (#7139) * llama: add compiler tags for cpu features (#7137) Eyad Issa 2024-11-01 02:20:51 +00:00
  • 3929e3a762 Accepting request 1219752 from science:machinelearning Dominique Leuenberger 2024-10-31 15:09:45 +00:00
  • 78bc1eed9a Accepting request 1219233 from home:cabelo:branches:science:machinelearning Eyad Issa 2024-10-31 01:55:17 +00:00
  • b2e945fdb7 Accepting request 1207827 from science:machinelearning leap-16.1 leap-16.0 Dominique Leuenberger 2024-10-14 11:07:54 +00:00
  • 9ac289025f Accepting request 1207827 from science:machinelearning Dominique Leuenberger 2024-10-14 11:07:54 +00:00
  • 332862e2b8 - Update to version 0.3.13: * New safety models: ~ Llama Guard 3: a series of models by Meta, fine-tuned for content safety classification of LLM inputs and responses. ~ ShieldGemma: ShieldGemma is set of instruction tuned models from Google DeepMind for evaluating the safety of text prompt input and text output responses against a set of defined safety policies. * Fixed issue where ollama pull would leave connections when encountering an error * ollama rm will now stop a model if it is running prior to deleting it Guillaume GARDET 2024-10-14 07:28:18 +00:00
  • 374f7878c3 Accepting request 1207484 from home:VaiTon:branches:science:machinelearning Guillaume GARDET 2024-10-14 07:28:18 +00:00
  • bd7fc28fe4 Accepting request 1204591 from science:machinelearning Ana Guerrero 2024-09-30 13:40:27 +00:00
  • 56b72bf587 Accepting request 1204591 from science:machinelearning Ana Guerrero 2024-09-30 13:40:27 +00:00
  • 2808304cf4 - Update to version 0.3.12: * Llama 3.2: Meta's Llama 3.2 goes small with 1B and 3B models. * Qwen 2.5 Coder: The latest series of Code-Specific Qwen models, with significant improvements in code generation, code reasoning, and code fixing. * Ollama now supports ARM Windows machines * Fixed rare issue where Ollama would report a missing .dll file on Windows * Fixed performance issue for Windows without GPUs Eyad Issa 2024-09-29 21:30:54 +00:00
  • e29af59180 Accepting request 1204394 from home:cabelo:branches:science:machinelearning Eyad Issa 2024-09-29 21:30:54 +00:00
  • f7aaf9b2af Accepting request 1202264 from science:machinelearning Ana Guerrero 2024-09-22 09:06:09 +00:00
  • edcd06fbe6 Accepting request 1202264 from science:machinelearning Ana Guerrero 2024-09-22 09:06:09 +00:00
  • 5bb20bbdee - Update to version 0.3.11: * llm: add solar pro (preview) (#6846) * server: add tool parsing support for nemotron-mini (#6849) * make patches git am-able * CI: dist directories no longer present (#6834) * CI: clean up naming, fix tagging latest (#6832) * CI: set platform build build_linux script to keep buildx happy (#6829) * readme: add Agents-Flex to community integrations (#6788) * fix typo in import docs (#6828) * readme: add vim-intelligence-bridge to Terminal section (#6818) * readme: add Obsidian Quiz Generator plugin to community integrations (#6789) * Fix incremental builds on linux (#6780) * Use GOARCH for build dirs (#6779) * Optimize container images for startup (#6547) * examples: updated requirements.txt for privategpt example * examples: polish loganalyzer example (#6744) * readme: add ollama_moe to community integrations (#6752) * runner: Flush pending responses before returning * add "stop" command (#6739) * refactor show ouput * readme: add QodeAssist to community integrations (#6754) * Verify permissions for AMD GPU (#6736) * add *_proxy for debugging * docs: update examples to use llama3.1 (#6718) * Quiet down dockers new lint warnings (#6716) * catch when model vocab size is set correctly (#6714) * readme: add crewAI to community integrations (#6699) * readme: add crewAI with mesop to community integrations Eyad Issa 2024-09-20 20:29:36 +00:00
  • b79ecf65ca Accepting request 1202133 from home:adrianSuSE:branches:science:machinelearning Eyad Issa 2024-09-20 20:29:36 +00:00
  • e5b1fec77c Accepting request 1201962 from science:machinelearning Ana Guerrero 2024-09-19 19:17:44 +00:00