1e48ea9d8a- Add patch 01-build-verbose.patch to add the -v option to go build - Update to version 0.4.1: * runner.go: Check for zero length images * docs: update langchainpy.md with proper model name (#7527) * Set macos min version for all architectures (#7579) * win: remove preview title from installer (#7529) * Workaround buggy P2P ROCm copy on windows (#7466) * Debug logging for nvcuda init (#7532) * Align rocm compiler flags (#7467) * Be explicit for gpu library link dir (#7560) * docs: OLLAMA_NEW_RUNNERS no longer exists * runner.go: Remove unused arguments * sched: Lift parallel restriction for multimodal models except mllamaEyad Issa2024-11-11 14:46:13 +0000
d62d34d2acAccepting request 1222485 from science:machinelearning
Dominique Leuenberger
2024-11-07 17:17:05 +0000
65708a6764- Update to version 0.4.0: * Update README.md (#7516) * One corrupt manifest should not wedge model operations (#7515) * prompt: Use a single token when estimating mllama context size * readme: add Hexabot to the list of community integrations * Quiet down debug log of image payload (#7454)Guillaume GARDET2024-11-07 15:09:03 +0000
da3e66a886- Update to version 0.4.0-rc6: * Refine default thread selection for NUMA systems (#7322) * runner.go: Better abstract vision model integration * Soften windows clang requirement (#7428) * Remove submodule and shift to Go server - 0.4.0 (#7157) * Move windows app out of preview (#7347) * windows: Support alt install paths, fit and finish (#6967) * add more tests for getting the optimal tiled canvas (#7411) * Switch windows to clang (#7407) * tests: Add test for Unicode processing * runner.go: Better handle return NULL values from llama.cpp * add mllama image processing to the generate handler (#7384) * Bump to latest Go 1.22 patch (#7379) * Fix deepseek deseret regex (#7369) * Better support for AMD multi-GPU on linux (#7212) * Fix unicode output on windows with redirect to file (#7358) * Fix incremental build file deps (#7361) * Improve dependency gathering logic (#7345) * fix#7247 - invalid image input (#7249) * integration: harden embedding test (#7306) * default to "FROM ." if a Modelfile isn't present (#7250) * Fix rocm windows build and clean up dependency gathering (#7305) * runner.go: Merge partial unicode characters before sending * readme: add Ollama for Swift to the community integrations (#7295) * server: allow vscode-webview origin (#7273) * image processing for llama3.2 (#6963) * llama: Decouple patching script from submodule (#7139) * llama: add compiler tags for cpu features (#7137)Eyad Issa2024-11-01 02:20:51 +0000
785127e35dAccepting request 1219752 from science:machinelearning
Dominique Leuenberger
2024-10-31 15:09:45 +0000
5a882751e3- Update to version 0.3.14: * New Models + Granite 3 MoE: The IBM Granite 1B and 3B models are the first mixture of experts (MoE) Granite models from IBM designed for low latency usage. + Granite 3 Dense: The IBM Granite 2B and 8B models are designed to support tool-based use cases and support for retrieval augmented generation (RAG), streamlining code generation, translation and bug fixing.Eyad Issa2024-10-31 01:55:17 +0000
b2e945fdb7Accepting request 1207827 from science:machinelearning
Dominique Leuenberger
2024-10-14 11:07:54 +0000
332862e2b8- Update to version 0.3.13: * New safety models: ~ Llama Guard 3: a series of models by Meta, fine-tuned for content safety classification of LLM inputs and responses. ~ ShieldGemma: ShieldGemma is set of instruction tuned models from Google DeepMind for evaluating the safety of text prompt input and text output responses against a set of defined safety policies. * Fixed issue where ollama pull would leave connections when encountering an error * ollama rm will now stop a model if it is running prior to deleting itGuillaume GARDET2024-10-14 07:28:18 +0000
bd7fc28fe4Accepting request 1204591 from science:machinelearning
Ana Guerrero
2024-09-30 13:40:27 +0000
2808304cf4- Update to version 0.3.12: * Llama 3.2: Meta's Llama 3.2 goes small with 1B and 3B models. * Qwen 2.5 Coder: The latest series of Code-Specific Qwen models, with significant improvements in code generation, code reasoning, and code fixing. * Ollama now supports ARM Windows machines * Fixed rare issue where Ollama would report a missing .dll file on Windows * Fixed performance issue for Windows without GPUsEyad Issa2024-09-29 21:30:54 +0000
f7aaf9b2afAccepting request 1202264 from science:machinelearning
Ana Guerrero
2024-09-22 09:06:09 +0000
5bb20bbdee- Update to version 0.3.11: * llm: add solar pro (preview) (#6846) * server: add tool parsing support for nemotron-mini (#6849) * make patches git am-able * CI: dist directories no longer present (#6834) * CI: clean up naming, fix tagging latest (#6832) * CI: set platform build build_linux script to keep buildx happy (#6829) * readme: add Agents-Flex to community integrations (#6788) * fix typo in import docs (#6828) * readme: add vim-intelligence-bridge to Terminal section (#6818) * readme: add Obsidian Quiz Generator plugin to community integrations (#6789) * Fix incremental builds on linux (#6780) * Use GOARCH for build dirs (#6779) * Optimize container images for startup (#6547) * examples: updated requirements.txt for privategpt example * examples: polish loganalyzer example (#6744) * readme: add ollama_moe to community integrations (#6752) * runner: Flush pending responses before returning * add "stop" command (#6739) * refactor show ouput * readme: add QodeAssist to community integrations (#6754) * Verify permissions for AMD GPU (#6736) * add *_proxy for debugging * docs: update examples to use llama3.1 (#6718) * Quiet down dockers new lint warnings (#6716) * catch when model vocab size is set correctly (#6714) * readme: add crewAI to community integrations (#6699) * readme: add crewAI with mesop to community integrationsEyad Issa2024-09-20 20:29:36 +0000
e5b1fec77cAccepting request 1201962 from science:machinelearning
Ana Guerrero
2024-09-19 19:17:44 +0000
c97461a42d- Update to version 0.3.10: * openai: align chat temperature and frequency_penalty options with completion (#6688) * docs: improve linux install documentation (#6683) * openai: don't scale temperature or frequency_penalty (#6514) * readme: add Archyve to community integrations (#6680) * readme: add Plasmoid Ollama Control to community integrations (#6681) * Improve logging on GPU too small (#6666) * openai: fix "presence_penalty" typo and add test (#6665) * Fix gemma2 2b conversion (#6645) * Document uninstall on windows (#6663) * Revert "Detect running in a container (#6495)" (#6662) * llm: make load time stall duration configurable via OLLAMA_LOAD_TIMEOUT * Introduce GPU Overhead env var (#5922) * Detect running in a container (#6495) * readme: add AiLama to the list of community integrations (#4957) * Update gpu.md: Add RTX 3050 Ti and RTX 3050 Ti (#5888) * server: fix blob download when receiving a 200 response (#6656) * readme: add Gentoo package manager entry to community integrations (#5714) * Update install.sh:Replace "command -v" with encapsulated functionality (#6035) * readme: include Enchanted for Apple Vision Pro (#4949) * readme: add lsp-ai to community integrations (#5063) * readme: add ollama-php library to community integrations (#6361) * readme: add vnc-lm discord bot community integration (#6644) * llm: use json.hpp from common (#6642) * readme: add confichat to community integrations (#6378) * docs: add group to manual Linux isntructions and verify service is running (#6430) * readme: add gollm to the list of community libraries (#6099) * readme: add Cherry Studio to community integrations (#6633) * readme: add Go fun package (#6421) * docs: fix spelling error (#6391)Eyad Issa2024-09-19 08:48:38 +0000
e290a75d6eAccepting request 1194354 from science:machinelearning
Dominique Leuenberger
2024-08-17 10:41:30 +0000
5a2110e469- Update to version 0.3.6: * Fixed issue where /api/embed would return an error instead of loading the model when the input field was not provided. * ollama create can now import Phi-3 models from Safetensors * Added progress information to ollama create when importing GGUF files * Ollama will now import GGUF files faster by minimizing file copies - Update to version 0.3.6: * Fixed issue where temporary files would not be cleaned up * Fix rare error when Ollama would start up due to invalid model dataEyad Issa2024-08-15 19:06:50 +0000
aa82c484e7- Update to version 0.3.4: * New embedding models - BGE-M3: a large embedding model from BAAI distinguished for its versatility in Multi-Functionality, Multi-Linguality, and Multi-Granularity. - BGE-Large: a large embedding model trained in english. - Paraphrase-Multilingual: A multilingual embedding model trained on parallel data for 50+ languages. * New embedding API with batch support - Ollama now supports a new API endpoint /api/embed for embedding generation: * This API endpoint supports new features: - Batches: generate embeddings for several documents in one request - Normalized embeddings: embeddings are now normalized, improving similarity results - Truncation: a new truncate parameter that will error if set to false - Metrics: responses include load_duration, total_duration and prompt_eval_count metricsEyad Issa2024-08-15 18:56:53 +0000
20accf24c0Accepting request 1191409 from science:machinelearning
Dominique Leuenberger
2024-08-03 18:06:55 +0000
9f8f3754a4Accepting request 1190824 from science:machinelearning
Dominique Leuenberger
2024-08-01 20:04:47 +0000
708b172f7cAccepting request 1190611 from home:adrianSuSEEyad Issa2024-07-31 22:03:04 +0000
fd7709c23aAccepting request 1189982 from science:machinelearning
Dominique Leuenberger
2024-07-28 15:20:22 +0000
8b38454cf5- Update to version 0.3.0: * Ollama now supports tool calling with popular models such as Llama 3.1. This enables a model to answer a given prompt using tool(s) it knows about, making it possible for models to perform more complex tasks or interact with the outside world. * New models: ~ Llama 3.1 ~ Mistral Large 2 ~ Firefunction v2 ~ Llama-3-Groq-Tool-Use * Fixed duplicate error message when running ollama createEyad Issa2024-07-28 11:46:59 +0000
973749ceecAccepting request 1189591 from science:machinelearning
Dominique Leuenberger
2024-07-26 14:15:22 +0000
808a0b582d- Update to version 0.2.8: * api embed docs (#5282) * convert: capture head_dim for mistral (#5818) * Update llama.cpp submodule commit to d94c6e0c (#5805) * server: collect nested tool call objects when parsing (#5824) * Remove no longer supported max vram var * Refine error reporting for subprocess crash * Remove out of space test temporarily (#5825) * llm: consider head_dim in llama arch (#5817) * Adjust windows ROCm discovery * add patch for tekken (#5807) * preserve last assistant message (#5802) * Fix generate test flakyness (#5804) * server: validate template (#5734) * OpenAI: Function Based Testing (#5752) * adjust openai chat msg processing (#5729) * fix parsing tool calls * server: check for empty tools array too (#5779) * always provide content even if empty (#5778) * server: only parse tool calls if tools are provided (#5771) * Fix context exhaustion integration test for small gpus * Refine scheduler unit tests for reliabilityEyad Issa2024-07-25 11:03:50 +0000
e1464e1fa0Accepting request 1188404 from science:machinelearning
Ana Guerrero
2024-07-19 13:27:51 +0000
44981711f9- Fixed issue with shared librariesEyad Issa2024-07-18 13:09:42 +0000
3e72c81bf1- Added %check section - Use -v when buildingEyad Issa2024-07-18 12:28:24 +0000
8d6b930083- Update to version 0.2.6: * New models: MathΣtral is a 7B model designed for math reasoning and scientific discovery by Mistral AI. * Fixed issue where uppercase roles such as USER would no longer work in the chat endpoints * Fixed issue where empty system message would be included in the promptEyad Issa2024-07-18 12:13:25 +0000
b2ca9b9e96Accepting request 1187407 from science:machinelearning
Ana Guerrero
2024-07-15 17:49:07 +0000
3ddd383b3c- Update to version 0.2.5: - Update to version 0.2.4: - Update to version 0.2.3: - Update to version 0.2.2: - Update to version 0.2.1: - Update to version 0.2.0:Eyad Issa2024-07-14 18:09:05 +0000
1202fb05d0Accepting request 1186033 from science:machinelearning
Ana Guerrero
2024-07-08 17:08:25 +0000
3eccb0320d- Update to version 0.1.48: * Fixed issue where Gemma 2 would continuously output when reaching context limits * Fixed out of memory and core dump errors when running Gemma 2 * /show info will now show additional model information in ollama run * Fixed issue where ollama show would result in an error on certain vision models - Update to version 0.1.48: * Added support for Google Gemma 2 models (9B and 27B) * Fixed issues with ollama create when importing from SafetensorsEyad Issa2024-07-07 19:20:28 +0000
f2bb339091Accepting request 1183991 from science:machinelearning
Dominique Leuenberger
2024-07-01 09:19:34 +0000
71c9d43509Accepting request 1175901 from home:mslacken:mlEyad Issa2024-05-22 17:56:10 +0000
13d8dc2da7Accepting request 1175546 from openSUSE:Factory:RISCVEyad Issa2024-05-21 11:57:06 +0000
b5d48c330cAccepting request 1174685 from science:machinelearning
Ana Guerrero
2024-05-17 18:05:30 +0000
32d8d25838Accepting request 1174682 from home:VaiTon:branches:science:machinelearning
Loren Burkholder
2024-05-16 20:34:11 +0000
7d10104dd4Accepting request 1173543 from science:machinelearning
Ana Guerrero
2024-05-13 15:57:59 +0000
cf9cf2a4dfAccepting request 1173521 from home:VaiTon:branches:science:machinelearning
Loren Burkholder
2024-05-13 03:27:47 +0000
cc201b9d5dAccepting request 1173462 from science:machinelearning
Ana Guerrero
2024-05-12 20:54:11 +0000
dfc1a9fa3aAccepting request 1173461 from home:VaiTon:branches:science:machinelearning
Loren Burkholder
2024-05-12 01:58:52 +0000
c712805838Accepting request 1169871 from science:machinelearning
Ana Guerrero
2024-04-23 16:57:20 +0000
a4111a1692Accepting request 1169791 from home:rrahl0:branches:science:machinelearning
Loren Burkholder
2024-04-23 12:00:24 +0000
10a34c1e5aAccepting request 1168439 from science:machinelearning
Dominique Leuenberger
2024-04-17 12:45:50 +0000
8ef2b26afeAccepting request 1168020 from home:bmwiedemann:branches:science:machinelearning
Loren Burkholder
2024-04-17 00:53:52 +0000
9c6d1dfa92- Update to version 0.1.28: * Fix embeddings load model behavior (#2848) * Add Community Integration: NextChat (#2780) * prepend image tags (#2789) * fix: print usedMemory size right (#2827) * bump submodule to 87c91c07663b707e831c59ec373b5e665ff9d64a (#2828) * Add ollama user to video group * Add env var so podman will map cuda GPUs * Omit build date from gzip headers * Log unexpected server errors checking for update * Refine container image build script * Bump llama.cpp to b2276 * Determine max VRAM on macOS using recommendedMaxWorkingSetSize (#2354) * Update types.go (#2744) * Update langchain python tutorial (#2737) * no extra disk space for windows installation (#2739) * clean up go.mod * remove format/openssh.go * Add Community Integration: Chatbox * better directory cleanup in ollama.iss * restore windows build flags and compression
Loren Burkholder
2024-03-06 23:53:38 +0000
3e267704d5Accepting request 1152310 from science:machinelearning
Ana Guerrero
2024-02-27 21:49:45 +0000
e775b54f22Accepting request 1152042 from home:jengelh:branches:science:machinelearning
Loren Burkholder
2024-02-27 12:34:45 +0000
14adc14d9dRemove the shadow dependency as it is not needed
Loren Burkholder
2024-02-27 12:33:42 +0000
b61d3cac10Apply some suggested changes to the user configuration
Loren Burkholder
2024-02-27 03:54:15 +0000