Commit Graph

  • 6de51226b7 Accepting request 1225993 from science:machinelearning factory Ana Guerrero 2024-11-24 10:04:51 +0000
  • 414040b708 - Update to version 0.4.2: devel Eyad Issa 2024-11-16 16:22:34 +0000
  • 1e48ea9d8a - Add patch 01-build-verbose.patch to add the -v option to go build - Update to version 0.4.1: * runner.go: Check for zero length images * docs: update langchainpy.md with proper model name (#7527) * Set macos min version for all architectures (#7579) * win: remove preview title from installer (#7529) * Workaround buggy P2P ROCm copy on windows (#7466) * Debug logging for nvcuda init (#7532) * Align rocm compiler flags (#7467) * Be explicit for gpu library link dir (#7560) * docs: OLLAMA_NEW_RUNNERS no longer exists * runner.go: Remove unused arguments * sched: Lift parallel restriction for multimodal models except mllama Eyad Issa 2024-11-11 14:46:13 +0000
  • d62d34d2ac Accepting request 1222485 from science:machinelearning Dominique Leuenberger 2024-11-07 17:17:05 +0000
  • 65708a6764 - Update to version 0.4.0: * Update README.md (#7516) * One corrupt manifest should not wedge model operations (#7515) * prompt: Use a single token when estimating mllama context size * readme: add Hexabot to the list of community integrations * Quiet down debug log of image payload (#7454) Guillaume GARDET 2024-11-07 15:09:03 +0000
  • 7cc332278e - Update to version 0.4.0-rc8: Eyad Issa 2024-11-06 12:33:58 +0000
  • d5f830b976 - Remove enable-lto.patch Eyad Issa 2024-11-01 02:38:53 +0000
  • 70b43d49c1 OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=60 Eyad Issa 2024-11-01 02:23:50 +0000
  • da3e66a886 - Update to version 0.4.0-rc6: * Refine default thread selection for NUMA systems (#7322) * runner.go: Better abstract vision model integration * Soften windows clang requirement (#7428) * Remove submodule and shift to Go server - 0.4.0 (#7157) * Move windows app out of preview (#7347) * windows: Support alt install paths, fit and finish (#6967) * add more tests for getting the optimal tiled canvas (#7411) * Switch windows to clang (#7407) * tests: Add test for Unicode processing * runner.go: Better handle return NULL values from llama.cpp * add mllama image processing to the generate handler (#7384) * Bump to latest Go 1.22 patch (#7379) * Fix deepseek deseret regex (#7369) * Better support for AMD multi-GPU on linux (#7212) * Fix unicode output on windows with redirect to file (#7358) * Fix incremental build file deps (#7361) * Improve dependency gathering logic (#7345) * fix #7247 - invalid image input (#7249) * integration: harden embedding test (#7306) * default to "FROM ." if a Modelfile isn't present (#7250) * Fix rocm windows build and clean up dependency gathering (#7305) * runner.go: Merge partial unicode characters before sending * readme: add Ollama for Swift to the community integrations (#7295) * server: allow vscode-webview origin (#7273) * image processing for llama3.2 (#6963) * llama: Decouple patching script from submodule (#7139) * llama: add compiler tags for cpu features (#7137) Eyad Issa 2024-11-01 02:20:51 +0000
  • 785127e35d Accepting request 1219752 from science:machinelearning Dominique Leuenberger 2024-10-31 15:09:45 +0000
  • 5a882751e3 - Update to version 0.3.14: * New Models + Granite 3 MoE: The IBM Granite 1B and 3B models are the first mixture of experts (MoE) Granite models from IBM designed for low latency usage. + Granite 3 Dense: The IBM Granite 2B and 8B models are designed to support tool-based use cases and support for retrieval augmented generation (RAG), streamlining code generation, translation and bug fixing. Eyad Issa 2024-10-31 01:55:17 +0000
  • b2e945fdb7 Accepting request 1207827 from science:machinelearning Dominique Leuenberger 2024-10-14 11:07:54 +0000
  • 332862e2b8 - Update to version 0.3.13: * New safety models: ~ Llama Guard 3: a series of models by Meta, fine-tuned for content safety classification of LLM inputs and responses. ~ ShieldGemma: ShieldGemma is set of instruction tuned models from Google DeepMind for evaluating the safety of text prompt input and text output responses against a set of defined safety policies. * Fixed issue where ollama pull would leave connections when encountering an error * ollama rm will now stop a model if it is running prior to deleting it Guillaume GARDET 2024-10-14 07:28:18 +0000
  • bd7fc28fe4 Accepting request 1204591 from science:machinelearning Ana Guerrero 2024-09-30 13:40:27 +0000
  • 2808304cf4 - Update to version 0.3.12: * Llama 3.2: Meta's Llama 3.2 goes small with 1B and 3B models. * Qwen 2.5 Coder: The latest series of Code-Specific Qwen models, with significant improvements in code generation, code reasoning, and code fixing. * Ollama now supports ARM Windows machines * Fixed rare issue where Ollama would report a missing .dll file on Windows * Fixed performance issue for Windows without GPUs Eyad Issa 2024-09-29 21:30:54 +0000
  • f7aaf9b2af Accepting request 1202264 from science:machinelearning Ana Guerrero 2024-09-22 09:06:09 +0000
  • 5bb20bbdee - Update to version 0.3.11: * llm: add solar pro (preview) (#6846) * server: add tool parsing support for nemotron-mini (#6849) * make patches git am-able * CI: dist directories no longer present (#6834) * CI: clean up naming, fix tagging latest (#6832) * CI: set platform build build_linux script to keep buildx happy (#6829) * readme: add Agents-Flex to community integrations (#6788) * fix typo in import docs (#6828) * readme: add vim-intelligence-bridge to Terminal section (#6818) * readme: add Obsidian Quiz Generator plugin to community integrations (#6789) * Fix incremental builds on linux (#6780) * Use GOARCH for build dirs (#6779) * Optimize container images for startup (#6547) * examples: updated requirements.txt for privategpt example * examples: polish loganalyzer example (#6744) * readme: add ollama_moe to community integrations (#6752) * runner: Flush pending responses before returning * add "stop" command (#6739) * refactor show ouput * readme: add QodeAssist to community integrations (#6754) * Verify permissions for AMD GPU (#6736) * add *_proxy for debugging * docs: update examples to use llama3.1 (#6718) * Quiet down dockers new lint warnings (#6716) * catch when model vocab size is set correctly (#6714) * readme: add crewAI to community integrations (#6699) * readme: add crewAI with mesop to community integrations Eyad Issa 2024-09-20 20:29:36 +0000
  • e5b1fec77c Accepting request 1201962 from science:machinelearning Ana Guerrero 2024-09-19 19:17:44 +0000
  • c97461a42d - Update to version 0.3.10: * openai: align chat temperature and frequency_penalty options with completion (#6688) * docs: improve linux install documentation (#6683) * openai: don't scale temperature or frequency_penalty (#6514) * readme: add Archyve to community integrations (#6680) * readme: add Plasmoid Ollama Control to community integrations (#6681) * Improve logging on GPU too small (#6666) * openai: fix "presence_penalty" typo and add test (#6665) * Fix gemma2 2b conversion (#6645) * Document uninstall on windows (#6663) * Revert "Detect running in a container (#6495)" (#6662) * llm: make load time stall duration configurable via OLLAMA_LOAD_TIMEOUT * Introduce GPU Overhead env var (#5922) * Detect running in a container (#6495) * readme: add AiLama to the list of community integrations (#4957) * Update gpu.md: Add RTX 3050 Ti and RTX 3050 Ti (#5888) * server: fix blob download when receiving a 200 response (#6656) * readme: add Gentoo package manager entry to community integrations (#5714) * Update install.sh:Replace "command -v" with encapsulated functionality (#6035) * readme: include Enchanted for Apple Vision Pro (#4949) * readme: add lsp-ai to community integrations (#5063) * readme: add ollama-php library to community integrations (#6361) * readme: add vnc-lm discord bot community integration (#6644) * llm: use json.hpp from common (#6642) * readme: add confichat to community integrations (#6378) * docs: add group to manual Linux isntructions and verify service is running (#6430) * readme: add gollm to the list of community libraries (#6099) * readme: add Cherry Studio to community integrations (#6633) * readme: add Go fun package (#6421) * docs: fix spelling error (#6391) Eyad Issa 2024-09-19 08:48:38 +0000
  • e290a75d6e Accepting request 1194354 from science:machinelearning Dominique Leuenberger 2024-08-17 10:41:30 +0000
  • 5a2110e469 - Update to version 0.3.6: * Fixed issue where /api/embed would return an error instead of loading the model when the input field was not provided. * ollama create can now import Phi-3 models from Safetensors * Added progress information to ollama create when importing GGUF files * Ollama will now import GGUF files faster by minimizing file copies - Update to version 0.3.6: * Fixed issue where temporary files would not be cleaned up * Fix rare error when Ollama would start up due to invalid model data Eyad Issa 2024-08-15 19:06:50 +0000
  • aa82c484e7 - Update to version 0.3.4: * New embedding models - BGE-M3: a large embedding model from BAAI distinguished for its versatility in Multi-Functionality, Multi-Linguality, and Multi-Granularity. - BGE-Large: a large embedding model trained in english. - Paraphrase-Multilingual: A multilingual embedding model trained on parallel data for 50+ languages. * New embedding API with batch support - Ollama now supports a new API endpoint /api/embed for embedding generation: * This API endpoint supports new features: - Batches: generate embeddings for several documents in one request - Normalized embeddings: embeddings are now normalized, improving similarity results - Truncation: a new truncate parameter that will error if set to false - Metrics: responses include load_duration, total_duration and prompt_eval_count metrics Eyad Issa 2024-08-15 18:56:53 +0000
  • 20accf24c0 Accepting request 1191409 from science:machinelearning Dominique Leuenberger 2024-08-03 18:06:55 +0000
  • b0a216702e - Update to version 0.3.2: Eyad Issa 2024-08-03 10:59:29 +0000
  • f24370d51a - Update to version 0.3.1: Eyad Issa 2024-08-03 09:43:59 +0000
  • 9f8f3754a4 Accepting request 1190824 from science:machinelearning Dominique Leuenberger 2024-08-01 20:04:47 +0000
  • 708b172f7c Accepting request 1190611 from home:adrianSuSE Eyad Issa 2024-07-31 22:03:04 +0000
  • fd7709c23a Accepting request 1189982 from science:machinelearning Dominique Leuenberger 2024-07-28 15:20:22 +0000
  • 8b38454cf5 - Update to version 0.3.0: * Ollama now supports tool calling with popular models such as Llama 3.1. This enables a model to answer a given prompt using tool(s) it knows about, making it possible for models to perform more complex tasks or interact with the outside world. * New models: ~ Llama 3.1 ~ Mistral Large 2 ~ Firefunction v2 ~ Llama-3-Groq-Tool-Use * Fixed duplicate error message when running ollama create Eyad Issa 2024-07-28 11:46:59 +0000
  • 973749ceec Accepting request 1189591 from science:machinelearning Dominique Leuenberger 2024-07-26 14:15:22 +0000
  • 808a0b582d - Update to version 0.2.8: * api embed docs (#5282) * convert: capture head_dim for mistral (#5818) * Update llama.cpp submodule commit to d94c6e0c (#5805) * server: collect nested tool call objects when parsing (#5824) * Remove no longer supported max vram var * Refine error reporting for subprocess crash * Remove out of space test temporarily (#5825) * llm: consider head_dim in llama arch (#5817) * Adjust windows ROCm discovery * add patch for tekken (#5807) * preserve last assistant message (#5802) * Fix generate test flakyness (#5804) * server: validate template (#5734) * OpenAI: Function Based Testing (#5752) * adjust openai chat msg processing (#5729) * fix parsing tool calls * server: check for empty tools array too (#5779) * always provide content even if empty (#5778) * server: only parse tool calls if tools are provided (#5771) * Fix context exhaustion integration test for small gpus * Refine scheduler unit tests for reliability Eyad Issa 2024-07-25 11:03:50 +0000
  • e1464e1fa0 Accepting request 1188404 from science:machinelearning Ana Guerrero 2024-07-19 13:27:51 +0000
  • 44981711f9 - Fixed issue with shared libraries Eyad Issa 2024-07-18 13:09:42 +0000
  • 3e72c81bf1 - Added %check section - Use -v when building Eyad Issa 2024-07-18 12:28:24 +0000
  • 8d6b930083 - Update to version 0.2.6: * New models: MathΣtral is a 7B model designed for math reasoning and scientific discovery by Mistral AI. * Fixed issue where uppercase roles such as USER would no longer work in the chat endpoints * Fixed issue where empty system message would be included in the prompt Eyad Issa 2024-07-18 12:13:25 +0000
  • b2ca9b9e96 Accepting request 1187407 from science:machinelearning Ana Guerrero 2024-07-15 17:49:07 +0000
  • 3ddd383b3c - Update to version 0.2.5: - Update to version 0.2.4: - Update to version 0.2.3: - Update to version 0.2.2: - Update to version 0.2.1: - Update to version 0.2.0: Eyad Issa 2024-07-14 18:09:05 +0000
  • 1202fb05d0 Accepting request 1186033 from science:machinelearning Ana Guerrero 2024-07-08 17:08:25 +0000
  • 3eccb0320d - Update to version 0.1.48: * Fixed issue where Gemma 2 would continuously output when reaching context limits * Fixed out of memory and core dump errors when running Gemma 2 * /show info will now show additional model information in ollama run * Fixed issue where ollama show would result in an error on certain vision models - Update to version 0.1.48: * Added support for Google Gemma 2 models (9B and 27B) * Fixed issues with ollama create when importing from Safetensors Eyad Issa 2024-07-07 19:20:28 +0000
  • f2bb339091 Accepting request 1183991 from science:machinelearning Dominique Leuenberger 2024-07-01 09:19:34 +0000
  • 5b2fad2c49 - Update to version 0.1.45: Eyad Issa 2024-06-22 12:15:28 +0000
  • 1abab79606 Accepting request 1181491 from science:machinelearning Ana Guerrero 2024-06-18 20:52:16 +0000
  • d49ff0ffe7 - Added documentation files to .spec - Update to version 0.1.44 Eyad Issa 2024-06-18 12:14:31 +0000
  • 9f709033f9 Accepting request 1178089 from science:machinelearning Ana Guerrero 2024-06-03 15:43:01 +0000
  • f55037f0ec - Update to version 0.1.40: Eyad Issa 2024-06-01 21:20:45 +0000
  • 69d8d59dcd Accepting request 1175956 from science:machinelearning Ana Guerrero 2024-05-23 13:35:11 +0000
  • 45b2d8c6ae OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/ollama?expand=0&rev=21 Eyad Issa 2024-05-22 20:22:14 +0000
  • 87f951cc3a - Added 15.6 build - Added patch to build on RISC-V arch Eyad Issa 2024-05-22 18:08:19 +0000
  • bc9c8a7092 - Added 15.6 build Eyad Issa 2024-05-22 18:06:13 +0000
  • 71c9d43509 Accepting request 1175901 from home:mslacken:ml Eyad Issa 2024-05-22 17:56:10 +0000
  • 13d8dc2da7 Accepting request 1175546 from openSUSE:Factory:RISCV Eyad Issa 2024-05-21 11:57:06 +0000
  • b5d48c330c Accepting request 1174685 from science:machinelearning Ana Guerrero 2024-05-17 18:05:30 +0000
  • 32d8d25838 Accepting request 1174682 from home:VaiTon:branches:science:machinelearning Loren Burkholder 2024-05-16 20:34:11 +0000
  • 7d10104dd4 Accepting request 1173543 from science:machinelearning Ana Guerrero 2024-05-13 15:57:59 +0000
  • cf9cf2a4df Accepting request 1173521 from home:VaiTon:branches:science:machinelearning Loren Burkholder 2024-05-13 03:27:47 +0000
  • cc201b9d5d Accepting request 1173462 from science:machinelearning Ana Guerrero 2024-05-12 20:54:11 +0000
  • dfc1a9fa3a Accepting request 1173461 from home:VaiTon:branches:science:machinelearning Loren Burkholder 2024-05-12 01:58:52 +0000
  • c712805838 Accepting request 1169871 from science:machinelearning Ana Guerrero 2024-04-23 16:57:20 +0000
  • a4111a1692 Accepting request 1169791 from home:rrahl0:branches:science:machinelearning Loren Burkholder 2024-04-23 12:00:24 +0000
  • 10a34c1e5a Accepting request 1168439 from science:machinelearning Dominique Leuenberger 2024-04-17 12:45:50 +0000
  • 8ef2b26afe Accepting request 1168020 from home:bmwiedemann:branches:science:machinelearning Loren Burkholder 2024-04-17 00:53:52 +0000
  • 9c6d1dfa92 - Update to version 0.1.28: * Fix embeddings load model behavior (#2848) * Add Community Integration: NextChat (#2780) * prepend image tags (#2789) * fix: print usedMemory size right (#2827) * bump submodule to 87c91c07663b707e831c59ec373b5e665ff9d64a (#2828) * Add ollama user to video group * Add env var so podman will map cuda GPUs * Omit build date from gzip headers * Log unexpected server errors checking for update * Refine container image build script * Bump llama.cpp to b2276 * Determine max VRAM on macOS using recommendedMaxWorkingSetSize (#2354) * Update types.go (#2744) * Update langchain python tutorial (#2737) * no extra disk space for windows installation (#2739) * clean up go.mod * remove format/openssh.go * Add Community Integration: Chatbox * better directory cleanup in ollama.iss * restore windows build flags and compression Loren Burkholder 2024-03-06 23:53:38 +0000
  • 3e267704d5 Accepting request 1152310 from science:machinelearning Ana Guerrero 2024-02-27 21:49:45 +0000
  • e775b54f22 Accepting request 1152042 from home:jengelh:branches:science:machinelearning Loren Burkholder 2024-02-27 12:34:45 +0000
  • 14adc14d9d Remove the shadow dependency as it is not needed Loren Burkholder 2024-02-27 12:33:42 +0000
  • b61d3cac10 Apply some suggested changes to the user configuration Loren Burkholder 2024-02-27 03:54:15 +0000
  • 5a3ae9ab21 Accepting request 1150495 from home:LorenDB Guillaume GARDET 2024-02-26 09:11:49 +0000