diff --git a/_service b/_service index ad2edaa..78dd84c 100644 --- a/_service +++ b/_service @@ -3,9 +3,9 @@ https://github.com/ollama/ollama.git git - v0.4.0-rc6 + v0.4.0-rc8 @PARENT_TAG@ - v(.*)-rc6 + v(.*)-rc8 enable enable macapp diff --git a/_servicedata b/_servicedata index b86fe6c..d82d9b9 100644 --- a/_servicedata +++ b/_servicedata @@ -1,4 +1,4 @@ https://github.com/ollama/ollama.git - 16f4eabe2d409b2b8a6e50fa08c8ce3a2a3b18d1 \ No newline at end of file + 046054fa3bba6d6511bcf46ca53f3ee8bc972df6 \ No newline at end of file diff --git a/ollama-0.4.0.obscpio b/ollama-0.4.0.obscpio index ec540f0..915955c 100644 --- a/ollama-0.4.0.obscpio +++ b/ollama-0.4.0.obscpio @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:a170a1b1dad3a0414739389095d73562aa9e2038357f4fc1a4c5db344f836547 -size 16461325 +oid sha256:bcbcbb9aa1cdde96a51a5bbf25f1b6a3cb97d71ed7589a17e68f9a30287bd450 +size 16564237 diff --git a/ollama.changes b/ollama.changes index d0f3512..730e32f 100644 --- a/ollama.changes +++ b/ollama.changes @@ -1,5 +1,23 @@ ------------------------------------------------------------------- -Fri Nov 01 02:18:50 UTC 2024 - eyadlorenzo@gmail.com +Wed Nov 06 12:31:53 UTC 2024 - Eyad Issa + +- Update to version 0.4.0-rc8: + * CI: Switch to v13 macos runner (#7498) + * CI: matrix strategy fix (#7496) + * Sign windows arm64 official binaries (#7493) + * readme: add TextCraft to community integrations (#7377) + * nvidia libs have inconsistent ordering (#7473) + * CI: omit unused tools for faster release builds (#7432) + * llama: Improve error handling + * runner.go: Only allocate 1 element embedding batches for mllama + * refactor kv estimation + * mllama cross attention + * Add basic mllama integration tests (#7455) + * runner.go: Don't set cross attention before sending embeddings + * Give unicode test more time to run (#7437) + +------------------------------------------------------------------- +Fri Nov 01 02:18:50 UTC 2024 - Eyad Issa - Remove enable-lto.patch @@ -37,12 +55,12 @@ Wed Oct 30 01:47:37 UTC 2024 - Alessandro de Oliveira Faria - Update to version 0.3.12: - * Llama 3.2: Meta's Llama 3.2 goes small with 1B and 3B + * Llama 3.2: Meta's Llama 3.2 goes small with 1B and 3B models. - * Qwen 2.5 Coder: The latest series of Code-Specific Qwen - models, with significant improvements in code generation, + * Qwen 2.5 Coder: The latest series of Code-Specific Qwen + models, with significant improvements in code generation, code reasoning, and code fixing. * Ollama now supports ARM Windows machines * Fixed rare issue where Ollama would report a missing .dll @@ -241,23 +259,23 @@ Sun Aug 11 02:40:06 UTC 2024 - Alessandro de Oliveira Faria @@ -343,16 +361,16 @@ Wed Jul 24 14:28:08 UTC 2024 - adrian@suse.de ------------------------------------------------------------------- Thu Jul 18 13:09:10 UTC 2024 - Eyad Issa -- Fixed issue with shared libraries +- Fixed issue with shared libraries ------------------------------------------------------------------- Thu Jul 18 12:27:54 UTC 2024 - Eyad Issa - Added %check section -- Use -v when building +- Use -v when building - Update to version 0.2.6: - * New models: MathΣtral is a 7B model designed for math + * New models: MathΣtral is a 7B model designed for math reasoning and scientific discovery by Mistral AI. * Fixed issue where uppercase roles such as USER would no longer work in the chat endpoints @@ -366,62 +384,62 @@ Sun Jul 14 17:48:36 UTC 2024 - eyadlorenzo@gmail.com * Fixed issue where a model's SYSTEM message not be applied - Update to version 0.2.4: - * Fixed issue where context, load_duration and total_duration + * Fixed issue where context, load_duration and total_duration fields would not be set in the /api/generate endpoint. - * Ollama will no longer error if loading models larger than + * Ollama will no longer error if loading models larger than system memory if disk space is available - Update to version 0.2.3: * Fix issue where system prompt would not be applied - Update to version 0.2.2: - * Fixed errors that occurred when using Ollama with Nvidia V100 + * Fixed errors that occurred when using Ollama with Nvidia V100 GPUs * glm4 models will no longer fail to load from out of memory errors - * Fixed error that would occur when running deepseek-v2 and + * Fixed error that would occur when running deepseek-v2 and deepseek-coder-v2 models * Fixed a series of out of memory issues when using Nvidia GPUs - * Fixed a series of errors that would occur when using multiple + * Fixed a series of errors that would occur when using multiple Radeon GPUs - Update to version 0.2.1: - * Fixed issue where setting OLLAMA_NUM_PARALLEL would cause + * Fixed issue where setting OLLAMA_NUM_PARALLEL would cause models to be reloaded after each request - Update to version 0.2.0: - * Ollama 0.2.0 is now available with concurrency support. + * Ollama 0.2.0 is now available with concurrency support. This unlocks 2 specific features: ~ Ollama can now serve multiple requests at the same time ~ Ollama now supports loading different models at the same time - * New models: GLM-4: A strong multi-lingual general language + * New models: GLM-4: A strong multi-lingual general language model with competitive performance to Llama 3. - * New models: CodeGeeX4: A versatile model for AI software + * New models: CodeGeeX4: A versatile model for AI software development scenarios, including code completion. - * New models: Gemma 2: Improved output quality and base text + * New models: Gemma 2: Improved output quality and base text generation models now available - * Ollama will now show a better error if a model architecture + * Ollama will now show a better error if a model architecture isn't supported * Improved handling of quotes and spaces in Modelfile FROM lines - * Ollama will now return an error if the system does not have + * Ollama will now return an error if the system does not have enough memory to run a model on Linux ------------------------------------------------------------------- Sun Jul 07 19:18:11 UTC 2024 - Eyad Issa - Update to version 0.1.48: - * Fixed issue where Gemma 2 would continuously output when + * Fixed issue where Gemma 2 would continuously output when reaching context limits * Fixed out of memory and core dump errors when running Gemma 2 * /show info will now show additional model information in ollama run - * Fixed issue where ollama show would result in an error on + * Fixed issue where ollama show would result in an error on certain vision models - Update to version 0.1.48: * Added support for Google Gemma 2 models (9B and 27B) * Fixed issues with ollama create when importing from Safetensors - + ------------------------------------------------------------------- Mon Jun 24 10:11:17 UTC 2024 - Eyad Issa @@ -456,44 +474,44 @@ Sat Jun 22 10:08:00 UTC 2024 - Eyad Issa ------------------------------------------------------------------- Tue Jun 18 12:12:41 UTC 2024 - Eyad Issa -- Added documentation files to .spec +- Added documentation files to .spec - Update to version 0.1.44: - * Fixed issue where unicode characters such as emojis would not + * Fixed issue where unicode characters such as emojis would not be loaded correctly when running ollama create * Fixed certain cases where Nvidia GPUs would not be detected and reported as compute capability 1.0 devices - Update to version 0.1.43: - * New import.md guide for converting and importing models to + * New import.md guide for converting and importing models to Ollama - * Fixed issue where embedding vectors resulting from + * Fixed issue where embedding vectors resulting from /api/embeddings would not be accurate - * JSON mode responses will no longer include invalid escape + * JSON mode responses will no longer include invalid escape characters - * Removing a model will no longer show incorrect File not found + * Removing a model will no longer show incorrect File not found errors - * Fixed issue where running ollama create would result in an + * Fixed issue where running ollama create would result in an error on Windows with certain file formatting - Update to version 0.1.42: - * New models: Qwen 2: a new series of large language models + * New models: Qwen 2: a new series of large language models from Alibaba group - * Qwen 2: a new series of large language models from Alibaba + * Qwen 2: a new series of large language models from Alibaba group - * ollama pull is now faster if it detects a model is already + * ollama pull is now faster if it detects a model is already downloaded * ollama create will now automatically detect prompt templates - for popular model architectures such as Llama, Gemma, Phi and + for popular model architectures such as Llama, Gemma, Phi and more. - * Ollama can now be accessed from local apps built with Electron + * Ollama can now be accessed from local apps built with Electron and Tauri, as well as in developing apps in local html files * Update welcome prompt in Windows to llama3 - * Fixed issues where /api/ps and /api/tags would show invalid + * Fixed issues where /api/ps and /api/tags would show invalid timestamps in responses - Update to version 0.1.41: - * Fixed issue on Windows 10 and 11 with Intel CPUs with + * Fixed issue on Windows 10 and 11 with Intel CPUs with integrated GPUs where Ollama would encounter an error ------------------------------------------------------------------- @@ -503,12 +521,12 @@ Sat Jun 01 21:12:20 UTC 2024 - Eyad Issa * New model: Codestral: Codestral is Mistral AI’s first-ever code model designed for code generation tasks. * New model: IBM Granite Code: now in 3B and 8B parameter sizes. - * New model: Deepseek V2: A Strong, Economical, and Efficient + * New model: Deepseek V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model - * Fixed out of memory and incorrect token issues when running + * Fixed out of memory and incorrect token issues when running Codestral on 16GB Macs - * Fixed issue where full-width characters (e.g. Japanese, - Chinese, Russian) were deleted at end of the line when using + * Fixed issue where full-width characters (e.g. Japanese, + Chinese, Russian) were deleted at end of the line when using ollama run ------------------------------------------------------------------- @@ -517,9 +535,9 @@ Wed May 29 11:38:26 UTC 2024 - Eyad Issa - Update to version 0.1.39: * New model: Cohere Aya 23: A new state-of-the-art, multilingual LLM covering 23 different languages. - * New model: Mistral 7B 0.3: A new version of Mistral 7B with + * New model: Mistral 7B 0.3: A new version of Mistral 7B with initial support for function calling. - * New model: Phi-3 Medium: a 14B parameters, lightweight, + * New model: Phi-3 Medium: a 14B parameters, lightweight, state-of-the-art open model by Microsoft. * New model: Phi-3 Mini 128K and Phi-3 Medium 128K: versions of the Phi-3 models that support a context window size of 128K @@ -527,7 +545,7 @@ Wed May 29 11:38:26 UTC 2024 - Eyad Issa IBM for Code Intelligence * It is now possible to import and quantize Llama 3 and its finetunes from Safetensors format to Ollama. - * Full changelog at + * Full changelog at https://github.com/ollama/ollama/releases/tag/v0.1.39 ------------------------------------------------------------------- @@ -541,7 +559,7 @@ Thu May 16 19:55:51 UTC 2024 - Eyad Issa - Update to version 0.1.38: * New model: Falcon 2: A new 11B parameters causal decoder-only model built by TII and trained over 5T tokens. - * New model: Yi 1.5: A new high-performing version of Yi, now + * New model: Yi 1.5: A new high-performing version of Yi, now licensed as Apache 2.0. Available in 6B, 9B and 34B sizes. * Added ollama ps command * Added /clear command @@ -566,7 +584,7 @@ Sun May 12 19:05:53 UTC 2024 - Eyad Issa Sun May 12 15:20:28 UTC 2024 - Eyad Issa - Use obs_scm service instead of the deprecated tar_scm -- Use zstd for vendor tarball compression +- Use zstd for vendor tarball compression ------------------------------------------------------------------- Sun May 12 01:39:26 UTC 2024 - Eyad Issa @@ -604,11 +622,11 @@ Sun May 12 01:39:26 UTC 2024 - Eyad Issa * New model: CodeGemma 1.1 * New model: StableLM2 12B * New model: Moondream 2 - * Fixed issues with LLaVa models where they would respond + * Fixed issues with LLaVa models where they would respond incorrectly after the first request - * Fixed out of memory errors when running large models such as + * Fixed out of memory errors when running large models such as Llama 3 70B - * Fixed various issues with Nvidia GPU discovery on Linux and + * Fixed various issues with Nvidia GPU discovery on Linux and Windows * Fixed a series of Modelfile errors when running ollama create * Fixed no slots available error that occurred when cancelling a @@ -626,13 +644,13 @@ Sun May 12 01:39:26 UTC 2024 - Eyad Issa * New model: Llama 3 Gradient 1048K * New model: Dolphin Llama 3 * New model: Qwen 110B - * Fixed issues where the model would not terminate, causing the + * Fixed issues where the model would not terminate, causing the API to hang. * Fixed a series of out of memory errors on Apple Silicon Macs - * Fixed out of memory errors when running Mixtral architecture + * Fixed out of memory errors when running Mixtral architecture models * Aded experimental concurrency features: - ~ OLLAMA_NUM_PARALLEL: Handle multiple requests simultaneously + ~ OLLAMA_NUM_PARALLEL: Handle multiple requests simultaneously for a single model ~ OLLAMA_MAX_LOADED_MODELS: Load multiple models simultaneously diff --git a/ollama.obsinfo b/ollama.obsinfo index 7096003..876df31 100644 --- a/ollama.obsinfo +++ b/ollama.obsinfo @@ -1,4 +1,4 @@ name: ollama version: 0.4.0 -mtime: 1730325945 -commit: 16f4eabe2d409b2b8a6e50fa08c8ce3a2a3b18d1 +mtime: 1730754127 +commit: 046054fa3bba6d6511bcf46ca53f3ee8bc972df6 diff --git a/vendor.tar.zstd b/vendor.tar.zstd index 1ed4711..25e89b2 100644 --- a/vendor.tar.zstd +++ b/vendor.tar.zstd @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:f66046626b5f525abb5373c347640f617eb7bd368d2a7f7f5b298a1db6e2b7b7 -size 5367921 +oid sha256:a465edc1e925c1c066e9a5923c6dd3b0534f3cef0ee0d32646b84d1347a126e6 +size 5367830