* Preview: Mixture of Experts (MoE) models optimized for CPUs and GPUs, validated for GPT-OSS 20B model. How to convert model: optimum-cli export openvino -m "openai/gpt-oss-20b" out_dir --weight-format int4 * Fixed issue ID 174531: Accuracy regression of Mistral-7b-instruct-v0.2 and Mistral-7b-instruct-v0.3 on all devices when executed with OpenVINO GenAI. As a workaround, use the IR converted with OpenVINO 2025.3. * Fixed issue ID 176777: Using the callback parameter with the Python API call generate() in Text2ImagePipeline, Image2ImagePipeline, InpaintingPipeline may cause the process to hang. As a workaround, do not use the callback parameter. C++ implementations was not affected. * Resolved an issue in the NPU plugin where the Level Zero (L0) context was implemented as a static global object and only destroyed during DLL unload, even after unload_plugin() was called. This behavior prevented the driver from spawning threads required for certain optimizations and features. OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/openvino?expand=0&rev=47
17 lines
585 B
Plaintext
17 lines
585 B
Plaintext
<services>
|
|
<service name="obs_scm" mode="manual">
|
|
<param name="url">https://github.com/openvinotoolkit/openvino.git</param>
|
|
<param name="scm">git</param>
|
|
<param name="revision">2025.4.1</param>
|
|
<param name="version">2025.4.1</param>
|
|
<param name="submodules">enable</param>
|
|
<param name="filename">openvino</param>
|
|
<param name="exclude">.git</param>
|
|
</service>
|
|
<service name="tar" mode="buildtime" />
|
|
<service name="recompress" mode="buildtime">
|
|
<param name="file">*.tar</param>
|
|
<param name="compression">zstd</param>
|
|
</service>
|
|
</services>
|