From 580d68900e4cee87e965cdd9224972690d04071923476b5a37bb31f7b635d765 Mon Sep 17 00:00:00 2001 From: Guillaume GARDET Date: Thu, 20 Jun 2024 13:47:15 +0000 Subject: [PATCH] =?UTF-8?q?-=20Update=20to=202024.2.0=20-=20More=20Gen=20A?= =?UTF-8?q?I=20coverage=20and=20framework=20integrations=20to=20minimize?= =?UTF-8?q?=20code=20=20=20changes=20=20=20*=20Llama=203=20optimizations?= =?UTF-8?q?=20for=20CPUs,=20built-in=20GPUs,=20and=20discrete=20=20=20=20?= =?UTF-8?q?=20GPUs=20for=20improved=20performance=20and=20efficient=20memo?= =?UTF-8?q?ry=20usage.=20=20=20*=20Support=20for=20Phi-3-mini,=20a=20famil?= =?UTF-8?q?y=20of=20AI=20models=20that=20leverages=20=20=20=20=20the=20pow?= =?UTF-8?q?er=20of=20small=20language=20models=20for=20faster,=20more=20ac?= =?UTF-8?q?curate=20=20=20=20=20and=20cost-effective=20text=20processing.?= =?UTF-8?q?=20=20=20*=20Python=20Custom=20Operation=20is=20now=20enabled?= =?UTF-8?q?=20in=20OpenVINO=20making=20it=20=20=20=20=20easier=20for=20Pyt?= =?UTF-8?q?hon=20developers=20to=20code=20their=20custom=20operations=20?= =?UTF-8?q?=20=20=20=20instead=20of=20using=20C++=20custom=20operations=20?= =?UTF-8?q?(also=20supported).=20=20=20=20=20Python=20Custom=20Operation?= =?UTF-8?q?=20empowers=20users=20to=20implement=20their=20own=20=20=20=20?= =?UTF-8?q?=20specialized=20operations=20into=20any=20model.=20=20=20*=20N?= =?UTF-8?q?otebooks=20expansion=20to=20ensure=20better=20coverage=20for=20?= =?UTF-8?q?new=20models.=20=20=20=20=20Noteworthy=20notebooks=20added:=20D?= =?UTF-8?q?ynamiCrafter,=20YOLOv10,=20Chatbot=20=20=20=20=20notebook=20wit?= =?UTF-8?q?h=20Phi-3,=20and=20QWEN2.=20-=20Broader=20Large=20Language=20Mo?= =?UTF-8?q?del=20(LLM)=20support=20and=20more=20model=20=20=20compression?= =?UTF-8?q?=20techniques.=20=20=20*=20GPTQ=20method=20for=204-bit=20weight?= =?UTF-8?q?=20compression=20added=20to=20NNCF=20for=20=20=20=20=20more=20e?= =?UTF-8?q?fficient=20inference=20and=20improved=20performance=20of=20=20?= =?UTF-8?q?=20=20=20compressed=20LLMs.=20=20=20*=20Significant=20LLM=20per?= =?UTF-8?q?formance=20improvements=20and=20reduced=20latency=20=20=20=20?= =?UTF-8?q?=20for=20both=20built-in=20GPUs=20and=20discrete=20GPUs.=20=20?= =?UTF-8?q?=20*=20Significant=20improvement=20in=202nd=20token=20latency?= =?UTF-8?q?=20and=20memory=20=20=20=20=20footprint=20of=20FP16=20weight=20?= =?UTF-8?q?LLMs=20on=20AVX2=20(13th=20Gen=20Intel=C2=AE=20Core=E2=84=A2=20?= =?UTF-8?q?=20=20=20=20processors)=20and=20AVX512=20(3rd=20Gen=20Intel?= =?UTF-8?q?=C2=AE=20Xeon=C2=AE=20Scalable=20=20=20=20=20Processors)=20base?= =?UTF-8?q?d=20CPU=20platforms,=20particularly=20for=20small=20=20=20=20?= =?UTF-8?q?=20batch=20sizes.=20-=20More=20portability=20and=20performance?= =?UTF-8?q?=20to=20run=20AI=20at=20the=20edge,=20in=20the=20=20=20cloud,?= =?UTF-8?q?=20or=20locally.=20=20=20*=20Model=20Serving=20Enhancements:=20?= =?UTF-8?q?=20=20*=20Preview:=20OpenVINO=20Model=20Server=20(OVMS)=20now?= =?UTF-8?q?=20supports=20=20=20=20=20OpenAI-compatible=20API=20along=20wit?= =?UTF-8?q?h=20Continuous=20Batching=20and=20=20=20=20=20PagedAttention,?= =?UTF-8?q?=20enabling=20significantly=20higher=20throughput=20=20=20=20?= =?UTF-8?q?=20for=20parallel=20inferencing,=20especially=20on=20Intel?= =?UTF-8?q?=C2=AE=20Xeon=C2=AE=20=20=20=20=20processors,=20when=20serving?= =?UTF-8?q?=20LLMs=20to=20many=20concurrent=20users.=20=20=20*=20OpenVINO?= =?UTF-8?q?=20backend=20for=20Triton=20Server=20now=20supports=20built-in?= =?UTF-8?q?=20=20=20=20=20GPUs=20and=20discrete=20GPUs,=20in=20addition=20?= =?UTF-8?q?to=20dynamic=20=20=20=20=20shapes=20support.=20=20=20*=20Integr?= =?UTF-8?q?ation=20of=20TorchServe=20through=20torch.compile=20OpenVINO=20?= =?UTF-8?q?=20=20=20=20backend=20for=20easy=20model=20deployment,=20provis?= =?UTF-8?q?ioning=20to=20=20=20=20=20multiple=20instances,=20model=20versi?= =?UTF-8?q?oning,=20and=20maintenance.=20=20=20*=20Preview:=20addition=20o?= =?UTF-8?q?f=20the=20Generate=20API,=20a=20simplified=20API=20=20=20=20=20?= =?UTF-8?q?for=20text=20generation=20using=20large=20language=20models=20w?= =?UTF-8?q?ith=20only=20=20=20=20=20a=20few=20lines=20of=20code.=20The=20A?= =?UTF-8?q?PI=20is=20available=20through=20the=20newly=20=20=20=20=20launc?= =?UTF-8?q?hed=20OpenVINO=20GenAI=20package.=20=20=20*=20Support=20for=20I?= =?UTF-8?q?ntel=20Atom=C2=AE=20Processor=20X=20Series.=20For=20more=20deta?= =?UTF-8?q?ils,=20=20=20=20=20see=20System=20Requirements.=20=20=20*=20Pre?= =?UTF-8?q?view:=20Support=20for=20Intel=C2=AE=20Xeon=C2=AE=206=20processo?= =?UTF-8?q?r.=20-=20Support=20Change=20and=20Deprecation=20Notices=20=20?= =?UTF-8?q?=20*=20Using=20deprecated=20features=20and=20components=20is=20?= =?UTF-8?q?not=20advised.=20=20=20=20=20They=20are=20available=20to=20enab?= =?UTF-8?q?le=20a=20smooth=20transition=20to=20new=20=20=20=20=20solutions?= =?UTF-8?q?=20and=20will=20be=20discontinued=20in=20the=20future.=20=20=20?= =?UTF-8?q?=20=20To=20keep=20using=20discontinued=20features,=20you=20will?= =?UTF-8?q?=20have=20to=20revert=20=20=20=20=20to=20the=20last=20LTS=20Ope?= =?UTF-8?q?nVINO=20version=20supporting=20them.=20For=20more=20=20=20=20?= =?UTF-8?q?=20details,=20refer=20to=20the=20OpenVINO=20Legacy=20Features?= =?UTF-8?q?=20and=20=20=20=20=20Components=20page.=20=20=20*=20Discontinue?= =?UTF-8?q?d=20in=202024.0:=20=20=20=20=20+=20Runtime=20components:=20=20?= =?UTF-8?q?=20=20=20=20=20-=20Intel=C2=AE=20Gaussian=20&=20Neural=20Accele?= =?UTF-8?q?rator=20(Intel=C2=AE=E2=80=AFGNA).=20=20=20=20=20=20=20=20=20Co?= =?UTF-8?q?nsider=20using=20the=20Neural=20Processing=20Unit=20(NPU)=20for?= =?UTF-8?q?=20=20=20=20=20=20=20=20=20low-powered=20systems=20like=20Intel?= =?UTF-8?q?=C2=AE=20Core=E2=84=A2=20Ultra=20or=2014th=20=20=20=20=20=20=20?= =?UTF-8?q?=20=20generation=20and=20beyond.=20=20=20=20=20=20=20-=20OpenVI?= =?UTF-8?q?NO=20C++/C/Python=201.0=20APIs=20(see=202023.3=20API=20=20=20?= =?UTF-8?q?=20=20=20=20=20=20transition=20guide=20for=20reference).=20=20?= =?UTF-8?q?=20=20=20=20=20-=20All=20ONNX=20Frontend=20legacy=20API=20(know?= =?UTF-8?q?n=20as=20ONNX=5FIMPORTER=5FAPI)=20=20=20=20=20=20=20-=20'Perfom?= =?UTF-8?q?anceMode.UNDEFINED'=20property=20as=20part=20of=20the=20=20=20?= =?UTF-8?q?=20=20=20=20=20=20OpenVINO=20Python=20API=20=20=20=20=20+=20Too?= =?UTF-8?q?ls:=20=20=20=20=20=20=20-=20Deployment=20Manager.=20See=20insta?= =?UTF-8?q?llation=20and=20deployment=20=20=20=20=20=20=20=20=20guides=20f?= =?UTF-8?q?or=20current=20distribution=20options.=20=20=20=20=20=20=20-=20?= =?UTF-8?q?Accuracy=20Checker.=20=20=20=20=20=20=20-=20Post-Training=20Opt?= =?UTF-8?q?imization=20Tool=20(POT).=E2=80=AFNeural=20Network=20=20=20=20?= =?UTF-8?q?=20=20=20=20=20Compression=20Framework=20(NNCF)=20should=20be?= =?UTF-8?q?=20used=20instead.=20=20=20=20=20=20=20-=20A=20Git=20patch?= =?UTF-8?q?=E2=80=AFfor=20NNCF=20integration=20with=E2=80=AF=20=20=20=20?= =?UTF-8?q?=20=20=20=20=20huggingface/transformers.=20The=20recommended=20?= =?UTF-8?q?approach=E2=80=AF=20=20=20=20=20=20=20=20=20is=20to=20use?= =?UTF-8?q?=E2=80=AFhuggingface/optimum-intel=E2=80=AFfor=20applying=20NNC?= =?UTF-8?q?F=20=20=20=20=20=20=20=20=20optimization=20on=20top=20of=20mode?= =?UTF-8?q?ls=20from=20Hugging=20Face.=20=20=20=20=20=20=20-=20Support=20f?= =?UTF-8?q?or=20Apache=20MXNet,=20Caffe,=20and=20Kaldi=20model=20formats.?= =?UTF-8?q?=20=20=20=20=20=20=20=20=20Conversion=20to=20ONNX=20may=20be=20?= =?UTF-8?q?used=20as=20a=20solution.=20=20=20*=20Deprecated=20and=20to=20b?= =?UTF-8?q?e=20removed=20in=20the=20future:=20=20=20=20=20+=20The=20OpenVI?= =?UTF-8?q?NO=E2=84=A2=20Development=20Tools=20package=20(pip=20install=20?= =?UTF-8?q?=20=20=20=20=20=20openvino-dev)=20will=20be=20removed=20from=20?= =?UTF-8?q?installation=20options=20=20=20=20=20=20=20and=20distribution?= =?UTF-8?q?=20channels=20beginning=20with=20OpenVINO=202025.0.=20=20=20=20?= =?UTF-8?q?=20+=20Model=20Optimizer=20will=20be=20discontinued=20with=20Op?= =?UTF-8?q?enVINO=202025.0.=20=20=20=20=20=20=20Consider=20using=20the=20n?= =?UTF-8?q?ew=20conversion=20methods=20instead.=20For=20=20=20=20=20=20=20?= =?UTF-8?q?more=20details,=20see=20the=20model=20conversion=20transition?= =?UTF-8?q?=20guide.=20=20=20=20=20+=20OpenVINO=20property=20Affinity=20AP?= =?UTF-8?q?I=20will=20be=20discontinued=20with=20=20=20=20=20=20=20OpenVIN?= =?UTF-8?q?O=202025.0.=20It=20will=20be=20replaced=20with=20CPU=20binding?= =?UTF-8?q?=20=20=20=20=20=20=20configurations=20(ov::hint::enable=5Fcpu?= =?UTF-8?q?=5Fpinning).=20=20=20=20=20+=20OpenVINO=20Model=20Server=20comp?= =?UTF-8?q?onents:=20=20=20=20=20+=20=E2=80=9Cauto=20shape=E2=80=9D=20and?= =?UTF-8?q?=20=E2=80=9Cauto=20batch=20size=E2=80=9D=20(reshaping=20a=20mod?= =?UTF-8?q?el=20in=20=20=20=20=20=20=20runtime)=20will=20be=20removed=20in?= =?UTF-8?q?=20the=20future.=20OpenVINO=E2=80=99s=20dynamic=20=20=20=20=20?= =?UTF-8?q?=20=20shape=20models=20are=20recommended=20instead.=20=20=20=20?= =?UTF-8?q?=20+=20A=20number=20of=20notebooks=20have=20been=20deprecated.?= =?UTF-8?q?=20For=20an=20=20=20=20=20=20=20up-to-date=20listing=20of=20ava?= =?UTF-8?q?ilable=20notebooks,=20refer=20to=20the=20=20=20=20=20=20=20Open?= =?UTF-8?q?VINO=E2=84=A2=20Notebook=20index=20(openvinotoolkit.github.io).?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit OBS-URL: https://build.opensuse.org/package/show/science:machinelearning/openvino?expand=0&rev=9 --- _service | 4 +- openvino-2024.1.0.obscpio | 3 -- openvino-2024.2.0.obscpio | 3 ++ openvino.changes | 101 ++++++++++++++++++++++++++++++++++++++ openvino.obsinfo | 6 +-- openvino.spec | 57 ++++++++++++++------- 6 files changed, 147 insertions(+), 27 deletions(-) delete mode 100644 openvino-2024.1.0.obscpio create mode 100644 openvino-2024.2.0.obscpio diff --git a/_service b/_service index 4b0502a..98812ba 100644 --- a/_service +++ b/_service @@ -2,8 +2,8 @@ https://github.com/openvinotoolkit/openvino.git git - 2024.1.0 - 2024.1.0 + 2024.2.0 + 2024.2.0 enable openvino .git diff --git a/openvino-2024.1.0.obscpio b/openvino-2024.1.0.obscpio deleted file mode 100644 index 2b53a62..0000000 --- a/openvino-2024.1.0.obscpio +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:556da89cbc03dd30dad270a1c1796598932c488a696a87056399e74a4d688680 -size 865282063 diff --git a/openvino-2024.2.0.obscpio b/openvino-2024.2.0.obscpio new file mode 100644 index 0000000..ef6fbe9 --- /dev/null +++ b/openvino-2024.2.0.obscpio @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50b3efee39ea06430456d49db9b4173b22999d8b3e7547dc658bb37df82b0f1b +size 1036420623 diff --git a/openvino.changes b/openvino.changes index 9bf1e9e..0874d36 100644 --- a/openvino.changes +++ b/openvino.changes @@ -1,3 +1,104 @@ +------------------------------------------------------------------- +Wed Jun 19 21:36:01 UTC 2024 - Alessandro de Oliveira Faria + +- Update to 2024.2.0 +- More Gen AI coverage and framework integrations to minimize code + changes + * Llama 3 optimizations for CPUs, built-in GPUs, and discrete + GPUs for improved performance and efficient memory usage. + * Support for Phi-3-mini, a family of AI models that leverages + the power of small language models for faster, more accurate + and cost-effective text processing. + * Python Custom Operation is now enabled in OpenVINO making it + easier for Python developers to code their custom operations + instead of using C++ custom operations (also supported). + Python Custom Operation empowers users to implement their own + specialized operations into any model. + * Notebooks expansion to ensure better coverage for new models. + Noteworthy notebooks added: DynamiCrafter, YOLOv10, Chatbot + notebook with Phi-3, and QWEN2. +- Broader Large Language Model (LLM) support and more model + compression techniques. + * GPTQ method for 4-bit weight compression added to NNCF for + more efficient inference and improved performance of + compressed LLMs. + * Significant LLM performance improvements and reduced latency + for both built-in GPUs and discrete GPUs. + * Significant improvement in 2nd token latency and memory + footprint of FP16 weight LLMs on AVX2 (13th Gen Intel® Core™ + processors) and AVX512 (3rd Gen Intel® Xeon® Scalable + Processors) based CPU platforms, particularly for small + batch sizes. +- More portability and performance to run AI at the edge, in the + cloud, or locally. + * Model Serving Enhancements: + * Preview: OpenVINO Model Server (OVMS) now supports + OpenAI-compatible API along with Continuous Batching and + PagedAttention, enabling significantly higher throughput + for parallel inferencing, especially on Intel® Xeon® + processors, when serving LLMs to many concurrent users. + * OpenVINO backend for Triton Server now supports built-in + GPUs and discrete GPUs, in addition to dynamic + shapes support. + * Integration of TorchServe through torch.compile OpenVINO + backend for easy model deployment, provisioning to + multiple instances, model versioning, and maintenance. + * Preview: addition of the Generate API, a simplified API + for text generation using large language models with only + a few lines of code. The API is available through the newly + launched OpenVINO GenAI package. + * Support for Intel Atom® Processor X Series. For more details, + see System Requirements. + * Preview: Support for Intel® Xeon® 6 processor. +- Support Change and Deprecation Notices + * Using deprecated features and components is not advised. + They are available to enable a smooth transition to new + solutions and will be discontinued in the future. + To keep using discontinued features, you will have to revert + to the last LTS OpenVINO version supporting them. For more + details, refer to the OpenVINO Legacy Features and + Components page. + * Discontinued in 2024.0: + + Runtime components: + - Intel® Gaussian & Neural Accelerator (Intel® GNA). + Consider using the Neural Processing Unit (NPU) for + low-powered systems like Intel® Core™ Ultra or 14th + generation and beyond. + - OpenVINO C++/C/Python 1.0 APIs (see 2023.3 API + transition guide for reference). + - All ONNX Frontend legacy API (known as ONNX_IMPORTER_API) + - 'PerfomanceMode.UNDEFINED' property as part of the + OpenVINO Python API + + Tools: + - Deployment Manager. See installation and deployment + guides for current distribution options. + - Accuracy Checker. + - Post-Training Optimization Tool (POT). Neural Network + Compression Framework (NNCF) should be used instead. + - A Git patch for NNCF integration with  + huggingface/transformers. The recommended approach  + is to use huggingface/optimum-intel for applying NNCF + optimization on top of models from Hugging Face. + - Support for Apache MXNet, Caffe, and Kaldi model formats. + Conversion to ONNX may be used as a solution. + * Deprecated and to be removed in the future: + + The OpenVINO™ Development Tools package (pip install + openvino-dev) will be removed from installation options + and distribution channels beginning with OpenVINO 2025.0. + + Model Optimizer will be discontinued with OpenVINO 2025.0. + Consider using the new conversion methods instead. For + more details, see the model conversion transition guide. + + OpenVINO property Affinity API will be discontinued with + OpenVINO 2025.0. It will be replaced with CPU binding + configurations (ov::hint::enable_cpu_pinning). + + OpenVINO Model Server components: + + “auto shape” and “auto batch size” (reshaping a model in + runtime) will be removed in the future. OpenVINO’s dynamic + shape models are recommended instead. + + A number of notebooks have been deprecated. For an + up-to-date listing of available notebooks, refer to the + OpenVINO™ Notebook index (openvinotoolkit.github.io). + ------------------------------------------------------------------- Thu May 9 22:56:53 UTC 2024 - Alessandro de Oliveira Faria diff --git a/openvino.obsinfo b/openvino.obsinfo index 3bfef0a..b9796a6 100644 --- a/openvino.obsinfo +++ b/openvino.obsinfo @@ -1,4 +1,4 @@ name: openvino -version: 2024.1.0 -mtime: 1713778234 -commit: f4afc983258bcb2592d999ed6700043fdb58ad78 +version: 2024.2.0 +mtime: 1716981469 +commit: 5c0f38f83f62fdabcdc980fa6dc3ed1ea16c8a05 diff --git a/openvino.spec b/openvino.spec index 4c2b4fd..643be5e 100644 --- a/openvino.spec +++ b/openvino.spec @@ -21,12 +21,12 @@ # Compilation takes ~1 hr on OBS for a single python, don't try all supported flavours %define pythons python3 %define __builder ninja -%define so_ver 2410 +%define so_ver 2420 %define shlib lib%{name}%{so_ver} %define shlib_c lib%{name}_c%{so_ver} %define prj_name OpenVINO Name: openvino -Version: 2024.1.0 +Version: 2024.2.0 Release: 0 Summary: A toolkit for optimizing and deploying AI inference # Let's be safe and put all third party licenses here, no matter that we use specific thirdparty libs or not @@ -83,9 +83,10 @@ OpenVINO is an open-source toolkit for optimizing and deploying AI inference. + + ## Main shared libs and devel pkg ## # - %package -n %{shlib} Summary: Shared library for OpenVINO toolkit @@ -96,8 +97,9 @@ This package provides the shared library for OpenVINO. -# + +# %package -n %{shlib_c} Summary: Shared C library for OpenVINO toolkit @@ -106,8 +108,9 @@ This package provides the C library for OpenVINO. -# + +# %package -n %{name}-devel Summary: Headers and sources for OpenVINO toolkit Requires: %{shlib_c} = %{version} @@ -140,9 +143,10 @@ OpenVINO. + + ## Plugins ## # - %package -n %{name}-arm-cpu-plugin Summary: Intel CPU plugin for OpenVINO toolkit @@ -153,8 +157,9 @@ This package provides the ARM CPU plugin for OpenVINO on %{arm64} archs. -# + +# %package -n %{name}-auto-plugin Summary: Auto / Multi software plugin for OpenVINO toolkit @@ -165,8 +170,9 @@ This package provides the Auto / Multi software plugin for OpenVINO. -# + +# %package -n %{name}-auto-batch-plugin Summary: Automatic batch software plugin for OpenVINO toolkit @@ -177,8 +183,9 @@ This package provides the automatic batch software plugin for OpenVINO. -# + +# %package -n %{name}-hetero-plugin Summary: Hetero frontend for Intel OpenVINO toolkit @@ -189,8 +196,9 @@ This package provides the hetero frontend for OpenVINO. -# + +# %package -n %{name}-intel-cpu-plugin Summary: Intel CPU plugin for OpenVINO toolkit @@ -201,8 +209,9 @@ This package provides the intel CPU plugin for OpenVINO for %{x86_64} archs. -# + +# %package -n %{name}-intel-npu-plugin Summary: Intel NPU plugin for OpenVINO toolkit @@ -213,9 +222,10 @@ This package provides the intel NPU plugin for OpenVINO for %{x86_64} archs. + + ## Frontend shared libs ## # - %package -n lib%{name}_ir_frontend%{so_ver} Summary: Paddle frontend for Intel OpenVINO toolkit @@ -226,8 +236,9 @@ This package provides the ir frontend for OpenVINO. -# + +# %package -n lib%{name}_onnx_frontend%{so_ver} Summary: Onnx frontend for OpenVINO toolkit @@ -238,8 +249,9 @@ This package provides the onnx frontend for OpenVINO. -# + +# %package -n lib%{name}_paddle_frontend%{so_ver} Summary: Paddle frontend for Intel OpenVINO toolkit @@ -250,8 +262,9 @@ This package provides the paddle frontend for OpenVINO. -# + +# %package -n lib%{name}_pytorch_frontend%{so_ver} Summary: PyTorch frontend for OpenVINO toolkit @@ -262,8 +275,9 @@ This package provides the pytorch frontend for OpenVINO. -# + +# %package -n lib%{name}_tensorflow_frontend%{so_ver} Summary: TensorFlow frontend for OpenVINO toolkit @@ -274,8 +288,9 @@ This package provides the tensorflow frontend for OpenVINO. -# + +# %package -n lib%{name}_tensorflow_lite_frontend%{so_ver} Summary: TensorFlow Lite frontend for OpenVINO toolkit @@ -286,9 +301,10 @@ This package provides the tensorflow-lite frontend for OpenVINO. + + ## Python module ## # - %package -n python-openvino Summary: Python module for openVINO toolkit Requires: python-numpy < 2 @@ -302,9 +318,10 @@ This package provides a Python module for interfacing with openVINO toolkit. + + ## Samples/examples ## # - %package -n %{name}-sample Summary: Samples for use with OpenVINO toolkit BuildArch: noarch @@ -316,7 +333,9 @@ This package provides some samples for use with openVINO. + # + %prep %autosetup -p1