OpenVINO Quick Start Guide
OpenVINO Quick Start Guide
TM
Bring AI everywhere with OpenVINO™: enabling developers to quickly optimize, deploy, and scale AI
applications across hardware device types with cutting-edge compression features and advanced performance
capabilities.
What is OpenVINO ?
TM
OpenVINO is an open-source toolkit for optimizing and deploying deep learning models. Deploy AI across
devices (from PC to cloud) with automatic acceleration!
AI acceleration by default
OpenVINO Installation
Linux install Windows install macOS install
PyPI example for Linux, macOS & Windows: #set up python venv
python -m pip install openvino
The install table also has: APT, YUM, Conda, vcpkg, Homebrew, Docker, Conan, & npm
Performance Features
OpenVINO can do automatic performance enhancements at runtime customized to your hardware (preserving
model accuracy), including:
Asynchronous execution, batch processing, tensor fusion, load balancing, dynamic inference parallelism,
automatic BF16 conversion, and more.
Automatic Batching ad-hoc groups inference re uests for max memory core utilizatio
q /
D ynamic hapes reshapes models to accept arbitrarily-sized inputs, for data exibilit
S fl
Supported ardware H
The plugin architecture of OpenVINO enables development and plug-independent inference solutions
dedicated to different devices. Learn more about the Plugin, OpenVINO Plugin Library, and how to build one
with CMake.
Additional community-supported plugins for Nvidia, ava and ust can be found here.
J R
OpenVINO can Accelerate as a Backend
If you want to stay in another framework API, OpenVINO provides accelerating backends:
import openvino.torch
#compile PyTorch model as usual with PyTorch
PyTorch
compiled_model = torch.compile(model, backend="openvino", options =
{"device" : "CPU"})
onnx_model = onnx.load("model.onnx")
ONNX Runtime onnx.save_model(onnx_model, 'saved_model.onnx’)
sess.set_providers([‘OpenVINOExecutionProvider’])
ov_llm=HuggingFacePipeline.from_model_id(…backend="openvino",
model_kwargs={"device":"CPU","ov_config": ov_config})
LangChain
ov_chain = prompt | ov_llm
print(ov_chain.invoke({"question":“what is neurobiology?”}))
Notices & Disclaimers: Intel does not control or audit third-party data. You should consult other sources to evaluate accuracy. © Intel
Corporation. Intel, the Intel logo, and other Intel marks are trademarks of Intel Corporation or its subsidiaries. Other names and brands may
be claimed as the property of others. Legal Notices and Disclaimers