OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference. The OpenVINO™ Runtime supports various hardware devices including x86 and ARM CPUs, and Intel GPUs. It can help to boost deep learning performance in Computer Vision, Automatic Speech Recognition, Natural Language Processing and other common tasks. Hugging Face embedding model can be supported by OpenVINO throughDocumentation Index
Fetch the complete documentation index at: https://docs.langchain.com/llms.txt
Use this file to discover all available pages before exploring further.
OpenVINOEmbeddings class. If you have an Intel GPU, you can specify model_kwargs={"device": "GPU"} to run inference on it.
Export IR model
It is possible to export your embedding model to the OpenVINO IR format withOVModelForFeatureExtraction, and load the model from local folder.
BGE with OpenVINO
We can also access BGE embedding models via theOpenVINOBgeEmbeddings class with OpenVINO.
- OpenVINO LLM guide.
- OpenVINO Documentation.
- OpenVINO Get Started Guide.
- RAG Notebook with LangChain.
Connect these docs to Claude, VSCode, and more via MCP for real-time answers.

