IPEX-LLM is a PyTorch library for running LLM on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) with very low latency.This example goes over how to use LangChain to conduct embedding tasks with
ipex-llm
optimizations on Intel CPU. This would be helpful in applications such as RAG, document QA, etc.
sentence-transformers
.
Note For Windows users,--extra-index-url https://download.pytorch.org/whl/cpu
when installipex-llm
is not required.