3
of the node-llama-cpp module to communicate with your local model.
node-llama-cpp
is tuned for running on a MacOS platform with support for the Metal GPU of Apple M-series of processors. If you need to turn this off or need support for the CUDA architecture then refer to the documentation at node-llama-cpp.
For advice on getting and preparing llama3
see the documentation for the LLM version of this module.
A note to LangChain.js contributors: if you want to run the tests associated with this module you will need to put the path to your local model in the environment variable LLAMA_PATH
.
llama_cpp
module a system message will cause the creation of a new session.
llama3
such as the 70B version.
invoke
method, we can also achieve stream generation, and use signal
to abort the generation.