Using LLaMA.cpp
Step 1 - Clone the repo
git clone https://github.com/ggerganov/llama.cpp
cd llama.cppStep 2 - Download the model
Step 3 - Build the server
make llama-serverStep 4 - Run the server
./llama-server -t 4 -c 4096 -ngl 35 -b 512 --mlock -m models/openchat_3.5.Q5_K_M.ggufStep 5 - Enable the server in the client
settings -> ChatBot -> ChatBot Backend -> LLaMA.cppLast updated