LlamaCpp
By default, NeoGPT uses LlamaCpp to connect to local large language models.
Simply run python main.py
in the command line to start NeoGPT. You can also use --model-type
flag to specifically use LlamaCpp.
By default, we use Mistral-7B-v0.2
model. You can change the model by using the following command:
Metal (M1) Support
Llama.cpp supports Apple silicon first-class citizen - optimized via ARM NEON, Accelerate and Metal frameworks. Use the FORCE_CMAKE=1
environment variable to force the use of cmake and install the pip package for the Metal support
Windows Notes
If you run into issues where it complains it can’t find ‘nmake’ ’?’ or CMAKE_C_COMPILER, you can extract w64devkit as mentioned in llama.cpp repo and add those manually to CMAKE_ARGS before running pip install: