llama.cpp
Port of Facebook's LLaMA model in C/C++
About llama.cpp
Inference of LLaMA model in pure C/C++. The main goal is to run the model using 4-bit quantization on a MacBook.
Supported platforms:
- Mac OS
- Linux
- Windows (via CMake)
- Docker
Supported models:
- LLaMA 🦙
- Alpaca
- GPT4All
- Chinese LLaMA / Alpaca
- Vigogne (French)
- Vicuna