最終更新:2025-01-24 (金) 10:42:38 (85d)
llama.cpp
Top / llama.cpp
Inference of LLaMA model in pure C/C++
https://github.com/ggerganov/llama.cpp
モデル
- llama.cpp requires the model to be stored in the GGUF file format.
- Models in other data formats can be converted to GGUF using the convert_*.py Python scripts in this repo.
対応モデル
- LLaMA 🦙
- Llama 2 🦙🦙
- Llama 3 🦙🦙🦙
- Mistral 7B
- Mixtral MoE?
- Falcon
- Alpaca
- GPT4All
- Chinese LLaMA? / Alpaca and Chinese LLaMA-2? / Alpaca-2?
- Vigogne? (French)
- Vicuna
- Koala?
- OpenBuddy 🐶 (Multilingual)
- Pygmalion?/Metharme?
- WizardLM
- Baichuan? 1 & 2 + derivations
- Aquila? 1 & 2
- Starcoder? models
- Mistral AI v0.1
- Refact?
- Persimmon 8B?
- MPT
- Bloom?
フロントエンド
バックエンド
速度
https://github.com/ggerganov/llama.cpp/discussions/4167
https://www.reddit.com/r/LocalLLaMA/comments/16z3jps/comment/k3f78x7/