最終更新:2024-04-10 (水) 13:24:12 (279d)
llama.cpp
Top / llama.cpp
Inference of LLaMA model in pure C/C++
https://github.com/ggerganov/llama.cpp
メモ
対応モデル
- LLaMA 🦙
- Llama 2 🦙🦙
- Falcon
- Alpaca
- GPT4All
- Chinese LLaMA? / Alpaca and Chinese LLaMA-2? / Alpaca-2?
- Vigogne? (French)
- Vicuna
- Koala?
- OpenBuddy 🐶 (Multilingual)
- Pygmalion?/Metharme?
- WizardLM
- Baichuan? 1 & 2 + derivations
- Aquila? 1 & 2
- Starcoder? models
- Mistral AI v0.1
- Refact?
- Persimmon 8B?
- MPT
- Bloom?
速度
https://github.com/ggerganov/llama.cpp/discussions/4167
https://www.reddit.com/r/LocalLLaMA/comments/16z3jps/comment/k3f78x7/