Skip to content
OpenCatalogcurated by FLOSSK
AI & Machine Learning

llama.cpp

Plain C/C++ inference for LLaMA-class models with broad community backends.

Why it is included

Reference-quality local inference stack powering countless GUIs and servers.

Best for

Embedding LLMs into apps, edge devices, and research sandboxes.

If you use Windows, Mac, or paid tools

Local LLM alternative to OpenAI, Anthropic, and Google cloud APIs when you run models on your hardware.

Strengths

  • Performance focus
  • Quantization ecosystem
  • Hardware breadth

Limitations

  • You must comply with each model’s license

Good alternatives

Ollama · MLC

Related tools