Skip to content
OpenCatalogcurated by FLOSSK
AI & Machine Learning

llamafile

Single-file distributable LLM weights + llama.cpp runtime: run large models from one executable with broad OS CPU/GPU support.

Why it is included

Mozilla-backed experiment in trivial distribution of open models for local users.

Best for

Demos, air-gapped USB sticks, and users who want zero pip installs.

Strengths

  • One binary
  • llama.cpp inside
  • Cosmopolitan libc story

Limitations

  • Large artifacts; not a cluster scheduler

Good alternatives

Ollama · llama.cpp

Related tools