Top pick
Local LLM runner and model library with simple CLI and API for workstation inference.
llmlocalinference
Filter by platform, license text, maturity, maintenance cadence, and editorial tags like privacy-focused or self-hosted. Search matches names, summaries, tags, and use cases.
5 tools match your filters
Local LLM runner and model library with simple CLI and API for workstation inference.
Plain C/C++ inference for LLaMA-class models with broad community backends.
Apple MLX-based LLM inference and training on Apple silicon: efficient Metal-backed transformers and examples for local chat models.
Single-file distributable LLM weights + llama.cpp runtime: run large models from one executable with broad OS CPU/GPU support.
Memory-efficient CUDA inference kernels for quantized Llama-class models—popular in consumer GPU chat UIs.