Skip to content
OpenCatalogcurated by FLOSSK
AI & Machine Learning

MLC LLM

Universal deployment stack compiling models to Vulkan, Metal, CUDA, and WebGPU via TVM/Unity for phones, browsers, and servers.

Why it is included

Unique open angle for edge and WebGPU LLM inference beyond desktop CUDA defaults.

Best for

Teams shipping LLMs to mobile, WebGPU, or heterogeneous devices.

Strengths

  • Multi-backend compilation
  • WebGPU path
  • MLC ecosystem

Limitations

  • Compile pipeline learning curve

Good alternatives

llama.cpp · ONNX Runtime

Related tools