Optimized fine-tuning library claiming 2× faster LoRA/QLoRA with less VRAM via custom kernels and Hugging Face compatibility.
Browse & filter
Filter by platform, license text, maturity, maintenance cadence, and editorial tags like privacy-focused or self-hosted. Search matches names, summaries, tags, and use cases.
5 tools match your filters
Also strong
llmfine-tuningloratrainingoptimization
Cross-platform inference accelerator for ONNX models: CPU, GPU, and mobile execution providers with graph optimizations.
inferenceonnxdeploymentoptimization
Also strong
Intel toolkit to optimize and deploy deep learning on Intel CPUs, GPUs, and NPUs with model conversion and runtime APIs.
inferenceinteledgeoptimization
Also strong
Automatic hyperparameter optimization framework with pruning, distributed search, and lightweight integration hooks.
hyperparameter-tuningautomlpythonoptimization
Also strong
CTranslate2 reimplementation of Whisper for faster CPU/GPU inference with lower memory use than reference PyTorch.
speechasrinferenceoptimization
