Axolotl
YAML-configured fine-tuning for LLMs: LoRA, QLoRA, FSDP, and many architectures on top of Hugging Face trainers.
Why it is included
Most-used open recipe runner for affordable adaptation of open weights on consumer GPUs.
Best for
ML engineers shipping domain adapters and instruction-tuned variants.
Strengths
- Config-driven
- LoRA/QLoRA
- Community configs
Limitations
- You must still comply with base model licenses
Good alternatives
Unsloth · LLaMA Factory
Related tools
AI & Machine Learning
Hugging Face Transformers
State-of-the-art pretrained models for PyTorch, TensorFlow, and JAX.
AI & Machine Learning
PyTorch
Deep learning framework with strong research-to-production paths.
AI & Machine Learning
Unsloth
Optimized fine-tuning library claiming 2× faster LoRA/QLoRA with less VRAM via custom kernels and Hugging Face compatibility.
AI & Machine Learning
PEFT
Parameter-efficient fine-tuning methods (LoRA, adapters, prompt tuning) integrated with Transformers models.
AI & Machine Learning
OLMo
Allen AI fully open LLM **pipeline**: weights, training code, data mixes, and evaluation—research transparency flagship.
AI & Machine Learning
GPT-NeoX
EleutherAI framework and 20B-class models for training large autoregressive LMs with 3D parallelism—Apache-2.0 training stack.
