Skip to content
OpenCatalogcurated by FLOSSK
AI & Machine Learning

OpenAI CLIP

Contrastive vision–language pretraining reference implementation: map images and text to a shared embedding space.

Why it is included

Landmark MIT-licensed research code featured in TAAFT’s machine-learning repository index.

Best for

Zero-shot classification, retrieval, and multimodal prototypes built on CLIP embeddings.

Strengths

  • Simple powerful baseline
  • Huge downstream use
  • Open weights story

Limitations

  • Newer VLMs often outperform on hard benchmarks

Good alternatives

OpenCLIP · SigLIP · LLaVA-style models

Related tools