Kimi-VL-A3B Collection Moonshot's efficient MoE VLMs, exceptional on agent, long-context, and thinking • 6 items • Updated 4 days ago • 58
Gemma 3 QAT Collection Quantization Aware Trained (QAT) Gemma 3 checkpoints. The model preserves similar quality as half precision while using 3x less memory • 8 items • Updated 13 days ago • 116
Open-Qwen2VL: Compute-Efficient Pre-Training of Fully-Open Multimodal LLMs on Academic Resources Paper • 2504.00595 • Published 15 days ago • 34
ShieldGemma Collection ShieldGemma is a family of models for text and image content moderation. • 4 items • Updated 13 days ago • 6
view article Article Training and Finetuning Reranker Models with Sentence Transformers v4 21 days ago • 110
SmolDocling: An ultra-compact vision-language model for end-to-end multi-modal document conversion Paper • 2503.11576 • Published Mar 14 • 94
Vision Language Models Quantization Collection Vision Language Models (VLMs) quantized by Neural Magic • 20 items • Updated Mar 4 • 6
LLM2CLIP Collection LLM2CLIP makes SOTA pretrained CLIP modal more SOTA ever. • 11 items • Updated Mar 12 • 60
view article Article Welcome Gemma 3: Google's all new multimodal, multilingual, long context open LLM Mar 12 • 387
olmOCR Collection olmOCR is a document recognition pipeline for efficiently converting documents into plain text. olmocr.allenai.org • 4 items • Updated 27 days ago • 104
Phi-4 Collection Phi-4 family of small language and multi-modal models. • 7 items • Updated Mar 3 • 115
Ovis2 Collection Our latest advancement in multi-modal large language models (MLLMs) • 15 items • Updated 22 days ago • 59
Sa2VA Model Zoo Collection Huggingace Model Zoo For Sa2VA: Marrying SAM2 with LLaVA for Dense Grounded Understanding of Images and Videos By Bytedance Seed CV Research • 4 items • Updated Feb 9 • 34