Kimi-VL-A3B Collection Moonshot's efficient MoE VLMs, exceptional on agent, long-context, and thinking • 6 items • Updated 9 days ago • 61
Gemma 3 QAT Collection Quantization Aware Trained (QAT) Gemma 3 checkpoints. The model preserves similar quality as half precision while using 3x less memory • 15 items • Updated 3 days ago • 147
Open-Qwen2VL: Compute-Efficient Pre-Training of Fully-Open Multimodal LLMs on Academic Resources Paper • 2504.00595 • Published 20 days ago • 34
ShieldGemma Collection ShieldGemma is a family of models for text and image content moderation. • 4 items • Updated 18 days ago • 6
view article Article Training and Finetuning Reranker Models with Sentence Transformers v4 26 days ago • 112
SmolDocling: An ultra-compact vision-language model for end-to-end multi-modal document conversion Paper • 2503.11576 • Published Mar 14 • 96
Vision Language Models Quantization Collection Vision Language Models (VLMs) quantized by Neural Magic • 20 items • Updated Mar 4 • 6
LLM2CLIP Collection LLM2CLIP makes SOTA pretrained CLIP modal more SOTA ever. • 11 items • Updated 3 days ago • 60
view article Article Welcome Gemma 3: Google's all new multimodal, multilingual, long context open LLM Mar 12 • 392
olmOCR Collection olmOCR is a document recognition pipeline for efficiently converting documents into plain text. olmocr.allenai.org • 4 items • Updated Mar 19 • 105
Phi-4 Collection Phi-4 family of small language and multi-modal models. • 9 items • Updated 3 days ago • 116
Ovis2 Collection Our latest advancement in multi-modal large language models (MLLMs) • 15 items • Updated 27 days ago • 59