Gemma 3 QAT Collection Quantization Aware Trained (QAT) Gemma 3 checkpoints. The model preserves similar quality as half precision while using 3x less memory • 8 items • Updated 2 days ago • 88
Nomic Embed Multimodal Collection Multimodal models allowing you to search over interleaved text, PDFs, charts, and images! • 14 items • Updated 1 day ago • 15
Qwen2.5-Omni Collection End-to-End Omni (text, audio, image, video, and natural speech interaction) model based Qwen2.5 • 3 items • Updated 10 days ago • 77
GME Models Collection General Multimodal Embedding Models Released by Tongyi Lab of Alibaba Group • 3 items • Updated Dec 24, 2024 • 6
Feature-Level Insights into Artificial Text Detection with Sparse Autoencoders Paper • 2503.03601 • Published Mar 5 • 226
C4AI Aya Vision Collection Aya Vision is a state-of-the-art family of vision models that brings multimodal capabilities to 23 languages. • 5 items • Updated Mar 4 • 68
Babel Collection Open Multilingual Large Language Models Serving Over 90% of Global Speakers • 7 items • Updated 3 days ago • 17
view article Article A Deepdive into Aya Vision: Advancing the Frontier of Multilingual Multimodality Mar 4 • 72
Ultravox v0.5 Collection Ultravox is a multimodal Speech LLM built around different pretrained LLMs (frozen) and the whisper-large-v3-turbo (fine-tuned) backbone. • 3 items • Updated Feb 10 • 11
Qwen2.5-1M Collection The long-context version of Qwen2.5, supporting 1M-token context lengths • 3 items • Updated Feb 26 • 111
Qwen2.5-VL Collection Vision-language model series based on Qwen2.5 • 11 items • Updated 5 days ago • 436