Open LLM Leaderboard
Track, rank and evaluate open LLMs and chatbots
Excellent SLM (small language models) and SVLM (small vison language models).
Track, rank and evaluate open LLMs and chatbots
A Leaderboard that demonstrates LMM reasoning capabilities
Note 0.5B Size. Qwen2.5 Technical Report, https://huggingface.co/papers/2412.15115
Note 1.5B Size. Qwen2.5 Technical Report, https://huggingface.co/papers/2412.15115
Note 3B Size. Qwen2.5 Technical Report, https://huggingface.co/papers/2412.15115
Note EXAONE Deep Released ━ Setting a New Standard for Reasoning AI, https://www.lgresearch.ai/news/view?seq=543
Note Welcome Gemma 3: Google's all new multimodal, multilingual, long context open LLM, https://huggingface.co/blog/gemma3 Gemma 3 Technical Report, https://huggingface.co/papers/2503.19786
Note Empowering innovation: The next generation of the Phi family,https://azure.microsoft.com/en-us/blog/empowering-innovation-the-next-generation-of-the-phi-family/ Phi-4-Mini Technical Report: Compact yet Powerful Multimodal Language Models via Mixture-of-LoRAs, https://huggingface.co/papers/2503.01743
Note 135M Size. SmolLM2: When Smol Goes Big -- Data-Centric Training of a Small Language Model, https://huggingface.co/papers/2502.02737
Note 360M Size. SmolLM2: When Smol Goes Big -- Data-Centric Training of a Small Language Model, https://huggingface.co/papers/2502.02737
Note 1.7B Size. SmolLM2: When Smol Goes Big -- Data-Centric Training of a Small Language Model, https://huggingface.co/papers/2502.02737
Note Granite-3.2-2B-Instruct is an 2-billion-parameter, long-context AI model fine-tuned for thinking capabilities. Tutorials: https://www.ibm.com/granite/docs/ Cookbook: https://github.com/ibm-granite-community/granite-snack-cookbook/tree/main
Note TxGemma: Efficient and Agentic LLMs for Therapeutics paper: https://storage.googleapis.com/research-media/txgemma/txgemma-report.pdf GitHub repository (supporting code, Colab notebooks, discussions, and issues): https://github.com/google-gemini/gemma-cookbook/tree/main/TxGemma
VLMEvalKit Evaluation Results Collection
Note SmolDocling: An ultra-compact vision-language model for end-to-end multi-modal document conversion, https://huggingface.co/papers/2503.11576
Note Empowering innovation: The next generation of the Phi family,https://azure.microsoft.com/en-us/blog/empowering-innovation-the-next-generation-of-the-phi-family/ Phi-4-Mini Technical Report: Compact yet Powerful Multimodal Language Models via Mixture-of-LoRAs, https://huggingface.co/papers/2503.01743
Note Welcome Gemma 3: Google's all new multimodal, multilingual, long context open LLM, https://huggingface.co/blog/gemma3 Gemma 3 Technical Report, https://huggingface.co/papers/2503.19786
Note Safer and Multimodal: Responsible AI with Gemma, https://developers.googleblog.com/en/safer-and-multimodal-responsible-ai-with-gemma/
Note 256M Size. SmolVLM - small yet mighty Vision Language Model, https://huggingface.co/blog/smolvlm SmolVLM Grows Smaller – Introducing the 250M & 500M Models, https://huggingface.co/blog/smolervlm
Note 500M Size. SmolVLM - small yet mighty Vision Language Model, https://huggingface.co/blog/smolvlm SmolVLM Grows Smaller – Introducing the 250M & 500M Models, https://huggingface.co/blog/smolervlm
Note 2.2B SIze. SmolVLM - small yet mighty Vision Language Model, https://huggingface.co/blog/smolvlm
Note 256M Size. SmolVLM2: Bringing Video Understanding to Every Device, https://huggingface.co/blog/smolvlm2
Note 500M Size. SmolVLM2: Bringing Video Understanding to Every Device, https://huggingface.co/blog/smolvlm2
Note 2.2B Size. SmolVLM2: Bringing Video Understanding to Every Device, https://huggingface.co/blog/smolvlm2
Note Qwen2.5-VL Technical Report, https://huggingface.co/papers/2502.13923 Grounding. https://qwenlm.github.io/blog/qwen2.5-vl/
Note Granite-vision-3.2-2b is a compact and efficient vision-language model, specifically designed for visual document understanding. Tutorials: https://www.ibm.com/granite/docs/models/vision/ Paper: https://arxiv.org/abs/2502.09927
Note NVIDIA Isaac GR00T N1 is the world's first open foundation model for generalized humanoid robot reasoning and skills. https://github.com/NVIDIA/Isaac-GR00T/
Note Moondream is a small vision language model designed to run efficiently on edge devices. Blog: https://moondream.ai/blog/moondream-2025-03-27-release Github: https://github.com/vikhyat/moondream