Kimi-K2 Collection Moonshot's MoE LLMs with 1 trillion parameters, exceptional on agentic intellegence • 2 items • Updated about 20 hours ago • 41
Jamba 1.7 Collection The AI21 Jamba family of models are hybrid SSM-Transformer foundation models, blending speed, efficient long context processing, and accuracy. • 4 items • Updated 10 days ago • 10
Vera 1.0 - Preview Collection Vera 1.0 - Preview (11B & 2.6B) sont des assistants IA polyvalents. Ils excellent dans les conversations fluides et l'assistance utilisateur. • 4 items • Updated 1 day ago • 1
WhiteRabbitNeo-V3 Collection The latest and most capable cybersecurity model we've ever created • 1 item • Updated 17 days ago • 2
OctoThinker-Llama-8B Family Collection What makes a base language model suitable for RL? Through controlled experiments, we identify key factors then leverage them to scale up mid-training. • 3 items • Updated 6 days ago • 2
Avey 1 Research Preview Collection 1.5B preview models trained on 100B tokens of FineWeb, and an instruct-tuned version on smoltalk. • 3 items • Updated 26 days ago • 6
sarvam-m Collection Collection of all variations of the sarvam-m model • 3 items • Updated May 24 • 16
Lingshu MLLMs Collection Lingshu: A Generalist Foundation Model for Unified Multimodal Medical Understanding and Reasoning • 3 items • Updated 29 days ago • 17
Athena-3 Collection Athena is a high-performance LLM that is designed to excel in most STEM areas as well as general NLP tasks! • 11 items • Updated 1 day ago • 2
Model Optimizer Collection A collection of generative models quantized and optimized with TensorRT Model Optimizer. • 21 items • Updated 1 day ago • 23
The Lessons of Developing Process Reward Models in Mathematical Reasoning Paper • 2501.07301 • Published Jan 13 • 99
Tucan — Tool using and function calling in Bulgarian Collection A series of open-source Bulgarian language models fine-tuned for function calling and tool use. 2.6B, 9B, and 27B parameter variants. • 12 items • Updated 11 days ago • 2