EARLY SNEAK PREVIEW: get a first look at the Celestia 3 science-reasoning dataset, built with DeepSeek's newest R1-0528 reasoning model! Subjects include physics, chemistry, biology, computer science, Earth science, astronomy, and information theory.
Coming up we'll have more dataset releases, including some novel reasoning and analysis methods - we think an important role for open source researchers is experimenting with new response styles on top of the increasingly excellent base models available to finetune.
- A full-stack software assistant: a reasoning finetune focused on coding, architecture, and DevOps using the Titanium and Tachibana datasets! - Improved general and creative reasoning skills, powered by the Raiden dataset.
Updates for the week: - released some new merge models using ValiantLabs/Qwen3-14B-Esper3 and other Qwen 3 14b finetunes - these merges include math, Web3, uncensored, and general mix. depending on your use case for Esper 3 these may be helpful to you! find them at @sequelbox - coming up we'll have more model sizes for Esper 3 and Cobalt 2, releasing soon! - also super excited for more dataset releases with the newly released deepseek-ai/DeepSeek-R1-0528
I am so happy to share to all that I’ve just completed the first unit of the new MCP course on Hugging Face and earned my certificate! The AI acceleration track is intense and fast-paced, but I’m doing my best to keep up. Excited for what’s ahead!