Do you know what I was planning to do this time last week?
I was preparing to write a report declaring that Jan Nano was a failed project because the benchmark results didn’t meet expectations.
But I thought — it can’t be. When loading the model into the app, the performance clearly felt better. So why were the benchmark results worse?
That’s when I reviewed the entire benchmark codebase and realized something fundamental: agentic or workflow-based approaches introduce a huge gap and variation when benchmarking. Jan-nano was trained with an agentic setup — it simply can’t be benchmarked using a rigid workflow-based method.
I made the necessary changes, and the model ended up performing even better than before the issues arose. Turns out the previous benchmarking method conflicted with the way the model was trained.
What if I had given up? That would’ve meant 1.5 months of training and a huge amount of company resources wasted.
But now, this is officially the most successful and biggest release for the whole team — all thanks to Jan-nano.
dots.llm1.base 🪐 a 142B MoE model with only 14B active params.
rednote-hilab/dotsllm1-68246aaaaba3363374a8aa7c ✨ Base & Instruct - MIT license ✨ Trained on 11.2T non-synthetic high-quality data ✨ Competitive with Qwen2.5/3 on reasoning, code, alignment
I'm collecting llama-bench results for inference with a llama 3.1 8B q4 and q8 reference models on varoius GPUs. The results are average of 5 executions. The system varies (different motherboard and CPU ... but that probably that has little effect on the inference performance).
There seems to multiple paid apps shared here that are based on models on hf, but some ppl sell their wrappers as "products" and promote them here. For a long time, hf was the best and only platform to do oss model stuff but with the recent AI website builders anyone can create a product (really crappy ones btw) and try to sell it with no contribution to oss stuff. Please dont do this, or try finetuning the models you use... Sorry for filling yall feed with this bs but yk...