Is this an MOE?

#5
by AlgorithmicKing - opened

if it is then how many active parameters does it have?

It is not, it is Llama 3.3 70B that has been fine-tuned using data generated by deepseek-R1

oh. thanks for pointing it out

AlgorithmicKing changed discussion status to closed

Sign up or log in to comment