Is this an MOE?
#5
by
AlgorithmicKing
- opened
if it is then how many active parameters does it have?
It is not, it is Llama 3.3 70B that has been fine-tuned using data generated by deepseek-R1
oh. thanks for pointing it out
AlgorithmicKing
changed discussion status to
closed