torch and llama.cpp integration

#1
by TobDeBer - opened

I just tried with latest torch/transformers and llama.cpp and inference failed.
Are there upstream branches I can use?

ValueError: The checkpoint you are trying to load has model type granitemoe but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.

@TobDeBer you need this specific PR: https://github.com/huggingface/transformers/pull/33207/
the code for the MoE model is not merged into HF transformers main branch yet

Thanks!
It has been merged last night.

TobDeBer changed discussion status to closed

awesome

Sign up or log in to comment