Error: llama runner process has terminated: exit status 2 when running ollama

#1
by DuyDoanLearning - opened

I run this command:
ollama run hf.co/openbmb/MiniCPM-o-2_6-gguf:Q4_K_M

and has this error:
Error: llama runner process has terminated: exit status 2

Specs:

  • CUDA: v12
  • GPU T4 Tesla
  • Running in SageMaker instance (AWS)
  • Python 3.11.11

Model has been downloaded but failed to run. I tried ollama list and confirmed that model has been downloaded:
hf.co/openbmb/MiniCPM-o-2_6-gguf:Q4_K_M 577cda2e3886 5.7 GB 21 seconds ago

I also tried running on llama-cpp-python but it also has a lot of errors.

Please help. I am desperate.
Thank you very much.

I get an error message as well running this on Ollama. I'm not desperate though, I hope this can be solved in the coming weeks. πŸ™‚

Sign up or log in to comment