This is just a reupload of the quantized 4Q K M version of Gemma 3 1b by Google and unsloth. It is used for benchmark purposes on the Jetson Nano and llama.cpp with CUDA support:
- https://github.com/kreier/llama.cpp-jetson - ๐ณ instructions to compile llama.cpp with CUDA support
- https://github.com/kreier/llama.cpp-jetson.nano - ๐ precompiled versions that can be installed on the Jetson in minutes
- https://github.com/kreier/jetson a few other insights and results
- Downloads last month
- 2
Hardware compatibility
Log In
to view the estimation
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support