QuantFactory/Biggie-SmoLlm-0.4B-GGUF
This is quantized version of nisten/Biggie-SmoLlm-0.4B created using llama.cpp
Original Model Card
###Coherent Frankenstein of smolLm-0.36b upped to 0.4b
This took about 5 hours of semi-automated continuous merging to figure out the recipe. Model is smarter, and UNTRAINED. Uploaded it for training. Yet it performs well as is even quantized to 8bit. 8bit gguf included for testing.
wget https://huggingface.co/nisten/Biggie-SmoLlm-0.4B/resolve/main/Biggie_SmolLM_400M_q8_0.gguf
./llama-cli -ngl 99 -co --temp 0 -p "How to build a city on Mars via calculating Aldrin-Cycler orbits?" -m Biggie_SmolLM_400M_q8_0.gguf -cnv -fa --keep -1
- Downloads last month
- 132
Hardware compatibility
Log In
to view the estimation
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for QuantFactory/Biggie-SmoLlm-0.4B-GGUF
Base model
HuggingFaceTB/SmolLM-360M
Quantized
HuggingFaceTB/SmolLM-360M-Instruct