medmekk/Llama-3.2-1B-ao-int8wo-gs128 (Quantized)
Description
This model is a quantized version of the original model medmekk/Llama-3.2-1B-ao-int8wo-gs128
.
It's quantized using the TorchAO library using the torchao-my-repo space.
Quantization Details
- Quantization Type: int8_weight_only
- Group Size: 128
- Downloads last month
- 1
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for medmekk/Llama-3.2-1B-ao-int8wo-gs128
Unable to build the model tree, the base model loops to the model itself. Learn more.