Using turboderp-org's ExLlamaV3 v0.0.1 for quantization.
2.00 bits per weight
2.25 bits per weight
2.50 bits per weight
3.00 bits per weight
3.50 bits per weight
4.00 bits per weight
5.00 bits per weight
6.00 bits per weight
8.00 bits per weight
8.00 bits per weight / H8
ELYZA-Shortcut-1.0-Qwen-7B-exl3
- Model creator: elyza
- Original model: ELYZA-Shortcut-1.0-Qwen-7B
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for RioShiina/ELYZA-Shortcut-1.0-Qwen-7B-exl3
Base model
Qwen/Qwen2.5-7B
Finetuned
Qwen/Qwen2.5-7B-Instruct
Finetuned
elyza/ELYZA-Shortcut-1.0-Qwen-7B