Training procedure
This adapter has been fine-tuned using quantization-aware LoRA (QA-LoRA).
More details on the training procedure here: Fine-tune Quantized Llama 2 on Your GPU with QA-LoRA
The base model was quantized with AutoGPTQ INT4. You can find it here:
kaitchup/Llama-2-7b-4bit-32g-autogptq
Framework versions
- PEFT 0.4.0
- Downloads last month
- 6
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support