Training procedure

This adapter has been fine-tuned using quantization-aware LoRA (QA-LoRA).

More details on the training procedure here: Fine-tune Quantized Llama 2 on Your GPU with QA-LoRA

The base model was quantized with AutoGPTQ INT4. You can find it here:

kaitchup/Llama-2-7b-4bit-32g-autogptq

Framework versions

  • PEFT 0.4.0
Downloads last month
6
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support