LLaMA3-2.1B QLoRA fine-tuned on PubMedQA
This model is a 4-bit quantized, QLoRA fine-tuned version of meta-llama/Llama-3.2-1B
, trained on the PubMedQA dataset for medical question classification (yes
, no
, maybe
). It was optimized using PEFT with LoRA adapters, and is designed for efficient inference on resource-constrained hardware.
Training Details
- Base model:
meta-llama/Llama-3.2-1B
- Dataset:
pubmed_qa/pqa_labeled
- Method: QLoRA (4-bit NF4)
- LoRA target modules:
q_proj
,v_proj
- Epochs: 5
- Batch size: 4
- Downloads last month
- 4
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for vannishh/llama3-2.1B-4bit-finetuned
Base model
meta-llama/Llama-3.2-1B