LLaMA3-2.1B QLoRA fine-tuned on PubMedQA

This model is a 4-bit quantized, QLoRA fine-tuned version of meta-llama/Llama-3.2-1B, trained on the PubMedQA dataset for medical question classification (yes, no, maybe). It was optimized using PEFT with LoRA adapters, and is designed for efficient inference on resource-constrained hardware.

Training Details

  • Base model: meta-llama/Llama-3.2-1B
  • Dataset: pubmed_qa/pqa_labeled
  • Method: QLoRA (4-bit NF4)
  • LoRA target modules: q_proj, v_proj
  • Epochs: 5
  • Batch size: 4
Downloads last month
4
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for vannishh/llama3-2.1B-4bit-finetuned

Adapter
(475)
this model