Llama 3.1 8B - MedQuad Medical QA Model

This is a fine-tuned version of unsloth/Meta-Llama-3.1-8B-bnb-4bit, trained on 7000 medical QA pairs from the MedQuAD dataset using LoRA and Unsloth for efficient adaptation.

🧠 Use Case

Designed for medical question answering, this model is capable of responding to complex patient queries in a conversational and informative style.

📊 Training

  • Model: Meta-Llama-3.1-8B-bnb-4bit
  • Dataset: MedQuAD
  • LoRA: Applied to all attention and FFN layers
  • Max tokens: 1024
  • Finetuned using: Unsloth

🔧 How to Use

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("iamsouvik/Llama-3.1-8B-Medquad-main")
model = AutoModelForCausalLM.from_pretrained("iamsouvik/Llama-3.1-8B-Medquad-main")

prompt = '''
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.

### Instruction:
You are an expert medical doctor capable of accurately answering any patient's question.

### Input:
What is Progressive Supranuclear Palsy?

### Response:
'''

inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=128)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
42
GGUF
Model size
8.03B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 1 Ask for provider support

Model tree for iamsouvik/Llama-3.1-8B-Medquad-gguf

Quantized
(239)
this model

Dataset used to train iamsouvik/Llama-3.1-8B-Medquad-gguf