Model Card for Model ID

Finetuned variant of Meta's Llama-3.2-3B-Instruct model for therapy-oriented, empathetic dialogue based on psychological principles.

Model Description

This model is designed for:

  1. Therapy-style chatbot assistants
  2. Educational tools in psychology and emotional support
  3. Empathy-enhanced dialogue agents
  4. Prompting for mental wellness and reflective dialogue

This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.

  • Developed by: lavanyamurugesan123
  • Model type: Causal Language Model
  • Language(s) (NLP): English
  • Finetuned from model [optional]: meta-llama/Llama-3.2-3B-Instruct

Uses

This model is designed for:

  1. Therapy-style chatbot assistants
  2. Educational tools in psychology and emotional support
  3. Empathy-enhanced dialogue agents
  4. Prompting for mental wellness and reflective dialogue

How to use

import torch
from transformers import AutoTokenizer, AutoModelForCausalLM

# Load model and tokenizer
model_id = "lavanyamurugesan123/Llama3.2-3B-Instruct-finetuned-Therapy-oriented"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32)
model.to("cuda" if torch.cuda.is_available() else "cpu")

# Define user message and prompt
user_message = "I've been feeling anxious lately. What should I do?"
prompt = f"""<|begin_of_text|><|start_header_id|>system<|end_header_id|>
You are a Psychology Assistant, designed to answer users' questions in a kind, empathetic, and respectful manner, drawing from psychological principles and research to provide thoughtful support.DO NOT USE THE NAME OF THE PERSON IN YOUR RESPONSE<|eot_id|><|start_header_id|>user<|end_header_id|>
{user_message}<|eot_id|><|start_header_id|>assistant<|end_header_id|>"""

# Tokenize input
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)

# Generate response
with torch.no_grad():
    outputs = model.generate(
        **inputs,
        max_new_tokens=200,
        do_sample=True,
        temperature=0.7,
        pad_token_id=tokenizer.eos_token_id
    )

# Decode and clean up
full_output = tokenizer.decode(outputs[0], skip_special_tokens=False)

# Extract only assistant's response
assistant_response = full_output.split("<|end_header_id|>")[-1].strip()

print(assistant_response)
Downloads last month
168
Safetensors
Model size
3.21B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for lavanyamurugesan123/Llama3.2-3B-Instruct-finetuned-Therapy-oriented

Finetuned
(499)
this model

Dataset used to train lavanyamurugesan123/Llama3.2-3B-Instruct-finetuned-Therapy-oriented