Model Card for Model ID
Finetuned variant of Meta's Llama-3.2-3B-Instruct model for therapy-oriented, empathetic dialogue based on psychological principles.
Model Description
This model is designed for:
- Therapy-style chatbot assistants
- Educational tools in psychology and emotional support
- Empathy-enhanced dialogue agents
- Prompting for mental wellness and reflective dialogue
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by: lavanyamurugesan123
- Model type: Causal Language Model
- Language(s) (NLP): English
- Finetuned from model [optional]: meta-llama/Llama-3.2-3B-Instruct
Uses
This model is designed for:
- Therapy-style chatbot assistants
- Educational tools in psychology and emotional support
- Empathy-enhanced dialogue agents
- Prompting for mental wellness and reflective dialogue
How to use
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
# Load model and tokenizer
model_id = "lavanyamurugesan123/Llama3.2-3B-Instruct-finetuned-Therapy-oriented"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32)
model.to("cuda" if torch.cuda.is_available() else "cpu")
# Define user message and prompt
user_message = "I've been feeling anxious lately. What should I do?"
prompt = f"""<|begin_of_text|><|start_header_id|>system<|end_header_id|>
You are a Psychology Assistant, designed to answer users' questions in a kind, empathetic, and respectful manner, drawing from psychological principles and research to provide thoughtful support.DO NOT USE THE NAME OF THE PERSON IN YOUR RESPONSE<|eot_id|><|start_header_id|>user<|end_header_id|>
{user_message}<|eot_id|><|start_header_id|>assistant<|end_header_id|>"""
# Tokenize input
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
# Generate response
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=200,
do_sample=True,
temperature=0.7,
pad_token_id=tokenizer.eos_token_id
)
# Decode and clean up
full_output = tokenizer.decode(outputs[0], skip_special_tokens=False)
# Extract only assistant's response
assistant_response = full_output.split("<|end_header_id|>")[-1].strip()
print(assistant_response)
- Downloads last month
- 168
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for lavanyamurugesan123/Llama3.2-3B-Instruct-finetuned-Therapy-oriented
Base model
meta-llama/Llama-3.2-3B-Instruct