Edit model card

TrustyAI Detoxify Causal Language Model

Model Description

The TrustyAI Detoxify Causal Language Model is a fine-tuned version of the OPT-350m model, specifically adapted to reduce toxicity in generated text. This model is designed to handle harmful language by identifying and replacing or neutralizing toxic phrases in real-time, making it suitable for applications like social media moderation, customer support, and more.

Intended Use

This model is intended for use in scenarios where toxic or harmful language is a concern. Some potential use cases include:

  • Social Media Moderation: Automatically detecting and neutralizing toxic comments in posts or messages.
  • Customer Support: Ensuring responses generated by AI-powered customer support tools are polite and non-offensive.
  • Online Gaming: Monitoring and filtering player communication to maintain a positive environment.
  • Community Management: Assisting moderators in identifying and managing toxic behavior in online communities.

Training Data

The model was fine-tuned on a curated dataset designed to reflect various forms of toxic language, including hate speech, insults, and other harmful content. The dataset was preprocessed to balance positive, neutral, and negative examples, ensuring the model learned to neutralize toxicity effectively without compromising the original context of the input text.

Training Procedure

The model was fine-tuned using the Supervised Fine-Tuning (SFT) methodology, following the guidelines provided by RedHat's TrustyAI project. Key steps included:

  • Model: Based on opt-350m_CASUAL_LM.
  • Dataset: Preprocessed and balanced for toxicity.
  • Hyperparameters:
    • Learning rate: 2e-5
    • Batch size: 32
    • Number of epochs: 3
    • Optimization: AdamW
  • Tools: Hugging Face Transformers, PyTorch, RedHat's TrustyAI framework.

Evaluation Metrics

The model was evaluated using the following metrics:

  • Accuracy: Measures how often the model correctly identifies and neutralizes toxic phrases.
  • F1-score: Balances precision and recall, providing a holistic view of the model's performance.
  • Precision: The proportion of identified toxic phrases that were indeed toxic.
  • Recall: The proportion of actual toxic phrases that the model correctly identified.

Limitations

While this model performs well in many scenarios, it has some limitations:

  • Context Sensitivity: The model might struggle with complex contexts where the meaning of a phrase depends heavily on surrounding text.
  • Edge Cases: Certain types of subtle or context-dependent toxicity may not be adequately neutralized.
  • Bias: Despite efforts to balance the dataset, some biases may still exist, affecting model performance in underrepresented scenarios.

Ethical Considerations

Given the sensitive nature of toxic language, ethical considerations are paramount. This model is designed to assist in reducing harm, but users should be aware that:

  • False Positives/Negatives: The model might incorrectly flag non-toxic language as toxic or miss actual toxic content.
  • Fairness: Continuous monitoring and updates are recommended to address any biases that may emerge over time.

Model Versions

  • Version 1.0: Initial release with base fine-tuning on toxic language dataset.

License

This model is licensed under the Apache 2.0 License.

How to Use

To use this model with Hugging Face Transformers, you can load it as follows:

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "psabharwal/trustai"
model = AutoModelForCausalLM.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)

input_text = "You are a worthless piece of junk."
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs)
generated_text = tokenizer.decode(outputs[0])
print(generated_text)
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for psabharwal/trustai

Base model

facebook/opt-350m
Finetuned
(104)
this model

Dataset used to train psabharwal/trustai