MuRIL_WR / README.md
Raj411's picture
Update README.md
949b25e verified
---
license: cc-by-4.0
tags:
- sentiment-classification
- telugu
- muril
- indian-languages
- baseline
- tesent
language: te
datasets:
- DSL-13-SRMAP/TeSent_Benchmark-Dataset
model_name: MuRIL_WR
---
# MuRIL_WR: MuRIL Telugu Sentiment Classification Model (With Rationale)
## Model Overview
**MuRIL_WR** is a Telugu sentiment classification model based on **MuRIL (Multilingual Representations for Indian Languages)**, a transformer-based BERT model specifically designed to support 17+ Indian languages, including Telugu and English.
The "WR" in the model name stands for "**With Rationale**", indicating that this model is trained using both sentiment labels and **human-annotated rationales** from the TeSent_Benchmark-Dataset.
---
## Model Details
- **Architecture:** MuRIL (BERT-base for Indian languages, pre-trained on 17+ languages)
- **Pretraining Data:** Large corpus of Telugu sentences from web, religious scripts, news data, etc.
- **Pretraining Objectives:** Masked Language Modeling (MLM) and Translation Language Modeling (TLM) tasks
- **Fine-tuning Data:** [TeSent_Benchmark-Dataset](https://github.com/DSL-13-SRMAP/TeSent_Benchmark-Dataset), using both sentence-level sentiment labels (positive, negative, neutral) and rationale annotations
- **Task:** Sentence-level sentiment classification (3-way)
- **Rationale Usage:** **Used** during training and/or inference ("WR" = With Rationale)
---
## Intended Use
- **Primary Use:** Benchmarking Telugu sentiment classification on the TeSent_Benchmark-Dataset, especially as a **baseline** for models trained with and without rationales
- **Research Setting:** Recommended for academic research in low-resource and explainable NLP settings, especially for informal, social media, or conversational Telugu data
---
## Why MuRIL?
MuRIL is specifically pre-trained on Indian languages and offers better understanding of Telugu morphology and syntax compared to general multilingual models like mBERT and XLM-R.
As the pre-training data favors informal texts from the web, MuRIL is especially effective for informal, social media, or conversational NLP tasks in Telugu. For formal or classical Telugu, performance may be lower.
---
## Performance and Limitations
**Strengths:**
- Superior understanding of Telugu compared to general multilingual models
- Excels in informal, web, or conversational Telugu sentiment tasks
- Provides **explicit rationales** for predictions, aiding explainability
- Strong baseline for Telugu sentiment classification
**Limitations:**
- May underperform on formal or classical Telugu tasks due to pre-training corpus
- Applicability limited to Telugu analysis; not ideal for highly formal text processing
- Requires sufficient labeled Telugu data and rationale annotations for best performance
---
## Training Data
- **Dataset:** [TeSent_Benchmark-Dataset](https://github.com/DSL-13-SRMAP/TeSent_Benchmark-Dataset)
- **Data Used:** The **Content** (Telugu sentence), **Label** (sentiment label), and **Rationale** (human-annotated rationale) columns are used for MuRIL_WR training
---
## Language Coverage
- **Language:** Telugu (`te`)
- **Model Scope:** This implementation and evaluation focus strictly on Telugu sentiment classification
---
## Citation and More Details
For detailed experimental setup, evaluation metrics, and comparisons with rationale-based models, **please refer to our paper**.
---
## License
Released under [CC BY 4.0](LICENSE).