--- license: cc-by-4.0 tags: - sentiment-classification - telugu - multilingual - mbert - baseline language: te datasets: - DSL-13-SRMAP/TeSent_Benchmark-Dataset model_name: mBERT_WOR --- # mBERT_WOR: Telugu Sentiment Classification Model (Without Rationale) ## Model Overview **mBERT_WOR** is a Telugu sentiment classification model based on Google's mBERT (BERT-base-multilingual-cased), trained specifically for sentence-level sentiment analysis **without rationale supervision**. The acronym "WOR" stands for "Without Rationale," indicating that this model was trained using only the sentiment labels and not the human-annotated rationales provided in the TeSent_Benchmark-Dataset. --- ## Model Details - **Architecture:** mBERT (BERT-base-multilingual-cased, 12 layers, ~100M parameters) - **Pretraining Data:** Wikipedia articles in 104 languages (including Telugu), using Masked Language Modeling (MLM) and Next Sentence Prediction (NSP) objectives. - **Fine-tuning Data:** TeSent_Benchmark-Dataset (Telugu only), using only the sentence-level sentiment labels (positive, negative, neutral); rationale annotations are not used in training. - **Task:** Sentence-level sentiment classification (3-way) - **Rationale Usage:** Not used during training or inference --- ## Intended Use - **Primary Use:** Benchmarking Telugu sentiment classification on the TeSent_Benchmark-Dataset, especially as a **baseline** for models trained with and without rationales. - **Research Setting:** Designed for academic research, particularly in low-resource and explainable NLP settings. --- ## Performance and Limitations - **Strengths:** - Leverages shared multilingual representations, enabling cross-lingual transfer and reasonable performance for Telugu even with limited labeled data. - Serves as a robust baseline for Telugu sentiment tasks. - **Limitations:** - Not specifically optimized for Telugu morphology or syntax, which may impact its ability to capture fine-grained, language-specific sentiment cues. - May underperform compared to Telugu-specialized models such as IndicBERT or L3Cube-Telugu-BERT, especially for nuanced or idiomatic expressions. - Since rationales are not used, the model cannot provide explicit explanations for its predictions. --- ## Training Data - **Dataset:** [TeSent_Benchmark-Dataset](https://github.com/DSL-13-SRMAP/TeSent_Benchmark-Dataset) - **Data Used:** Only the **Content** (Telugu sentence) and **Label** (sentiment label) columns; **rationale** annotations are ignored for mBERT_WOR training. --- ## Language Coverage - **Language:** Telugu (the only language in the dataset) - **Note:** While mBERT is a multilingual model, this implementation and evaluation are strictly for Telugu sentiment classification. --- ## Citation and More Details For detailed experimental setup, evaluation metrics, and comparisons with rationale-based models, **please refer to our paper**. --- ## License Released under [CC BY 4.0](LICENSE).