Raj411 commited on
Commit
f5664d8
·
verified ·
1 Parent(s): 1c5898d

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +89 -0
README.md ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-4.0
3
+ tags:
4
+ - sentiment-classification
5
+ - telugu
6
+ - muril
7
+ - indian-languages
8
+ - baseline
9
+ - tesent
10
+ language: te
11
+ datasets:
12
+ - TeSent_Benchmark-Dataset
13
+ model_name: MuRIL_WR
14
+ ---
15
+
16
+ # MuRIL_WR: MuRIL Telugu Sentiment Classification Model (With Rationale)
17
+
18
+ ## Model Overview
19
+
20
+ **MuRIL_WR** is a Telugu sentiment classification model based on **MuRIL (Multilingual Representations for Indian Languages)**, a transformer-based BERT model specifically designed to support 17+ Indian languages, including Telugu and English.
21
+ The "WR" in the model name stands for "**With Rationale**", indicating that this model is trained using both sentiment labels and **human-annotated rationales** from the TeSent_Benchmark-Dataset.
22
+
23
+ ---
24
+
25
+ ## Model Details
26
+
27
+ - **Architecture:** MuRIL (BERT-base for Indian languages, pre-trained on 17+ languages)
28
+ - **Pretraining Data:** Large corpus of Telugu sentences from web, religious scripts, news data, etc.
29
+ - **Pretraining Objectives:** Masked Language Modeling (MLM) and Translation Language Modeling (TLM) tasks
30
+ - **Fine-tuning Data:** [TeSent_Benchmark-Dataset](https://github.com/DSL-13-SRMAP/TeSent_Benchmark-Dataset), using both sentence-level sentiment labels (positive, negative, neutral) and rationale annotations
31
+ - **Task:** Sentence-level sentiment classification (3-way)
32
+ - **Rationale Usage:** **Used** during training and/or inference ("WR" = With Rationale)
33
+
34
+ ---
35
+
36
+ ## Intended Use
37
+
38
+ - **Primary Use:** Benchmarking Telugu sentiment classification on the TeSent_Benchmark-Dataset, especially as a **baseline** for models trained with and without rationales
39
+ - **Research Setting:** Recommended for academic research in low-resource and explainable NLP settings, especially for informal, social media, or conversational Telugu data
40
+
41
+ ---
42
+
43
+ ## Why MuRIL?
44
+
45
+ MuRIL is specifically pre-trained on Indian languages and offers better understanding of Telugu morphology and syntax compared to general multilingual models like mBERT and XLM-R.
46
+ As the pre-training data favors informal texts from the web, MuRIL is especially effective for informal, social media, or conversational NLP tasks in Telugu. For formal or classical Telugu, performance may be lower.
47
+
48
+ ---
49
+
50
+ ## Performance and Limitations
51
+
52
+ **Strengths:**
53
+ - Superior understanding of Telugu compared to general multilingual models
54
+ - Excels in informal, web, or conversational Telugu sentiment tasks
55
+ - Provides **explicit rationales** for predictions, aiding explainability
56
+ - Strong baseline for Telugu sentiment classification
57
+
58
+ **Limitations:**
59
+ - May underperform on formal or classical Telugu tasks due to pre-training corpus
60
+ - Applicability limited to Telugu analysis; not ideal for highly formal text processing
61
+ - Requires sufficient labeled Telugu data and rationale annotations for best performance
62
+
63
+ ---
64
+
65
+ ## Training Data
66
+
67
+ - **Dataset:** [TeSent_Benchmark-Dataset](https://github.com/DSL-13-SRMAP/TeSent_Benchmark-Dataset)
68
+ - **Data Used:** The **Content** (Telugu sentence), **Label** (sentiment label), and **Rationale** (human-annotated rationale) columns are used for MuRIL_WR training
69
+
70
+ ---
71
+
72
+ ## Language Coverage
73
+
74
+ - **Language:** Telugu (`te`)
75
+ - **Model Scope:** This implementation and evaluation focus strictly on Telugu sentiment classification
76
+
77
+ ---
78
+
79
+ ## Citation and More Details
80
+
81
+ For detailed experimental setup, evaluation metrics, and comparisons with rationale-based models, **please refer to our paper**.
82
+
83
+
84
+
85
+ ---
86
+
87
+ ## License
88
+
89
+ Released under [CC BY 4.0](LICENSE).