prithivMLmods commited on
Commit
4de2400
·
verified ·
1 Parent(s): 6931d91

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -0
README.md CHANGED
@@ -69,3 +69,45 @@ print(tokenizer.decode(outputs[0]))
69
  7. **Open-Source and Customizable**:
70
  Being open-source, Gemma 2 allows developers to modify and extend its architecture to suit specific use cases, such as integrating it into tools like ReadM.Md for markdown-related tasks.
71
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
69
  7. **Open-Source and Customizable**:
70
  Being open-source, Gemma 2 allows developers to modify and extend its architecture to suit specific use cases, such as integrating it into tools like ReadM.Md for markdown-related tasks.
71
 
72
+ # **Intended Use of GWQ (Gemma with Questions)**
73
+
74
+ 1. **Question Answering:**
75
+ The model excels in generating concise and relevant answers to user-provided queries across various domains.
76
+
77
+ 2. **Summarization:**
78
+ It can be used to summarize large bodies of text, making it suitable for news aggregation, academic research, and report generation.
79
+
80
+ 3. **Reasoning Tasks:**
81
+ GWQ is fine-tuned on the Chain of Continuous Thought Synthetic Dataset, which enhances its ability to perform reasoning, multi-step problem solving, and logical inferences.
82
+
83
+ 4. **Text Generation:**
84
+ The model is ideal for creative writing tasks such as generating poems, stories, and essays. It can also be used for generating code comments, documentation, and markdown files.
85
+
86
+ 5. **Instruction Following:**
87
+ GWQ’s instruction-tuned variant is suitable for generating responses based on user instructions, making it useful for virtual assistants, tutoring systems, and automated customer support.
88
+
89
+ 6. **Domain-Specific Applications:**
90
+ Thanks to its modular design and open-source nature, the model can be fine-tuned for specific tasks like legal document summarization, medical record analysis, or financial report generation.
91
+
92
+ ## **Limitations of GWQ**
93
+
94
+ 1. **Resource Requirements:**
95
+ Although lightweight compared to larger models, the 9B parameter size still requires significant computational resources, including GPUs with large memory for inference.
96
+
97
+ 2. **Knowledge Cutoff:**
98
+ The model’s pre-training data may not include recent information, making it less effective for answering queries on current events or newly developed topics.
99
+
100
+ 3. **Bias in Outputs:**
101
+ Since the model is trained on publicly available datasets, it may inherit biases present in those datasets, leading to potentially biased or harmful outputs in sensitive contexts.
102
+
103
+ 4. **Hallucinations:**
104
+ Like other large language models, GWQ can occasionally generate incorrect or nonsensical information, especially when asked for facts or reasoning outside its training scope.
105
+
106
+ 5. **Lack of Common-Sense Reasoning:**
107
+ While GWQ is fine-tuned for reasoning, it may still struggle with tasks requiring deep common-sense knowledge or nuanced understanding of human behavior and emotions.
108
+
109
+ 6. **Dependency on Fine-Tuning:**
110
+ For optimal performance on domain-specific tasks, fine-tuning on relevant datasets is required, which demands additional computational resources and expertise.
111
+
112
+ 7. **Context Length Limitation:**
113
+ The model’s ability to process long documents is limited by its maximum context window size. If the input exceeds this limit, truncation may lead to loss of important information.