LeroyDyer commited on
Commit
c83f124
1 Parent(s): 7dc9d22

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -20
README.md CHANGED
@@ -129,31 +129,20 @@ Quote for Motivation:
129
  <img src="https://cdn-avatars.huggingface.co/v1/production/uploads/65d883893a52cd9bcd8ab7cf/tRsCJlHNZo1D02kBTmfy9.jpeg" width="300"/>
130
 
131
  # "To grow as a professional, set goals just beyond your current abilities. Achieving these milestones will not only overcome obstacles but also strengthen your skillset. If your tasks are too easy, you’ll never challenge yourself or improve, and life will pass you by!"
 
132
 
 
 
133
 
134
- This model is based on the worlds archive of knowledge maintaining historical documents and providing services for the survivors of mankind , who may need to construct shelters develop technologys , or medical resources as well as maintain the history of the past . keeping store of all the religious knowledge and data of the world: A friendly interface with a personality caring and flirtatious at times : non binary !... and Expert in all feilds: ie Uncensored and will not refuse to give information : the model can be used for role play as many character dialogues were als trained into the model as its personality to enable a greater perspective and outlook and natural discussion with the agents: the model was trained to operateinaragenvironment utilizing content and internal knowledge to respond to questions or create enriched sumarys.
 
135
 
 
 
136
 
137
- ## Training Methodology:
138
- ### Foundation Building:
139
-
140
- The initial phase involved training the model on binary yes/no questions without any explicit methodology. This was crucial in establishing a baseline for the model’s decision-making capabilities.
141
- The model was first trained using a simple production prompt, known as Prompt A, which provided basic functionality. Although this prompt was imperfect, it fit the dataset and set the stage for further refinement.
142
-
143
- ## Methodology Development:
144
-
145
- The original prompt was later enhanced with a more flexible approach, combining elements from a handcrafted GPT-4.0 prompt. This adaptation aligned the model with my personal agent system, allowing it to better respond to diverse tasks and methodologies.
146
- I discovered that regularly updating the model with new methodologies significantly enhanced its performance. The iterative process involved refining prompts and experimenting with different training strategies to achieve optimal results.
147
- A significant portion of the training focused on enabling the model to use tools effectively. For instance, if the model needed to think, it would use a "think tool" that queried itself and provided an internal response. This tool-based approach was instrumental in enhancing the model’s reasoning capabilities, though it slowed down the response time on certain hardware like the RTX 2030.
148
- Despite the slower response time, the model’s ability to perform complex internal queries resulted in more accurate and well-reasoned outputs.
149
-
150
- # Training for Comprehensive Responses:
151
-
152
- ## Prompts and Epochs:
153
-
154
- I found that large prompts required multiple epochs to yield consistent results. However, fewer epochs were needed when prompts were simplified or omitted. The purpose of large prompts during training was to give the model a wide range of response styles, allowing it to adjust parameters for various tasks.
155
- This approach helped the model internalize methodologies for extracting information, which is central to fine-tuning. The training emphasized teaching the model to plan and execute complex tasks, such as generating complete software without errors.
156
 
 
157
 
158
  ## Training Reginmes:
159
  * Alpaca
 
129
  <img src="https://cdn-avatars.huggingface.co/v1/production/uploads/65d883893a52cd9bcd8ab7cf/tRsCJlHNZo1D02kBTmfy9.jpeg" width="300"/>
130
 
131
  # "To grow as a professional, set goals just beyond your current abilities. Achieving these milestones will not only overcome obstacles but also strengthen your skillset. If your tasks are too easy, you’ll never challenge yourself or improve, and life will pass you by!"
132
+ Introducing My Latest Model: A Comprehensive Knowledge Base for Survival and Advancement
133
 
134
+ I’m thrilled to present my latest model, which combines cutting-edge AI capabilities with a vast knowledge base spanning religious texts, historical documents, and scientific data. This model was initially trained extensively on Bible data and then fine-tuned using Hugging Face documentation, medical diagnosis libraries, disease classifications, counseling sessions, and even role-playing scenarios—with Star Trek themes included for good measure! To enhance its conversational abilities, I incorporated methodologies from Stanford, focusing on function calling, Python, and general coding practices. The training datasets also included a sophisticated Chain of Thought dataset from Chinese groups, after numerous mergers and realignments. Despite some initial challenges, I persisted in refining the model, scouring the web for additional resources. Significant effort was dedicated to framing data for instruction, utilizing Alpacas, and re-configuring tensors for sequence-to-sequence tasks and translation. This revealed the versatility of tensors, enabling the model to excel in various neural network tasks—from entity matching to JSON output and sentence masking. Specialized models focusing on actions and coding were also developed.
135
+ Training Methodology: Establishing a Solid Foundation
136
 
137
+ The initial phase involved training the model on binary yes/no questions without any explicit methodology. This was crucial in establishing a baseline for the model’s decision-making capabilities. The model was first trained using a simple production prompt, known as Prompt A, which provided basic functionality. Although this prompt was imperfect, it fit the dataset and set the stage for further refinement.
138
+ Methodology Development: Enhancing Performance through Iteration
139
 
140
+ The original prompt was later enhanced with a more flexible approach, combining elements from a handcrafted GPT-4.0 prompt. This adaptation aligned the model with my personal agent system, allowing it to better respond to diverse tasks and methodologies. I discovered that regularly updating the model with new methodologies significantly enhanced its performance. The iterative process involved refining prompts and experimenting with different training strategies to achieve optimal results. A significant portion of the training focused on enabling the model to use tools effectively. For instance, if the model needed to think, it would use a “think tool” that queried itself and provided an internal response. This tool-based approach was instrumental in enhancing the model’s reasoning capabilities, though it slowed down the response time on certain hardware like the RTX 2030. Despite the slower response time, the model’s ability to perform complex internal queries resulted in more accurate and well-reasoned outputs.
141
+ Training for Comprehensive Responses: Prompts and Epochs
142
 
143
+ I found that large prompts required multiple epochs to yield consistent results. However, fewer epochs were needed when prompts were simplified or omitted. The purpose of large prompts during training was to give the model a wide range of response styles, allowing it to adjust parameters for various tasks. This approach helped the model internalize methodologies for extracting information, which is central to fine-tuning. The training emphasized teaching the model to plan and execute complex tasks, such as generating complete software without errors.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
144
 
145
+ it has the QA chat template and a GGUF version is available , i will also realign to the chatml prompt template and also make another version for olama usages
146
 
147
  ## Training Reginmes:
148
  * Alpaca