TheBloke commited on
Commit
02dd938
·
1 Parent(s): 1b2f7dc

Updating model files

Browse files
Files changed (1) hide show
  1. README.md +28 -6
README.md CHANGED
@@ -17,6 +17,17 @@ language:
17
  library_name: transformers
18
  pipeline_tag: text-generation
19
  ---
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  # Manticore 13B Chat GGML
22
 
@@ -62,23 +73,34 @@ GGML models can be loaded into text-generation-webui by installing the llama.cpp
62
  Further instructions here: [text-generation-webui/docs/llama.cpp-models.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/llama.cpp-models.md).
63
 
64
 
 
 
 
 
 
 
 
 
 
 
 
65
  # Original model card - Manticore 13B Chat
66
 
67
- Manticore 13B Chat builds on Manticore with new datasets, including a de-duped subset of the Pygmalion dataset. It also removes all Alpaca style prompts using `###` in favor of
68
  chat only style prompts using `USER:`,`ASSISTANT:` as well as [pygmalion/metharme prompting](https://huggingface.co/PygmalionAI/metharme-7b#prompting) using `<|system|>, <|user|> and <|model|>` tokens.
69
 
70
  Questions, comments, feedback, looking to donate, or want to help? Reach out on our [Discord](https://discord.gg/EqrvvehG) or email [[email protected]](mailto:[email protected])
71
 
72
  # Training Datasets
73
 
74
- Manticore 13B Chat is a Llama 13B model fine-tuned on the following datasets along with the datasets from the original Manticore 13B.
75
 
76
  **Manticore 13B Chat was trained on 25% of the datasets below. The datasets were merged, shuffled, and then sharded into 4 parts.**
77
 
78
  - de-duped pygmalion dataset, filtered down to RP data
79
- - [riddle_sense](https://huggingface.co/datasets/riddle_sense) - instruct augmented
80
  - hellaswag, updated for detailed explanations w 30K+ rows
81
- - [gsm8k](https://huggingface.co/datasets/gsm8k) - instruct augmented
82
  - [ewof/code-alpaca-instruct-unfiltered](https://huggingface.co/datasets/ewof/code-alpaca-instruct-unfiltered)
83
 
84
  Manticore 13B
@@ -110,8 +132,8 @@ Try out the model in HF Spaces. The demo uses a quantized GGML version of the mo
110
 
111
  ## Build
112
 
113
- Manticore was built with [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) on 8xA100 80GB
114
- - 3 epochs taking approximately 8 hours. No further epochs will be released.
115
  - The configuration to duplicate this build is provided in this repo's [/config folder](https://huggingface.co/openaccess-ai-collective/manticore-13b/tree/main/configs).
116
 
117
  ## Bias, Risks, and Limitations
 
17
  library_name: transformers
18
  pipeline_tag: text-generation
19
  ---
20
+ <div style="width: 100%;">
21
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
22
+ </div>
23
+ <div style="display: flex; justify-content: space-between; width: 100%;">
24
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
25
+ <p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
26
+ </div>
27
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
28
+ <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
29
+ </div>
30
+ </div>
31
 
32
  # Manticore 13B Chat GGML
33
 
 
73
  Further instructions here: [text-generation-webui/docs/llama.cpp-models.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/llama.cpp-models.md).
74
 
75
 
76
+ ## Want to support my work?
77
+
78
+ I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
79
+
80
+ So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
81
+
82
+ Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
83
+
84
+ * Patreon: coming soon! (just awaiting approval)
85
+ * Ko-Fi: https://ko-fi.com/TheBlokeAI
86
+ * Discord: https://discord.gg/UBgz4VXf
87
  # Original model card - Manticore 13B Chat
88
 
89
+ Manticore 13B Chat builds on Manticore with new datasets, including a de-duped subset of the Pygmalion dataset. It also removes all Alpaca style prompts using `###` in favor of
90
  chat only style prompts using `USER:`,`ASSISTANT:` as well as [pygmalion/metharme prompting](https://huggingface.co/PygmalionAI/metharme-7b#prompting) using `<|system|>, <|user|> and <|model|>` tokens.
91
 
92
  Questions, comments, feedback, looking to donate, or want to help? Reach out on our [Discord](https://discord.gg/EqrvvehG) or email [[email protected]](mailto:[email protected])
93
 
94
  # Training Datasets
95
 
96
+ Manticore 13B Chat is a Llama 13B model fine-tuned on the following datasets along with the datasets from the original Manticore 13B.
97
 
98
  **Manticore 13B Chat was trained on 25% of the datasets below. The datasets were merged, shuffled, and then sharded into 4 parts.**
99
 
100
  - de-duped pygmalion dataset, filtered down to RP data
101
+ - [riddle_sense](https://huggingface.co/datasets/riddle_sense) - instruct augmented
102
  - hellaswag, updated for detailed explanations w 30K+ rows
103
+ - [gsm8k](https://huggingface.co/datasets/gsm8k) - instruct augmented
104
  - [ewof/code-alpaca-instruct-unfiltered](https://huggingface.co/datasets/ewof/code-alpaca-instruct-unfiltered)
105
 
106
  Manticore 13B
 
132
 
133
  ## Build
134
 
135
+ Manticore was built with [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) on 8xA100 80GB
136
+ - 3 epochs taking approximately 8 hours. No further epochs will be released.
137
  - The configuration to duplicate this build is provided in this repo's [/config folder](https://huggingface.co/openaccess-ai-collective/manticore-13b/tree/main/configs).
138
 
139
  ## Bias, Risks, and Limitations