Triangle104 commited on
Commit
fc20bff
·
verified ·
1 Parent(s): 14c2c94

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -0
README.md CHANGED
@@ -110,6 +110,47 @@ model-index:
110
  This model was converted to GGUF format from [`anthracite-org/magnum-v4-12b`](https://huggingface.co/anthracite-org/magnum-v4-12b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
111
  Refer to the [original model card](https://huggingface.co/anthracite-org/magnum-v4-12b) for more details on the model.
112
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
113
  ## Use with llama.cpp
114
  Install llama.cpp through brew (works on Mac and Linux)
115
 
 
110
  This model was converted to GGUF format from [`anthracite-org/magnum-v4-12b`](https://huggingface.co/anthracite-org/magnum-v4-12b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
111
  Refer to the [original model card](https://huggingface.co/anthracite-org/magnum-v4-12b) for more details on the model.
112
 
113
+ ---
114
+ Model details:
115
+ -
116
+ This is a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus.
117
+
118
+ This model is fine-tuned on top of mistralai/Mistral-Nemo-Instruct-2407.
119
+
120
+ Prompting
121
+ -
122
+ A typical input would look like this:
123
+
124
+ <s>[INST] SYSTEM MESSAGE
125
+ USER MESSAGE[/INST] ASSISTANT MESSAGE</s>[INST] USER MESSAGE[/INST]
126
+
127
+ SillyTavern templates
128
+ -
129
+ Below are Instruct and Context templates for use within SillyTavern.
130
+
131
+ context template
132
+
133
+ instruct template
134
+
135
+ Credits
136
+ -
137
+ We'd like to thank Recursal / Featherless for sponsoring the compute for this train, Featherless has been hosting our Magnum models since the first 72 B and has given thousands of people access to our models and helped us grow.
138
+
139
+ We would also like to thank all members of Anthracite who made this finetune possible.
140
+ Datasets
141
+
142
+ anthracite-org/c2_logs_32k_llama3_qwen2_v1.2_no_system
143
+ anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system
144
+ anthracite-org/kalo-opus-instruct-3k-filtered-no-system
145
+ anthracite-org/nopm_claude_writing_fixed
146
+ anthracite-org/kalo_opus_misc_240827_no_system
147
+ anthracite-org/kalo_misc_part2_no_system
148
+
149
+ Training
150
+ -
151
+ The training was done for 2 epochs. We used 8xH100s GPUs graciously provided by Recursal AI / Featherless AI for the full-parameter fine-tuning of the model.
152
+
153
+ ---
154
  ## Use with llama.cpp
155
  Install llama.cpp through brew (works on Mac and Linux)
156