v000000 commited on
Commit
b3805b0
·
verified ·
1 Parent(s): 3493bd2

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -0
README.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - mistralai/Mixtral-8x7B-v0.1
4
+ library_name: transformers
5
+ tags:
6
+ - merge
7
+ - mixtral
8
+ - mistral
9
+ - llama.cpp
10
+ license: cc-by-nc-4.0
11
+ ---
12
+
13
+ Mixtral 8x7B
14
+
15
+ ***
16
+
17
+ Typhon - A Custom Experimental Mixtral Merge
18
+
19
+ An experimental Merge I tried for fun. Honestly did not expect it to work for Mixtral at all considering how its an MoE and the gates and all would be fucked by this custom merge.
20
+
21
+ From my testing it was able to handle SFW <--> NSFW scenarios fine, handle 1st and 3rd person roleplays fine, and seemed fairly smart.
22
+
23
+ It did pretty well for non NSFW tasks so that's a win.
24
+
25
+ Due to the nature of the merge, and Mixtral itself, it is sensitive to Prompts, does follow them well. Sampler settings are fine. i stuck with universal-light and was okay at up to 16k context during testing.
26
+ ***
27
+
28
+ Recipe Below:
29
+
30
+ ```
31
+ base_model: mistralai/Mixtral-8x7B-v0.1
32
+ models:
33
+ - model: mistralai/Mixtral-8x7B-v0.1
34
+ # no parameters necessary for base model
35
+ - model: smelborp/MixtralOrochi8x7B
36
+ parameters:
37
+ weight: 0.30
38
+ density: 0.47
39
+ - model: notstoic/Nous-Hermes-2-Mixtruct-v0.1-8x7B-DPO-DARE_TIES
40
+ parameters:
41
+ weight: 0.31
42
+ density: 0.56
43
+ - model: Sao10K/Solstice-Mixtral-v1
44
+ parameters:
45
+ weight: 0.36
46
+ density: 0.64
47
+ - model: Sao10K/Frostwind-Mixtral-v1
48
+ parameters:
49
+ weight: 0.22
50
+ density: 0.44
51
+ - model: KoboldAI/Mixtral-8x7B-Holodeck-v1
52
+ parameters:
53
+ weight: 0.21
54
+ density: 0.36
55
+ merge_method: dare_ties
56
+ dtype: bfloat16
57
+ ```