InferenceIllusionist commited on
Commit
fe927e8
·
verified ·
1 Parent(s): dac091f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -0
README.md CHANGED
@@ -1,3 +1,49 @@
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
4
+
5
+ # Model Card for Envoid_Mixtral-Instruct-ITR-8x7B-GGUF
6
+
7
+ <!-- Provide a quick summary of what the model is/does. -->
8
+
9
+ Envoid_Mixtral-Instruct-ITR-8x7B quantized with love.
10
+
11
+ Starting out with Q4_K_M, future plans for imatrix IQ quants.
12
+
13
+ Original model card below for reference.
14
+
15
+ ---
16
+ license: cc-by-nc-4.0
17
+ ---
18
+ # Caution this model may be unpredictable
19
+ ![](https://files.catbox.moe/y8nv86.jpg)
20
+ ## Mixtral-Instruct-ITR (Interpolative Training Regression)
21
+
22
+ We have to go back, edition.
23
+
24
+ For this model I took what I learned in the making of [Cat-8x7B](https://huggingface.co/Envoid/Cat-8x7B) and went back to the very beginning and SLERP merged [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) onto [mistralai/Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1)
25
+
26
+ While the results aren't perfect the model feels more creative and less overcooked than Mixtral Instruct is often accused of being.
27
+
28
+ The hopes are that this should also have left the model much more receptive to additional finetuning and I am interested to see what comes of it so please feel free to download it and have fun.
29
+
30
+ Apologies about the small shard size (keep forgetting to change the mergekit config back)
31
+
32
+ ## The model is a lot less likely to refuse certain requests in this state:
33
+
34
+ so if you are going to apply additional finetuning to the model you may need to bolster its alignment depending on your use case.
35
+
36
+ The model still responds well to [INST] Thingie [/INST] formatting quite well.
37
+
38
+ Or if preferred this can easily be reproduced if you have both base and instruct models handy using mergekit (mixtral branch) with the following config
39
+ ```
40
+ models:
41
+ - model: ./mistralai_Mixtral-8x7B-Instruct-v0.1
42
+ - model: ./mistralai_Mixtral-8x7B-v0.1
43
+ merge_method: slerp
44
+ base_model: ./mistralai_Mixtral-8x7B-v0.1
45
+ parameters:
46
+ t:
47
+ - value: 0.5
48
+ dtype: float16
49
+ ```