InferenceIllusionist commited on
Commit
bf9e151
·
verified ·
1 Parent(s): b107893

Create readme.md

Browse files
Files changed (1) hide show
  1. README.md +80 -0
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: []
3
+ tags:
4
+ - gguf
5
+ ---
6
+
7
+ # friend-8x7B-GGUF
8
+
9
+ Quantized from fp16 with love.
10
+
11
+ Uploading Q5_K_M for starters, other sizes available upon request.
12
+
13
+ See original model card details below.
14
+
15
+
16
+ ---
17
+ # friend-8x7B-hf
18
+ ![friends](https://files.catbox.moe/qlwy3m.png)
19
+
20
+ Several-staged merge this time. Actually decent, from my testings.
21
+
22
+ Use ChatML or Alpaca, both seemed to work though I liked the outputs from ChatML more.
23
+
24
+ ### temp-output-base:
25
+ ```yaml
26
+ models:
27
+ - model: mistralai/Mixtral-8x7B-v0.1+/ai/LLM/tmp/pefts/daybreak-peft/mixtral-8x7b
28
+ parameters:
29
+ weight: 0.65
30
+ - model: mistralai/Mixtral-8x7B-v0.1+retrieval-bar/Mixtral-8x7B-v0.1_case-briefs
31
+ parameters:
32
+ weight: 0.25
33
+ merge_method: task_arithmetic
34
+ base_model: mistralai/Mixtral-8x7B-v0.1
35
+ dtype: float16
36
+ ```
37
+
38
+ ### temp-output-instruct:
39
+ ```yaml
40
+ models:
41
+ - model: mistralai/Mixtral-8x7B-Instruct-v0.1+SeanWu25/Mixtral_8x7b_Medicine
42
+ parameters:
43
+ weight: 0.33
44
+ - model: NeverSleep/Noromaid-v0.4-Mixtral-Instruct-8x7b-Zloss
45
+ parameters:
46
+ weight: 0.15
47
+ merge_method: task_arithmetic
48
+ base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
49
+ dtype: float16
50
+ ```
51
+
52
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
53
+
54
+ ## Merge Details
55
+ ### Merge Method
56
+
57
+ This model was merged using the SLERP merge method.
58
+
59
+ ### Models Merged
60
+
61
+ The following models were included in the merge:
62
+ * ./temp-output-base
63
+ * ./temp-output-instruct
64
+
65
+ ### Configuration
66
+
67
+ The following YAML configuration was used to produce this model:
68
+
69
+ ```yaml
70
+ models:
71
+ - model: ./temp-output-base
72
+ - model: ./temp-output-instruct
73
+ merge_method: slerp
74
+ base_model: ./temp-output-base
75
+ parameters:
76
+ t:
77
+ - value: 0.5
78
+ dtype: float16
79
+
80
+ ```