Darkhn commited on
Commit
14b54e4
·
verified ·
1 Parent(s): c6ce022

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +58 -3
README.md CHANGED
@@ -1,3 +1,58 @@
1
- ---
2
- license: llama3.3
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model_relation: quantized
3
+ library_name: transformers
4
+ tags:
5
+ - mergekit
6
+ - merge
7
+ base_model:
8
+ - TareksLab/Deviant-EXPERIMENTAL-V3-70B
9
+ ---
10
+ # Deviant
11
+
12
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
13
+
14
+ ## Merge Details
15
+ ### Merge Method
16
+
17
+ This model was merged using the [Linear DELLA](https://arxiv.org/abs/2406.11617) merge method using [huihui-ai/Llama-3.3-70B-Instruct-abliterated](https://huggingface.co/huihui-ai/Llama-3.3-70B-Instruct-abliterated) as a base.
18
+
19
+ ### Models Merged
20
+
21
+ The following models were included in the merge:
22
+ * [ReadyArt/Forgotten-Safeword-70B-3.6](https://huggingface.co/ReadyArt/Forgotten-Safeword-70B-3.6)
23
+ * [allura-org/Bigger-Body-70b](https://huggingface.co/allura-org/Bigger-Body-70b)
24
+ * [SicariusSicariiStuff/Negative_LLAMA_70B](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B)
25
+ * [TheDrummer/Fallen-Llama-3.3-R1-70B-v1](https://huggingface.co/TheDrummer/Fallen-Llama-3.3-R1-70B-v1)
26
+
27
+ ### Configuration
28
+
29
+ The following YAML configuration was used to produce this model:
30
+
31
+ ```yaml
32
+ models:
33
+ - model: TheDrummer/Fallen-Llama-3.3-R1-70B-v1
34
+ parameters:
35
+ weight: 0.25
36
+ density: 0.7
37
+ - model: ReadyArt/Forgotten-Safeword-70B-3.6
38
+ parameters:
39
+ weight: 0.25
40
+ density: 0.7
41
+ - model: allura-org/Bigger-Body-70b
42
+ parameters:
43
+ weight: 0.25
44
+ density: 0.7
45
+ - model: SicariusSicariiStuff/Negative_LLAMA_70B
46
+ parameters:
47
+ weight: 0.25
48
+ density: 0.7
49
+ merge_method: della_linear
50
+ base_model: huihui-ai/Llama-3.3-70B-Instruct-abliterated
51
+ parameters:
52
+ epsilon: 0.2
53
+ lambda: 1.1
54
+ int8_mask: true
55
+ dtype: bfloat16
56
+ tokenizer:
57
+ source: base
58
+ ```