CorticalStack
commited on
Commit
•
a36dd64
1
Parent(s):
c3648ba
Update README.md
Browse files
README.md
CHANGED
@@ -3,6 +3,7 @@ license: apache-2.0
|
|
3 |
tags:
|
4 |
- merge
|
5 |
- mergekit
|
|
|
6 |
- mlabonne/AlphaMonarch-7B
|
7 |
- mlabonne/NeuralMonarch-7B
|
8 |
- macadeliccc/MBX-7B-v3-DPO
|
@@ -13,6 +14,7 @@ tags:
|
|
13 |
# pastiche-crown-clown-7B-dare
|
14 |
|
15 |
pastiche-crown-clown-7B-dare is a DARE merge of the following models using [mergekit](https://github.com/cg123/mergekit):
|
|
|
16 |
* [mlabonne/AlphaMonarch-7B](https://huggingface.co/mlabonne/AlphaMonarch-7B)
|
17 |
* [mlabonne/NeuralMonarch-7B](https://huggingface.co/mlabonne/NeuralMonarch-7B)
|
18 |
* [macadeliccc/MBX-7B-v3-DPO](https://huggingface.co/macadeliccc/MBX-7B-v3-DPO)
|
@@ -24,7 +26,6 @@ See the paper [Language Models are Super Mario: Absorbing Abilities from Homolog
|
|
24 |
```yaml
|
25 |
models:
|
26 |
- model: bardsai/jaskier-7b-dpo-v5.6
|
27 |
-
# No parameters necessary for base model
|
28 |
- model: mlabonne/AlphaMonarch-7B
|
29 |
parameters:
|
30 |
density: 0.53
|
|
|
3 |
tags:
|
4 |
- merge
|
5 |
- mergekit
|
6 |
+
- bardsai/jaskier-7b-dpo-v5.6
|
7 |
- mlabonne/AlphaMonarch-7B
|
8 |
- mlabonne/NeuralMonarch-7B
|
9 |
- macadeliccc/MBX-7B-v3-DPO
|
|
|
14 |
# pastiche-crown-clown-7B-dare
|
15 |
|
16 |
pastiche-crown-clown-7B-dare is a DARE merge of the following models using [mergekit](https://github.com/cg123/mergekit):
|
17 |
+
* [bardsai/jaskier-7b-dpo-v5.6](https://huggingface.co/bardsai/jaskier-7b-dpo-v5.6)
|
18 |
* [mlabonne/AlphaMonarch-7B](https://huggingface.co/mlabonne/AlphaMonarch-7B)
|
19 |
* [mlabonne/NeuralMonarch-7B](https://huggingface.co/mlabonne/NeuralMonarch-7B)
|
20 |
* [macadeliccc/MBX-7B-v3-DPO](https://huggingface.co/macadeliccc/MBX-7B-v3-DPO)
|
|
|
26 |
```yaml
|
27 |
models:
|
28 |
- model: bardsai/jaskier-7b-dpo-v5.6
|
|
|
29 |
- model: mlabonne/AlphaMonarch-7B
|
30 |
parameters:
|
31 |
density: 0.53
|