lunarflu HF Staff commited on
Commit
516f273
·
verified ·
1 Parent(s): 0e2d32c

add credit to original author

Browse files
Files changed (1) hide show
  1. README.md +62 -60
README.md CHANGED
@@ -1,61 +1,63 @@
1
- ---
2
- license: llama2
3
- language:
4
- - en
5
- pipeline_tag: conversational
6
- tags:
7
- - merge
8
- ---
9
- # Goliath 120B
10
-
11
- An auto-regressive causal LM created by combining 2x finetuned [Llama-2 70B](https://huggingface.co/meta-llama/llama-2-70b-hf) into one.
12
-
13
- Please check out the quantized formats provided by [@TheBloke](https:///huggingface.co/TheBloke) and [@Panchovix](https://huggingface.co/Panchovix):
14
-
15
- - [GGUF](https://huggingface.co/TheBloke/goliath-120b-GGUF) (llama.cpp)
16
- - [GPTQ](https://huggingface.co/TheBloke/goliath-120b-GPTQ) (KoboldAI, TGW, Aphrodite)
17
- - [AWQ](https://huggingface.co/TheBloke/goliath-120b-AWQ) (TGW, Aphrodite, vLLM)
18
- - [Exllamav2](https://huggingface.co/Panchovix/goliath-120b-exl2) (TGW, KoboldAI)
19
-
20
- # Prompting Format
21
-
22
- Both Vicuna and Alpaca will work, but due the initial and final layers belonging primarily to Xwin, I expect Vicuna to work the best.
23
-
24
- # Merge process
25
-
26
- The models used in the merge are [Xwin](https://huggingface.co/Xwin-LM/Xwin-LM-70B-V0.1) and [Euryale](https://huggingface.co/Sao10K/Euryale-1.3-L2-70B).
27
-
28
- The layer ranges used are as follows:
29
-
30
- ```yaml
31
- - range 0, 16
32
- Xwin
33
- - range 8, 24
34
- Euryale
35
- - range 17, 32
36
- Xwin
37
- - range 25, 40
38
- Euryale
39
- - range 33, 48
40
- Xwin
41
- - range 41, 56
42
- Euryale
43
- - range 49, 64
44
- Xwin
45
- - range 57, 72
46
- Euryale
47
- - range 65, 80
48
- Xwin
49
- ```
50
-
51
- # Screenshots
52
-
53
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/635567189c72a7e742f1419c/Cat8_Rimaz6Ni7YhQiiGB.png)
54
-
55
- # Benchmarks
56
- Coming soon.
57
-
58
- # Acknowledgements
59
- Credits goes to [@chargoddard](https://huggingface.co/chargoddard) for developing the framework used to merge the model - [mergekit](https://github.com/cg123/mergekit).
60
-
 
 
61
  Special thanks to [@Undi95](https://huggingface.co/Undi95) for helping with the merge ratios.
 
1
+ ---
2
+ license: llama2
3
+ language:
4
+ - en
5
+ pipeline_tag: conversational
6
+ tags:
7
+ - merge
8
+ ---
9
+ # Goliath 120B
10
+
11
+ Credit to [alpindale](https://huggingface.co/alpindale), the original model author!
12
+
13
+ An auto-regressive causal LM created by combining 2x finetuned [Llama-2 70B](https://huggingface.co/meta-llama/llama-2-70b-hf) into one.
14
+
15
+ Please check out the quantized formats provided by [@TheBloke](https:///huggingface.co/TheBloke) and [@Panchovix](https://huggingface.co/Panchovix):
16
+
17
+ - [GGUF](https://huggingface.co/TheBloke/goliath-120b-GGUF) (llama.cpp)
18
+ - [GPTQ](https://huggingface.co/TheBloke/goliath-120b-GPTQ) (KoboldAI, TGW, Aphrodite)
19
+ - [AWQ](https://huggingface.co/TheBloke/goliath-120b-AWQ) (TGW, Aphrodite, vLLM)
20
+ - [Exllamav2](https://huggingface.co/Panchovix/goliath-120b-exl2) (TGW, KoboldAI)
21
+
22
+ # Prompting Format
23
+
24
+ Both Vicuna and Alpaca will work, but due the initial and final layers belonging primarily to Xwin, the original author expects Vicuna to work the best.
25
+
26
+ # Merge process
27
+
28
+ The models used in the merge are [Xwin](https://huggingface.co/Xwin-LM/Xwin-LM-70B-V0.1) and [Euryale](https://huggingface.co/Sao10K/Euryale-1.3-L2-70B).
29
+
30
+ The layer ranges used are as follows:
31
+
32
+ ```yaml
33
+ - range 0, 16
34
+ Xwin
35
+ - range 8, 24
36
+ Euryale
37
+ - range 17, 32
38
+ Xwin
39
+ - range 25, 40
40
+ Euryale
41
+ - range 33, 48
42
+ Xwin
43
+ - range 41, 56
44
+ Euryale
45
+ - range 49, 64
46
+ Xwin
47
+ - range 57, 72
48
+ Euryale
49
+ - range 65, 80
50
+ Xwin
51
+ ```
52
+
53
+ # Screenshots
54
+
55
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/635567189c72a7e742f1419c/Cat8_Rimaz6Ni7YhQiiGB.png)
56
+
57
+ # Benchmarks
58
+ Coming soon.
59
+
60
+ # Acknowledgements
61
+ Credits goes to [@chargoddard](https://huggingface.co/chargoddard) for developing the framework used to merge the model - [mergekit](https://github.com/cg123/mergekit).
62
+
63
  Special thanks to [@Undi95](https://huggingface.co/Undi95) for helping with the merge ratios.