Upload folder using huggingface_hub
Browse files- README.md +8 -8
- config.json +2 -9
- mergekit_config.yml +4 -4
- model-00001-of-00005.safetensors +1 -1
- model-00002-of-00005.safetensors +1 -1
- model-00003-of-00005.safetensors +1 -1
- model-00004-of-00005.safetensors +1 -1
- model-00005-of-00005.safetensors +1 -1
- special_tokens_map.json +23 -30
- tokenizer.json +2 -2
- tokenizer_config.json +0 -0
README.md
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
---
|
2 |
base_model:
|
3 |
-
-
|
4 |
-
- pot99rta/
|
5 |
library_name: transformers
|
6 |
tags:
|
7 |
- mergekit
|
@@ -15,12 +15,12 @@ This is a merge of pre-trained language models created using [mergekit](https://
|
|
15 |
## Merge Details
|
16 |
### Merge Method
|
17 |
|
18 |
-
This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [
|
19 |
|
20 |
### Models Merged
|
21 |
|
22 |
The following models were included in the merge:
|
23 |
-
* [
|
24 |
|
25 |
### Configuration
|
26 |
|
@@ -28,19 +28,19 @@ The following YAML configuration was used to produce this model:
|
|
28 |
|
29 |
```yaml
|
30 |
models:
|
31 |
-
- model:
|
32 |
#no parameters necessary for base model
|
33 |
-
- model:
|
34 |
parameters:
|
35 |
density: 0.5
|
36 |
weight: 0.5
|
37 |
-
- model:
|
38 |
parameters:
|
39 |
density: 0.5
|
40 |
weight: 0.5
|
41 |
|
42 |
merge_method: ties
|
43 |
-
base_model:
|
44 |
parameters:
|
45 |
normalize: false
|
46 |
int8_mask: true
|
|
|
1 |
---
|
2 |
base_model:
|
3 |
+
- Nitral-AI/Captain_BMO-12B
|
4 |
+
- pot99rta/CaptainMaid-12B-VioletMell-V0.420
|
5 |
library_name: transformers
|
6 |
tags:
|
7 |
- mergekit
|
|
|
15 |
## Merge Details
|
16 |
### Merge Method
|
17 |
|
18 |
+
This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [pot99rta/CaptainMaid-12B-VioletMell-V0.420](https://huggingface.co/pot99rta/CaptainMaid-12B-VioletMell-V0.420) as a base.
|
19 |
|
20 |
### Models Merged
|
21 |
|
22 |
The following models were included in the merge:
|
23 |
+
* [Nitral-AI/Captain_BMO-12B](https://huggingface.co/Nitral-AI/Captain_BMO-12B)
|
24 |
|
25 |
### Configuration
|
26 |
|
|
|
28 |
|
29 |
```yaml
|
30 |
models:
|
31 |
+
- model: pot99rta/CaptainMaid-12B-VioletMell-V0.420
|
32 |
#no parameters necessary for base model
|
33 |
+
- model: pot99rta/CaptainMaid-12B-VioletMell-V0.420
|
34 |
parameters:
|
35 |
density: 0.5
|
36 |
weight: 0.5
|
37 |
+
- model: Nitral-AI/Captain_BMO-12B
|
38 |
parameters:
|
39 |
density: 0.5
|
40 |
weight: 0.5
|
41 |
|
42 |
merge_method: ties
|
43 |
+
base_model: pot99rta/CaptainMaid-12B-VioletMell-V0.420
|
44 |
parameters:
|
45 |
normalize: false
|
46 |
int8_mask: true
|
config.json
CHANGED
@@ -10,7 +10,7 @@
|
|
10 |
"hidden_size": 5120,
|
11 |
"initializer_range": 0.02,
|
12 |
"intermediate_size": 14336,
|
13 |
-
"max_position_embeddings":
|
14 |
"model_type": "mistral",
|
15 |
"num_attention_heads": 32,
|
16 |
"num_hidden_layers": 40,
|
@@ -19,15 +19,8 @@
|
|
19 |
"rope_theta": 1000000.0,
|
20 |
"sliding_window": null,
|
21 |
"tie_word_embeddings": false,
|
22 |
-
"tokenizer": {
|
23 |
-
"add_bos_token": true,
|
24 |
-
"add_eos_token": false,
|
25 |
-
"pad_token": "</s>",
|
26 |
-
"padding_side": "right"
|
27 |
-
},
|
28 |
"torch_dtype": "float16",
|
29 |
"transformers_version": "4.51.3",
|
30 |
-
"
|
31 |
-
"use_cache": false,
|
32 |
"vocab_size": 131072
|
33 |
}
|
|
|
10 |
"hidden_size": 5120,
|
11 |
"initializer_range": 0.02,
|
12 |
"intermediate_size": 14336,
|
13 |
+
"max_position_embeddings": 1024000,
|
14 |
"model_type": "mistral",
|
15 |
"num_attention_heads": 32,
|
16 |
"num_hidden_layers": 40,
|
|
|
19 |
"rope_theta": 1000000.0,
|
20 |
"sliding_window": null,
|
21 |
"tie_word_embeddings": false,
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
"torch_dtype": "float16",
|
23 |
"transformers_version": "4.51.3",
|
24 |
+
"use_cache": true,
|
|
|
25 |
"vocab_size": 131072
|
26 |
}
|
mergekit_config.yml
CHANGED
@@ -1,17 +1,17 @@
|
|
1 |
models:
|
2 |
-
- model:
|
3 |
#no parameters necessary for base model
|
4 |
-
- model:
|
5 |
parameters:
|
6 |
density: 0.5
|
7 |
weight: 0.5
|
8 |
-
- model:
|
9 |
parameters:
|
10 |
density: 0.5
|
11 |
weight: 0.5
|
12 |
|
13 |
merge_method: ties
|
14 |
-
base_model:
|
15 |
parameters:
|
16 |
normalize: false
|
17 |
int8_mask: true
|
|
|
1 |
models:
|
2 |
+
- model: pot99rta/CaptainMaid-12B-VioletMell-V0.420
|
3 |
#no parameters necessary for base model
|
4 |
+
- model: pot99rta/CaptainMaid-12B-VioletMell-V0.420
|
5 |
parameters:
|
6 |
density: 0.5
|
7 |
weight: 0.5
|
8 |
+
- model: Nitral-AI/Captain_BMO-12B
|
9 |
parameters:
|
10 |
density: 0.5
|
11 |
weight: 0.5
|
12 |
|
13 |
merge_method: ties
|
14 |
+
base_model: pot99rta/CaptainMaid-12B-VioletMell-V0.420
|
15 |
parameters:
|
16 |
normalize: false
|
17 |
int8_mask: true
|
model-00001-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4865489296
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f4a7718bd794395287f2f8cb67aac109c492c875fd49aaa1d48c8e4562b7995c
|
3 |
size 4865489296
|
model-00002-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4907529376
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6411e28f7169e465931a099ed80973106fa11617dccddbe83604f58bf1200241
|
3 |
size 4907529376
|
model-00003-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4907529384
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:270e617502ac6b340d8ea44e99980eabc4a6f589b5573b14aec113f169a43773
|
3 |
size 4907529384
|
model-00004-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4907529376
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:89cf962167d791e00cbd33159ad5fbca7ef04bb72986035190d5fe162675727d
|
3 |
size 4907529376
|
model-00005-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4907529312
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:98f648a92463037d396f16fe314f2e93aa41da8c95635f246508e744c0022a01
|
3 |
size 4907529312
|
special_tokens_map.json
CHANGED
@@ -1,30 +1,23 @@
|
|
1 |
-
{
|
2 |
-
"bos_token": {
|
3 |
-
"content": "<s>",
|
4 |
-
"lstrip": false,
|
5 |
-
"normalized": false,
|
6 |
-
"rstrip": false,
|
7 |
-
"single_word": false
|
8 |
-
},
|
9 |
-
"eos_token": {
|
10 |
-
"content": "</s>",
|
11 |
-
"lstrip": false,
|
12 |
-
"normalized": false,
|
13 |
-
"rstrip": false,
|
14 |
-
"single_word": false
|
15 |
-
},
|
16 |
-
"
|
17 |
-
"content": "
|
18 |
-
"lstrip": false,
|
19 |
-
"normalized": false,
|
20 |
-
"rstrip": false,
|
21 |
-
"single_word": false
|
22 |
-
}
|
23 |
-
|
24 |
-
"content": "<unk>",
|
25 |
-
"lstrip": false,
|
26 |
-
"normalized": false,
|
27 |
-
"rstrip": false,
|
28 |
-
"single_word": false
|
29 |
-
}
|
30 |
-
}
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"unk_token": {
|
17 |
+
"content": "<unk>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
}
|
23 |
+
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
tokenizer.json
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f0a6f254fae49eaf5d67dfdb8b895c261220dee4df2b9cf1002ff95ee7b3c0a
|
3 |
+
size 9674069
|
tokenizer_config.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|