pot99rta commited on
Commit
b9221c5
·
verified ·
1 Parent(s): 8168c6f

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  base_model:
3
- - ReadyArt/Omega-Darker_The-Final-Directive-12B
4
- - pot99rta/MagcarpMell-ThinkandReasoner-12B
5
  library_name: transformers
6
  tags:
7
  - mergekit
@@ -15,12 +15,12 @@ This is a merge of pre-trained language models created using [mergekit](https://
15
  ## Merge Details
16
  ### Merge Method
17
 
18
- This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [ReadyArt/Omega-Darker_The-Final-Directive-12B](https://huggingface.co/ReadyArt/Omega-Darker_The-Final-Directive-12B) as a base.
19
 
20
  ### Models Merged
21
 
22
  The following models were included in the merge:
23
- * [pot99rta/MagcarpMell-ThinkandReasoner-12B](https://huggingface.co/pot99rta/MagcarpMell-ThinkandReasoner-12B)
24
 
25
  ### Configuration
26
 
@@ -28,19 +28,19 @@ The following YAML configuration was used to produce this model:
28
 
29
  ```yaml
30
  models:
31
- - model: ReadyArt/Omega-Darker_The-Final-Directive-12B
32
  #no parameters necessary for base model
33
- - model: ReadyArt/Omega-Darker_The-Final-Directive-12B
34
  parameters:
35
  density: 0.5
36
  weight: 0.5
37
- - model: pot99rta/MagcarpMell-ThinkandReasoner-12B
38
  parameters:
39
  density: 0.5
40
  weight: 0.5
41
 
42
  merge_method: ties
43
- base_model: ReadyArt/Omega-Darker_The-Final-Directive-12B
44
  parameters:
45
  normalize: false
46
  int8_mask: true
 
1
  ---
2
  base_model:
3
+ - Nitral-AI/Captain_BMO-12B
4
+ - pot99rta/CaptainMaid-12B-VioletMell-V0.420
5
  library_name: transformers
6
  tags:
7
  - mergekit
 
15
  ## Merge Details
16
  ### Merge Method
17
 
18
+ This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [pot99rta/CaptainMaid-12B-VioletMell-V0.420](https://huggingface.co/pot99rta/CaptainMaid-12B-VioletMell-V0.420) as a base.
19
 
20
  ### Models Merged
21
 
22
  The following models were included in the merge:
23
+ * [Nitral-AI/Captain_BMO-12B](https://huggingface.co/Nitral-AI/Captain_BMO-12B)
24
 
25
  ### Configuration
26
 
 
28
 
29
  ```yaml
30
  models:
31
+ - model: pot99rta/CaptainMaid-12B-VioletMell-V0.420
32
  #no parameters necessary for base model
33
+ - model: pot99rta/CaptainMaid-12B-VioletMell-V0.420
34
  parameters:
35
  density: 0.5
36
  weight: 0.5
37
+ - model: Nitral-AI/Captain_BMO-12B
38
  parameters:
39
  density: 0.5
40
  weight: 0.5
41
 
42
  merge_method: ties
43
+ base_model: pot99rta/CaptainMaid-12B-VioletMell-V0.420
44
  parameters:
45
  normalize: false
46
  int8_mask: true
config.json CHANGED
@@ -10,7 +10,7 @@
10
  "hidden_size": 5120,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 14336,
13
- "max_position_embeddings": 131072,
14
  "model_type": "mistral",
15
  "num_attention_heads": 32,
16
  "num_hidden_layers": 40,
@@ -19,15 +19,8 @@
19
  "rope_theta": 1000000.0,
20
  "sliding_window": null,
21
  "tie_word_embeddings": false,
22
- "tokenizer": {
23
- "add_bos_token": true,
24
- "add_eos_token": false,
25
- "pad_token": "</s>",
26
- "padding_side": "right"
27
- },
28
  "torch_dtype": "float16",
29
  "transformers_version": "4.51.3",
30
- "trust_remote_code": true,
31
- "use_cache": false,
32
  "vocab_size": 131072
33
  }
 
10
  "hidden_size": 5120,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 14336,
13
+ "max_position_embeddings": 1024000,
14
  "model_type": "mistral",
15
  "num_attention_heads": 32,
16
  "num_hidden_layers": 40,
 
19
  "rope_theta": 1000000.0,
20
  "sliding_window": null,
21
  "tie_word_embeddings": false,
 
 
 
 
 
 
22
  "torch_dtype": "float16",
23
  "transformers_version": "4.51.3",
24
+ "use_cache": true,
 
25
  "vocab_size": 131072
26
  }
mergekit_config.yml CHANGED
@@ -1,17 +1,17 @@
1
  models:
2
- - model: ReadyArt/Omega-Darker_The-Final-Directive-12B
3
  #no parameters necessary for base model
4
- - model: ReadyArt/Omega-Darker_The-Final-Directive-12B
5
  parameters:
6
  density: 0.5
7
  weight: 0.5
8
- - model: pot99rta/MagcarpMell-ThinkandReasoner-12B
9
  parameters:
10
  density: 0.5
11
  weight: 0.5
12
 
13
  merge_method: ties
14
- base_model: ReadyArt/Omega-Darker_The-Final-Directive-12B
15
  parameters:
16
  normalize: false
17
  int8_mask: true
 
1
  models:
2
+ - model: pot99rta/CaptainMaid-12B-VioletMell-V0.420
3
  #no parameters necessary for base model
4
+ - model: pot99rta/CaptainMaid-12B-VioletMell-V0.420
5
  parameters:
6
  density: 0.5
7
  weight: 0.5
8
+ - model: Nitral-AI/Captain_BMO-12B
9
  parameters:
10
  density: 0.5
11
  weight: 0.5
12
 
13
  merge_method: ties
14
+ base_model: pot99rta/CaptainMaid-12B-VioletMell-V0.420
15
  parameters:
16
  normalize: false
17
  int8_mask: true
model-00001-of-00005.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a20a8dc3344250471eda8d13cea9cfae2f0ebcec5267ed94e8a7efc871d46d1
3
  size 4865489296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4a7718bd794395287f2f8cb67aac109c492c875fd49aaa1d48c8e4562b7995c
3
  size 4865489296
model-00002-of-00005.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0eb97d1269afb7cd054adff202d34dbaf98031551bad6a2ed61b1cac437a3f70
3
  size 4907529376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6411e28f7169e465931a099ed80973106fa11617dccddbe83604f58bf1200241
3
  size 4907529376
model-00003-of-00005.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b69ef88d502280c953ab211ff76efcbe8afb0732a241aef28fd8863bfd28d92
3
  size 4907529384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:270e617502ac6b340d8ea44e99980eabc4a6f589b5573b14aec113f169a43773
3
  size 4907529384
model-00004-of-00005.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34ae7a91f375a7d0d9ce996172a73878a47762def29fe261df89fc68bc780ac1
3
  size 4907529376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89cf962167d791e00cbd33159ad5fbca7ef04bb72986035190d5fe162675727d
3
  size 4907529376
model-00005-of-00005.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79c669b384fa754445be5bdbfe2321f5878c17b79ba7c6307c86ce8188c71326
3
  size 4907529312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98f648a92463037d396f16fe314f2e93aa41da8c95635f246508e744c0022a01
3
  size 4907529312
special_tokens_map.json CHANGED
@@ -1,30 +1,23 @@
1
- {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "</s>",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "</s>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "unk_token": {
24
- "content": "<unk>",
25
- "lstrip": false,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- }
30
- }
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
 
 
 
 
 
 
 
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0240ce510f08e6c2041724e9043e33be9d251d1e4a4d94eb68cd47b954b61d2
3
- size 17078292
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f0a6f254fae49eaf5d67dfdb8b895c261220dee4df2b9cf1002ff95ee7b3c0a
3
+ size 9674069
tokenizer_config.json CHANGED
The diff for this file is too large to render. See raw diff