danielhanchen commited on
Commit
25a809a
·
verified ·
1 Parent(s): efc696d

Add files using upload-large-folder tool

Browse files
Qwen3-30B-A3B-128K-UD-IQ1_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39399820bc5b7a865b8918b0aba1ab49502bf73ce2f8663ff40080aff33b2c2f
3
- size 9043301408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c29ae173f2f87c22b7600ace863d3660910cb6affc6f458860e9778919e2c834
3
+ size 9019970080
Qwen3-30B-A3B-128K-UD-IQ2_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:45a05db956461a302f4932bb9054f3d656cba13e5c65eacd086b7f838554b8ca
3
- size 10866889760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ef27af0a6f0855d51396c1589c3a1383bbb307a0311304114b50a83f882cfb0
3
+ size 10865578528
Qwen3-30B-A3B-128K-UD-Q2_K_XL.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:451f22a03f828f608c13a30e34ff83c2de6cab199fb147125e7f3a90bd3212bc
3
- size 11815326752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1defc6d4a7b640a648b9bf89d4f568952006c5f4f6daccda27317dc1428e2965
3
+ size 11814277664
Qwen3-30B-A3B-128K-UD-Q4_K_XL.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7be9b102fddcea8b575dfd732df08e0063c020dbaf20fbe4adf22d6f3819b833
3
- size 17715663904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b03db3ded6d0542b70cc734cd90835c8fb9f35b5b58f2c3a4b7c1e0b3659a5d
3
+ size 17714614816
config.json CHANGED
@@ -4,7 +4,6 @@
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
- "bos_token_id": 151643,
8
  "decoder_sparse_step": 1,
9
  "eos_token_id": 151645,
10
  "head_dim": 128,
@@ -12,7 +11,7 @@
12
  "hidden_size": 2048,
13
  "initializer_range": 0.02,
14
  "intermediate_size": 6144,
15
- "max_position_embeddings": 131072,
16
  "max_window_layers": 48,
17
  "mlp_only_layers": [],
18
  "model_type": "qwen3_moe",
@@ -24,6 +23,7 @@
24
  "num_hidden_layers": 48,
25
  "num_key_value_heads": 4,
26
  "output_router_logits": false,
 
27
  "rms_norm_eps": 1e-06,
28
  "rope_scaling": {
29
  "factor": 4.0,
@@ -36,8 +36,9 @@
36
  "sliding_window": null,
37
  "tie_word_embeddings": false,
38
  "torch_dtype": "bfloat16",
39
- "transformers_version": "4.51.0",
 
40
  "use_cache": true,
41
  "use_sliding_window": false,
42
  "vocab_size": 151936
43
- }
 
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
 
7
  "decoder_sparse_step": 1,
8
  "eos_token_id": 151645,
9
  "head_dim": 128,
 
11
  "hidden_size": 2048,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 6144,
14
+ "max_position_embeddings": 40960,
15
  "max_window_layers": 48,
16
  "mlp_only_layers": [],
17
  "model_type": "qwen3_moe",
 
23
  "num_hidden_layers": 48,
24
  "num_key_value_heads": 4,
25
  "output_router_logits": false,
26
+ "pad_token_id": 151654,
27
  "rms_norm_eps": 1e-06,
28
  "rope_scaling": {
29
  "factor": 4.0,
 
36
  "sliding_window": null,
37
  "tie_word_embeddings": false,
38
  "torch_dtype": "bfloat16",
39
+ "transformers_version": "4.52.0.dev0",
40
+ "unsloth_fixed": true,
41
  "use_cache": true,
42
  "use_sliding_window": false,
43
  "vocab_size": 151936
44
+ }