danielhanchen commited on
Commit
024ae12
·
verified ·
1 Parent(s): 3fbadf0

Add files using upload-large-folder tool

Browse files
Qwen3-30B-A3B-128K-Q4_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c3e51a3733f557b08e2fe5c919ce76ae75fcd9fa0d548c725459c59956133f7c
3
- size 18556686976
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f80077489eb6816e8f0178bea4779087822a6c364ebb9af23600001cd13029fe
3
+ size 18556687136
Qwen3-30B-A3B-128K-UD-Q2_K_XL.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c6dd7e90a87c64b99610b339134f4b6b65a55a78ae0383eb4824f29febcbb4a3
3
- size 11814277760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cf3f35e28e33d04fb02487f717c5f46141cb856985b834c41b6c4d1926f8e5d
3
+ size 11814277920
Qwen3-30B-A3B-128K-UD-Q3_K_XL.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2d73ac5491af3c7cb239eaf72468bc51167e29ee7cf9a1465f89cc0dffa2eac
3
- size 13833048704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:246ac4707b56e32d4f572a3fe4cd8ea338e62a9d5b46f1f47007efd6122a1a21
3
+ size 13833048864
Qwen3-30B-A3B-128K-UD-Q4_K_XL.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2eeedaea0aeedc3ada77304de6b673cb723a65bdc3c9a4e4015ce5e3c594b43c
3
- size 17715663488
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae835231266d900166b35ea1e6c065729a34316ce5c86335e08e209a601e7ecc
3
+ size 17715663648
Qwen3-30B-A3B-128K-UD-Q5_K_XL.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6253babd83cb843516b8df9bf36fc182627f50c6fb12f5ed216cd4b1a1f24544
3
- size 21740302976
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e006e1cbeae428bd90bb77fb2df6acda82ab925c05ab0ecd23bee61a74c55f04
3
+ size 21740303136
config.json CHANGED
@@ -36,7 +36,7 @@
36
  "sliding_window": null,
37
  "tie_word_embeddings": false,
38
  "torch_dtype": "bfloat16",
39
- "transformers_version": "4.52.0.dev0",
40
  "unsloth_fixed": true,
41
  "use_cache": true,
42
  "use_sliding_window": false,
 
36
  "sliding_window": null,
37
  "tie_word_embeddings": false,
38
  "torch_dtype": "bfloat16",
39
+ "transformers_version": "4.51.3",
40
  "unsloth_fixed": true,
41
  "use_cache": true,
42
  "use_sliding_window": false,