saumyaop commited on
Commit
111612d
·
verified ·
1 Parent(s): cc833b7

Upload Gemma3ForCausalLM

Browse files
config.json CHANGED
@@ -30,7 +30,7 @@
30
  "rope_theta": 1000000.0,
31
  "sliding_window": 1024,
32
  "sliding_window_pattern": 6,
33
- "torch_dtype": "float16",
34
  "transformers_version": "4.50.0.dev0",
35
  "use_cache": true,
36
  "vocab_size": 262208
 
30
  "rope_theta": 1000000.0,
31
  "sliding_window": 1024,
32
  "sliding_window_pattern": 6,
33
+ "torch_dtype": "bfloat16",
34
  "transformers_version": "4.50.0.dev0",
35
  "use_cache": true,
36
  "vocab_size": 262208
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:796c67c5874b0daaf29f4d367799834fc64b076c87d3aa52c88a9f985008a47f
3
- size 4960531096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7706149128ec37847418a18b23f2d1dbff47793ee4a57320d35197c0f1403c0
3
+ size 4960531344
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1fb86438fcfa7d073ca9119ef3a27ec2c80cc646eb7d021b1e1910edbf3ec992
3
- size 2800046480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbab1b243288d204b8ceeab729e343e3af2ccf9d5fd244c239e2d01eec012dd7
3
+ size 2800046672