RaushanTurganbay HF Staff commited on
Commit
39cdd30
·
verified ·
1 Parent(s): c1ecab2

Upload CosmosForConditionalGeneration

Browse files
config.json CHANGED
@@ -3,6 +3,7 @@
3
  "CosmosForConditionalGeneration"
4
  ],
5
  "image_token_id": 64000,
 
6
  "model_type": "cosmos",
7
  "prompt_encoder": {
8
  "d_ff": 65536,
@@ -15,7 +16,26 @@
15
  "num_layers": 24
16
  },
17
  "text_config": {
18
- "cross_attn_hidden_size": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  "model_type": "cosmos_text_model",
20
  "rope_latent_shape": [
21
  5,
@@ -25,7 +45,8 @@
25
  "rope_scaling": {
26
  "original_max_position_embeddings": 8192,
27
  "rope_type": "3d"
28
- }
 
29
  },
30
  "torch_dtype": "bfloat16",
31
  "transformers_version": "4.50.0.dev0",
 
3
  "CosmosForConditionalGeneration"
4
  ],
5
  "image_token_id": 64000,
6
+ "is_encoder_decoder": true,
7
  "model_type": "cosmos",
8
  "prompt_encoder": {
9
  "d_ff": 65536,
 
16
  "num_layers": 24
17
  },
18
  "text_config": {
19
+ "apply_abs_pos_emb": true,
20
+ "insert_cross_attn_layers": [
21
+ 0,
22
+ 1,
23
+ 2,
24
+ 3,
25
+ 4,
26
+ 5,
27
+ 6,
28
+ 7,
29
+ 8,
30
+ 9,
31
+ 10,
32
+ 11,
33
+ 12,
34
+ 13,
35
+ 14,
36
+ 15
37
+ ],
38
+ "is_video_to_world": true,
39
  "model_type": "cosmos_text_model",
40
  "rope_latent_shape": [
41
  5,
 
45
  "rope_scaling": {
46
  "original_max_position_embeddings": 8192,
47
  "rope_type": "3d"
48
+ },
49
+ "vocab_size": 64064
50
  },
51
  "torch_dtype": "bfloat16",
52
  "transformers_version": "4.50.0.dev0",
model-00001-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b6371999200570e90e77d2b0fc742ad671c07bd31b2d5ac1de8b82bc891a5bc
3
+ size 4975203704
model-00002-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07a2a1f136172805d6a6462149b7d526d406324deefb472d63d97c4c368c4195
3
+ size 4963223972
model-00003-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cabc19d4805efac77b65526b0f3a2e05fc9e18d303a8726e25c2204bd40f7774
3
+ size 4966116112
model-00004-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2990e629c7a411f9b7155637a9c8a94534d73c7a1ba296d0d4a2a4fabd662e98
3
+ size 4831893920
model-00005-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93aeaf399e2f8a3aed7109b045d3d9c77022d1b06e6da36ba77790b8fa5af369
3
+ size 4831893944
model-00006-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:968426beb6551684ff2c8457e59c52a0a409086810fba96a3e576d064fb6e630
3
+ size 4807766952
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff