qgallouedec HF Staff commited on
Commit
0d62778
·
verified ·
1 Parent(s): 722c4a6

Upload Idefics3ForConditionalGeneration

Browse files
Files changed (3) hide show
  1. config.json +5 -3
  2. generation_config.json +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -2,6 +2,7 @@
2
  "architectures": [
3
  "Idefics3ForConditionalGeneration"
4
  ],
 
5
  "image_token_id": 128257,
6
  "model_type": "idefics3",
7
  "pad_token_id": 128002,
@@ -19,6 +20,7 @@
19
  "AutoModelForCausalLM": "modeling_idefics3.Idefics3ForVisionText2Text"
20
  },
21
  "bos_token_id": 128000,
 
22
  "eos_token_id": [
23
  128001,
24
  128008,
@@ -29,6 +31,7 @@
29
  "hidden_size": 16,
30
  "initializer_range": 0.02,
31
  "intermediate_size": 14336,
 
32
  "max_position_embeddings": 131072,
33
  "mlp_bias": false,
34
  "model_type": "llama",
@@ -120,17 +123,16 @@
120
  "rope_type": "llama3"
121
  },
122
  "rope_theta": 500000.0,
123
- "torch_dtype": "bfloat16",
124
  "use_cache": true,
125
  "use_resampler": false,
126
  "vocab_size": 128259
127
  },
128
  "tie_word_embeddings": false,
129
- "torch_dtype": "bfloat16",
130
- "transformers_version": "4.56.0.dev0",
131
  "use_cache": true,
132
  "vision_config": {
133
  "attention_dropout": 0.0,
 
134
  "hidden_act": "gelu_pytorch_tanh",
135
  "hidden_size": 16,
136
  "image_size": 364,
 
2
  "architectures": [
3
  "Idefics3ForConditionalGeneration"
4
  ],
5
+ "dtype": "bfloat16",
6
  "image_token_id": 128257,
7
  "model_type": "idefics3",
8
  "pad_token_id": 128002,
 
20
  "AutoModelForCausalLM": "modeling_idefics3.Idefics3ForVisionText2Text"
21
  },
22
  "bos_token_id": 128000,
23
+ "dtype": "bfloat16",
24
  "eos_token_id": [
25
  128001,
26
  128008,
 
31
  "hidden_size": 16,
32
  "initializer_range": 0.02,
33
  "intermediate_size": 14336,
34
+ "layer_types": null,
35
  "max_position_embeddings": 131072,
36
  "mlp_bias": false,
37
  "model_type": "llama",
 
123
  "rope_type": "llama3"
124
  },
125
  "rope_theta": 500000.0,
 
126
  "use_cache": true,
127
  "use_resampler": false,
128
  "vocab_size": 128259
129
  },
130
  "tie_word_embeddings": false,
131
+ "transformers_version": "4.57.0.dev0",
 
132
  "use_cache": true,
133
  "vision_config": {
134
  "attention_dropout": 0.0,
135
+ "embed_dim": 32,
136
  "hidden_act": "gelu_pytorch_tanh",
137
  "hidden_size": 16,
138
  "image_size": 364,
generation_config.json CHANGED
@@ -7,5 +7,5 @@
7
  128009
8
  ],
9
  "pad_token_id": 128002,
10
- "transformers_version": "4.56.0.dev0"
11
  }
 
7
  128009
8
  ],
9
  "pad_token_id": 128002,
10
+ "transformers_version": "4.57.0.dev0"
11
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a661eb4c15afccb9c68e712860b057ce8fdf886c5a0ee655c130a44d406dfc07
3
  size 11682096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b27ea2b8b1c0a8ec9a57b425e85c05f8832cfcf202d0945a879ffd675143c3db
3
  size 11682096