piercemaloney
commited on
Commit
•
8a7a7a4
1
Parent(s):
6e7b100
Upload MllamaForConditionalGeneration
Browse files
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "meta-llama/Llama-3.2-11B-Vision",
|
3 |
"architectures": [
|
4 |
"MllamaForConditionalGeneration"
|
5 |
],
|
@@ -30,7 +30,11 @@
|
|
30 |
"dropout": 0,
|
31 |
"early_stopping": false,
|
32 |
"encoder_no_repeat_ngram_size": 0,
|
33 |
-
"eos_token_id":
|
|
|
|
|
|
|
|
|
34 |
"exponential_decay_length_penalty": null,
|
35 |
"finetuning_task": null,
|
36 |
"forced_bos_token_id": null,
|
@@ -126,7 +130,7 @@
|
|
126 |
"0": "LABEL_0",
|
127 |
"1": "LABEL_1"
|
128 |
},
|
129 |
-
"image_size":
|
130 |
"initializer_range": 0.02,
|
131 |
"intermediate_layers_indices": [
|
132 |
3,
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "meta-llama/Llama-3.2-11B-Vision-Instruct",
|
3 |
"architectures": [
|
4 |
"MllamaForConditionalGeneration"
|
5 |
],
|
|
|
30 |
"dropout": 0,
|
31 |
"early_stopping": false,
|
32 |
"encoder_no_repeat_ngram_size": 0,
|
33 |
+
"eos_token_id": [
|
34 |
+
128001,
|
35 |
+
128008,
|
36 |
+
128009
|
37 |
+
],
|
38 |
"exponential_decay_length_penalty": null,
|
39 |
"finetuning_task": null,
|
40 |
"forced_bos_token_id": null,
|
|
|
130 |
"0": "LABEL_0",
|
131 |
"1": "LABEL_1"
|
132 |
},
|
133 |
+
"image_size": 560,
|
134 |
"initializer_range": 0.02,
|
135 |
"intermediate_layers_indices": [
|
136 |
3,
|
generation_config.json
CHANGED
@@ -1,7 +1,13 @@
|
|
1 |
{
|
2 |
-
"_from_model_config": true,
|
3 |
"bos_token_id": 128000,
|
4 |
-
"
|
|
|
|
|
|
|
|
|
|
|
5 |
"pad_token_id": 128004,
|
|
|
|
|
6 |
"transformers_version": "4.45.1"
|
7 |
}
|
|
|
1 |
{
|
|
|
2 |
"bos_token_id": 128000,
|
3 |
+
"do_sample": true,
|
4 |
+
"eos_token_id": [
|
5 |
+
128001,
|
6 |
+
128008,
|
7 |
+
128009
|
8 |
+
],
|
9 |
"pad_token_id": 128004,
|
10 |
+
"temperature": 0.6,
|
11 |
+
"top_p": 0.9,
|
12 |
"transformers_version": "4.45.1"
|
13 |
}
|
model-00001-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d183d799f0476061adc580ccc0a24abba5ab274d610f02e73aff7576e581118c
|
3 |
+
size 4992622346
|
model-00002-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:51a7afac7fbc248a8b038709ec3fb5bae4e5588ea84130056c8ed4e5a5144b40
|
3 |
+
size 4966251712
|
model-00003-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4915919704
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:24149e6a1922551067d64d78d8a41ac8393e502d6b64f551a4f73bbacbb5bf90
|
3 |
size 4915919704
|
model-00004-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999823980
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:caf17798db8a5de475bc7580706956032533cd7a9dfb588ad8488229c745bbad
|
3 |
size 4999823980
|
model-00005-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1465943128
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:656566a9aeee1115ca174c16ea3c51a2839b3a7f14cc349805aa0ef5c7433e7b
|
3 |
size 1465943128
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"language_model.lm_head.weight": "model-00005-of-00005.safetensors",
|
@@ -353,10 +353,10 @@
|
|
353 |
"language_model.model.layers.5.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
354 |
"language_model.model.layers.5.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
355 |
"language_model.model.layers.5.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
356 |
-
"language_model.model.layers.5.self_attn.k_proj.weight": "model-
|
357 |
-
"language_model.model.layers.5.self_attn.o_proj.weight": "model-
|
358 |
"language_model.model.layers.5.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
359 |
-
"language_model.model.layers.5.self_attn.v_proj.weight": "model-
|
360 |
"language_model.model.layers.6.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
361 |
"language_model.model.layers.6.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
362 |
"language_model.model.layers.6.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 21340441670
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"language_model.lm_head.weight": "model-00005-of-00005.safetensors",
|
|
|
353 |
"language_model.model.layers.5.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
354 |
"language_model.model.layers.5.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
355 |
"language_model.model.layers.5.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
356 |
+
"language_model.model.layers.5.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
357 |
+
"language_model.model.layers.5.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
358 |
"language_model.model.layers.5.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
359 |
+
"language_model.model.layers.5.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
360 |
"language_model.model.layers.6.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
361 |
"language_model.model.layers.6.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
362 |
"language_model.model.layers.6.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|