Dorjzodovsuren commited on
Commit
8109d9b
·
verified ·
1 Parent(s): d3a94c0

Delete config.json

Browse files
Files changed (1) hide show
  1. config.json +0 -227
config.json DELETED
@@ -1,227 +0,0 @@
1
- {
2
- "_name_or_path": "meta-llama/Llama-3.2-11B-Vision-Instruct",
3
- "architectures": [
4
- "MllamaForConditionalGeneration"
5
- ],
6
- "image_token_index": 128256,
7
- "model_type": "mllama",
8
- "text_config": {
9
- "_attn_implementation_autoset": false,
10
- "_name_or_path": "",
11
- "add_cross_attention": false,
12
- "architectures": null,
13
- "bad_words_ids": null,
14
- "begin_suppress_tokens": null,
15
- "bos_token_id": 128000,
16
- "chunk_size_feed_forward": 0,
17
- "cross_attention_hidden_size": null,
18
- "cross_attention_layers": [
19
- 3,
20
- 8,
21
- 13,
22
- 18,
23
- 23,
24
- 28,
25
- 33,
26
- 38
27
- ],
28
- "decoder_start_token_id": null,
29
- "diversity_penalty": 0.0,
30
- "do_sample": false,
31
- "dropout": 0,
32
- "early_stopping": false,
33
- "encoder_no_repeat_ngram_size": 0,
34
- "eos_token_id": [
35
- 128001,
36
- 128008,
37
- 128009
38
- ],
39
- "exponential_decay_length_penalty": null,
40
- "finetuning_task": null,
41
- "forced_bos_token_id": null,
42
- "forced_eos_token_id": null,
43
- "hidden_act": "silu",
44
- "hidden_size": 4096,
45
- "id2label": {
46
- "0": "LABEL_0",
47
- "1": "LABEL_1"
48
- },
49
- "initializer_range": 0.02,
50
- "intermediate_size": 14336,
51
- "is_decoder": false,
52
- "is_encoder_decoder": false,
53
- "label2id": {
54
- "LABEL_0": 0,
55
- "LABEL_1": 1
56
- },
57
- "length_penalty": 1.0,
58
- "max_length": 20,
59
- "max_position_embeddings": 131072,
60
- "min_length": 0,
61
- "model_type": "mllama_text_model",
62
- "no_repeat_ngram_size": 0,
63
- "num_attention_heads": 32,
64
- "num_beam_groups": 1,
65
- "num_beams": 1,
66
- "num_hidden_layers": 40,
67
- "num_key_value_heads": 8,
68
- "num_return_sequences": 1,
69
- "output_attentions": false,
70
- "output_hidden_states": false,
71
- "output_scores": false,
72
- "pad_token_id": 128004,
73
- "prefix": null,
74
- "problem_type": null,
75
- "pruned_heads": {},
76
- "remove_invalid_values": false,
77
- "repetition_penalty": 1.0,
78
- "return_dict": true,
79
- "return_dict_in_generate": false,
80
- "rms_norm_eps": 1e-05,
81
- "rope_scaling": {
82
- "factor": 8.0,
83
- "high_freq_factor": 4.0,
84
- "low_freq_factor": 1.0,
85
- "original_max_position_embeddings": 8192,
86
- "rope_type": "llama3"
87
- },
88
- "rope_theta": 500000.0,
89
- "sep_token_id": null,
90
- "suppress_tokens": null,
91
- "task_specific_params": null,
92
- "temperature": 1.0,
93
- "tf_legacy_loss": false,
94
- "tie_encoder_decoder": false,
95
- "tie_word_embeddings": false,
96
- "tokenizer_class": null,
97
- "top_k": 50,
98
- "top_p": 1.0,
99
- "torch_dtype": "bfloat16",
100
- "torchscript": false,
101
- "typical_p": 1.0,
102
- "use_bfloat16": false,
103
- "use_cache": true,
104
- "vocab_size": 128256
105
- },
106
- "torch_dtype": "bfloat16",
107
- "transformers_version": "4.46.3",
108
- "unsloth_fixed": true,
109
- "vision_config": {
110
- "_attn_implementation_autoset": false,
111
- "_name_or_path": "",
112
- "add_cross_attention": false,
113
- "architectures": null,
114
- "attention_heads": 16,
115
- "bad_words_ids": null,
116
- "begin_suppress_tokens": null,
117
- "bos_token_id": null,
118
- "chunk_size_feed_forward": 0,
119
- "cross_attention_hidden_size": null,
120
- "decoder_start_token_id": null,
121
- "diversity_penalty": 0.0,
122
- "do_sample": false,
123
- "early_stopping": false,
124
- "encoder_no_repeat_ngram_size": 0,
125
- "eos_token_id": null,
126
- "exponential_decay_length_penalty": null,
127
- "finetuning_task": null,
128
- "forced_bos_token_id": null,
129
- "forced_eos_token_id": null,
130
- "hidden_act": "gelu",
131
- "hidden_size": 1280,
132
- "id2label": {
133
- "0": "LABEL_0",
134
- "1": "LABEL_1"
135
- },
136
- "image_size": 560,
137
- "initializer_range": 0.02,
138
- "intermediate_layers_indices": [
139
- 3,
140
- 7,
141
- 15,
142
- 23,
143
- 30
144
- ],
145
- "intermediate_size": 5120,
146
- "is_decoder": false,
147
- "is_encoder_decoder": false,
148
- "label2id": {
149
- "LABEL_0": 0,
150
- "LABEL_1": 1
151
- },
152
- "length_penalty": 1.0,
153
- "max_length": 20,
154
- "max_num_tiles": 4,
155
- "min_length": 0,
156
- "model_type": "mllama_vision_model",
157
- "no_repeat_ngram_size": 0,
158
- "norm_eps": 1e-05,
159
- "num_beam_groups": 1,
160
- "num_beams": 1,
161
- "num_channels": 3,
162
- "num_global_layers": 8,
163
- "num_hidden_layers": 32,
164
- "num_return_sequences": 1,
165
- "output_attentions": false,
166
- "output_hidden_states": false,
167
- "output_scores": false,
168
- "pad_token_id": null,
169
- "patch_size": 14,
170
- "prefix": null,
171
- "problem_type": null,
172
- "pruned_heads": {},
173
- "remove_invalid_values": false,
174
- "repetition_penalty": 1.0,
175
- "return_dict": true,
176
- "return_dict_in_generate": false,
177
- "sep_token_id": null,
178
- "supported_aspect_ratios": [
179
- [
180
- 1,
181
- 1
182
- ],
183
- [
184
- 1,
185
- 2
186
- ],
187
- [
188
- 1,
189
- 3
190
- ],
191
- [
192
- 1,
193
- 4
194
- ],
195
- [
196
- 2,
197
- 1
198
- ],
199
- [
200
- 2,
201
- 2
202
- ],
203
- [
204
- 3,
205
- 1
206
- ],
207
- [
208
- 4,
209
- 1
210
- ]
211
- ],
212
- "suppress_tokens": null,
213
- "task_specific_params": null,
214
- "temperature": 1.0,
215
- "tf_legacy_loss": false,
216
- "tie_encoder_decoder": false,
217
- "tie_word_embeddings": true,
218
- "tokenizer_class": null,
219
- "top_k": 50,
220
- "top_p": 1.0,
221
- "torch_dtype": "bfloat16",
222
- "torchscript": false,
223
- "typical_p": 1.0,
224
- "use_bfloat16": false,
225
- "vision_output_dim": 7680
226
- }
227
- }