RickyChen-Infinirc commited on
Commit
7f62e33
·
verified ·
1 Parent(s): 5dc5720

Delete .ipynb_checkpoints

Browse files
.ipynb_checkpoints/config-checkpoint.json DELETED
@@ -1,225 +0,0 @@
1
- {
2
- "_name_or_path": "/home/rickychen/\u684c\u9762/llm/models/Llama-3.2-11B-Vision-Instruct",
3
- "architectures": [
4
- "MllamaForConditionalGeneration"
5
- ],
6
- "image_token_index": 128256,
7
- "model_type": "mllama",
8
- "text_config": {
9
- "_name_or_path": "",
10
- "add_cross_attention": false,
11
- "architectures": null,
12
- "bad_words_ids": null,
13
- "begin_suppress_tokens": null,
14
- "bos_token_id": 128000,
15
- "chunk_size_feed_forward": 0,
16
- "cross_attention_hidden_size": null,
17
- "cross_attention_layers": [
18
- 3,
19
- 8,
20
- 13,
21
- 18,
22
- 23,
23
- 28,
24
- 33,
25
- 38
26
- ],
27
- "decoder_start_token_id": null,
28
- "diversity_penalty": 0.0,
29
- "do_sample": false,
30
- "dropout": 0,
31
- "early_stopping": false,
32
- "encoder_no_repeat_ngram_size": 0,
33
- "eos_token_id": [
34
- 128001,
35
- 128008,
36
- 128009
37
- ],
38
- "exponential_decay_length_penalty": null,
39
- "finetuning_task": null,
40
- "forced_bos_token_id": null,
41
- "forced_eos_token_id": null,
42
- "hidden_act": "silu",
43
- "hidden_size": 4096,
44
- "id2label": {
45
- "0": "LABEL_0",
46
- "1": "LABEL_1"
47
- },
48
- "initializer_range": 0.02,
49
- "intermediate_size": 14336,
50
- "is_decoder": false,
51
- "is_encoder_decoder": false,
52
- "label2id": {
53
- "LABEL_0": 0,
54
- "LABEL_1": 1
55
- },
56
- "length_penalty": 1.0,
57
- "max_length": 20,
58
- "max_position_embeddings": 131072,
59
- "min_length": 0,
60
- "model_type": "mllama_text_model",
61
- "no_repeat_ngram_size": 0,
62
- "num_attention_heads": 32,
63
- "num_beam_groups": 1,
64
- "num_beams": 1,
65
- "num_hidden_layers": 40,
66
- "num_key_value_heads": 8,
67
- "num_return_sequences": 1,
68
- "output_attentions": false,
69
- "output_hidden_states": false,
70
- "output_scores": false,
71
- "pad_token_id": 128004,
72
- "prefix": null,
73
- "problem_type": null,
74
- "pruned_heads": {},
75
- "remove_invalid_values": false,
76
- "repetition_penalty": 1.0,
77
- "return_dict": true,
78
- "return_dict_in_generate": false,
79
- "rms_norm_eps": 1e-05,
80
- "rope_scaling": {
81
- "factor": 8.0,
82
- "high_freq_factor": 4.0,
83
- "low_freq_factor": 1.0,
84
- "original_max_position_embeddings": 8192,
85
- "rope_type": "llama3"
86
- },
87
- "rope_theta": 500000.0,
88
- "sep_token_id": null,
89
- "suppress_tokens": null,
90
- "task_specific_params": null,
91
- "temperature": 1.0,
92
- "tf_legacy_loss": false,
93
- "tie_encoder_decoder": false,
94
- "tie_word_embeddings": false,
95
- "tokenizer_class": null,
96
- "top_k": 50,
97
- "top_p": 1.0,
98
- "torch_dtype": "bfloat16",
99
- "torchscript": false,
100
- "typical_p": 1.0,
101
- "use_bfloat16": false,
102
- "use_cache": true,
103
- "vocab_size": 128256
104
- },
105
- "torch_dtype": "bfloat16",
106
- "transformers_version": "4.45.0",
107
- "use_cache": true,
108
- "vision_config": {
109
- "_name_or_path": "",
110
- "add_cross_attention": false,
111
- "architectures": null,
112
- "attention_heads": 16,
113
- "bad_words_ids": null,
114
- "begin_suppress_tokens": null,
115
- "bos_token_id": null,
116
- "chunk_size_feed_forward": 0,
117
- "cross_attention_hidden_size": null,
118
- "decoder_start_token_id": null,
119
- "diversity_penalty": 0.0,
120
- "do_sample": false,
121
- "early_stopping": false,
122
- "encoder_no_repeat_ngram_size": 0,
123
- "eos_token_id": null,
124
- "exponential_decay_length_penalty": null,
125
- "finetuning_task": null,
126
- "forced_bos_token_id": null,
127
- "forced_eos_token_id": null,
128
- "hidden_act": "gelu",
129
- "hidden_size": 1280,
130
- "id2label": {
131
- "0": "LABEL_0",
132
- "1": "LABEL_1"
133
- },
134
- "image_size": 560,
135
- "initializer_range": 0.02,
136
- "intermediate_layers_indices": [
137
- 3,
138
- 7,
139
- 15,
140
- 23,
141
- 30
142
- ],
143
- "intermediate_size": 5120,
144
- "is_decoder": false,
145
- "is_encoder_decoder": false,
146
- "label2id": {
147
- "LABEL_0": 0,
148
- "LABEL_1": 1
149
- },
150
- "length_penalty": 1.0,
151
- "max_length": 20,
152
- "max_num_tiles": 4,
153
- "min_length": 0,
154
- "model_type": "mllama_vision_model",
155
- "no_repeat_ngram_size": 0,
156
- "norm_eps": 1e-05,
157
- "num_beam_groups": 1,
158
- "num_beams": 1,
159
- "num_channels": 3,
160
- "num_global_layers": 8,
161
- "num_hidden_layers": 32,
162
- "num_return_sequences": 1,
163
- "output_attentions": false,
164
- "output_hidden_states": false,
165
- "output_scores": false,
166
- "pad_token_id": null,
167
- "patch_size": 14,
168
- "prefix": null,
169
- "problem_type": null,
170
- "pruned_heads": {},
171
- "remove_invalid_values": false,
172
- "repetition_penalty": 1.0,
173
- "return_dict": true,
174
- "return_dict_in_generate": false,
175
- "sep_token_id": null,
176
- "supported_aspect_ratios": [
177
- [
178
- 1,
179
- 1
180
- ],
181
- [
182
- 1,
183
- 2
184
- ],
185
- [
186
- 1,
187
- 3
188
- ],
189
- [
190
- 1,
191
- 4
192
- ],
193
- [
194
- 2,
195
- 1
196
- ],
197
- [
198
- 2,
199
- 2
200
- ],
201
- [
202
- 3,
203
- 1
204
- ],
205
- [
206
- 4,
207
- 1
208
- ]
209
- ],
210
- "suppress_tokens": null,
211
- "task_specific_params": null,
212
- "temperature": 1.0,
213
- "tf_legacy_loss": false,
214
- "tie_encoder_decoder": false,
215
- "tie_word_embeddings": true,
216
- "tokenizer_class": null,
217
- "top_k": 50,
218
- "top_p": 1.0,
219
- "torch_dtype": "bfloat16",
220
- "torchscript": false,
221
- "typical_p": 1.0,
222
- "use_bfloat16": false,
223
- "vision_output_dim": 7680
224
- }
225
- }