hardlyworking commited on
Commit
78d5d6e
·
verified ·
1 Parent(s): 4b84b31

Upload 11 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,162 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ license: gemma
4
+ base_model: huihui-ai/Huihui-gemma-3n-E4B-it-abliterated
5
+ tags:
6
+ - axolotl
7
+ - base_model:adapter:huihui-ai/Huihui-gemma-3n-E4B-it-abliterated
8
+ - lora
9
+ - transformers
10
+ datasets:
11
+ - hardlyworking/HardlyRPv2-10k
12
+ pipeline_tag: text-generation
13
+ model-index:
14
+ - name: outputs/out
15
+ results: []
16
+ ---
17
+
18
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
19
+ should probably proofread and complete it, then remove this comment. -->
20
+
21
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
22
+ <details><summary>See axolotl config</summary>
23
+
24
+ axolotl version: `0.12.0.dev0`
25
+ ```yaml
26
+ base_model: huihui-ai/Huihui-gemma-3n-E4B-it-abliterated
27
+
28
+ # Automatically upload checkpoint and final model to HF
29
+ # hub_model_id: username/custom_model_name
30
+
31
+ plugins:
32
+ - axolotl.integrations.liger.LigerPlugin
33
+ - axolotl.integrations.cut_cross_entropy.CutCrossEntropyPlugin
34
+ liger_rope: true
35
+ liger_rms_norm: true
36
+ liger_layer_norm: true
37
+ liger_glu_activation: true
38
+ liger_fused_linear_cross_entropy: false
39
+ cut_cross_entropy: true
40
+
41
+ load_in_8bit: false
42
+ load_in_4bit: true
43
+
44
+ # for use with fft to only train on language model layers
45
+ # unfrozen_parameters:
46
+ # - model.language_model.*
47
+ # - lm_head
48
+ # - embed_tokens
49
+
50
+
51
+ chat_template: gemma3n
52
+ eot_tokens:
53
+ - <end_of_turn>
54
+ datasets:
55
+ - path: hardlyworking/HardlyRPv2-10k
56
+ type: chat_template
57
+ split: train
58
+ field_messages: conversations
59
+ message_property_mappings:
60
+ role: from
61
+ content: value
62
+
63
+ val_set_size: 0.0
64
+ output_dir: ./outputs/out
65
+
66
+ adapter: qlora
67
+ lora_r: 128
68
+ lora_alpha: 64
69
+ lora_dropout: 0.05
70
+ # lora_target_linear: # Does not work with gemma3n currently
71
+ lora_target_modules:
72
+ - self_attn.q_proj
73
+ - self_attn.k_proj
74
+ - self_attn.v_proj
75
+ - self_attn.o_proj
76
+ - mlp.gate_proj
77
+ - mlp.up_proj
78
+ - mlp.down_proj
79
+
80
+
81
+ sequence_len: 8192
82
+ sample_packing: true
83
+ eval_sample_packing: true
84
+ pad_to_sequence_len: true
85
+
86
+ wandb_project:
87
+ wandb_entity:
88
+ wandb_watch:
89
+ wandb_name:
90
+ wandb_log_model:
91
+
92
+ gradient_accumulation_steps: 8
93
+ micro_batch_size: 4
94
+ num_epochs: 2
95
+ optimizer: adamw_bnb_8bit
96
+ lr_scheduler: cosine
97
+ learning_rate: 0.0002
98
+
99
+ bf16: auto
100
+ tf32: true
101
+
102
+ gradient_checkpointing: true
103
+ gradient_checkpointing_kwargs:
104
+ use_reentrant: false
105
+ unsloth: true
106
+ resume_from_checkpoint:
107
+ logging_steps: 1
108
+ # flash_attention: true # Any attention impl does not work with gemma3n now
109
+
110
+ warmup_ratio: 0.1
111
+ evals_per_epoch:
112
+ saves_per_epoch: 1
113
+ weight_decay: 0.0
114
+ special_tokens:
115
+
116
+ ```
117
+
118
+ </details><br>
119
+
120
+ # outputs/out
121
+
122
+ This model is a fine-tuned version of [huihui-ai/Huihui-gemma-3n-E4B-it-abliterated](https://huggingface.co/huihui-ai/Huihui-gemma-3n-E4B-it-abliterated) on the hardlyworking/HardlyRPv2-10k dataset.
123
+
124
+ ## Model description
125
+
126
+ More information needed
127
+
128
+ ## Intended uses & limitations
129
+
130
+ More information needed
131
+
132
+ ## Training and evaluation data
133
+
134
+ More information needed
135
+
136
+ ## Training procedure
137
+
138
+ ### Training hyperparameters
139
+
140
+ The following hyperparameters were used during training:
141
+ - learning_rate: 0.0002
142
+ - train_batch_size: 4
143
+ - eval_batch_size: 4
144
+ - seed: 42
145
+ - gradient_accumulation_steps: 8
146
+ - total_train_batch_size: 32
147
+ - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
148
+ - lr_scheduler_type: cosine
149
+ - lr_scheduler_warmup_steps: 13
150
+ - training_steps: 132
151
+
152
+ ### Training results
153
+
154
+
155
+
156
+ ### Framework versions
157
+
158
+ - PEFT 0.17.0
159
+ - Transformers 4.55.0
160
+ - Pytorch 2.7.1+cu126
161
+ - Datasets 4.0.0
162
+ - Tokenizers 0.21.4
adapter_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "huihui-ai/Huihui-gemma-3n-E4B-it-abliterated",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": null,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 64,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 128,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "mlp.down_proj",
29
+ "self_attn.k_proj",
30
+ "mlp.gate_proj",
31
+ "self_attn.v_proj",
32
+ "mlp.up_proj",
33
+ "self_attn.q_proj",
34
+ "self_attn.o_proj"
35
+ ],
36
+ "target_parameters": [],
37
+ "task_type": "CAUSAL_LM",
38
+ "trainable_token_indices": null,
39
+ "use_dora": false,
40
+ "use_qalora": false,
41
+ "use_rslora": false
42
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f38331ac1f5a6d581ae831ac492378274bf83c47250b0f3b675b18288789be8
3
+ size 1229529600
chat_template.jinja ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {{ bos_token }}
2
+ {%- if messages[0]['role'] == 'system' -%}
3
+ {%- if messages[0]['content'] is string -%}
4
+ {%- set first_user_prefix = messages[0]['content'] + '
5
+
6
+ ' -%}
7
+ {%- else -%}
8
+ {%- set first_user_prefix = messages[0]['content'][0]['text'] + '
9
+
10
+ ' -%}
11
+ {%- endif -%}
12
+ {%- set loop_messages = messages[1:] -%}
13
+ {%- else -%}
14
+ {%- set first_user_prefix = "" -%}
15
+ {%- set loop_messages = messages -%}
16
+ {%- endif -%}
17
+ {%- for message in loop_messages -%}
18
+ {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}
19
+ {{ raise_exception("Conversation roles must alternate user/assistant/user/assistant/...") }}
20
+ {%- endif -%}
21
+ {%- if (message['role'] == 'assistant') -%}
22
+ {%- set role = "model" -%}
23
+ {%- else -%}
24
+ {%- set role = message['role'] -%}
25
+ {%- endif -%}
26
+ {{ '<start_of_turn>' + role + '
27
+ ' + (first_user_prefix if loop.first else "") }}
28
+ {%- if message['content'] is string -%}
29
+ {{ message['content'] | trim }}
30
+ {%- elif message['content'] is iterable -%}
31
+ {%- for item in message['content'] -%}
32
+ {%- if item['type'] == 'audio' -%}
33
+ {{ '<audio_soft_token>' }}
34
+ {%- elif item['type'] == 'image' -%}
35
+ {{ '<image_soft_token>' }}
36
+ {%- elif item['type'] == 'text' -%}
37
+ {{ item['text'] | trim }}
38
+ {%- endif -%}
39
+ {%- endfor -%}
40
+ {%- else -%}
41
+ {{ raise_exception("Invalid content type") }}
42
+ {%- endif -%}
43
+ {{ '<end_of_turn>
44
+ ' }}
45
+ {%- endfor -%}
46
+ {%- if add_generation_prompt -%}
47
+ {{'<start_of_turn>model
48
+ '}}
49
+ {%- endif -%}
config.json ADDED
@@ -0,0 +1,239 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Gemma3nForConditionalGeneration"
4
+ ],
5
+ "audio_config": {
6
+ "conf_attention_chunk_size": 12,
7
+ "conf_attention_context_left": 13,
8
+ "conf_attention_context_right": 0,
9
+ "conf_attention_logit_cap": 50.0,
10
+ "conf_conv_kernel_size": 5,
11
+ "conf_num_attention_heads": 8,
12
+ "conf_num_hidden_layers": 12,
13
+ "conf_positional_bias_size": 256,
14
+ "conf_reduction_factor": 4,
15
+ "conf_residual_weight": 0.5,
16
+ "gradient_clipping": 10000000000.0,
17
+ "hidden_size": 1536,
18
+ "input_feat_size": 128,
19
+ "model_type": "gemma3n_audio",
20
+ "rms_norm_eps": 1e-06,
21
+ "sscp_conv_channel_size": [
22
+ 128,
23
+ 32
24
+ ],
25
+ "sscp_conv_eps": 0.001,
26
+ "sscp_conv_group_norm_eps": 0.001,
27
+ "sscp_conv_kernel_size": [
28
+ [
29
+ 3,
30
+ 3
31
+ ],
32
+ [
33
+ 3,
34
+ 3
35
+ ]
36
+ ],
37
+ "sscp_conv_stride_size": [
38
+ [
39
+ 2,
40
+ 2
41
+ ],
42
+ [
43
+ 2,
44
+ 2
45
+ ]
46
+ ],
47
+ "torch_dtype": "bfloat16",
48
+ "vocab_offset": 262272,
49
+ "vocab_size": 128
50
+ },
51
+ "audio_soft_tokens_per_image": 188,
52
+ "audio_token_id": 262273,
53
+ "boa_token_id": 256000,
54
+ "boi_token_id": 255999,
55
+ "eoa_token_id": 262272,
56
+ "eoi_token_id": 262144,
57
+ "eos_token_id": 1,
58
+ "image_token_id": 262145,
59
+ "initializer_range": 0.02,
60
+ "model_type": "gemma3n",
61
+ "quantization_config": {
62
+ "_load_in_4bit": true,
63
+ "_load_in_8bit": false,
64
+ "bnb_4bit_compute_dtype": "bfloat16",
65
+ "bnb_4bit_quant_storage": "bfloat16",
66
+ "bnb_4bit_quant_type": "nf4",
67
+ "bnb_4bit_use_double_quant": true,
68
+ "llm_int8_enable_fp32_cpu_offload": false,
69
+ "llm_int8_has_fp16_weight": false,
70
+ "llm_int8_skip_modules": null,
71
+ "llm_int8_threshold": 6.0,
72
+ "load_in_4bit": true,
73
+ "load_in_8bit": false,
74
+ "quant_method": "bitsandbytes"
75
+ },
76
+ "text_config": {
77
+ "activation_sparsity_pattern": [
78
+ 0.95,
79
+ 0.95,
80
+ 0.95,
81
+ 0.95,
82
+ 0.95,
83
+ 0.95,
84
+ 0.95,
85
+ 0.95,
86
+ 0.95,
87
+ 0.95,
88
+ 0.0,
89
+ 0.0,
90
+ 0.0,
91
+ 0.0,
92
+ 0.0,
93
+ 0.0,
94
+ 0.0,
95
+ 0.0,
96
+ 0.0,
97
+ 0.0,
98
+ 0.0,
99
+ 0.0,
100
+ 0.0,
101
+ 0.0,
102
+ 0.0,
103
+ 0.0,
104
+ 0.0,
105
+ 0.0,
106
+ 0.0,
107
+ 0.0,
108
+ 0.0,
109
+ 0.0,
110
+ 0.0,
111
+ 0.0,
112
+ 0.0
113
+ ],
114
+ "altup_active_idx": 0,
115
+ "altup_coef_clip": 120.0,
116
+ "altup_correct_scale": true,
117
+ "altup_lr_multiplier": 1.0,
118
+ "altup_num_inputs": 4,
119
+ "attention_bias": false,
120
+ "attention_dropout": 0.0,
121
+ "final_logit_softcapping": 30.0,
122
+ "head_dim": 256,
123
+ "hidden_activation": "gelu_pytorch_tanh",
124
+ "hidden_size": 2048,
125
+ "hidden_size_per_layer_input": 256,
126
+ "initializer_range": 0.02,
127
+ "intermediate_size": [
128
+ 16384,
129
+ 16384,
130
+ 16384,
131
+ 16384,
132
+ 16384,
133
+ 16384,
134
+ 16384,
135
+ 16384,
136
+ 16384,
137
+ 16384,
138
+ 16384,
139
+ 16384,
140
+ 16384,
141
+ 16384,
142
+ 16384,
143
+ 16384,
144
+ 16384,
145
+ 16384,
146
+ 16384,
147
+ 16384,
148
+ 16384,
149
+ 16384,
150
+ 16384,
151
+ 16384,
152
+ 16384,
153
+ 16384,
154
+ 16384,
155
+ 16384,
156
+ 16384,
157
+ 16384,
158
+ 16384,
159
+ 16384,
160
+ 16384,
161
+ 16384,
162
+ 16384
163
+ ],
164
+ "laurel_rank": 64,
165
+ "layer_types": [
166
+ "sliding_attention",
167
+ "sliding_attention",
168
+ "sliding_attention",
169
+ "sliding_attention",
170
+ "full_attention",
171
+ "sliding_attention",
172
+ "sliding_attention",
173
+ "sliding_attention",
174
+ "sliding_attention",
175
+ "full_attention",
176
+ "sliding_attention",
177
+ "sliding_attention",
178
+ "sliding_attention",
179
+ "sliding_attention",
180
+ "full_attention",
181
+ "sliding_attention",
182
+ "sliding_attention",
183
+ "sliding_attention",
184
+ "sliding_attention",
185
+ "full_attention",
186
+ "sliding_attention",
187
+ "sliding_attention",
188
+ "sliding_attention",
189
+ "sliding_attention",
190
+ "full_attention",
191
+ "sliding_attention",
192
+ "sliding_attention",
193
+ "sliding_attention",
194
+ "sliding_attention",
195
+ "full_attention",
196
+ "sliding_attention",
197
+ "sliding_attention",
198
+ "sliding_attention",
199
+ "sliding_attention",
200
+ "full_attention"
201
+ ],
202
+ "max_position_embeddings": 32768,
203
+ "model_type": "gemma3n_text",
204
+ "num_attention_heads": 8,
205
+ "num_hidden_layers": 35,
206
+ "num_key_value_heads": 2,
207
+ "num_kv_shared_layers": 15,
208
+ "query_pre_attn_scalar": 256,
209
+ "rms_norm_eps": 1e-06,
210
+ "rope_local_base_freq": 10000.0,
211
+ "rope_scaling": null,
212
+ "rope_theta": 1000000.0,
213
+ "sliding_window": 512,
214
+ "torch_dtype": "bfloat16",
215
+ "use_cache": false,
216
+ "vocab_size": 262400,
217
+ "vocab_size_per_layer_input": 262144
218
+ },
219
+ "torch_dtype": "bfloat16",
220
+ "transformers_version": "4.55.0",
221
+ "vision_config": {
222
+ "architecture": "mobilenetv5_300m_enc",
223
+ "do_pooling": true,
224
+ "hidden_size": 2048,
225
+ "initializer_range": 0.02,
226
+ "label_names": [
227
+ "LABEL_0",
228
+ "LABEL_1"
229
+ ],
230
+ "model_args": null,
231
+ "model_type": "gemma3n_vision",
232
+ "num_classes": 2,
233
+ "rms_norm_eps": 1e-06,
234
+ "torch_dtype": "bfloat16",
235
+ "vocab_offset": 262144,
236
+ "vocab_size": 128
237
+ },
238
+ "vision_soft_tokens_per_image": 256
239
+ }
preprocessor_config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": null,
3
+ "data_format": "channels_first",
4
+ "default_to_square": false,
5
+ "device": null,
6
+ "disable_grouping": null,
7
+ "dither": 0.0,
8
+ "do_center_crop": null,
9
+ "do_convert_rgb": null,
10
+ "do_normalize": false,
11
+ "do_rescale": true,
12
+ "do_resize": true,
13
+ "feature_size": 128,
14
+ "fft_length": 1024,
15
+ "fft_overdrive": true,
16
+ "frame_length": 512,
17
+ "hop_length": 160,
18
+ "image_mean": [
19
+ 0.5,
20
+ 0.5,
21
+ 0.5
22
+ ],
23
+ "image_processor_type": "SiglipImageProcessorFast",
24
+ "image_seq_length": 256,
25
+ "image_std": [
26
+ 0.5,
27
+ 0.5,
28
+ 0.5
29
+ ],
30
+ "input_data_format": null,
31
+ "input_scale_factor": 1.0,
32
+ "max_frequency": 7600.0,
33
+ "mel_floor": 1e-05,
34
+ "min_frequency": 125.0,
35
+ "padding_side": "right",
36
+ "padding_value": 0.0,
37
+ "per_bin_mean": null,
38
+ "per_bin_stddev": null,
39
+ "preemphasis": 0.97,
40
+ "preemphasis_htk_flavor": true,
41
+ "processor_class": "Gemma3nProcessor",
42
+ "resample": 2,
43
+ "rescale_factor": 0.00392156862745098,
44
+ "return_attention_mask": false,
45
+ "return_tensors": null,
46
+ "sampling_rate": 16000,
47
+ "size": {
48
+ "height": 768,
49
+ "width": 768
50
+ }
51
+ }
processor_config.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "audio_seq_length": 188,
3
+ "image_seq_length": 256,
4
+ "processor_class": "Gemma3nProcessor"
5
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "audio_token": "<audio_soft_token>",
3
+ "boa_token": "<start_of_audio>",
4
+ "boi_token": "<start_of_image>",
5
+ "bos_token": {
6
+ "content": "<bos>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "eoa_token": "<end_of_audio>",
13
+ "eoi_token": "<end_of_image>",
14
+ "eos_token": {
15
+ "content": "<eos>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "image_token": "<image_soft_token>",
22
+ "pad_token": {
23
+ "content": "<pad>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false
28
+ },
29
+ "unk_token": {
30
+ "content": "<unk>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false
35
+ }
36
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6c35ee648c07754b44cd9e371c75d4caa05c4504910b7ad29b1847ee9d8ba5d
3
+ size 33442553
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea5f0cc48abfbfc04d14562270a32e02149a3e7035f368cc5a462786f4a59961
3
+ size 4696020
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff