m-nzr commited on
Commit
1c7ac9c
·
verified ·
1 Parent(s): 8ecad13

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "/kaggle/input/gemma-3/transformers/gemma-3-1b-pt/1",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
@@ -27,13 +27,13 @@
27
  "rank_pattern": {},
28
  "revision": null,
29
  "target_modules": [
30
- "q_proj",
31
- "gate_proj",
32
  "down_proj",
33
  "o_proj",
34
- "up_proj",
35
  "v_proj",
36
- "k_proj"
 
 
37
  ],
38
  "task_type": "CAUSAL_LM",
39
  "trainable_token_indices": null,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "google/gemma-3-1b-it",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
 
27
  "rank_pattern": {},
28
  "revision": null,
29
  "target_modules": [
30
+ "k_proj",
 
31
  "down_proj",
32
  "o_proj",
 
33
  "v_proj",
34
+ "gate_proj",
35
+ "q_proj",
36
+ "up_proj"
37
  ],
38
  "task_type": "CAUSAL_LM",
39
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7280c5df88ef9c952d660cb51887df8088a9628c18f5bf541b938e5bc05fbc84
3
  size 2468152240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6996d238bc8d562967db4e2f636d2a75f8f41ce70c86526ef7891eec8a186d9
3
  size 2468152240
runs/Jun10_19-32-52_5f903ee70bcc/events.out.tfevents.1749584003.5f903ee70bcc.1757.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a0a7095ef53f1cbf2165078602d7f00469cd2cfd0aa2081964fe569d493684c
3
+ size 14440
tokenizer_config.json CHANGED
@@ -51325,7 +51325,7 @@
51325
  },
51326
  "boi_token": "<start_of_image>",
51327
  "bos_token": "<bos>",
51328
- "chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- set first_user_prefix = messages[0]['content'][0]['text'] + '\n\n' -%}\n {%- set loop_messages = messages[1:] -%}\n{%- else -%}\n {%- set first_user_prefix = \"\" -%}\n {%- set loop_messages = messages -%}\n{%- endif -%}\n{%- for message in loop_messages -%}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}\n {{ raise_exception(\"Conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif -%}\n {%- if (message['role'] == 'assistant') -%}\n {%- set role = \"model\" -%}\n {%- else -%}\n {%- set role = message['role'] -%}\n {%- endif -%}\n {{ '<start_of_turn>' + role + '\n' + (first_user_prefix if loop.first else \"\") }}\n {%- if message['content'] is string -%}\n {{ message['content'] | trim }}\n {%- elif message['content'] is iterable -%}\n {%- for item in message['content'] -%}\n {%- if item['type'] == 'image' -%}\n {{ '<start_of_image>' }}\n {%- elif item['type'] == 'text' -%}\n {{ item['text'] | trim }}\n {%- endif -%}\n {%- endfor -%}\n {%- else -%}\n {{ raise_exception(\"Invalid content type\") }}\n {%- endif -%}\n {{ '<end_of_turn>\n' }}\n{%- endfor -%}\n{%- if add_generation_prompt -%}\n {{'<start_of_turn>model\n'}}\n{%- endif -%}\n",
51329
  "clean_up_tokenization_spaces": false,
51330
  "eoi_token": "<end_of_image>",
51331
  "eos_token": "<eos>",
@@ -51337,6 +51337,7 @@
51337
  "image_token": "<image_soft_token>",
51338
  "model_max_length": 1000000000000000019884624838656,
51339
  "pad_token": "<pad>",
 
51340
  "sp_model_kwargs": null,
51341
  "spaces_between_special_tokens": false,
51342
  "tokenizer_class": "GemmaTokenizer",
 
51325
  },
51326
  "boi_token": "<start_of_image>",
51327
  "bos_token": "<bos>",
51328
+ "chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n {%- set first_user_prefix = messages[0]['content'][0]['text'] + '\n\n' -%}\n {%- endif -%}\n {%- set loop_messages = messages[1:] -%}\n{%- else -%}\n {%- set first_user_prefix = \"\" -%}\n {%- set loop_messages = messages -%}\n{%- endif -%}\n{%- for message in loop_messages -%}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}\n {{ raise_exception(\"Conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif -%}\n {%- if (message['role'] == 'assistant') -%}\n {%- set role = \"model\" -%}\n {%- else -%}\n {%- set role = message['role'] -%}\n {%- endif -%}\n {{ '<start_of_turn>' + role + '\n' + (first_user_prefix if loop.first else \"\") }}\n {%- if message['content'] is string -%}\n {{ message['content'] | trim }}\n {%- elif message['content'] is iterable -%}\n {%- for item in message['content'] -%}\n {%- if item['type'] == 'image' -%}\n {{ '<start_of_image>' }}\n {%- elif item['type'] == 'text' -%}\n {{ item['text'] | trim }}\n {%- endif -%}\n {%- endfor -%}\n {%- else -%}\n {{ raise_exception(\"Invalid content type\") }}\n {%- endif -%}\n {{ '<end_of_turn>\n' }}\n{%- endfor -%}\n{%- if add_generation_prompt -%}\n {{'<start_of_turn>model\n'}}\n{%- endif -%}\n",
51329
  "clean_up_tokenization_spaces": false,
51330
  "eoi_token": "<end_of_image>",
51331
  "eos_token": "<eos>",
 
51337
  "image_token": "<image_soft_token>",
51338
  "model_max_length": 1000000000000000019884624838656,
51339
  "pad_token": "<pad>",
51340
+ "processor_class": "Gemma3Processor",
51341
  "sp_model_kwargs": null,
51342
  "spaces_between_special_tokens": false,
51343
  "tokenizer_class": "GemmaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:737927ea039d72d5c7657cdaa5cf8ac32c502610170d403942bb38898b91be4b
3
  size 5688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de74554c0849c63bf66496cdb57beddf0681e70193c714facba1994f1cc8b305
3
  size 5688