Taehyun34 commited on
Commit
eeacbc8
·
verified ·
1 Parent(s): cf0daa6

End of training

Browse files
README.md CHANGED
@@ -16,9 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # output_gptreasoning
18
 
19
- This model is a fine-tuned version of [HuggingFaceM4/idefics2-8b](https://huggingface.co/HuggingFaceM4/idefics2-8b) on the None dataset.
20
- It achieves the following results on the evaluation set:
21
- - Loss: 0.5066
22
 
23
  ## Model description
24
 
@@ -41,11 +39,11 @@ The following hyperparameters were used during training:
41
  - train_batch_size: 1
42
  - eval_batch_size: 1
43
  - seed: 42
44
- - gradient_accumulation_steps: 16
45
- - total_train_batch_size: 16
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: cosine
48
- - num_epochs: 3
49
  - mixed_precision_training: Native AMP
50
 
51
  ### Training results
@@ -54,8 +52,8 @@ The following hyperparameters were used during training:
54
 
55
  ### Framework versions
56
 
57
- - PEFT 0.12.0
58
- - Transformers 4.42.4
59
- - Pytorch 2.4.0+cu121
60
- - Datasets 2.21.0
61
  - Tokenizers 0.19.1
 
16
 
17
  # output_gptreasoning
18
 
19
+ This model is a fine-tuned version of [HuggingFaceM4/idefics2-8b](https://huggingface.co/HuggingFaceM4/idefics2-8b) on an unknown dataset.
 
 
20
 
21
  ## Model description
22
 
 
39
  - train_batch_size: 1
40
  - eval_batch_size: 1
41
  - seed: 42
42
+ - gradient_accumulation_steps: 32
43
+ - total_train_batch_size: 32
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
+ - num_epochs: 10
47
  - mixed_precision_training: Native AMP
48
 
49
  ### Training results
 
52
 
53
  ### Framework versions
54
 
55
+ - PEFT 0.13.0
56
+ - Transformers 4.44.2
57
+ - Pytorch 2.4.1+cu121
58
+ - Datasets 3.0.1
59
  - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -21,8 +21,8 @@
21
  "revision": null,
22
  "target_modules": [
23
  "q_proj",
24
- "k_proj",
25
  "o_proj",
 
26
  "v_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
 
21
  "revision": null,
22
  "target_modules": [
23
  "q_proj",
 
24
  "o_proj",
25
+ "k_proj",
26
  "v_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d026e45012755890dc5fc5d8688452acde57a9a87af2066d4b9f19adc731a535
3
  size 35242256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e433b9fe0891916bd86f074463fc1834c88a4e7b3944a581170993652469109e
3
  size 35242256
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a28f16faf459d1881162647cbb26f4d1828890629ed9be9ff7d9df51c93f8a36
3
- size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a83a71f962db632f6cdb6dd6b220c37d92250caa1a89544b2d9a754f1d35c3c2
3
+ size 5496