bluuluu commited on
Commit
2c69336
·
verified ·
1 Parent(s): acc18ab

Model save

Browse files
README.md CHANGED
@@ -1,10 +1,9 @@
1
  ---
2
- datasets: MMInstruction/Clevr_CoGenT_TrainA_R1
3
  library_name: transformers
4
  model_name: Qwen2-VL-2B-Instruct-SFT
5
  tags:
6
  - generated_from_trainer
7
- - R1-V
8
  - trl
9
  - sft
10
  licence: license
@@ -12,7 +11,7 @@ licence: license
12
 
13
  # Model Card for Qwen2-VL-2B-Instruct-SFT
14
 
15
- This model is a fine-tuned version of [None](https://huggingface.co/None) on the [MMInstruction/Clevr_CoGenT_TrainA_R1](https://huggingface.co/datasets/MMInstruction/Clevr_CoGenT_TrainA_R1) dataset.
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
17
 
18
  ## Quick start
@@ -36,9 +35,9 @@ This model was trained with SFT.
36
  ### Framework versions
37
 
38
  - TRL: 0.14.0
39
- - Transformers: 4.51.1
40
  - Pytorch: 2.5.1
41
- - Datasets: 3.2.0
42
  - Tokenizers: 0.21.1
43
 
44
  ## Citations
 
1
  ---
2
+ base_model: Qwen/Qwen2-VL-2B-Instruct
3
  library_name: transformers
4
  model_name: Qwen2-VL-2B-Instruct-SFT
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
  - sft
9
  licence: license
 
11
 
12
  # Model Card for Qwen2-VL-2B-Instruct-SFT
13
 
14
+ This model is a fine-tuned version of [Qwen/Qwen2-VL-2B-Instruct](https://huggingface.co/Qwen/Qwen2-VL-2B-Instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
35
  ### Framework versions
36
 
37
  - TRL: 0.14.0
38
+ - Transformers: 4.52.0.dev0
39
  - Pytorch: 2.5.1
40
+ - Datasets: 3.6.0
41
  - Tokenizers: 0.21.1
42
 
43
  ## Citations
all_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 1.9998919736415686,
3
- "total_flos": 2.1613118366416896e+16,
4
- "train_loss": 0.0653228780393709,
5
- "train_runtime": 98246.8444,
6
- "train_samples": 222165,
7
- "train_samples_per_second": 4.523,
8
- "train_steps_per_second": 0.071
9
  }
 
1
  {
2
+ "epoch": 0.9998197093715069,
3
+ "total_flos": 4977616761913344.0,
4
+ "train_loss": 0.6325558101381102,
5
+ "train_runtime": 63848.9812,
6
+ "train_samples": 221862,
7
+ "train_samples_per_second": 3.475,
8
+ "train_steps_per_second": 0.054
9
  }
generation_config.json CHANGED
@@ -10,6 +10,6 @@
10
  "temperature": 0.01,
11
  "top_k": 1,
12
  "top_p": 0.001,
13
- "transformers_version": "4.51.1",
14
  "use_cache": false
15
  }
 
10
  "temperature": 0.01,
11
  "top_k": 1,
12
  "top_p": 0.001,
13
+ "transformers_version": "4.52.0.dev0",
14
  "use_cache": false
15
  }
preprocessor_config.json CHANGED
@@ -1,29 +1,29 @@
1
  {
 
 
 
 
 
2
  "do_convert_rgb": true,
3
  "do_normalize": true,
4
  "do_rescale": true,
5
  "do_resize": true,
6
  "image_mean": [
7
- 0.48145466,
8
- 0.4578275,
9
- 0.40821073
10
  ],
11
- "image_processor_type": "Qwen2VLImageProcessor",
12
  "image_std": [
13
- 0.26862954,
14
- 0.26130258,
15
- 0.27577711
16
  ],
17
- "max_pixels": 12845056,
18
- "merge_size": 2,
19
- "min_pixels": 3136,
20
- "patch_size": 14,
21
  "processor_class": "Qwen2VLProcessor",
22
  "resample": 3,
23
  "rescale_factor": 0.00392156862745098,
24
  "size": {
25
- "longest_edge": 12845056,
26
- "shortest_edge": 3136
27
- },
28
- "temporal_patch_size": 2
29
  }
 
1
  {
2
+ "crop_size": {
3
+ "height": 518,
4
+ "width": 518
5
+ },
6
+ "do_center_crop": true,
7
  "do_convert_rgb": true,
8
  "do_normalize": true,
9
  "do_rescale": true,
10
  "do_resize": true,
11
  "image_mean": [
12
+ 0.5307,
13
+ 0.5307,
14
+ 0.5307
15
  ],
16
+ "image_processor_type": "BitImageProcessor",
17
  "image_std": [
18
+ 0.2583,
19
+ 0.2583,
20
+ 0.2583
21
  ],
22
+ "merge_size": 1,
 
 
 
23
  "processor_class": "Qwen2VLProcessor",
24
  "resample": 3,
25
  "rescale_factor": 0.00392156862745098,
26
  "size": {
27
+ "shortest_edge": 518
28
+ }
 
 
29
  }
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 1.9998919736415686,
3
- "total_flos": 2.1613118366416896e+16,
4
- "train_loss": 0.0653228780393709,
5
- "train_runtime": 98246.8444,
6
- "train_samples": 222165,
7
- "train_samples_per_second": 4.523,
8
- "train_steps_per_second": 0.071
9
  }
 
1
  {
2
+ "epoch": 0.9998197093715069,
3
+ "total_flos": 4977616761913344.0,
4
+ "train_loss": 0.6325558101381102,
5
+ "train_runtime": 63848.9812,
6
+ "train_samples": 221862,
7
+ "train_samples_per_second": 3.475,
8
+ "train_steps_per_second": 0.054
9
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff