Model save
Browse files- README.md +4 -5
- all_results.json +7 -7
- generation_config.json +1 -1
- preprocessor_config.json +15 -15
- train_results.json +7 -7
- trainer_state.json +0 -0
README.md
CHANGED
@@ -1,10 +1,9 @@
|
|
1 |
---
|
2 |
-
|
3 |
library_name: transformers
|
4 |
model_name: Qwen2-VL-2B-Instruct-SFT
|
5 |
tags:
|
6 |
- generated_from_trainer
|
7 |
-
- R1-V
|
8 |
- trl
|
9 |
- sft
|
10 |
licence: license
|
@@ -12,7 +11,7 @@ licence: license
|
|
12 |
|
13 |
# Model Card for Qwen2-VL-2B-Instruct-SFT
|
14 |
|
15 |
-
This model is a fine-tuned version of [
|
16 |
It has been trained using [TRL](https://github.com/huggingface/trl).
|
17 |
|
18 |
## Quick start
|
@@ -36,9 +35,9 @@ This model was trained with SFT.
|
|
36 |
### Framework versions
|
37 |
|
38 |
- TRL: 0.14.0
|
39 |
-
- Transformers: 4.
|
40 |
- Pytorch: 2.5.1
|
41 |
-
- Datasets: 3.
|
42 |
- Tokenizers: 0.21.1
|
43 |
|
44 |
## Citations
|
|
|
1 |
---
|
2 |
+
base_model: Qwen/Qwen2-VL-2B-Instruct
|
3 |
library_name: transformers
|
4 |
model_name: Qwen2-VL-2B-Instruct-SFT
|
5 |
tags:
|
6 |
- generated_from_trainer
|
|
|
7 |
- trl
|
8 |
- sft
|
9 |
licence: license
|
|
|
11 |
|
12 |
# Model Card for Qwen2-VL-2B-Instruct-SFT
|
13 |
|
14 |
+
This model is a fine-tuned version of [Qwen/Qwen2-VL-2B-Instruct](https://huggingface.co/Qwen/Qwen2-VL-2B-Instruct).
|
15 |
It has been trained using [TRL](https://github.com/huggingface/trl).
|
16 |
|
17 |
## Quick start
|
|
|
35 |
### Framework versions
|
36 |
|
37 |
- TRL: 0.14.0
|
38 |
+
- Transformers: 4.52.0.dev0
|
39 |
- Pytorch: 2.5.1
|
40 |
+
- Datasets: 3.6.0
|
41 |
- Tokenizers: 0.21.1
|
42 |
|
43 |
## Citations
|
all_results.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
-
"total_flos":
|
4 |
-
"train_loss": 0.
|
5 |
-
"train_runtime":
|
6 |
-
"train_samples":
|
7 |
-
"train_samples_per_second":
|
8 |
-
"train_steps_per_second": 0.
|
9 |
}
|
|
|
1 |
{
|
2 |
+
"epoch": 0.9998197093715069,
|
3 |
+
"total_flos": 4977616761913344.0,
|
4 |
+
"train_loss": 0.6325558101381102,
|
5 |
+
"train_runtime": 63848.9812,
|
6 |
+
"train_samples": 221862,
|
7 |
+
"train_samples_per_second": 3.475,
|
8 |
+
"train_steps_per_second": 0.054
|
9 |
}
|
generation_config.json
CHANGED
@@ -10,6 +10,6 @@
|
|
10 |
"temperature": 0.01,
|
11 |
"top_k": 1,
|
12 |
"top_p": 0.001,
|
13 |
-
"transformers_version": "4.
|
14 |
"use_cache": false
|
15 |
}
|
|
|
10 |
"temperature": 0.01,
|
11 |
"top_k": 1,
|
12 |
"top_p": 0.001,
|
13 |
+
"transformers_version": "4.52.0.dev0",
|
14 |
"use_cache": false
|
15 |
}
|
preprocessor_config.json
CHANGED
@@ -1,29 +1,29 @@
|
|
1 |
{
|
|
|
|
|
|
|
|
|
|
|
2 |
"do_convert_rgb": true,
|
3 |
"do_normalize": true,
|
4 |
"do_rescale": true,
|
5 |
"do_resize": true,
|
6 |
"image_mean": [
|
7 |
-
0.
|
8 |
-
0.
|
9 |
-
0.
|
10 |
],
|
11 |
-
"image_processor_type": "
|
12 |
"image_std": [
|
13 |
-
0.
|
14 |
-
0.
|
15 |
-
0.
|
16 |
],
|
17 |
-
"
|
18 |
-
"merge_size": 2,
|
19 |
-
"min_pixels": 3136,
|
20 |
-
"patch_size": 14,
|
21 |
"processor_class": "Qwen2VLProcessor",
|
22 |
"resample": 3,
|
23 |
"rescale_factor": 0.00392156862745098,
|
24 |
"size": {
|
25 |
-
"
|
26 |
-
|
27 |
-
},
|
28 |
-
"temporal_patch_size": 2
|
29 |
}
|
|
|
1 |
{
|
2 |
+
"crop_size": {
|
3 |
+
"height": 518,
|
4 |
+
"width": 518
|
5 |
+
},
|
6 |
+
"do_center_crop": true,
|
7 |
"do_convert_rgb": true,
|
8 |
"do_normalize": true,
|
9 |
"do_rescale": true,
|
10 |
"do_resize": true,
|
11 |
"image_mean": [
|
12 |
+
0.5307,
|
13 |
+
0.5307,
|
14 |
+
0.5307
|
15 |
],
|
16 |
+
"image_processor_type": "BitImageProcessor",
|
17 |
"image_std": [
|
18 |
+
0.2583,
|
19 |
+
0.2583,
|
20 |
+
0.2583
|
21 |
],
|
22 |
+
"merge_size": 1,
|
|
|
|
|
|
|
23 |
"processor_class": "Qwen2VLProcessor",
|
24 |
"resample": 3,
|
25 |
"rescale_factor": 0.00392156862745098,
|
26 |
"size": {
|
27 |
+
"shortest_edge": 518
|
28 |
+
}
|
|
|
|
|
29 |
}
|
train_results.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
-
"total_flos":
|
4 |
-
"train_loss": 0.
|
5 |
-
"train_runtime":
|
6 |
-
"train_samples":
|
7 |
-
"train_samples_per_second":
|
8 |
-
"train_steps_per_second": 0.
|
9 |
}
|
|
|
1 |
{
|
2 |
+
"epoch": 0.9998197093715069,
|
3 |
+
"total_flos": 4977616761913344.0,
|
4 |
+
"train_loss": 0.6325558101381102,
|
5 |
+
"train_runtime": 63848.9812,
|
6 |
+
"train_samples": 221862,
|
7 |
+
"train_samples_per_second": 3.475,
|
8 |
+
"train_steps_per_second": 0.054
|
9 |
}
|
trainer_state.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|