Metal079 commited on
Commit
1a021d3
·
1 Parent(s): 308d90e

Upload 4 files

Browse files
sonic/config.json CHANGED
@@ -1,13 +1,22 @@
1
  {
2
- "_name_or_path": "facebook/deit-base-distilled-patch16-384",
3
  "architectures": [
4
- "DeiTForImageClassification"
5
  ],
6
- "attention_probs_dropout_prob": 0.0,
7
- "encoder_stride": 16,
 
 
 
 
 
8
  "hidden_act": "gelu",
9
- "hidden_dropout_prob": 0.0,
10
- "hidden_size": 768,
 
 
 
 
11
  "id2label": {
12
  "0": "Tails the fox",
13
  "1": "Amy rose",
@@ -27,7 +36,6 @@
27
  },
28
  "image_size": 384,
29
  "initializer_range": 0.02,
30
- "intermediate_size": 3072,
31
  "label2id": {
32
  "Amy rose": "1",
33
  "Big the cat": "14",
@@ -46,14 +54,12 @@
46
  "mobian": "12"
47
  },
48
  "layer_norm_eps": 1e-12,
49
- "model_type": "deit",
50
- "num_attention_heads": 12,
51
  "num_channels": 3,
52
- "num_hidden_layers": 12,
53
- "patch_size": 16,
54
  "problem_type": "multi_label_classification",
55
- "qkv_bias": true,
56
  "torch_dtype": "float32",
57
- "transformers_version": "4.25.1",
58
- "labels":[0,0,0,0,0,0,0,0]
59
  }
 
1
  {
2
+ "_name_or_path": "facebook/convnext-base-384-22k-1k",
3
  "architectures": [
4
+ "ConvNextForImageClassification"
5
  ],
6
+ "depths": [
7
+ 3,
8
+ 3,
9
+ 27,
10
+ 3
11
+ ],
12
+ "drop_path_rate": 0.0,
13
  "hidden_act": "gelu",
14
+ "hidden_sizes": [
15
+ 128,
16
+ 256,
17
+ 512,
18
+ 1024
19
+ ],
20
  "id2label": {
21
  "0": "Tails the fox",
22
  "1": "Amy rose",
 
36
  },
37
  "image_size": 384,
38
  "initializer_range": 0.02,
 
39
  "label2id": {
40
  "Amy rose": "1",
41
  "Big the cat": "14",
 
54
  "mobian": "12"
55
  },
56
  "layer_norm_eps": 1e-12,
57
+ "layer_scale_init_value": 1e-06,
58
+ "model_type": "convnext",
59
  "num_channels": 3,
60
+ "num_stages": 4,
61
+ "patch_size": 4,
62
  "problem_type": "multi_label_classification",
 
63
  "torch_dtype": "float32",
64
+ "transformers_version": "4.25.1"
 
65
  }
sonic/preprocessor_config.json CHANGED
@@ -1,19 +1,15 @@
1
  {
2
- "crop_size": {
3
- "height": 384,
4
- "width": 384
5
- },
6
- "do_center_crop": true,
7
  "do_normalize": true,
8
  "do_rescale": true,
9
  "do_resize": true,
10
- "feature_extractor_type": "DeiTFeatureExtractor",
11
  "image_mean": [
12
  0.485,
13
  0.456,
14
  0.406
15
  ],
16
- "image_processor_type": "DeiTImageProcessor",
17
  "image_std": [
18
  0.229,
19
  0.224,
@@ -22,7 +18,6 @@
22
  "resample": 3,
23
  "rescale_factor": 0.00392156862745098,
24
  "size": {
25
- "height": 438,
26
- "width": 438
27
  }
28
  }
 
1
  {
2
+ "crop_pct": 0.875,
 
 
 
 
3
  "do_normalize": true,
4
  "do_rescale": true,
5
  "do_resize": true,
6
+ "feature_extractor_type": "ConvNextFeatureExtractor",
7
  "image_mean": [
8
  0.485,
9
  0.456,
10
  0.406
11
  ],
12
+ "image_processor_type": "ConvNextImageProcessor",
13
  "image_std": [
14
  0.229,
15
  0.224,
 
18
  "resample": 3,
19
  "rescale_factor": 0.00392156862745098,
20
  "size": {
21
+ "shortest_edge": 384
 
22
  }
23
  }
sonic/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19a5c9f847dc16aca5d968b920576a77877b5ee936ec862475ae594d5cc717cd
3
- size 344482737
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd4d6897c65fc76cc210003c9552b14442dd4d5786bab4e5b323c52bafa2e877
3
+ size 350446061
sonic/trainer_state.json CHANGED
@@ -1,73 +1,177 @@
1
  {
2
- "best_metric": 0.09283134341239929,
3
- "best_model_checkpoint": "./SonicCharacterClassifier\\checkpoint-573",
4
- "epoch": 3.0,
5
- "global_step": 573,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
- "epoch": 0.52,
12
- "learning_rate": 4.738219895287958e-05,
13
- "loss": 0.2703,
14
  "step": 100
15
  },
16
  {
17
- "epoch": 1.0,
18
- "eval_accuracy": 0.7315789473684211,
19
- "eval_loss": 0.12457749992609024,
20
- "eval_runtime": 84.7067,
21
- "eval_samples_per_second": 8.972,
22
- "eval_steps_per_second": 1.122,
23
- "step": 191
24
  },
25
  {
26
- "epoch": 1.05,
27
- "learning_rate": 4.4764397905759164e-05,
28
- "loss": 0.1533,
 
 
 
 
 
29
  "step": 200
30
  },
31
  {
32
- "epoch": 1.57,
33
- "learning_rate": 4.2146596858638745e-05,
34
- "loss": 0.0971,
35
  "step": 300
36
  },
37
  {
38
- "epoch": 2.0,
39
- "eval_accuracy": 0.7355263157894737,
40
- "eval_loss": 0.10028258711099625,
41
- "eval_runtime": 80.347,
42
- "eval_samples_per_second": 9.459,
43
- "eval_steps_per_second": 1.182,
44
- "step": 382
45
  },
46
  {
47
- "epoch": 2.09,
48
- "learning_rate": 3.9528795811518326e-05,
49
- "loss": 0.0773,
 
 
 
 
 
50
  "step": 400
51
  },
52
  {
53
- "epoch": 2.62,
54
- "learning_rate": 3.691099476439791e-05,
55
- "loss": 0.04,
56
  "step": 500
57
  },
58
  {
59
- "epoch": 3.0,
60
- "eval_accuracy": 0.7447368421052631,
61
- "eval_loss": 0.09283134341239929,
62
- "eval_runtime": 80.5834,
63
- "eval_samples_per_second": 9.431,
64
- "eval_steps_per_second": 1.179,
65
- "step": 573
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
66
  }
67
  ],
68
- "max_steps": 1910,
69
  "num_train_epochs": 10,
70
- "total_flos": 4.182223007386829e+18,
71
  "trial_name": null,
72
  "trial_params": null
73
  }
 
1
  {
2
+ "best_metric": 0.09438519924879074,
3
+ "best_model_checkpoint": "./SonicCharacterClassifier\\checkpoint-1400",
4
+ "epoch": 4.093567251461988,
5
+ "global_step": 1400,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
+ "epoch": 0.29,
12
+ "learning_rate": 4.853801169590643e-05,
13
+ "loss": 0.3188,
14
  "step": 100
15
  },
16
  {
17
+ "epoch": 0.58,
18
+ "learning_rate": 4.707602339181287e-05,
19
+ "loss": 0.2087,
20
+ "step": 200
 
 
 
21
  },
22
  {
23
+ "epoch": 0.58,
24
+ "eval_accuracy": 0.9407407407407361,
25
+ "eval_loss": 0.16877754032611847,
26
+ "eval_precision": 0.6539702845075653,
27
+ "eval_recall": 0.5742313120602596,
28
+ "eval_runtime": 161.9664,
29
+ "eval_samples_per_second": 8.446,
30
+ "eval_steps_per_second": 1.056,
31
  "step": 200
32
  },
33
  {
34
+ "epoch": 0.88,
35
+ "learning_rate": 4.56140350877193e-05,
36
+ "loss": 0.1663,
37
  "step": 300
38
  },
39
  {
40
+ "epoch": 1.17,
41
+ "learning_rate": 4.4152046783625734e-05,
42
+ "loss": 0.1155,
43
+ "step": 400
 
 
 
44
  },
45
  {
46
+ "epoch": 1.17,
47
+ "eval_accuracy": 0.9615009746588669,
48
+ "eval_loss": 0.12015020102262497,
49
+ "eval_precision": 0.8305708768976313,
50
+ "eval_recall": 0.7801112001221648,
51
+ "eval_runtime": 161.8616,
52
+ "eval_samples_per_second": 8.452,
53
+ "eval_steps_per_second": 1.056,
54
  "step": 400
55
  },
56
  {
57
+ "epoch": 1.46,
58
+ "learning_rate": 4.269005847953216e-05,
59
+ "loss": 0.095,
60
  "step": 500
61
  },
62
  {
63
+ "epoch": 1.75,
64
+ "learning_rate": 4.12280701754386e-05,
65
+ "loss": 0.0932,
66
+ "step": 600
67
+ },
68
+ {
69
+ "epoch": 1.75,
70
+ "eval_accuracy": 0.964278752436644,
71
+ "eval_loss": 0.1112869530916214,
72
+ "eval_precision": 0.8358102748453625,
73
+ "eval_recall": 0.7573138768314205,
74
+ "eval_runtime": 160.6953,
75
+ "eval_samples_per_second": 8.513,
76
+ "eval_steps_per_second": 1.064,
77
+ "step": 600
78
+ },
79
+ {
80
+ "epoch": 2.05,
81
+ "learning_rate": 3.976608187134503e-05,
82
+ "loss": 0.087,
83
+ "step": 700
84
+ },
85
+ {
86
+ "epoch": 2.34,
87
+ "learning_rate": 3.8304093567251465e-05,
88
+ "loss": 0.0489,
89
+ "step": 800
90
+ },
91
+ {
92
+ "epoch": 2.34,
93
+ "eval_accuracy": 0.9684697855750467,
94
+ "eval_loss": 0.09839322417974472,
95
+ "eval_precision": 0.877861029889539,
96
+ "eval_recall": 0.8274951210258222,
97
+ "eval_runtime": 160.5107,
98
+ "eval_samples_per_second": 8.523,
99
+ "eval_steps_per_second": 1.065,
100
+ "step": 800
101
+ },
102
+ {
103
+ "epoch": 2.63,
104
+ "learning_rate": 3.6842105263157895e-05,
105
+ "loss": 0.0506,
106
+ "step": 900
107
+ },
108
+ {
109
+ "epoch": 2.92,
110
+ "learning_rate": 3.538011695906433e-05,
111
+ "loss": 0.0503,
112
+ "step": 1000
113
+ },
114
+ {
115
+ "epoch": 2.92,
116
+ "eval_accuracy": 0.9692495126705621,
117
+ "eval_loss": 0.09646110981702805,
118
+ "eval_precision": 0.8755529497305813,
119
+ "eval_recall": 0.8374415545468176,
120
+ "eval_runtime": 161.798,
121
+ "eval_samples_per_second": 8.455,
122
+ "eval_steps_per_second": 1.057,
123
+ "step": 1000
124
+ },
125
+ {
126
+ "epoch": 3.22,
127
+ "learning_rate": 3.391812865497076e-05,
128
+ "loss": 0.0297,
129
+ "step": 1100
130
+ },
131
+ {
132
+ "epoch": 3.51,
133
+ "learning_rate": 3.24561403508772e-05,
134
+ "loss": 0.0245,
135
+ "step": 1200
136
+ },
137
+ {
138
+ "epoch": 3.51,
139
+ "eval_accuracy": 0.9697855750487306,
140
+ "eval_loss": 0.09691683202981949,
141
+ "eval_precision": 0.8796307635671673,
142
+ "eval_recall": 0.8392919296428067,
143
+ "eval_runtime": 163.6772,
144
+ "eval_samples_per_second": 8.358,
145
+ "eval_steps_per_second": 1.045,
146
+ "step": 1200
147
+ },
148
+ {
149
+ "epoch": 3.8,
150
+ "learning_rate": 3.0994152046783626e-05,
151
+ "loss": 0.026,
152
+ "step": 1300
153
+ },
154
+ {
155
+ "epoch": 4.09,
156
+ "learning_rate": 2.9532163742690062e-05,
157
+ "loss": 0.0231,
158
+ "step": 1400
159
+ },
160
+ {
161
+ "epoch": 4.09,
162
+ "eval_accuracy": 0.9713937621832336,
163
+ "eval_loss": 0.09438519924879074,
164
+ "eval_precision": 0.8850769600769599,
165
+ "eval_recall": 0.8511938435184049,
166
+ "eval_runtime": 161.9688,
167
+ "eval_samples_per_second": 8.446,
168
+ "eval_steps_per_second": 1.056,
169
+ "step": 1400
170
  }
171
  ],
172
+ "max_steps": 3420,
173
  "num_train_epochs": 10,
174
+ "total_flos": 5.207113357118669e+18,
175
  "trial_name": null,
176
  "trial_params": null
177
  }