lmytimagedata commited on
Commit
cb14680
·
verified ·
1 Parent(s): af06679

Upload checkpoint-50000

Browse files
checkpoint-50000/config.json ADDED
@@ -0,0 +1,187 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "batch_size": 32,
3
+ "buffer_size": 64,
4
+ "eval_mix": "qiwu_test",
5
+ "frozen_keys": [
6
+ "*hf_model*"
7
+ ],
8
+ "lora": false,
9
+ "lora_config": {
10
+ "bias": "none",
11
+ "lora_alpha": 16,
12
+ "lora_dropout": 0.05,
13
+ "r": 8
14
+ },
15
+ "model": {
16
+ "heads": {
17
+ "action": {
18
+ "args": [],
19
+ "kwargs": {
20
+ "action_dim": 7,
21
+ "action_horizon": 4,
22
+ "dropout_rate": 0.0,
23
+ "n_diffusion_samples": 1,
24
+ "readout_key": "readout_action",
25
+ "token_embedding_size": 768,
26
+ "use_map": false
27
+ },
28
+ "module": "octo.model.components.action_heads",
29
+ "name": "DiffusionActionHead"
30
+ }
31
+ },
32
+ "max_horizon": 10,
33
+ "observation_tokenizers": {
34
+ "primary": {
35
+ "args": [],
36
+ "kwargs": {
37
+ "encoder": {
38
+ "args": [],
39
+ "kwargs": {
40
+ "in_features": 6
41
+ },
42
+ "module": "octo.model.components.vit_encoders",
43
+ "name": "SmallStem16"
44
+ },
45
+ "obs_stack_keys": [
46
+ "image_primary"
47
+ ],
48
+ "task_stack_keys": [
49
+ "image_primary"
50
+ ]
51
+ },
52
+ "module": "octo.model.components.tokenizers",
53
+ "name": "ImageTokenizer"
54
+ },
55
+ "state": {
56
+ "args": [],
57
+ "kwargs": {
58
+ "bin_type": "normal",
59
+ "high": 2.0,
60
+ "low": -2.0,
61
+ "n_bins": 256,
62
+ "obs_keys": [
63
+ "state"
64
+ ]
65
+ },
66
+ "module": "model.octo.components.tokenizers",
67
+ "name": "LowdimObsTokenizer"
68
+ },
69
+ "wrist": {
70
+ "args": [],
71
+ "kwargs": {
72
+ "encoder": {
73
+ "args": [],
74
+ "kwargs": {
75
+ "in_features": 6
76
+ },
77
+ "module": "octo.model.components.vit_encoders",
78
+ "name": "SmallStem16"
79
+ },
80
+ "obs_stack_keys": [
81
+ "image_wrist"
82
+ ],
83
+ "task_stack_keys": [
84
+ "image_wrist"
85
+ ]
86
+ },
87
+ "module": "octo.model.components.tokenizers",
88
+ "name": "ImageTokenizer"
89
+ }
90
+ },
91
+ "readouts": {
92
+ "action": 1
93
+ },
94
+ "repeat_task_tokens": true,
95
+ "task_tokenizers": {
96
+ "language": {
97
+ "args": [],
98
+ "kwargs": {
99
+ "encoder": "t5-base",
100
+ "finetune_encoder": false
101
+ },
102
+ "module": "octo.model.components.tokenizers",
103
+ "name": "LanguageTokenizer"
104
+ }
105
+ },
106
+ "token_embedding_size": 768,
107
+ "transformer_kwargs": {
108
+ "add_position_embedding": false,
109
+ "attention_dropout_rate": 0.0,
110
+ "dropout_rate": 0.0,
111
+ "mlp_dim": 3072,
112
+ "num_attention_heads": 12,
113
+ "num_layers": 12
114
+ },
115
+ "use_correct_attention": true
116
+ },
117
+ "no_state": false,
118
+ "obs_token_nums": {
119
+ "primary": 256,
120
+ "state": 7,
121
+ "wrist": 64
122
+ },
123
+ "output_dir": "ljp_qiwu_banana_4",
124
+ "pretrained": {
125
+ "kwargs": {
126
+ "subpath": "oxe-g2-checkpoint-300000"
127
+ },
128
+ "model": "hf://chuanmew/octo_torch"
129
+ },
130
+ "run_name": "qiwu_banana_4",
131
+ "sampler_num_samples": 2560000,
132
+ "seed": 42,
133
+ "subsample_length": 99999,
134
+ "text_processor": {
135
+ "args": [],
136
+ "kwargs": {
137
+ "encode_with_model": false,
138
+ "tokenizer_kwargs": {
139
+ "max_length": 16,
140
+ "padding": "max_length",
141
+ "return_tensors": "np",
142
+ "truncation": true
143
+ },
144
+ "tokenizer_name": "t5-base"
145
+ },
146
+ "module": "octo.components.text_processing",
147
+ "name": "HFTokenizer"
148
+ },
149
+ "train_mix": "ljp_qiwu_banana",
150
+ "training_arguments": {
151
+ "bf16": true,
152
+ "dataloader_pin_memory": true,
153
+ "ddp_find_unused_parameters": true,
154
+ "eval_steps": 50000,
155
+ "gradient_accumulation_steps": 1,
156
+ "learning_rate": 0.0001,
157
+ "logging_nan_inf_filter": false,
158
+ "logging_steps": 10000,
159
+ "lr_scheduler_kwargs": {
160
+ "decay_type": "cosine",
161
+ "min_lr_ratio": 0.05,
162
+ "num_decay_steps": 47500
163
+ },
164
+ "lr_scheduler_type": "warmup_stable_decay",
165
+ "max_grad_norm": 20,
166
+ "max_steps": 50000,
167
+ "optim": "adamw_torch_fused",
168
+ "per_device_eval_batch_size": 1,
169
+ "per_device_train_batch_size": 1,
170
+ "report_to": "wandb",
171
+ "save_safetensors": true,
172
+ "save_steps": 50000,
173
+ "torch_compile": true,
174
+ "warmup_steps": 2500,
175
+ "weight_decay": 0.1
176
+ },
177
+ "training_keys": [
178
+ "*action*",
179
+ "*observation_tokenizers*",
180
+ "*transformer.transformer*",
181
+ "*task_projections*",
182
+ "*obs_projections*",
183
+ "*readout_embeddings*",
184
+ "*task_pos_embeddings*",
185
+ "*obs_pos_embeddings*"
186
+ ]
187
+ }
checkpoint-50000/dataset_statistics.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "action": {
4
+ "mean": [
5
+ 0.0035762591287493706,
6
+ 0.0038505252450704575,
7
+ 0.0005051779444329441,
8
+ 0.004462657496333122,
9
+ 0.00569327874109149,
10
+ 0.00843613687902689,
11
+ 0.5597720146179199
12
+ ],
13
+ "std": [
14
+ 0.030762936919927597,
15
+ 0.022843312472105026,
16
+ 0.037356432527303696,
17
+ 0.07383003830909729,
18
+ 0.07964984327554703,
19
+ 0.10997529327869415,
20
+ 0.45039603114128113
21
+ ],
22
+ "min": [
23
+ -0.12550033628940582,
24
+ -0.08252930641174316,
25
+ -0.08629363775253296,
26
+ -0.5667345523834229,
27
+ -0.5142402648925781,
28
+ -1.2305364608764648,
29
+ 0.0
30
+ ],
31
+ "max": [
32
+ 0.15785643458366394,
33
+ 0.1229468286037445,
34
+ 0.15745647251605988,
35
+ 0.9040577411651611,
36
+ 0.43572473526000977,
37
+ 1.0832273960113525,
38
+ 1.0
39
+ ],
40
+ "mask": [
41
+ true,
42
+ true,
43
+ true,
44
+ true,
45
+ true,
46
+ true,
47
+ false
48
+ ]
49
+ },
50
+ "state": {
51
+ "mean": [
52
+ -1.147829294204712,
53
+ -1.018812894821167,
54
+ -1.342349648475647,
55
+ -0.603922963142395,
56
+ -0.9552030563354492,
57
+ 0.9721980690956116,
58
+ 0.5594711899757385
59
+ ],
60
+ "std": [
61
+ 0.2672520875930786,
62
+ 0.21161462366580963,
63
+ 0.19813641905784607,
64
+ 0.20742598176002502,
65
+ 0.2395419031381607,
66
+ 0.8849623203277588,
67
+ 0.45082923769950867
68
+ ],
69
+ "min": [
70
+ -1.749814748764038,
71
+ -1.5758228302001953,
72
+ -1.8777823448181152,
73
+ -1.1325790882110596,
74
+ -1.702690839767456,
75
+ -0.8983733057975769,
76
+ 0.0
77
+ ],
78
+ "max": [
79
+ -0.5615771412849426,
80
+ -0.41973423957824707,
81
+ -0.6838898062705994,
82
+ 0.18788468837738037,
83
+ -0.38179078698158264,
84
+ 3.117105722427368,
85
+ 1.0
86
+ ],
87
+ "mask": [
88
+ true,
89
+ true,
90
+ true,
91
+ true,
92
+ true,
93
+ true,
94
+ false
95
+ ]
96
+ }
97
+ }
98
+ ]
checkpoint-50000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:591d0e746a6743eb56fe8058da598ad85c535d6fd5e89abb8c393429c0c223c1
3
+ size 405079824
checkpoint-50000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f67351955e514f45c5287ad47d595585f54b901494501ce48226b556f1c1529
3
+ size 371730170
checkpoint-50000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f036448e1c518cbfba99f8dc6073173a8945e5e544e3be8b29689efe55c5d33c
3
+ size 14244
checkpoint-50000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c41ebb9a2c5635768a739e9e1b728bd4c30879d068859a65ed1c92f2838a08b
3
+ size 1064
checkpoint-50000/trainer_state.json ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.625,
6
+ "eval_steps": 50000,
7
+ "global_step": 50000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.125,
14
+ "grad_norm": 15.0625,
15
+ "learning_rate": 9.427649832377157e-05,
16
+ "loss": 0.5293,
17
+ "step": 10000
18
+ },
19
+ {
20
+ "epoch": 0.25,
21
+ "grad_norm": 17.375,
22
+ "learning_rate": 7.158340961603695e-05,
23
+ "loss": 0.2863,
24
+ "step": 20000
25
+ },
26
+ {
27
+ "epoch": 0.375,
28
+ "grad_norm": 11.875,
29
+ "learning_rate": 4.08424848470673e-05,
30
+ "loss": 0.2011,
31
+ "step": 30000
32
+ },
33
+ {
34
+ "epoch": 0.5,
35
+ "grad_norm": 16.25,
36
+ "learning_rate": 1.5017755491800084e-05,
37
+ "loss": 0.1537,
38
+ "step": 40000
39
+ },
40
+ {
41
+ "epoch": 0.625,
42
+ "grad_norm": 12.875,
43
+ "learning_rate": 5.000000103890576e-06,
44
+ "loss": 0.1406,
45
+ "step": 50000
46
+ },
47
+ {
48
+ "epoch": 0.625,
49
+ "eval_loss": 2.7418313026428223,
50
+ "eval_runtime": 50.5148,
51
+ "eval_samples_per_second": 400.219,
52
+ "eval_steps_per_second": 400.219,
53
+ "step": 50000
54
+ }
55
+ ],
56
+ "logging_steps": 10000,
57
+ "max_steps": 50000,
58
+ "num_input_tokens_seen": 0,
59
+ "num_train_epochs": 1,
60
+ "save_steps": 50000,
61
+ "stateful_callbacks": {
62
+ "TrainerControl": {
63
+ "args": {
64
+ "should_epoch_stop": false,
65
+ "should_evaluate": false,
66
+ "should_log": false,
67
+ "should_save": true,
68
+ "should_training_stop": true
69
+ },
70
+ "attributes": {}
71
+ }
72
+ },
73
+ "total_flos": 0.0,
74
+ "train_batch_size": 1,
75
+ "trial_name": null,
76
+ "trial_params": null
77
+ }
checkpoint-50000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64ce60c98eb243ba0fd307e9bcc15f6403531e42eace07bf6449d73b31ae1b71
3
+ size 5432