kbberendsen commited on
Commit
cde662d
·
verified ·
1 Parent(s): 863e739

Training in progress, epoch 1

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. model.safetensors +1 -1
  2. run-2/checkpoint-28/config.json +25 -0
  3. run-2/checkpoint-28/model.safetensors +3 -0
  4. run-2/checkpoint-28/optimizer.pt +3 -0
  5. run-2/checkpoint-28/rng_state.pth +3 -0
  6. run-2/checkpoint-28/scheduler.pt +3 -0
  7. run-2/checkpoint-28/special_tokens_map.json +7 -0
  8. run-2/checkpoint-28/tokenizer.json +0 -0
  9. run-2/checkpoint-28/tokenizer_config.json +55 -0
  10. run-2/checkpoint-28/trainer_state.json +44 -0
  11. run-2/checkpoint-28/training_args.bin +3 -0
  12. run-2/checkpoint-28/vocab.txt +0 -0
  13. run-2/checkpoint-42/config.json +25 -0
  14. run-2/checkpoint-42/model.safetensors +3 -0
  15. run-2/checkpoint-42/optimizer.pt +3 -0
  16. run-2/checkpoint-42/rng_state.pth +3 -0
  17. run-2/checkpoint-42/scheduler.pt +3 -0
  18. run-2/checkpoint-42/special_tokens_map.json +7 -0
  19. run-2/checkpoint-42/tokenizer.json +0 -0
  20. run-2/checkpoint-42/tokenizer_config.json +55 -0
  21. run-2/checkpoint-42/trainer_state.json +53 -0
  22. run-2/checkpoint-42/training_args.bin +3 -0
  23. run-2/checkpoint-42/vocab.txt +0 -0
  24. run-3/checkpoint-1070/config.json +25 -0
  25. run-3/checkpoint-1070/model.safetensors +3 -0
  26. run-3/checkpoint-1070/optimizer.pt +3 -0
  27. run-3/checkpoint-1070/rng_state.pth +3 -0
  28. run-3/checkpoint-1070/scheduler.pt +3 -0
  29. run-3/checkpoint-1070/special_tokens_map.json +7 -0
  30. run-3/checkpoint-1070/tokenizer.json +0 -0
  31. run-3/checkpoint-1070/tokenizer_config.json +55 -0
  32. run-3/checkpoint-1070/trainer_state.json +85 -0
  33. run-3/checkpoint-1070/training_args.bin +3 -0
  34. run-3/checkpoint-1070/vocab.txt +0 -0
  35. run-3/checkpoint-214/config.json +25 -0
  36. run-3/checkpoint-214/model.safetensors +3 -0
  37. run-3/checkpoint-214/optimizer.pt +3 -0
  38. run-3/checkpoint-214/rng_state.pth +3 -0
  39. run-3/checkpoint-214/scheduler.pt +3 -0
  40. run-3/checkpoint-214/special_tokens_map.json +7 -0
  41. run-3/checkpoint-214/tokenizer.json +0 -0
  42. run-3/checkpoint-214/tokenizer_config.json +55 -0
  43. run-3/checkpoint-214/trainer_state.json +35 -0
  44. run-3/checkpoint-214/training_args.bin +3 -0
  45. run-3/checkpoint-214/vocab.txt +0 -0
  46. run-3/checkpoint-428/config.json +25 -0
  47. run-3/checkpoint-428/model.safetensors +3 -0
  48. run-3/checkpoint-428/optimizer.pt +3 -0
  49. run-3/checkpoint-428/rng_state.pth +3 -0
  50. run-3/checkpoint-428/scheduler.pt +3 -0
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da46fb04c766dfa02bb30bf71844c048a72006148d901a09352bc389782e39f4
3
  size 267832560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:452d9ed5c0693a004691e67d1911da2f6d537364410446882d40da14ca1013e6
3
  size 267832560
run-2/checkpoint-28/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-uncased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "initializer_range": 0.02,
12
+ "max_position_embeddings": 512,
13
+ "model_type": "distilbert",
14
+ "n_heads": 12,
15
+ "n_layers": 6,
16
+ "pad_token_id": 0,
17
+ "problem_type": "single_label_classification",
18
+ "qa_dropout": 0.1,
19
+ "seq_classif_dropout": 0.2,
20
+ "sinusoidal_pos_embds": false,
21
+ "tie_weights_": true,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.38.1",
24
+ "vocab_size": 30522
25
+ }
run-2/checkpoint-28/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21c0bf68fbe224eda477f5cb8ee40f39c2c32264d62ce03ce7b0152c11516c21
3
+ size 267832560
run-2/checkpoint-28/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a6aeb3929ec3213116667d38a17155d6da94b0730cb49c94cb2bd3b250dfbed
3
+ size 535727290
run-2/checkpoint-28/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0b84025de59ee0b253f7a8c26c6d5e8bd41da85b344bd6d56e1c12a05a09bd3
3
+ size 14308
run-2/checkpoint-28/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:455a0d19ffe2f2ba49d7ccb8bb670aad81b5a404431b5ebb1f46d693bdc232ac
3
+ size 1064
run-2/checkpoint-28/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-2/checkpoint-28/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-2/checkpoint-28/tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "DistilBertTokenizer",
54
+ "unk_token": "[UNK]"
55
+ }
run-2/checkpoint-28/trainer_state.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.0,
3
+ "best_model_checkpoint": "distilbert-base-uncased-finetuned-cola/run-2/checkpoint-14",
4
+ "epoch": 2.0,
5
+ "eval_steps": 500,
6
+ "global_step": 28,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_loss": 0.6168319582939148,
14
+ "eval_matthews_correlation": 0.0,
15
+ "eval_runtime": 0.7436,
16
+ "eval_samples_per_second": 1402.622,
17
+ "eval_steps_per_second": 88.756,
18
+ "step": 14
19
+ },
20
+ {
21
+ "epoch": 2.0,
22
+ "eval_loss": 0.6112757921218872,
23
+ "eval_matthews_correlation": 0.0,
24
+ "eval_runtime": 0.8465,
25
+ "eval_samples_per_second": 1232.11,
26
+ "eval_steps_per_second": 77.967,
27
+ "step": 28
28
+ }
29
+ ],
30
+ "logging_steps": 500,
31
+ "max_steps": 42,
32
+ "num_input_tokens_seen": 0,
33
+ "num_train_epochs": 3,
34
+ "save_steps": 500,
35
+ "total_flos": 0,
36
+ "train_batch_size": 64,
37
+ "trial_name": null,
38
+ "trial_params": {
39
+ "learning_rate": 9.743882443444139e-06,
40
+ "num_train_epochs": 3,
41
+ "per_device_train_batch_size": 64,
42
+ "seed": 29
43
+ }
44
+ }
run-2/checkpoint-28/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c28ecd07ec5597508913119261774d29a3ed562f21aba4dfbe51170c935cb693
3
+ size 4984
run-2/checkpoint-28/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
run-2/checkpoint-42/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-uncased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "initializer_range": 0.02,
12
+ "max_position_embeddings": 512,
13
+ "model_type": "distilbert",
14
+ "n_heads": 12,
15
+ "n_layers": 6,
16
+ "pad_token_id": 0,
17
+ "problem_type": "single_label_classification",
18
+ "qa_dropout": 0.1,
19
+ "seq_classif_dropout": 0.2,
20
+ "sinusoidal_pos_embds": false,
21
+ "tie_weights_": true,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.38.1",
24
+ "vocab_size": 30522
25
+ }
run-2/checkpoint-42/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f3490dec0bd37f5669061aa44eb1d6db58d4103b547f0cebaa6f2375be9c535
3
+ size 267832560
run-2/checkpoint-42/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8e4963310780120ccb6093a48a643168debf2d7ef4f74bd067c76d4f39e6bf2
3
+ size 535727290
run-2/checkpoint-42/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1d1e04ec28dde90a91d7963cf2642d878d182f10435134b2229fe5b31152ebf
3
+ size 14308
run-2/checkpoint-42/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93fbfbcf3c8db8f663ddec070707e4161bf0eabc28bcd5ff140add4adc1e059d
3
+ size 1064
run-2/checkpoint-42/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-2/checkpoint-42/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-2/checkpoint-42/tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "DistilBertTokenizer",
54
+ "unk_token": "[UNK]"
55
+ }
run-2/checkpoint-42/trainer_state.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.0,
3
+ "best_model_checkpoint": "distilbert-base-uncased-finetuned-cola/run-2/checkpoint-14",
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 42,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_loss": 0.6168319582939148,
14
+ "eval_matthews_correlation": 0.0,
15
+ "eval_runtime": 0.7436,
16
+ "eval_samples_per_second": 1402.622,
17
+ "eval_steps_per_second": 88.756,
18
+ "step": 14
19
+ },
20
+ {
21
+ "epoch": 2.0,
22
+ "eval_loss": 0.6112757921218872,
23
+ "eval_matthews_correlation": 0.0,
24
+ "eval_runtime": 0.8465,
25
+ "eval_samples_per_second": 1232.11,
26
+ "eval_steps_per_second": 77.967,
27
+ "step": 28
28
+ },
29
+ {
30
+ "epoch": 3.0,
31
+ "eval_loss": 0.6096367835998535,
32
+ "eval_matthews_correlation": 0.0,
33
+ "eval_runtime": 0.7961,
34
+ "eval_samples_per_second": 1310.218,
35
+ "eval_steps_per_second": 82.909,
36
+ "step": 42
37
+ }
38
+ ],
39
+ "logging_steps": 500,
40
+ "max_steps": 42,
41
+ "num_input_tokens_seen": 0,
42
+ "num_train_epochs": 3,
43
+ "save_steps": 500,
44
+ "total_flos": 0,
45
+ "train_batch_size": 64,
46
+ "trial_name": null,
47
+ "trial_params": {
48
+ "learning_rate": 9.743882443444139e-06,
49
+ "num_train_epochs": 3,
50
+ "per_device_train_batch_size": 64,
51
+ "seed": 29
52
+ }
53
+ }
run-2/checkpoint-42/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c28ecd07ec5597508913119261774d29a3ed562f21aba4dfbe51170c935cb693
3
+ size 4984
run-2/checkpoint-42/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
run-3/checkpoint-1070/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-uncased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "initializer_range": 0.02,
12
+ "max_position_embeddings": 512,
13
+ "model_type": "distilbert",
14
+ "n_heads": 12,
15
+ "n_layers": 6,
16
+ "pad_token_id": 0,
17
+ "problem_type": "single_label_classification",
18
+ "qa_dropout": 0.1,
19
+ "seq_classif_dropout": 0.2,
20
+ "sinusoidal_pos_embds": false,
21
+ "tie_weights_": true,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.38.1",
24
+ "vocab_size": 30522
25
+ }
run-3/checkpoint-1070/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d71bebba48dab02a513130549649d850f8caa8c64ec08581b5bb93453e78b93e
3
+ size 267832560
run-3/checkpoint-1070/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ab04bb006da56612ffd2000113bba394febf3e04f74c9238aa4bff54060b1e2
3
+ size 535727290
run-3/checkpoint-1070/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81f198ffd7068bcb4af877a4fe1ef392884503fd4eb8a3427369f89fcc2851c1
3
+ size 14244
run-3/checkpoint-1070/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25c2e7eb26c50cafaf9591e622cab9d35507ed3cb084bb2df8019bb7073f41c9
3
+ size 1064
run-3/checkpoint-1070/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-3/checkpoint-1070/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-3/checkpoint-1070/tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "DistilBertTokenizer",
54
+ "unk_token": "[UNK]"
55
+ }
run-3/checkpoint-1070/trainer_state.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.3540182182677824,
3
+ "best_model_checkpoint": "distilbert-base-uncased-finetuned-cola/run-3/checkpoint-856",
4
+ "epoch": 5.0,
5
+ "eval_steps": 500,
6
+ "global_step": 1070,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_loss": 0.6083057522773743,
14
+ "eval_matthews_correlation": 0.0,
15
+ "eval_runtime": 0.7911,
16
+ "eval_samples_per_second": 1318.406,
17
+ "eval_steps_per_second": 83.427,
18
+ "step": 214
19
+ },
20
+ {
21
+ "epoch": 2.0,
22
+ "eval_loss": 0.5953883528709412,
23
+ "eval_matthews_correlation": 0.3250986825407545,
24
+ "eval_runtime": 0.8711,
25
+ "eval_samples_per_second": 1197.404,
26
+ "eval_steps_per_second": 75.771,
27
+ "step": 428
28
+ },
29
+ {
30
+ "epoch": 2.34,
31
+ "grad_norm": 3.379061222076416,
32
+ "learning_rate": 5.696646790637325e-06,
33
+ "loss": 0.5111,
34
+ "step": 500
35
+ },
36
+ {
37
+ "epoch": 3.0,
38
+ "eval_loss": 0.8310299515724182,
39
+ "eval_matthews_correlation": 0.3476302977665388,
40
+ "eval_runtime": 1.4001,
41
+ "eval_samples_per_second": 744.949,
42
+ "eval_steps_per_second": 47.14,
43
+ "step": 642
44
+ },
45
+ {
46
+ "epoch": 4.0,
47
+ "eval_loss": 1.0389337539672852,
48
+ "eval_matthews_correlation": 0.3540182182677824,
49
+ "eval_runtime": 1.434,
50
+ "eval_samples_per_second": 727.328,
51
+ "eval_steps_per_second": 46.025,
52
+ "step": 856
53
+ },
54
+ {
55
+ "epoch": 4.67,
56
+ "grad_norm": 0.10751748085021973,
57
+ "learning_rate": 6.995882023589696e-07,
58
+ "loss": 0.1674,
59
+ "step": 1000
60
+ },
61
+ {
62
+ "epoch": 5.0,
63
+ "eval_loss": 1.1440064907073975,
64
+ "eval_matthews_correlation": 0.3520719211856073,
65
+ "eval_runtime": 0.9814,
66
+ "eval_samples_per_second": 1062.811,
67
+ "eval_steps_per_second": 67.254,
68
+ "step": 1070
69
+ }
70
+ ],
71
+ "logging_steps": 500,
72
+ "max_steps": 1070,
73
+ "num_input_tokens_seen": 0,
74
+ "num_train_epochs": 5,
75
+ "save_steps": 500,
76
+ "total_flos": 16663726064916.0,
77
+ "train_batch_size": 4,
78
+ "trial_name": null,
79
+ "trial_params": {
80
+ "learning_rate": 1.069370537891568e-05,
81
+ "num_train_epochs": 5,
82
+ "per_device_train_batch_size": 4,
83
+ "seed": 40
84
+ }
85
+ }
run-3/checkpoint-1070/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a33fe80cc2d2bb9d1b804fc05ab4c4183d86de256301a43cb40d78354723a680
3
+ size 4984
run-3/checkpoint-1070/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
run-3/checkpoint-214/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-uncased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "initializer_range": 0.02,
12
+ "max_position_embeddings": 512,
13
+ "model_type": "distilbert",
14
+ "n_heads": 12,
15
+ "n_layers": 6,
16
+ "pad_token_id": 0,
17
+ "problem_type": "single_label_classification",
18
+ "qa_dropout": 0.1,
19
+ "seq_classif_dropout": 0.2,
20
+ "sinusoidal_pos_embds": false,
21
+ "tie_weights_": true,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.38.1",
24
+ "vocab_size": 30522
25
+ }
run-3/checkpoint-214/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f483026ae9d1934e62ea073317475abf52df3f44c430aaada8ab9ff767aea92
3
+ size 267832560
run-3/checkpoint-214/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ee916f9008266a791db70d8adf7b79de4ff6212a7d86b59cde920200cf0be6d
3
+ size 535727290
run-3/checkpoint-214/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a232c3a4f6a37e2a3be6bd8372d670bf2f7c0d0bcfd968173b612c27189b597
3
+ size 14244
run-3/checkpoint-214/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30e8bcf5aca0d33cc5cd0c46ef5c55dfe7516608dc2b6f1696884a05d28a62cd
3
+ size 1064
run-3/checkpoint-214/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-3/checkpoint-214/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-3/checkpoint-214/tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "DistilBertTokenizer",
54
+ "unk_token": "[UNK]"
55
+ }
run-3/checkpoint-214/trainer_state.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.0,
3
+ "best_model_checkpoint": "distilbert-base-uncased-finetuned-cola/run-3/checkpoint-214",
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 214,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_loss": 0.6083057522773743,
14
+ "eval_matthews_correlation": 0.0,
15
+ "eval_runtime": 0.7911,
16
+ "eval_samples_per_second": 1318.406,
17
+ "eval_steps_per_second": 83.427,
18
+ "step": 214
19
+ }
20
+ ],
21
+ "logging_steps": 500,
22
+ "max_steps": 1070,
23
+ "num_input_tokens_seen": 0,
24
+ "num_train_epochs": 5,
25
+ "save_steps": 500,
26
+ "total_flos": 0,
27
+ "train_batch_size": 4,
28
+ "trial_name": null,
29
+ "trial_params": {
30
+ "learning_rate": 1.069370537891568e-05,
31
+ "num_train_epochs": 5,
32
+ "per_device_train_batch_size": 4,
33
+ "seed": 40
34
+ }
35
+ }
run-3/checkpoint-214/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a33fe80cc2d2bb9d1b804fc05ab4c4183d86de256301a43cb40d78354723a680
3
+ size 4984
run-3/checkpoint-214/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
run-3/checkpoint-428/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-uncased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "initializer_range": 0.02,
12
+ "max_position_embeddings": 512,
13
+ "model_type": "distilbert",
14
+ "n_heads": 12,
15
+ "n_layers": 6,
16
+ "pad_token_id": 0,
17
+ "problem_type": "single_label_classification",
18
+ "qa_dropout": 0.1,
19
+ "seq_classif_dropout": 0.2,
20
+ "sinusoidal_pos_embds": false,
21
+ "tie_weights_": true,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.38.1",
24
+ "vocab_size": 30522
25
+ }
run-3/checkpoint-428/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1f1d2485c5eff9981dfaf5889f2dcb1ec67b0a1b3647c0401f976cdfc2cf3b6
3
+ size 267832560
run-3/checkpoint-428/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ae541aadf5536ea1beaf7c5be8a8dc8d07c5281e3a98d424e2b3c936f5249d8
3
+ size 535727290
run-3/checkpoint-428/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ae2cd9d962a0ab4552cc3511b460ead58646cc1258fd300323b7e26f30cdc1a
3
+ size 14244
run-3/checkpoint-428/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:655b7f6e59a00d1b17f98e0a2420f4ca42e697422e8c1a1b8a89a20982277d8a
3
+ size 1064