educhav commited on
Commit
d7418dd
·
1 Parent(s): 48e42d1

remove previous checkpoints

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. checkpoint-10500/config.json +0 -36
  2. checkpoint-10500/merges.txt +0 -0
  3. checkpoint-10500/optimizer.pt +0 -3
  4. checkpoint-10500/pytorch_model.bin +0 -3
  5. checkpoint-10500/scheduler.pt +0 -3
  6. checkpoint-10500/special_tokens_map.json +0 -1
  7. checkpoint-10500/tokenizer.json +0 -0
  8. checkpoint-10500/tokenizer_config.json +0 -1
  9. checkpoint-10500/training_args.bin +0 -3
  10. checkpoint-10500/vocab.json +0 -0
  11. checkpoint-14000/config.json +0 -36
  12. checkpoint-14000/merges.txt +0 -0
  13. checkpoint-14000/optimizer.pt +0 -3
  14. checkpoint-14000/pytorch_model.bin +0 -3
  15. checkpoint-14000/scheduler.pt +0 -3
  16. checkpoint-14000/special_tokens_map.json +0 -1
  17. checkpoint-14000/tokenizer.json +0 -0
  18. checkpoint-14000/tokenizer_config.json +0 -1
  19. checkpoint-14000/training_args.bin +0 -3
  20. checkpoint-14000/vocab.json +0 -0
  21. checkpoint-17500/config.json +0 -36
  22. checkpoint-17500/merges.txt +0 -0
  23. checkpoint-17500/optimizer.pt +0 -3
  24. checkpoint-17500/pytorch_model.bin +0 -3
  25. checkpoint-17500/scheduler.pt +0 -3
  26. checkpoint-17500/special_tokens_map.json +0 -1
  27. checkpoint-17500/tokenizer.json +0 -0
  28. checkpoint-17500/tokenizer_config.json +0 -1
  29. checkpoint-17500/training_args.bin +0 -3
  30. checkpoint-17500/vocab.json +0 -0
  31. checkpoint-21000/config.json +0 -36
  32. checkpoint-21000/merges.txt +0 -0
  33. checkpoint-21000/optimizer.pt +0 -3
  34. checkpoint-21000/pytorch_model.bin +0 -3
  35. checkpoint-21000/scheduler.pt +0 -3
  36. checkpoint-21000/special_tokens_map.json +0 -1
  37. checkpoint-21000/tokenizer.json +0 -0
  38. checkpoint-21000/tokenizer_config.json +0 -1
  39. checkpoint-21000/training_args.bin +0 -3
  40. checkpoint-21000/vocab.json +0 -0
  41. checkpoint-24500/config.json +0 -36
  42. checkpoint-24500/merges.txt +0 -0
  43. checkpoint-24500/optimizer.pt +0 -3
  44. checkpoint-24500/pytorch_model.bin +0 -3
  45. checkpoint-24500/scheduler.pt +0 -3
  46. checkpoint-24500/special_tokens_map.json +0 -1
  47. checkpoint-24500/tokenizer.json +0 -0
  48. checkpoint-24500/tokenizer_config.json +0 -1
  49. checkpoint-24500/training_args.bin +0 -3
  50. checkpoint-24500/vocab.json +0 -0
checkpoint-10500/config.json DELETED
@@ -1,36 +0,0 @@
1
- {
2
- "_name_or_path": "microsoft/DialoGPT-small",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPT2LMHeadModel"
6
- ],
7
- "attn_pdrop": 0.1,
8
- "bos_token_id": 50256,
9
- "embd_pdrop": 0.1,
10
- "eos_token_id": 50256,
11
- "initializer_range": 0.02,
12
- "layer_norm_epsilon": 1e-05,
13
- "model_type": "gpt2",
14
- "n_ctx": 1024,
15
- "n_embd": 768,
16
- "n_head": 12,
17
- "n_inner": null,
18
- "n_layer": 12,
19
- "n_positions": 1024,
20
- "resid_pdrop": 0.1,
21
- "scale_attn_weights": true,
22
- "summary_activation": null,
23
- "summary_first_dropout": 0.1,
24
- "summary_proj_to_labels": true,
25
- "summary_type": "cls_index",
26
- "summary_use_proj": true,
27
- "task_specific_params": {
28
- "conversational": {
29
- "max_length": 1000
30
- }
31
- },
32
- "torch_dtype": "float32",
33
- "transformers_version": "4.11.3",
34
- "use_cache": true,
35
- "vocab_size": 50257
36
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-10500/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-10500/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:984f039e1beb41ed40aa47cea1bd21059224be802f59e1430b66adef9192dfb2
3
- size 995604017
 
 
 
 
checkpoint-10500/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e3e9c8768a06551d877d3c2a26743fbd513843c0ce986fdadb751108b8d3ff74
3
- size 510403817
 
 
 
 
checkpoint-10500/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8fff69891f2fd5e808642c63d411c8f4998a823f1ca26b032bdcfed6564e01a6
3
- size 623
 
 
 
 
checkpoint-10500/special_tokens_map.json DELETED
@@ -1 +0,0 @@
1
- {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
 
 
checkpoint-10500/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-10500/tokenizer_config.json DELETED
@@ -1 +0,0 @@
1
- {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
 
 
checkpoint-10500/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bbc2f0658a595599fe8144230cef6c7198537be5440bd1e176a88913d8ca5be
3
- size 1327
 
 
 
 
checkpoint-10500/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-14000/config.json DELETED
@@ -1,36 +0,0 @@
1
- {
2
- "_name_or_path": "microsoft/DialoGPT-small",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPT2LMHeadModel"
6
- ],
7
- "attn_pdrop": 0.1,
8
- "bos_token_id": 50256,
9
- "embd_pdrop": 0.1,
10
- "eos_token_id": 50256,
11
- "initializer_range": 0.02,
12
- "layer_norm_epsilon": 1e-05,
13
- "model_type": "gpt2",
14
- "n_ctx": 1024,
15
- "n_embd": 768,
16
- "n_head": 12,
17
- "n_inner": null,
18
- "n_layer": 12,
19
- "n_positions": 1024,
20
- "resid_pdrop": 0.1,
21
- "scale_attn_weights": true,
22
- "summary_activation": null,
23
- "summary_first_dropout": 0.1,
24
- "summary_proj_to_labels": true,
25
- "summary_type": "cls_index",
26
- "summary_use_proj": true,
27
- "task_specific_params": {
28
- "conversational": {
29
- "max_length": 1000
30
- }
31
- },
32
- "torch_dtype": "float32",
33
- "transformers_version": "4.11.3",
34
- "use_cache": true,
35
- "vocab_size": 50257
36
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-14000/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-14000/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8940c4789f03b80fc2cdcd7ef9809542e0955a5b8525c1c6a24645ce8d7e674
3
- size 995604017
 
 
 
 
checkpoint-14000/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a0fcc1d097305b3bd9720acfdc19f45992ef20a7bb4f7847eb91dc0d1a9c698
3
- size 510403817
 
 
 
 
checkpoint-14000/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:17e7d1d03ff158d705870706827a602d9c9d1976549cf84bfd409e785d329382
3
- size 623
 
 
 
 
checkpoint-14000/special_tokens_map.json DELETED
@@ -1 +0,0 @@
1
- {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
 
 
checkpoint-14000/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-14000/tokenizer_config.json DELETED
@@ -1 +0,0 @@
1
- {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
 
 
checkpoint-14000/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bbc2f0658a595599fe8144230cef6c7198537be5440bd1e176a88913d8ca5be
3
- size 1327
 
 
 
 
checkpoint-14000/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-17500/config.json DELETED
@@ -1,36 +0,0 @@
1
- {
2
- "_name_or_path": "microsoft/DialoGPT-small",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPT2LMHeadModel"
6
- ],
7
- "attn_pdrop": 0.1,
8
- "bos_token_id": 50256,
9
- "embd_pdrop": 0.1,
10
- "eos_token_id": 50256,
11
- "initializer_range": 0.02,
12
- "layer_norm_epsilon": 1e-05,
13
- "model_type": "gpt2",
14
- "n_ctx": 1024,
15
- "n_embd": 768,
16
- "n_head": 12,
17
- "n_inner": null,
18
- "n_layer": 12,
19
- "n_positions": 1024,
20
- "resid_pdrop": 0.1,
21
- "scale_attn_weights": true,
22
- "summary_activation": null,
23
- "summary_first_dropout": 0.1,
24
- "summary_proj_to_labels": true,
25
- "summary_type": "cls_index",
26
- "summary_use_proj": true,
27
- "task_specific_params": {
28
- "conversational": {
29
- "max_length": 1000
30
- }
31
- },
32
- "torch_dtype": "float32",
33
- "transformers_version": "4.11.3",
34
- "use_cache": true,
35
- "vocab_size": 50257
36
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-17500/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-17500/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:63d22cabdf55d2810e68ed88728b69fd2426ed7463409304ce274a2620790078
3
- size 995604017
 
 
 
 
checkpoint-17500/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:108ee0a1637c015209b16b74179256e4cb56fba35c2ab99e167e18064faa6046
3
- size 510403817
 
 
 
 
checkpoint-17500/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c41d6a9b0411dfb84dce0e72be293925ed8de48ec90f410d7c30cd96fb72475d
3
- size 623
 
 
 
 
checkpoint-17500/special_tokens_map.json DELETED
@@ -1 +0,0 @@
1
- {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
 
 
checkpoint-17500/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-17500/tokenizer_config.json DELETED
@@ -1 +0,0 @@
1
- {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
 
 
checkpoint-17500/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bbc2f0658a595599fe8144230cef6c7198537be5440bd1e176a88913d8ca5be
3
- size 1327
 
 
 
 
checkpoint-17500/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-21000/config.json DELETED
@@ -1,36 +0,0 @@
1
- {
2
- "_name_or_path": "microsoft/DialoGPT-small",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPT2LMHeadModel"
6
- ],
7
- "attn_pdrop": 0.1,
8
- "bos_token_id": 50256,
9
- "embd_pdrop": 0.1,
10
- "eos_token_id": 50256,
11
- "initializer_range": 0.02,
12
- "layer_norm_epsilon": 1e-05,
13
- "model_type": "gpt2",
14
- "n_ctx": 1024,
15
- "n_embd": 768,
16
- "n_head": 12,
17
- "n_inner": null,
18
- "n_layer": 12,
19
- "n_positions": 1024,
20
- "resid_pdrop": 0.1,
21
- "scale_attn_weights": true,
22
- "summary_activation": null,
23
- "summary_first_dropout": 0.1,
24
- "summary_proj_to_labels": true,
25
- "summary_type": "cls_index",
26
- "summary_use_proj": true,
27
- "task_specific_params": {
28
- "conversational": {
29
- "max_length": 1000
30
- }
31
- },
32
- "torch_dtype": "float32",
33
- "transformers_version": "4.11.3",
34
- "use_cache": true,
35
- "vocab_size": 50257
36
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-21000/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-21000/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f87e8f49ddce78d89159d455203b1198ad252d837864ce31fd1aaad6e69e4aae
3
- size 995604017
 
 
 
 
checkpoint-21000/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:404434065b7d132319c1b43f95a3118ed73646f7656b0e3d03cc54be6c76d1d0
3
- size 510403817
 
 
 
 
checkpoint-21000/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4601444b71769d79ee47b8a0f2d4899682fc0b278b1ab2bbd05b8d13fd2d240
3
- size 623
 
 
 
 
checkpoint-21000/special_tokens_map.json DELETED
@@ -1 +0,0 @@
1
- {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
 
 
checkpoint-21000/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-21000/tokenizer_config.json DELETED
@@ -1 +0,0 @@
1
- {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
 
 
checkpoint-21000/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bbc2f0658a595599fe8144230cef6c7198537be5440bd1e176a88913d8ca5be
3
- size 1327
 
 
 
 
checkpoint-21000/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-24500/config.json DELETED
@@ -1,36 +0,0 @@
1
- {
2
- "_name_or_path": "microsoft/DialoGPT-small",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPT2LMHeadModel"
6
- ],
7
- "attn_pdrop": 0.1,
8
- "bos_token_id": 50256,
9
- "embd_pdrop": 0.1,
10
- "eos_token_id": 50256,
11
- "initializer_range": 0.02,
12
- "layer_norm_epsilon": 1e-05,
13
- "model_type": "gpt2",
14
- "n_ctx": 1024,
15
- "n_embd": 768,
16
- "n_head": 12,
17
- "n_inner": null,
18
- "n_layer": 12,
19
- "n_positions": 1024,
20
- "resid_pdrop": 0.1,
21
- "scale_attn_weights": true,
22
- "summary_activation": null,
23
- "summary_first_dropout": 0.1,
24
- "summary_proj_to_labels": true,
25
- "summary_type": "cls_index",
26
- "summary_use_proj": true,
27
- "task_specific_params": {
28
- "conversational": {
29
- "max_length": 1000
30
- }
31
- },
32
- "torch_dtype": "float32",
33
- "transformers_version": "4.11.3",
34
- "use_cache": true,
35
- "vocab_size": 50257
36
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-24500/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-24500/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:bbd679d33ce972d719e792af880c0733f3ce5970c3abe0271ee5afbc6812d557
3
- size 995604017
 
 
 
 
checkpoint-24500/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b995adb57a8d9f3ad4fcb18f54aff8845112f98f2c8413e417013e82df83155
3
- size 510403817
 
 
 
 
checkpoint-24500/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e636b0f06210f75fb94636cb7f63130cd9edc9acffde7ad24ef596946760391
3
- size 623
 
 
 
 
checkpoint-24500/special_tokens_map.json DELETED
@@ -1 +0,0 @@
1
- {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
 
 
checkpoint-24500/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-24500/tokenizer_config.json DELETED
@@ -1 +0,0 @@
1
- {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
 
 
checkpoint-24500/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bbc2f0658a595599fe8144230cef6c7198537be5440bd1e176a88913d8ca5be
3
- size 1327
 
 
 
 
checkpoint-24500/vocab.json DELETED
The diff for this file is too large to render. See raw diff