remove previous checkpoints
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- checkpoint-10500/config.json +0 -36
- checkpoint-10500/merges.txt +0 -0
- checkpoint-10500/optimizer.pt +0 -3
- checkpoint-10500/pytorch_model.bin +0 -3
- checkpoint-10500/scheduler.pt +0 -3
- checkpoint-10500/special_tokens_map.json +0 -1
- checkpoint-10500/tokenizer.json +0 -0
- checkpoint-10500/tokenizer_config.json +0 -1
- checkpoint-10500/training_args.bin +0 -3
- checkpoint-10500/vocab.json +0 -0
- checkpoint-14000/config.json +0 -36
- checkpoint-14000/merges.txt +0 -0
- checkpoint-14000/optimizer.pt +0 -3
- checkpoint-14000/pytorch_model.bin +0 -3
- checkpoint-14000/scheduler.pt +0 -3
- checkpoint-14000/special_tokens_map.json +0 -1
- checkpoint-14000/tokenizer.json +0 -0
- checkpoint-14000/tokenizer_config.json +0 -1
- checkpoint-14000/training_args.bin +0 -3
- checkpoint-14000/vocab.json +0 -0
- checkpoint-17500/config.json +0 -36
- checkpoint-17500/merges.txt +0 -0
- checkpoint-17500/optimizer.pt +0 -3
- checkpoint-17500/pytorch_model.bin +0 -3
- checkpoint-17500/scheduler.pt +0 -3
- checkpoint-17500/special_tokens_map.json +0 -1
- checkpoint-17500/tokenizer.json +0 -0
- checkpoint-17500/tokenizer_config.json +0 -1
- checkpoint-17500/training_args.bin +0 -3
- checkpoint-17500/vocab.json +0 -0
- checkpoint-21000/config.json +0 -36
- checkpoint-21000/merges.txt +0 -0
- checkpoint-21000/optimizer.pt +0 -3
- checkpoint-21000/pytorch_model.bin +0 -3
- checkpoint-21000/scheduler.pt +0 -3
- checkpoint-21000/special_tokens_map.json +0 -1
- checkpoint-21000/tokenizer.json +0 -0
- checkpoint-21000/tokenizer_config.json +0 -1
- checkpoint-21000/training_args.bin +0 -3
- checkpoint-21000/vocab.json +0 -0
- checkpoint-24500/config.json +0 -36
- checkpoint-24500/merges.txt +0 -0
- checkpoint-24500/optimizer.pt +0 -3
- checkpoint-24500/pytorch_model.bin +0 -3
- checkpoint-24500/scheduler.pt +0 -3
- checkpoint-24500/special_tokens_map.json +0 -1
- checkpoint-24500/tokenizer.json +0 -0
- checkpoint-24500/tokenizer_config.json +0 -1
- checkpoint-24500/training_args.bin +0 -3
- checkpoint-24500/vocab.json +0 -0
checkpoint-10500/config.json
DELETED
@@ -1,36 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_name_or_path": "microsoft/DialoGPT-small",
|
3 |
-
"activation_function": "gelu_new",
|
4 |
-
"architectures": [
|
5 |
-
"GPT2LMHeadModel"
|
6 |
-
],
|
7 |
-
"attn_pdrop": 0.1,
|
8 |
-
"bos_token_id": 50256,
|
9 |
-
"embd_pdrop": 0.1,
|
10 |
-
"eos_token_id": 50256,
|
11 |
-
"initializer_range": 0.02,
|
12 |
-
"layer_norm_epsilon": 1e-05,
|
13 |
-
"model_type": "gpt2",
|
14 |
-
"n_ctx": 1024,
|
15 |
-
"n_embd": 768,
|
16 |
-
"n_head": 12,
|
17 |
-
"n_inner": null,
|
18 |
-
"n_layer": 12,
|
19 |
-
"n_positions": 1024,
|
20 |
-
"resid_pdrop": 0.1,
|
21 |
-
"scale_attn_weights": true,
|
22 |
-
"summary_activation": null,
|
23 |
-
"summary_first_dropout": 0.1,
|
24 |
-
"summary_proj_to_labels": true,
|
25 |
-
"summary_type": "cls_index",
|
26 |
-
"summary_use_proj": true,
|
27 |
-
"task_specific_params": {
|
28 |
-
"conversational": {
|
29 |
-
"max_length": 1000
|
30 |
-
}
|
31 |
-
},
|
32 |
-
"torch_dtype": "float32",
|
33 |
-
"transformers_version": "4.11.3",
|
34 |
-
"use_cache": true,
|
35 |
-
"vocab_size": 50257
|
36 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
checkpoint-10500/merges.txt
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-10500/optimizer.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:984f039e1beb41ed40aa47cea1bd21059224be802f59e1430b66adef9192dfb2
|
3 |
-
size 995604017
|
|
|
|
|
|
|
|
checkpoint-10500/pytorch_model.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:e3e9c8768a06551d877d3c2a26743fbd513843c0ce986fdadb751108b8d3ff74
|
3 |
-
size 510403817
|
|
|
|
|
|
|
|
checkpoint-10500/scheduler.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:8fff69891f2fd5e808642c63d411c8f4998a823f1ca26b032bdcfed6564e01a6
|
3 |
-
size 623
|
|
|
|
|
|
|
|
checkpoint-10500/special_tokens_map.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
|
|
|
|
checkpoint-10500/tokenizer.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-10500/tokenizer_config.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
|
|
|
|
checkpoint-10500/training_args.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:3bbc2f0658a595599fe8144230cef6c7198537be5440bd1e176a88913d8ca5be
|
3 |
-
size 1327
|
|
|
|
|
|
|
|
checkpoint-10500/vocab.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-14000/config.json
DELETED
@@ -1,36 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_name_or_path": "microsoft/DialoGPT-small",
|
3 |
-
"activation_function": "gelu_new",
|
4 |
-
"architectures": [
|
5 |
-
"GPT2LMHeadModel"
|
6 |
-
],
|
7 |
-
"attn_pdrop": 0.1,
|
8 |
-
"bos_token_id": 50256,
|
9 |
-
"embd_pdrop": 0.1,
|
10 |
-
"eos_token_id": 50256,
|
11 |
-
"initializer_range": 0.02,
|
12 |
-
"layer_norm_epsilon": 1e-05,
|
13 |
-
"model_type": "gpt2",
|
14 |
-
"n_ctx": 1024,
|
15 |
-
"n_embd": 768,
|
16 |
-
"n_head": 12,
|
17 |
-
"n_inner": null,
|
18 |
-
"n_layer": 12,
|
19 |
-
"n_positions": 1024,
|
20 |
-
"resid_pdrop": 0.1,
|
21 |
-
"scale_attn_weights": true,
|
22 |
-
"summary_activation": null,
|
23 |
-
"summary_first_dropout": 0.1,
|
24 |
-
"summary_proj_to_labels": true,
|
25 |
-
"summary_type": "cls_index",
|
26 |
-
"summary_use_proj": true,
|
27 |
-
"task_specific_params": {
|
28 |
-
"conversational": {
|
29 |
-
"max_length": 1000
|
30 |
-
}
|
31 |
-
},
|
32 |
-
"torch_dtype": "float32",
|
33 |
-
"transformers_version": "4.11.3",
|
34 |
-
"use_cache": true,
|
35 |
-
"vocab_size": 50257
|
36 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
checkpoint-14000/merges.txt
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-14000/optimizer.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:b8940c4789f03b80fc2cdcd7ef9809542e0955a5b8525c1c6a24645ce8d7e674
|
3 |
-
size 995604017
|
|
|
|
|
|
|
|
checkpoint-14000/pytorch_model.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:3a0fcc1d097305b3bd9720acfdc19f45992ef20a7bb4f7847eb91dc0d1a9c698
|
3 |
-
size 510403817
|
|
|
|
|
|
|
|
checkpoint-14000/scheduler.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:17e7d1d03ff158d705870706827a602d9c9d1976549cf84bfd409e785d329382
|
3 |
-
size 623
|
|
|
|
|
|
|
|
checkpoint-14000/special_tokens_map.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
|
|
|
|
checkpoint-14000/tokenizer.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-14000/tokenizer_config.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
|
|
|
|
checkpoint-14000/training_args.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:3bbc2f0658a595599fe8144230cef6c7198537be5440bd1e176a88913d8ca5be
|
3 |
-
size 1327
|
|
|
|
|
|
|
|
checkpoint-14000/vocab.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-17500/config.json
DELETED
@@ -1,36 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_name_or_path": "microsoft/DialoGPT-small",
|
3 |
-
"activation_function": "gelu_new",
|
4 |
-
"architectures": [
|
5 |
-
"GPT2LMHeadModel"
|
6 |
-
],
|
7 |
-
"attn_pdrop": 0.1,
|
8 |
-
"bos_token_id": 50256,
|
9 |
-
"embd_pdrop": 0.1,
|
10 |
-
"eos_token_id": 50256,
|
11 |
-
"initializer_range": 0.02,
|
12 |
-
"layer_norm_epsilon": 1e-05,
|
13 |
-
"model_type": "gpt2",
|
14 |
-
"n_ctx": 1024,
|
15 |
-
"n_embd": 768,
|
16 |
-
"n_head": 12,
|
17 |
-
"n_inner": null,
|
18 |
-
"n_layer": 12,
|
19 |
-
"n_positions": 1024,
|
20 |
-
"resid_pdrop": 0.1,
|
21 |
-
"scale_attn_weights": true,
|
22 |
-
"summary_activation": null,
|
23 |
-
"summary_first_dropout": 0.1,
|
24 |
-
"summary_proj_to_labels": true,
|
25 |
-
"summary_type": "cls_index",
|
26 |
-
"summary_use_proj": true,
|
27 |
-
"task_specific_params": {
|
28 |
-
"conversational": {
|
29 |
-
"max_length": 1000
|
30 |
-
}
|
31 |
-
},
|
32 |
-
"torch_dtype": "float32",
|
33 |
-
"transformers_version": "4.11.3",
|
34 |
-
"use_cache": true,
|
35 |
-
"vocab_size": 50257
|
36 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
checkpoint-17500/merges.txt
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-17500/optimizer.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:63d22cabdf55d2810e68ed88728b69fd2426ed7463409304ce274a2620790078
|
3 |
-
size 995604017
|
|
|
|
|
|
|
|
checkpoint-17500/pytorch_model.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:108ee0a1637c015209b16b74179256e4cb56fba35c2ab99e167e18064faa6046
|
3 |
-
size 510403817
|
|
|
|
|
|
|
|
checkpoint-17500/scheduler.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:c41d6a9b0411dfb84dce0e72be293925ed8de48ec90f410d7c30cd96fb72475d
|
3 |
-
size 623
|
|
|
|
|
|
|
|
checkpoint-17500/special_tokens_map.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
|
|
|
|
checkpoint-17500/tokenizer.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-17500/tokenizer_config.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
|
|
|
|
checkpoint-17500/training_args.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:3bbc2f0658a595599fe8144230cef6c7198537be5440bd1e176a88913d8ca5be
|
3 |
-
size 1327
|
|
|
|
|
|
|
|
checkpoint-17500/vocab.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-21000/config.json
DELETED
@@ -1,36 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_name_or_path": "microsoft/DialoGPT-small",
|
3 |
-
"activation_function": "gelu_new",
|
4 |
-
"architectures": [
|
5 |
-
"GPT2LMHeadModel"
|
6 |
-
],
|
7 |
-
"attn_pdrop": 0.1,
|
8 |
-
"bos_token_id": 50256,
|
9 |
-
"embd_pdrop": 0.1,
|
10 |
-
"eos_token_id": 50256,
|
11 |
-
"initializer_range": 0.02,
|
12 |
-
"layer_norm_epsilon": 1e-05,
|
13 |
-
"model_type": "gpt2",
|
14 |
-
"n_ctx": 1024,
|
15 |
-
"n_embd": 768,
|
16 |
-
"n_head": 12,
|
17 |
-
"n_inner": null,
|
18 |
-
"n_layer": 12,
|
19 |
-
"n_positions": 1024,
|
20 |
-
"resid_pdrop": 0.1,
|
21 |
-
"scale_attn_weights": true,
|
22 |
-
"summary_activation": null,
|
23 |
-
"summary_first_dropout": 0.1,
|
24 |
-
"summary_proj_to_labels": true,
|
25 |
-
"summary_type": "cls_index",
|
26 |
-
"summary_use_proj": true,
|
27 |
-
"task_specific_params": {
|
28 |
-
"conversational": {
|
29 |
-
"max_length": 1000
|
30 |
-
}
|
31 |
-
},
|
32 |
-
"torch_dtype": "float32",
|
33 |
-
"transformers_version": "4.11.3",
|
34 |
-
"use_cache": true,
|
35 |
-
"vocab_size": 50257
|
36 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
checkpoint-21000/merges.txt
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-21000/optimizer.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:f87e8f49ddce78d89159d455203b1198ad252d837864ce31fd1aaad6e69e4aae
|
3 |
-
size 995604017
|
|
|
|
|
|
|
|
checkpoint-21000/pytorch_model.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:404434065b7d132319c1b43f95a3118ed73646f7656b0e3d03cc54be6c76d1d0
|
3 |
-
size 510403817
|
|
|
|
|
|
|
|
checkpoint-21000/scheduler.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:a4601444b71769d79ee47b8a0f2d4899682fc0b278b1ab2bbd05b8d13fd2d240
|
3 |
-
size 623
|
|
|
|
|
|
|
|
checkpoint-21000/special_tokens_map.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
|
|
|
|
checkpoint-21000/tokenizer.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-21000/tokenizer_config.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
|
|
|
|
checkpoint-21000/training_args.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:3bbc2f0658a595599fe8144230cef6c7198537be5440bd1e176a88913d8ca5be
|
3 |
-
size 1327
|
|
|
|
|
|
|
|
checkpoint-21000/vocab.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-24500/config.json
DELETED
@@ -1,36 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_name_or_path": "microsoft/DialoGPT-small",
|
3 |
-
"activation_function": "gelu_new",
|
4 |
-
"architectures": [
|
5 |
-
"GPT2LMHeadModel"
|
6 |
-
],
|
7 |
-
"attn_pdrop": 0.1,
|
8 |
-
"bos_token_id": 50256,
|
9 |
-
"embd_pdrop": 0.1,
|
10 |
-
"eos_token_id": 50256,
|
11 |
-
"initializer_range": 0.02,
|
12 |
-
"layer_norm_epsilon": 1e-05,
|
13 |
-
"model_type": "gpt2",
|
14 |
-
"n_ctx": 1024,
|
15 |
-
"n_embd": 768,
|
16 |
-
"n_head": 12,
|
17 |
-
"n_inner": null,
|
18 |
-
"n_layer": 12,
|
19 |
-
"n_positions": 1024,
|
20 |
-
"resid_pdrop": 0.1,
|
21 |
-
"scale_attn_weights": true,
|
22 |
-
"summary_activation": null,
|
23 |
-
"summary_first_dropout": 0.1,
|
24 |
-
"summary_proj_to_labels": true,
|
25 |
-
"summary_type": "cls_index",
|
26 |
-
"summary_use_proj": true,
|
27 |
-
"task_specific_params": {
|
28 |
-
"conversational": {
|
29 |
-
"max_length": 1000
|
30 |
-
}
|
31 |
-
},
|
32 |
-
"torch_dtype": "float32",
|
33 |
-
"transformers_version": "4.11.3",
|
34 |
-
"use_cache": true,
|
35 |
-
"vocab_size": 50257
|
36 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
checkpoint-24500/merges.txt
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-24500/optimizer.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:bbd679d33ce972d719e792af880c0733f3ce5970c3abe0271ee5afbc6812d557
|
3 |
-
size 995604017
|
|
|
|
|
|
|
|
checkpoint-24500/pytorch_model.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:6b995adb57a8d9f3ad4fcb18f54aff8845112f98f2c8413e417013e82df83155
|
3 |
-
size 510403817
|
|
|
|
|
|
|
|
checkpoint-24500/scheduler.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:6e636b0f06210f75fb94636cb7f63130cd9edc9acffde7ad24ef596946760391
|
3 |
-
size 623
|
|
|
|
|
|
|
|
checkpoint-24500/special_tokens_map.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
|
|
|
|
checkpoint-24500/tokenizer.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-24500/tokenizer_config.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
|
|
|
|
checkpoint-24500/training_args.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:3bbc2f0658a595599fe8144230cef6c7198537be5440bd1e176a88913d8ca5be
|
3 |
-
size 1327
|
|
|
|
|
|
|
|
checkpoint-24500/vocab.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|