donggyu commited on
Commit
efeee89
·
1 Parent(s): ec95005

mgpt model upload test

Browse files
added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<s>": 250100}
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/opt/ml/final-project-level3-nlp-01/temp",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
@@ -26,6 +26,6 @@
26
  "summary_use_proj": true,
27
  "torch_dtype": "float32",
28
  "transformers_version": "4.11.0",
29
- "use_cache": true,
30
  "vocab_size": 250101
31
  }
 
1
  {
2
+ "_name_or_path": "/opt/ml/final-project-level3-nlp-01/experiment/mnmt",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
 
26
  "summary_use_proj": true,
27
  "torch_dtype": "float32",
28
  "transformers_version": "4.11.0",
29
+ "use_cache": false,
30
  "vocab_size": 250101
31
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e8057319ac08840806c0cb007271ff823c26fa396f191c08965c889927bc804
3
  size 2263145763
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9560db7fb458cea01bf08c37d2b25f188d667c74b925062acfb4dc1939cb6402
3
  size 2263145763
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "additional_special_tokens": ["<s>"]}
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef78f86560d809067d12bac6c09f19a462cb3af3f54d2b8acbba26e1433125d6
3
+ size 4309802
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "extra_ids": 0, "additional_special_tokens": null, "sp_model_kwargs": {}, "special_tokens_map_file": "/opt/ml/.cache/huggingface/transformers/8283e5c59ea5a25d7e56f7afa647fbaaeb1f8f98ccd517f4ffeae883937f9357.f76030f3ec1b96a8199b2593390c610e76ca8028ef3d24680000619ffb646276", "tokenizer_file": null, "name_or_path": "THUMT/mGPT", "tokenizer_class": "T5Tokenizer"}