sciarrilli commited on
Commit
dfd101c
1 Parent(s): 031b1eb

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "sciarrilli/biobert-base-cased-v1.2-finetuned-ner",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
 
1
  {
2
+ "_name_or_path": "dmis-lab/biobert-base-cased-v1.2",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:201ab54b0ceb6904c3347b39e66e4358e78f007503ffdab3cc3941e260ee9443
3
  size 430995953
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f00e4f4907d9f3a38f22d2d61ea62dc2d9bf75c826263f95a3f78862d853bbe4
3
  size 430995953
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "special_tokens_map_file": null, "name_or_path": "sciarrilli/biobert-base-cased-v1.2-finetuned-ner", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "BertTokenizer"}
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "special_tokens_map_file": null, "name_or_path": "dmis-lab/biobert-base-cased-v1.2", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "BertTokenizer"}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84bee8f1efbc708d8ad4875b78282507eef3874cec2dc042b51d39a231a030a0
3
  size 2927
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f81312874bc78b343d2f5054f49f287b9f343368c9ac810f433ce2e3b2436c5a
3
  size 2927