jayanta commited on
Commit
75ff628
·
1 Parent(s): 8e31e4b

Training in progress, epoch 1

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -1,13 +1,12 @@
1
  {
2
- "_name_or_path": "cardiffnlp/twitter-roberta-base-sentiment-latest",
3
  "architectures": [
4
- "RobertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
8
  "classifier_dropout": null,
9
  "eos_token_id": 2,
10
- "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
  "hidden_size": 768,
@@ -27,9 +26,10 @@
27
  },
28
  "layer_norm_eps": 1e-05,
29
  "max_position_embeddings": 514,
30
- "model_type": "roberta",
31
  "num_attention_heads": 12,
32
  "num_hidden_layers": 12,
 
33
  "pad_token_id": 1,
34
  "position_embedding_type": "absolute",
35
  "problem_type": "single_label_classification",
@@ -37,5 +37,5 @@
37
  "transformers_version": "4.30.2",
38
  "type_vocab_size": 1,
39
  "use_cache": true,
40
- "vocab_size": 50265
41
  }
 
1
  {
2
+ "_name_or_path": "FacebookAI/xlm-roberta-base",
3
  "architectures": [
4
+ "XLMRobertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
8
  "classifier_dropout": null,
9
  "eos_token_id": 2,
 
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
  "hidden_size": 768,
 
26
  },
27
  "layer_norm_eps": 1e-05,
28
  "max_position_embeddings": 514,
29
+ "model_type": "xlm-roberta",
30
  "num_attention_heads": 12,
31
  "num_hidden_layers": 12,
32
+ "output_past": true,
33
  "pad_token_id": 1,
34
  "position_embedding_type": "absolute",
35
  "problem_type": "single_label_classification",
 
37
  "transformers_version": "4.30.2",
38
  "type_vocab_size": 1,
39
  "use_cache": true,
40
+ "vocab_size": 250002
41
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dbb883bfa47a6a6021cda7c7777005645cea5143ccf6760d89a9998edcc6d69b
3
- size 498668213
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efa33da3f55a77c4cd38708a8f7173eb0ea9ea994181059e3f9c2d1695e83576
3
+ size 1112260277
runs/Mar13_02-18-12_teesta/events.out.tfevents.1710276502.teesta.20405.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6fb6bdcebf4c66630d82fe79dc5c00781c458d2bc7362ea2ce56b7e705431ffc
3
- size 7860
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4a68158c506641d8dc0ae2e1fcbe27284ae80a215830c31c2d0902f58529be3
3
+ size 8214
runs/Mar13_03-03-04_teesta/events.out.tfevents.1710279214.teesta.29079.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99b2a4ca292e5ecca432246730de177dc3df07da39f102653670933409be8ce4
3
+ size 4806
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
+ size 5069051
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,31 +1,8 @@
1
  {
2
- "add_prefix_space": false,
3
- "bos_token": {
4
- "__type": "AddedToken",
5
- "content": "<s>",
6
- "lstrip": false,
7
- "normalized": true,
8
- "rstrip": false,
9
- "single_word": false
10
- },
11
  "clean_up_tokenization_spaces": true,
12
- "cls_token": {
13
- "__type": "AddedToken",
14
- "content": "<s>",
15
- "lstrip": false,
16
- "normalized": true,
17
- "rstrip": false,
18
- "single_word": false
19
- },
20
- "eos_token": {
21
- "__type": "AddedToken",
22
- "content": "</s>",
23
- "lstrip": false,
24
- "normalized": true,
25
- "rstrip": false,
26
- "single_word": false
27
- },
28
- "errors": "replace",
29
  "mask_token": {
30
  "__type": "AddedToken",
31
  "content": "<mask>",
@@ -34,31 +11,9 @@
34
  "rstrip": false,
35
  "single_word": false
36
  },
37
- "model_max_length": 1000000000000000019884624838656,
38
- "pad_token": {
39
- "__type": "AddedToken",
40
- "content": "<pad>",
41
- "lstrip": false,
42
- "normalized": true,
43
- "rstrip": false,
44
- "single_word": false
45
- },
46
- "sep_token": {
47
- "__type": "AddedToken",
48
- "content": "</s>",
49
- "lstrip": false,
50
- "normalized": true,
51
- "rstrip": false,
52
- "single_word": false
53
- },
54
- "tokenizer_class": "RobertaTokenizer",
55
- "trim_offsets": true,
56
- "unk_token": {
57
- "__type": "AddedToken",
58
- "content": "<unk>",
59
- "lstrip": false,
60
- "normalized": true,
61
- "rstrip": false,
62
- "single_word": false
63
- }
64
  }
 
1
  {
2
+ "bos_token": "<s>",
 
 
 
 
 
 
 
 
3
  "clean_up_tokenization_spaces": true,
4
+ "cls_token": "<s>",
5
+ "eos_token": "</s>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  "mask_token": {
7
  "__type": "AddedToken",
8
  "content": "<mask>",
 
11
  "rstrip": false,
12
  "single_word": false
13
  },
14
+ "model_max_length": 512,
15
+ "pad_token": "<pad>",
16
+ "sep_token": "</s>",
17
+ "tokenizer_class": "XLMRobertaTokenizer",
18
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c5edcaf40291d7551d3f6eafeeafd301d5847d4ba3264fe1005cd32ce866d2b5
3
  size 4027
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f181a52148b5685d649a6d289744dce3d8c84915ec6f266789faec8f1bb36e13
3
  size 4027