Theoreticallyhugo commited on
Commit
dd10324
1 Parent(s): a3cfccf

trainer: training complete at 2024-02-06 13:23:50.865104.

Browse files
Files changed (6) hide show
  1. README.md +15 -15
  2. config.json +1 -1
  3. model.safetensors +3 -0
  4. tokenizer.json +4 -4
  5. tokenizer_config.json +42 -0
  6. training_args.bin +2 -2
README.md CHANGED
@@ -17,13 +17,13 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.6725
21
- - Claim: {'precision': 0.4435483870967742, 'recall': 0.3819444444444444, 'f1-score': 0.4104477611940298, 'support': 144.0}
22
- - Majorclaim: {'precision': 0.6166666666666667, 'recall': 0.5138888888888888, 'f1-score': 0.5606060606060607, 'support': 72.0}
23
- - Premise: {'precision': 0.7976470588235294, 'recall': 0.8625954198473282, 'f1-score': 0.8288508557457213, 'support': 393.0}
24
- - Accuracy: 0.7077
25
- - Macro avg: {'precision': 0.6192873708623234, 'recall': 0.5861429177268872, 'f1-score': 0.5999682258486039, 'support': 609.0}
26
- - Weighted avg: {'precision': 0.6925225974705789, 'recall': 0.7077175697865353, 'f1-score': 0.6982044339632925, 'support': 609.0}
27
 
28
  ## Model description
29
 
@@ -52,15 +52,15 @@ The following hyperparameters were used during training:
52
 
53
  ### Training results
54
 
55
- | Training Loss | Epoch | Step | Validation Loss | Claim | Majorclaim | Premise | Accuracy | Macro avg | Weighted avg |
56
- |:-------------:|:-----:|:----:|:---------------:|:------------------------------------------------------------------------------------------------------------------:|:----------------------------------------------------------------------------------------------------------------:|:-----------------------------------------------------------------------------------------------------------------:|:--------:|:-----------------------------------------------------------------------------------------------------------------:|:-----------------------------------------------------------------------------------------------------------------:|
57
- | No log | 1.0 | 267 | 0.6970 | {'precision': 0.4307692307692308, 'recall': 0.19444444444444445, 'f1-score': 0.2679425837320574, 'support': 144.0} | {'precision': 0.5774647887323944, 'recall': 0.5694444444444444, 'f1-score': 0.5734265734265734, 'support': 72.0} | {'precision': 0.758985200845666, 'recall': 0.9134860050890585, 'f1-score': 0.8290993071593533, 'support': 393.0} | 0.7028 | {'precision': 0.589073073449097, 'recall': 0.5591249646593158, 'f1-score': 0.556822821439328, 'support': 609.0} | {'precision': 0.6599169424496689, 'recall': 0.7027914614121511, 'f1-score': 0.6661846848239006, 'support': 609.0} |
58
- | 0.7281 | 2.0 | 534 | 0.6725 | {'precision': 0.4435483870967742, 'recall': 0.3819444444444444, 'f1-score': 0.4104477611940298, 'support': 144.0} | {'precision': 0.6166666666666667, 'recall': 0.5138888888888888, 'f1-score': 0.5606060606060607, 'support': 72.0} | {'precision': 0.7976470588235294, 'recall': 0.8625954198473282, 'f1-score': 0.8288508557457213, 'support': 393.0} | 0.7077 | {'precision': 0.6192873708623234, 'recall': 0.5861429177268872, 'f1-score': 0.5999682258486039, 'support': 609.0} | {'precision': 0.6925225974705789, 'recall': 0.7077175697865353, 'f1-score': 0.6982044339632925, 'support': 609.0} |
59
 
60
 
61
  ### Framework versions
62
 
63
- - Transformers 4.33.0
64
- - Pytorch 2.0.1+cu118
65
- - Datasets 2.14.4
66
- - Tokenizers 0.13.3
 
17
 
18
  This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.6785
21
+ - Claim: {'precision': 0.47368421052631576, 'recall': 0.375, 'f1-score': 0.4186046511627907, 'support': 144.0}
22
+ - Majorclaim: {'precision': 0.6923076923076923, 'recall': 0.5, 'f1-score': 0.5806451612903226, 'support': 72.0}
23
+ - Premise: {'precision': 0.7900677200902935, 'recall': 0.8905852417302799, 'f1-score': 0.8373205741626795, 'support': 393.0}
24
+ - Accuracy: 0.7225
25
+ - Macro avg: {'precision': 0.6520198743081005, 'recall': 0.5885284139100934, 'f1-score': 0.6121901288719309, 'support': 609.0}
26
+ - Weighted avg: {'precision': 0.7036999904062868, 'recall': 0.722495894909688, 'f1-score': 0.707967991832969, 'support': 609.0}
27
 
28
  ## Model description
29
 
 
52
 
53
  ### Training results
54
 
55
+ | Training Loss | Epoch | Step | Validation Loss | Claim | Majorclaim | Premise | Accuracy | Macro avg | Weighted avg |
56
+ |:-------------:|:-----:|:----:|:---------------:|:-------------------------------------------------------------------------------------------------------------------:|:----------------------------------------------------------------------------------------------------------------:|:-----------------------------------------------------------------------------------------------------------------:|:--------:|:-----------------------------------------------------------------------------------------------------------------:|:----------------------------------------------------------------------------------------------------------------:|
57
+ | No log | 1.0 | 267 | 0.7119 | {'precision': 0.5072463768115942, 'recall': 0.24305555555555555, 'f1-score': 0.32863849765258213, 'support': 144.0} | {'precision': 0.5753424657534246, 'recall': 0.5833333333333334, 'f1-score': 0.5793103448275863, 'support': 72.0} | {'precision': 0.7687366167023555, 'recall': 0.9134860050890585, 'f1-score': 0.8348837209302326, 'support': 393.0} | 0.7159 | {'precision': 0.6171084864224582, 'recall': 0.5799582979926492, 'f1-score': 0.580944187803467, 'support': 609.0} | {'precision': 0.6840420790790506, 'recall': 0.715927750410509, 'f1-score': 0.6849648453450565, 'support': 609.0} |
58
+ | 0.7298 | 2.0 | 534 | 0.6785 | {'precision': 0.47368421052631576, 'recall': 0.375, 'f1-score': 0.4186046511627907, 'support': 144.0} | {'precision': 0.6923076923076923, 'recall': 0.5, 'f1-score': 0.5806451612903226, 'support': 72.0} | {'precision': 0.7900677200902935, 'recall': 0.8905852417302799, 'f1-score': 0.8373205741626795, 'support': 393.0} | 0.7225 | {'precision': 0.6520198743081005, 'recall': 0.5885284139100934, 'f1-score': 0.6121901288719309, 'support': 609.0} | {'precision': 0.7036999904062868, 'recall': 0.722495894909688, 'f1-score': 0.707967991832969, 'support': 609.0} |
59
 
60
 
61
  ### Framework versions
62
 
63
+ - Transformers 4.37.1
64
+ - Pytorch 2.1.2+cu121
65
+ - Datasets 2.16.1
66
+ - Tokenizers 0.15.1
config.json CHANGED
@@ -31,7 +31,7 @@
31
  "position_embedding_type": "absolute",
32
  "problem_type": "single_label_classification",
33
  "torch_dtype": "float32",
34
- "transformers_version": "4.33.0",
35
  "type_vocab_size": 1,
36
  "use_cache": true,
37
  "vocab_size": 50265
 
31
  "position_embedding_type": "absolute",
32
  "problem_type": "single_label_classification",
33
  "torch_dtype": "float32",
34
+ "transformers_version": "4.37.1",
35
  "type_vocab_size": 1,
36
  "use_cache": true,
37
  "vocab_size": 50265
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b255249fe64b38dcdc3ef3cc978c7ccddb34bb9caf791234d88d8e2ff294b46c
3
+ size 498615900
tokenizer.json CHANGED
@@ -14,7 +14,7 @@
14
  "single_word": false,
15
  "lstrip": false,
16
  "rstrip": false,
17
- "normalized": false,
18
  "special": true
19
  },
20
  {
@@ -23,7 +23,7 @@
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
- "normalized": false,
27
  "special": true
28
  },
29
  {
@@ -32,7 +32,7 @@
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
- "normalized": false,
36
  "special": true
37
  },
38
  {
@@ -41,7 +41,7 @@
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
44
- "normalized": false,
45
  "special": true
46
  },
47
  {
 
14
  "single_word": false,
15
  "lstrip": false,
16
  "rstrip": false,
17
+ "normalized": true,
18
  "special": true
19
  },
20
  {
 
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
+ "normalized": true,
27
  "special": true
28
  },
29
  {
 
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
+ "normalized": true,
36
  "special": true
37
  },
38
  {
 
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
44
+ "normalized": true,
45
  "special": true
46
  },
47
  {
tokenizer_config.json CHANGED
@@ -1,5 +1,47 @@
1
  {
2
  "add_prefix_space": true,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "bos_token": "<s>",
4
  "clean_up_tokenization_spaces": true,
5
  "cls_token": "<s>",
 
1
  {
2
  "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
  "bos_token": "<s>",
46
  "clean_up_tokenization_spaces": true,
47
  "cls_token": "<s>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d659f2c05c60f157d7a0ec6dc0d517af668613072c7024daf433342a786fe682
3
- size 4027
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a690db7c3d609d9dd4b2b8bbcdde8143de2fe10df2ac70b85dd0c33404a7370
3
+ size 4728