nguyenanh2803 commited on
Commit
ffd62ef
·
verified ·
1 Parent(s): 0f372ef

End of training

Browse files
Files changed (4) hide show
  1. README.md +2 -27
  2. config.json +1 -1
  3. special_tokens_map.json +35 -5
  4. tokenizer_config.json +7 -0
README.md CHANGED
@@ -1,13 +1,8 @@
1
  ---
2
  license: apache-2.0
3
- base_model: google-bert/bert-base-multilingual-cased
4
  tags:
5
  - generated_from_trainer
6
- metrics:
7
- - accuracy
8
- - precision
9
- - recall
10
- - f1
11
  model-index:
12
  - name: absa-train-service-gg-bert-multilingual
13
  results: []
@@ -19,13 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
19
  [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/cunho2803032003/absa-1721788682.8055813/runs/6du7w59h)
20
  # absa-train-service-gg-bert-multilingual
21
 
22
- This model is a fine-tuned version of [google-bert/bert-base-multilingual-cased](https://huggingface.co/google-bert/bert-base-multilingual-cased) on the None dataset.
23
- It achieves the following results on the evaluation set:
24
- - Loss: 1.9467
25
- - Accuracy: 0.4667
26
- - Precision: 0.4925
27
- - Recall: 0.4772
28
- - F1: 0.4218
29
 
30
  ## Model description
31
 
@@ -53,20 +42,6 @@ The following hyperparameters were used during training:
53
  - lr_scheduler_warmup_steps: 500
54
  - num_epochs: 8
55
 
56
- ### Training results
57
-
58
- | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 |
59
- |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|
60
- | 2.2865 | 1.0 | 438 | 2.2450 | 0.2067 | 0.1363 | 0.2097 | 0.1125 |
61
- | 2.2026 | 2.0 | 876 | 2.1433 | 0.3013 | 0.3069 | 0.3118 | 0.2477 |
62
- | 2.1325 | 3.0 | 1314 | 2.0787 | 0.3307 | 0.4191 | 0.3267 | 0.2936 |
63
- | 2.0705 | 4.0 | 1752 | 2.0219 | 0.4107 | 0.4532 | 0.4202 | 0.3533 |
64
- | 2.0617 | 5.0 | 2190 | 1.9910 | 0.4293 | 0.3812 | 0.4402 | 0.3826 |
65
- | 2.0226 | 6.0 | 2628 | 1.9646 | 0.4333 | 0.4792 | 0.4421 | 0.3745 |
66
- | 1.9997 | 7.0 | 3066 | 1.9522 | 0.4733 | 0.5002 | 0.4838 | 0.4294 |
67
- | 1.9882 | 8.0 | 3504 | 1.9467 | 0.4667 | 0.4925 | 0.4772 | 0.4218 |
68
-
69
-
70
  ### Framework versions
71
 
72
  - Transformers 4.43.1
 
1
  ---
2
  license: apache-2.0
3
+ base_model: nguyenanh2803/absa-train-service-gg-bert-multilingual
4
  tags:
5
  - generated_from_trainer
 
 
 
 
 
6
  model-index:
7
  - name: absa-train-service-gg-bert-multilingual
8
  results: []
 
14
  [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/cunho2803032003/absa-1721788682.8055813/runs/6du7w59h)
15
  # absa-train-service-gg-bert-multilingual
16
 
17
+ This model is a fine-tuned version of [nguyenanh2803/absa-train-service-gg-bert-multilingual](https://huggingface.co/nguyenanh2803/absa-train-service-gg-bert-multilingual) on the None dataset.
 
 
 
 
 
 
18
 
19
  ## Model description
20
 
 
42
  - lr_scheduler_warmup_steps: 500
43
  - num_epochs: 8
44
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  ### Framework versions
46
 
47
  - Transformers 4.43.1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "google-bert/bert-base-multilingual-cased",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
 
1
  {
2
+ "_name_or_path": "nguyenanh2803/absa-train-service-gg-bert-multilingual",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
special_tokens_map.json CHANGED
@@ -1,7 +1,37 @@
1
  {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  }
 
1
  {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
  }
tokenizer_config.json CHANGED
@@ -45,11 +45,18 @@
45
  "cls_token": "[CLS]",
46
  "do_lower_case": false,
47
  "mask_token": "[MASK]",
 
48
  "model_max_length": 512,
 
49
  "pad_token": "[PAD]",
 
 
50
  "sep_token": "[SEP]",
 
51
  "strip_accents": null,
52
  "tokenize_chinese_chars": true,
53
  "tokenizer_class": "BertTokenizer",
 
 
54
  "unk_token": "[UNK]"
55
  }
 
45
  "cls_token": "[CLS]",
46
  "do_lower_case": false,
47
  "mask_token": "[MASK]",
48
+ "max_length": 512,
49
  "model_max_length": 512,
50
+ "pad_to_multiple_of": null,
51
  "pad_token": "[PAD]",
52
+ "pad_token_type_id": 0,
53
+ "padding_side": "right",
54
  "sep_token": "[SEP]",
55
+ "stride": 0,
56
  "strip_accents": null,
57
  "tokenize_chinese_chars": true,
58
  "tokenizer_class": "BertTokenizer",
59
+ "truncation_side": "right",
60
+ "truncation_strategy": "longest_first",
61
  "unk_token": "[UNK]"
62
  }