gokulsrinivasagan commited on
Commit
7747589
·
verified ·
1 Parent(s): 953af8a

Model save

Browse files
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: google/bert_uncased_L-2_H-128_A-2
5
+ tags:
6
+ - generated_from_trainer
7
+ metrics:
8
+ - accuracy
9
+ - f1
10
+ model-index:
11
+ - name: bert_uncased_L-2_H-128_A-2_mrpc
12
+ results: []
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+ # bert_uncased_L-2_H-128_A-2_mrpc
19
+
20
+ This model is a fine-tuned version of [google/bert_uncased_L-2_H-128_A-2](https://huggingface.co/google/bert_uncased_L-2_H-128_A-2) on an unknown dataset.
21
+ It achieves the following results on the evaluation set:
22
+ - Loss: 0.5587
23
+ - Accuracy: 0.7377
24
+ - F1: 0.8283
25
+ - Combined Score: 0.7830
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
+
37
+ More information needed
38
+
39
+ ## Training procedure
40
+
41
+ ### Training hyperparameters
42
+
43
+ The following hyperparameters were used during training:
44
+ - learning_rate: 5e-05
45
+ - train_batch_size: 256
46
+ - eval_batch_size: 256
47
+ - seed: 10
48
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
49
+ - lr_scheduler_type: linear
50
+ - num_epochs: 50
51
+
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Combined Score |
55
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:--------------:|
56
+ | 0.6493 | 1.0 | 15 | 0.6227 | 0.6838 | 0.8122 | 0.7480 |
57
+ | 0.6257 | 2.0 | 30 | 0.6134 | 0.6838 | 0.8122 | 0.7480 |
58
+ | 0.6126 | 3.0 | 45 | 0.6052 | 0.6838 | 0.8122 | 0.7480 |
59
+ | 0.6036 | 4.0 | 60 | 0.5954 | 0.6961 | 0.8176 | 0.7569 |
60
+ | 0.5897 | 5.0 | 75 | 0.5879 | 0.6985 | 0.8167 | 0.7576 |
61
+ | 0.5781 | 6.0 | 90 | 0.5741 | 0.7034 | 0.8158 | 0.7596 |
62
+ | 0.5635 | 7.0 | 105 | 0.5711 | 0.7108 | 0.8201 | 0.7655 |
63
+ | 0.5429 | 8.0 | 120 | 0.5674 | 0.7132 | 0.8208 | 0.7670 |
64
+ | 0.5228 | 9.0 | 135 | 0.5685 | 0.7206 | 0.8252 | 0.7729 |
65
+ | 0.5057 | 10.0 | 150 | 0.5497 | 0.7304 | 0.8281 | 0.7793 |
66
+ | 0.4856 | 11.0 | 165 | 0.5438 | 0.7377 | 0.8293 | 0.7835 |
67
+ | 0.4657 | 12.0 | 180 | 0.5352 | 0.7328 | 0.8233 | 0.7781 |
68
+ | 0.4447 | 13.0 | 195 | 0.5435 | 0.7402 | 0.8323 | 0.7862 |
69
+ | 0.4175 | 14.0 | 210 | 0.5562 | 0.7402 | 0.8328 | 0.7865 |
70
+ | 0.4039 | 15.0 | 225 | 0.5759 | 0.7426 | 0.8357 | 0.7892 |
71
+ | 0.3964 | 16.0 | 240 | 0.5610 | 0.7377 | 0.8299 | 0.7838 |
72
+ | 0.3735 | 17.0 | 255 | 0.5587 | 0.7377 | 0.8283 | 0.7830 |
73
+
74
+
75
+ ### Framework versions
76
+
77
+ - Transformers 4.46.3
78
+ - Pytorch 2.2.1+cu118
79
+ - Datasets 2.17.0
80
+ - Tokenizers 0.20.3
logs/events.out.tfevents.1733343265.ki-g0008.1759256.20 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8c5f04adb531b3f79e7c227d609c100ecb575d42f39af216666f63fa18ecdeb
3
- size 14591
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95f43469a628862be49eb5d9d3904d54ad988d5d72efe226951e357ffd7cb2b5
3
+ size 16221
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:266051f30829aff586701af243276af320c750fffebe7d2b8826fec01a889f57
3
  size 17549312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c46c7548e4dca319ea54b381291335674153433a878f2375bbf0d1789d847ce
3
  size 17549312