SST-2-FULL_FT-seed62

This model is a fine-tuned version of roberta-base on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.2731
  • Accuracy: 0.9381

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 3e-05
  • train_batch_size: 32
  • eval_batch_size: 32
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Accuracy
0.4131 0.0950 200 0.2722 0.9117
0.318 0.1900 400 0.2142 0.9255
0.2652 0.2850 600 0.1965 0.9255
0.2432 0.3800 800 0.2342 0.9220
0.2278 0.4751 1000 0.3632 0.8716
0.2182 0.5701 1200 0.2370 0.9220
0.2142 0.6651 1400 0.2226 0.9358
0.2131 0.7601 1600 0.2025 0.9232
0.2084 0.8551 1800 0.2086 0.9335
0.1881 0.9501 2000 0.2530 0.9232
0.1837 1.0451 2200 0.2140 0.9300
0.141 1.1401 2400 0.2244 0.9289
0.1584 1.2352 2600 0.2290 0.9232
0.1457 1.3302 2800 0.2745 0.9300
0.155 1.4252 3000 0.2839 0.9278
0.1478 1.5202 3200 0.2952 0.9174
0.148 1.6152 3400 0.2036 0.9369
0.1406 1.7102 3600 0.2416 0.9289
0.1431 1.8052 3800 0.2523 0.9289
0.1547 1.9002 4000 0.2371 0.9335
0.1477 1.9952 4200 0.2435 0.9358
0.1147 2.0903 4400 0.2541 0.9243
0.1087 2.1853 4600 0.2867 0.9243
0.1167 2.2803 4800 0.2779 0.9289
0.1097 2.3753 5000 0.2468 0.9358
0.1108 2.4703 5200 0.2918 0.9300
0.1058 2.5653 5400 0.2734 0.9266
0.1228 2.6603 5600 0.2466 0.9323
0.1062 2.7553 5800 0.2699 0.9335
0.1057 2.8504 6000 0.2520 0.9335
0.1008 2.9454 6200 0.3001 0.9312
0.0955 3.0404 6400 0.3554 0.9278
0.0802 3.1354 6600 0.2838 0.9278
0.082 3.2304 6800 0.3157 0.9289
0.0939 3.3254 7000 0.2936 0.9197
0.0795 3.4204 7200 0.2614 0.9346
0.0825 3.5154 7400 0.2551 0.9369
0.0769 3.6105 7600 0.2780 0.9346
0.0879 3.7055 7800 0.2410 0.9358
0.0823 3.8005 8000 0.2587 0.9346
0.0812 3.8955 8200 0.2731 0.9381
0.0871 3.9905 8400 0.2679 0.9381
0.0651 4.0855 8600 0.3009 0.9369
0.0533 4.1805 8800 0.3355 0.9358
0.0663 4.2755 9000 0.3103 0.9346
0.0623 4.3705 9200 0.2901 0.9358
0.0603 4.4656 9400 0.3071 0.9369
0.0564 4.5606 9600 0.3191 0.9381
0.0663 4.6556 9800 0.2836 0.9369
0.0639 4.7506 10000 0.2740 0.9369
0.0588 4.8456 10200 0.2894 0.9381
0.0643 4.9406 10400 0.2818 0.9346

Framework versions

  • Transformers 4.54.1
  • Pytorch 2.5.1+cu121
  • Datasets 4.0.0
  • Tokenizers 0.21.4
Downloads last month
52
Safetensors
Model size
125M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for ekiprop/SST-2-FULL_FT-seed62

Finetuned
(1840)
this model