5_BiasDetctor_HTW_ANG.KI
This model is a fine-tuned version of roberta-large on the None dataset. It achieves the following results on the evaluation set:
- Loss: 0.6177
- Accuracy: 0.9114
- F1: 0.9116
- Precision: 0.9122
- Recall: 0.9114
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 16
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 16
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 300
- num_epochs: 7
- label_smoothing_factor: 0.1
Training results
Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
---|---|---|---|---|---|---|---|
0.962 | 0.2 | 250 | 0.8191 | 0.7846 | 0.7854 | 0.8076 | 0.7846 |
0.7921 | 0.4 | 500 | 0.7972 | 0.7982 | 0.8029 | 0.8436 | 0.7982 |
0.7592 | 0.6 | 750 | 0.6858 | 0.8614 | 0.8625 | 0.8748 | 0.8614 |
0.701 | 0.8 | 1000 | 0.6871 | 0.8666 | 0.8673 | 0.8731 | 0.8666 |
0.6595 | 1.0 | 1250 | 0.6236 | 0.8888 | 0.8895 | 0.8921 | 0.8888 |
0.5826 | 1.2 | 1500 | 0.6629 | 0.8898 | 0.8909 | 0.8952 | 0.8898 |
0.5983 | 1.4 | 1750 | 0.6282 | 0.8934 | 0.8933 | 0.8942 | 0.8934 |
0.594 | 1.6 | 2000 | 0.6249 | 0.894 | 0.8944 | 0.8956 | 0.894 |
0.6477 | 1.8 | 2250 | 0.6045 | 0.9014 | 0.9020 | 0.9034 | 0.9014 |
0.5968 | 2.0 | 2500 | 0.6189 | 0.8962 | 0.8964 | 0.8971 | 0.8962 |
0.5291 | 2.2 | 2750 | 0.6177 | 0.9114 | 0.9116 | 0.9122 | 0.9114 |
0.5225 | 2.4 | 3000 | 0.6260 | 0.9078 | 0.9080 | 0.9103 | 0.9078 |
0.5286 | 2.6 | 3250 | 0.6468 | 0.8978 | 0.8976 | 0.9004 | 0.8978 |
Framework versions
- Transformers 4.51.3
- Pytorch 2.1.2
- Datasets 3.5.1
- Tokenizers 0.21.1
training metrics:
TrainOutput(global_step=3250, training_loss=0.6634712289663461, metrics={'train_runtime': 728.1708, 'train_samples_per_second': 192.263, 'train_steps_per_second': 12.016, 'total_flos': 1.2115230526464e+16, 'train_loss': 0.6634712289663461, 'epoch': 2.6})
- Downloads last month
- 4
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for Ghazi-nak/5_BiasDetctor_HTW_ANG.KI
Base model
FacebookAI/roberta-large