kaixkhazaki's picture
Pushing of the new best model checkpoint
2e15e24 verified
|
raw
history blame
4.49 kB
metadata
library_name: transformers
license: mit
base_model: dbmdz/bert-base-turkish-cased
tags:
  - generated_from_trainer
metrics:
  - f1
model-index:
  - name: turkish-medical-question-answering
    results: []

turkish-medical-question-answering

This model is a fine-tuned version of dbmdz/bert-base-turkish-cased on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 1.2864
  • Exact Match: 55.1899
  • F1: 75.1246

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 16
  • eval_batch_size: 64
  • seed: 42
  • optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_steps: 1000
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Exact Match F1
5.9451 0.1166 50 5.9731 0.0 5.7871
5.8483 0.2331 100 5.8315 0.0 5.9124
5.6758 0.3497 150 5.6207 0.2519 6.2966
5.4079 0.4662 200 5.3396 0.2584 6.1159
5.1136 0.5828 250 5.0135 0.2755 7.7324
4.7974 0.6993 300 4.6818 1.2531 9.4001
4.4412 0.8159 350 4.3335 1.5656 10.9210
4.1566 0.9324 400 3.8584 7.0281 22.6904
3.599 1.0490 450 2.8534 23.4628 48.7349
3.0347 1.1655 500 2.4007 28.9362 53.4749
2.7626 1.2821 550 2.2064 33.4711 57.0427
2.5741 1.3986 600 2.0689 37.3297 60.0019
2.3917 1.5152 650 1.9761 38.6968 61.5842
2.3607 1.6317 700 1.9137 41.5344 63.5713
2.2972 1.7483 750 1.8294 44.0210 65.5393
2.1214 1.8648 800 1.7721 44.7028 65.9376
2.1775 1.9814 850 1.7058 46.1538 66.8453
1.9282 2.0979 900 1.6579 46.7784 67.7813
1.9428 2.2145 950 1.6366 46.8912 68.5467
1.8639 2.3310 1000 1.5751 49.4819 70.2320
1.8969 2.4476 1050 1.5519 48.9610 70.3424
1.7348 2.5641 1100 1.5173 50.4505 71.1044
1.7847 2.6807 1150 1.4999 50.5762 71.6186
1.7822 2.7972 1200 1.4566 53.0691 72.0005
1.7989 2.9138 1250 1.4300 51.6005 72.0121
1.7683 3.0303 1300 1.4319 52.0305 72.2366
1.5444 3.1469 1350 1.4277 51.7903 72.0603
1.5121 3.2634 1400 1.3861 53.6122 73.3486
1.6294 3.3800 1450 1.3830 52.6718 73.2456
1.514 3.4965 1500 1.3456 53.7389 73.4757
1.3778 3.6131 1550 1.3644 53.2319 73.5271
1.4502 3.7296 1600 1.3491 53.6030 73.8642
1.5388 3.8462 1650 1.3611 53.0380 73.1390
1.5244 3.9627 1700 1.3143 53.3587 74.0002
1.3127 4.0793 1750 1.3191 54.4767 74.6247
1.3819 4.1958 1800 1.2864 55.1899 75.1246
1.307 4.3124 1850 1.3762 54.1401 74.4158
1.2792 4.4289 1900 1.3156 53.4943 75.0122
1.289 4.5455 1950 1.2809 55.0063 74.6137

Framework versions

  • Transformers 4.48.0.dev0
  • Pytorch 2.4.1+cu121
  • Datasets 3.1.0
  • Tokenizers 0.21.0