timesformer-bert-video-captioning

This model is a fine-tuned version of on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.2821
  • Rouge1: 30.0468
  • Rouge2: 8.4998
  • Rougel: 29.0632
  • Rougelsum: 29.0231
  • Bleu: 4.8298
  • Gen Len: 9.5332

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 2

Training results

Training Loss Epoch Step Bleu Gen Len Validation Loss Rouge1 Rouge2 Rougel Rougelsum
2.4961 0.12 200 1.5879 9.5332 1.6548 25.4717 5.11 24.6679 24.6696
1.6561 0.25 400 2.3515 9.5332 1.5339 26.1748 5.9106 25.413 25.3958
1.5772 0.37 600 2.266 9.5332 1.4510 28.6891 6.0431 27.7387 27.8043
1.492 0.49 800 3.6517 9.5332 1.3760 29.0257 7.8515 28.3142 28.3036
1.4736 0.61 1000 3.4866 9.5332 1.3425 27.9774 6.2175 26.7783 26.7207
1.3856 0.74 1200 3.1649 9.5332 1.3118 27.3532 6.5569 26.4964 26.5087
1.3972 0.86 1400 3.5337 9.5332 1.2868 28.233 7.6471 27.3651 27.3354
1.374 0.98 1600 3.5737 9.5332 1.2571 28.8216 7.542 27.9166 27.9353
1.2207 1.1 1800 3.7983 9.5332 1.3362 29.9574 8.1088 28.8866 28.855
1.1861 1.23 2000 3.6521 9.5332 1.3295 30.072 7.7799 28.8417 28.864
1.1173 1.35 2200 3.9784 9.5332 1.3335 29.736 7.9661 28.6877 28.6974
1.1255 1.47 2400 4.3021 9.5332 1.3097 29.8176 8.4656 28.958 28.9571
1.0909 1.6 2600 1.3095 30.0233 8.4896 29.2562 29.2375 4.4782 9.5332
1.1205 1.72 2800 1.2992 29.7164 8.007 28.5027 28.5018 4.44 9.5332
1.1069 1.84 3000 1.2830 29.851 8.4312 28.8139 28.8205 4.6065 9.5332
1.076 1.96 3200 1.2821 30.0468 8.4998 29.0632 29.0231 4.8298 9.5332

Framework versions

  • Transformers 4.30.2
  • Pytorch 2.0.1+cu118
  • Datasets 2.13.1
  • Tokenizers 0.13.3
Downloads last month
51
Inference API
Inference API (serverless) does not yet support transformers models for this pipeline type.