flan-t5-large-samsum

This model is a fine-tuned version of google/flan-t5-large on the samsum dataset. It achieves the following results on the evaluation set:

  • Loss: 1.2066
  • Rouge1: 49.1053
  • Rouge2: 25.4565
  • Rougel: 41.9146
  • Rougelsum: 45.3592
  • Gen Len: 17.1380

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
1.1847 1.0 1844 1.2066 49.1053 25.4565 41.9146 45.3592 17.1380
1.0533 2.0 3688 1.2126 49.5421 26.1526 42.1131 45.5735 17.2564
0.9521 3.0 5532 1.2315 49.7252 26.1855 42.2726 45.747 17.3358
0.8746 4.0 7376 1.2510 49.4306 25.9048 41.9821 45.4322 17.4750
0.8334 5.0 9220 1.2631 49.4852 25.9416 42.0469 45.5014 17.3944

Framework versions

  • Transformers 4.34.1
  • Pytorch 2.1.0+cu118
  • Datasets 2.14.6
  • Tokenizers 0.14.1
Downloads last month
23
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for andgonzalez/flan-t5-large-samsum

Finetuned
(106)
this model

Dataset used to train andgonzalez/flan-t5-large-samsum

Evaluation results