Edit model card

bart-base-instructiongen-w-inputs

Use this text2text model to find out what LLM instruction (and inputs if relevant) might have generated <arbitrary input text>!

about

This model is a fine-tuned version of facebook/bart-base on the pszemraj/fleece2instructions-inputs-alpaca-cleaned dataset.

It achieves the following results on the evaluation set:

  • Loss: 0.9579
  • Rouge1: 62.3604
  • Rouge2: 39.5109
  • Rougel: 58.8843
  • Rougelsum: 60.4494
  • Gen Len: 24.9917

Example

base

Intended uses & limitations

This model is intended to be used to generate instructions from arbitrary text. You can then use these instructions + your data to fine-tune an LLM on instructions w.r.t. a specific domain. This model is primarily intended to enable low-resource domain adaptation, rather than "I want to generate even better prompts for the FLAN-V2 dataset!".

The fleece2instructions-inputs-alpaca-cleaned dataset, obtained from the alpaca-lora repo under the ODC-BY license, has been converted to a text2text format for use with language models. In this dataset, the original 'inputs' and 'instructions' columns are combined into a single 'instructions_inputs' column. To clearly separate the two types of content, each piece of text is prefixed with either an <instruction> or <inputs> token. These tokens not only facilitate model comprehension, but also allow for easy regex separation of model outputs during inference.

As such, users can expect the output of this model to be similarly structured with <instruction> and <inputs> tokens.

This is just the base model, for better performance (but slower/compute intensive) see the bart-large version. Further exploration/data may lead to even better models!

Training and evaluation data

Refer to the fleece2instructions-inputs-alpaca-cleaned dataset

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 8e-05
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • distributed_type: multi-GPU
  • gradient_accumulation_steps: 16
  • total_train_batch_size: 64
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.02
  • num_epochs: 2.0

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
1.1147 1.0 680 0.9901 61.8451 38.8293 58.3372 59.8658 25.2401
0.9565 2.0 1360 0.9579 62.3604 39.5109 58.8843 60.4494 24.9917
Downloads last month
4
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train pszemraj/bart-base-instructiongen-w-inputs