T5-Small Grammar Correction

A fine-tuned t5-small model for correcting grammar errors in English text. Given a sentence, the model generates a grammatically correct version using a text-to-text approach.

Model Details

  • Developed by: Harsha Vardhan N
  • Model type: Sequence-to-Sequence Transformer
  • Language(s): English
  • License: Apache 2.0
  • Finetuned from model: t5-small

Training Details

Training Data

The model was fine-tuned on the wiki_auto/auto_full_with_split dataset, a large-scale corpus designed for sentence-level grammatical and stylistic simplification. It contains aligned pairs of complex and simplified English sentences extracted from Wikipedia and Simple Wikipedia. For this task, the dataset was used to teach the model how to correct ungrammatical sentences into fluent and grammatically correct English.

Training Procedure

  • Epochs: 3
  • Training Duration: ~1 hour
  • Optimizer: AdamW (via Hugging Face Seq2SeqTrainer)
  • Learning Rate: 5e-5
  • Batch Size: 8
  • Environment: Google Colab GPU

Technical Specifications

Compute Infrastructure

Hardware

  • GPU: Google Colab-provided GPU (likely Tesla T4)

Software

  • Framework: Hugging Face Transformers, PyTorch
  • Trainer Used: Seq2SeqTrainer
Downloads last month
146
Safetensors
Model size
60.5M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Harshathemonster/t5-small-updated

Base model

google-t5/t5-small
Finetuned
(1949)
this model

Dataset used to train Harshathemonster/t5-small-updated