gpt-grug-125m / README.md
DarwinAnim8or's picture
Update README.md
16023fc
|
raw
history blame
712 Bytes
metadata
license: mit
datasets:
  - DarwinAnim8or/grug
language:
  - en
pipeline_tag: text-generation
tags:
  - grug
  - caveman
  - fun

GPT-Grug-125m

A finetuned version of GPT-Neo-125M on the 'grug' dataset.

Training Procedure

This was trained on the 'grug' dataset, using the "HappyTransformers" library on Google Colab. This model was trained for 4 epochs with learning rate 1e-2.

Biases & Limitations

This likely contains the same biases and limitations as the original GPT-Neo-125M that it is based on, and additionally heavy biases from the grug datasets.

Intended Use

This model is meant for fun, please do not take anything this caveman says seriously.