--- license: mit datasets: - DarwinAnim8or/grug language: - en pipeline_tag: text-generation tags: - grug - caveman - fun --- # GPT-Grug-125m A finetuned version of [GPT-Neo-125M](https://huggingface.co/EleutherAI/gpt-neo-125M) on the 'grug' dataset. # Training Procedure This was trained on the 'grug' dataset, using the "HappyTransformers" library on Google Colab. This model was trained for 4 epochs with learning rate 1e-2. # Biases & Limitations This likely contains the same biases and limitations as the original GPT-Neo-125M that it is based on, and additionally heavy biases from the grug datasets. # Intended Use This model is meant for fun, please do not take anything this caveman says seriously.