Not needed
Browse filesThe readme is for the old model name and package, for huggingface we will use a model card.
- Readme.txt +0 -7
Readme.txt
DELETED
@@ -1,7 +0,0 @@
|
|
1 |
-
This is my first attempt at finetuning the GPT AI.
|
2 |
-
Mia is based on GPT-Neo-125M and was trained on the original training data of the AI Dungeon model_v5 model.
|
3 |
-
It is a much smaller model than the original model_v5 and will run a lot quicker.
|
4 |
-
Training seemed to have helped with it not repeating commands back to you, but 125M is not large enough for proper gameplay.
|
5 |
-
|
6 |
-
The model was entirely trained on Google Colab (125M is the largest you can train) and ran for almost 3 epoch's (The last 5 mintues I ran out of diskspace but managed to export the model).
|
7 |
-
- Henk717
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|