Trained in 4-bit on pygmalion-6b as POC Uses the GPTeacher roleplay dataset. ``` INFO:Getting model ready... INFO:Prepping for training... INFO:Creating LoRA model... INFO:Starting training... {'loss': 12.5737, 'learning_rate': 0.0002926829268292683, 'epoch': 0.33} {'loss': 8.5515, 'learning_rate': 0.0002560975609756097, 'epoch': 0.67} {'loss': 7.5768, 'learning_rate': 0.0002195121951219512, 'epoch': 1.0} {'loss': 6.9769, 'learning_rate': 0.00018292682926829266, 'epoch': 1.33} {'loss': 6.6842, 'learning_rate': 0.00014634146341463414, 'epoch': 1.66} {'loss': 6.3925, 'learning_rate': 0.0001097560975609756, 'epoch': 2.0} {'loss': 6.041, 'learning_rate': 7.317073170731707e-05, 'epoch': 2.33} {'loss': 5.6818, 'learning_rate': 3.6585365853658535e-05, 'epoch': 2.66} {'loss': 5.4639, 'learning_rate': 0.0, 'epoch': 2.99} {'train_runtime': 960.7748, 'train_samples_per_second': 6.005, 'train_steps_per_second': 0.047, 'train_loss': 7.326934729682074, 'epoch': 2.99} INFO:LoRA training run is completed and saved. INFO:Training complete! ``` I used the electricity so might as well post it.