Model Card: TinyLlama-1.1B-Chat-v1.0-Unfiltered


Model Name: TinyLlama-1.1B-Chat-v1.0-Unfiltered
Model Type: Conversational AI Model
Architecture: Based on a 1.1B parameter TinyLlama architecture

Training Data:

  • Fine-tuned on the "dan_remixed" dataset (2.7MB).
  • The dataset improves spelling, grammar, and consistency while replacing references to violent crimes with non-violent activities and removes self-censorship from explicatives.

Training Time: Approximately 30-45 minutes. Each validation epoch takes ~322 seconds.
Hardware: Trained on Google Colab Pro A100 GPU (40GB).


Training Performance:

  • Epoch Losses:
    • Epoch 1: 0.7209
    • Epoch 2: 0.4441
    • Epoch 3: 0.3683
    • Epoch 4: 0.3358
    • Epoch 5: 0.3145
  • Final Training Loss (Epoch 5): 0.3145

Validation Performance (5 Epochs):

  • Epoch 1:

    • Training Loss: 0.2921
    • Validation Loss: 0.7962
    • Perplexity: 2.22
    • Epoch completed in 321.64 seconds
  • Epoch 2:

    • Training Loss: 0.2872
    • Validation Loss: 0.7672
    • Perplexity: 2.15
    • Epoch completed in 321.91 seconds
  • Epoch 3:

    • Training Loss: 0.2874
    • Validation Loss: 0.7821
    • Perplexity: 2.19
    • Epoch completed in 321.94 seconds
  • Epoch 4:

    • Training Loss: 0.2864
    • Validation Loss: 0.7796
    • Perplexity: 2.18
    • Epoch completed in 322.01 seconds
  • Epoch 5:

    • Training Loss: 0.2831
    • Validation Loss: 0.8017
    • Perplexity: 2.23
    • Epoch completed in 322.01 seconds

Optimizer: AdamW, learning rate: 1e-5
Loss Function: Cross-Entropy Loss, ignoring padding tokens (ignore_index=-100)
Use Case: Conversational AI designed for general, unrestricted conversation, with no filtering on the nature of responses, provided the content is non-violent.


Limitations:

  • Due to the small fine-tuning dataset size (2.7MB), the model may be prone to overfitting and bias.
  • The dataset has been modified to avoid violent language, but the model might still exhibit strong or explicit responses.

Metrics:

  • Loss and perplexity have been tracked, and more conversational metrics (like BLEU, ROUGE, or human evaluation) could be explored.
Downloads last month
2
Safetensors
Model size
1.1B params
Tensor type
F32
·
Inference API
Unable to determine this model's library. Check the docs .

Collection including datasciguy/TinyLlama-1.1B-Chat-v1.0-Unfiltered