Model Card: TinyLlama-1.1B-Chat-v1.0-Unfiltered


Model Name: TinyLlama-1.1B-Chat-v1.0-Unfiltered
Model Type: Conversational AI Model
Architecture: Based on a 1.1B parameter TinyLlama architecture

Training Data:

  • Fine-tuned on the "dan_remixed" dataset (2.7MB).
  • The dataset improves spelling, grammar, and consistency while replacing references to violent crimes with non-violent activities and removes self-censorship from explicatives.

Training Time: Approximately 30-45 minutes. Each validation epoch takes ~322 seconds.
Hardware: Trained on Google Colab Pro A100 GPU (40GB).


Training Performance:

  • Epoch Losses:
    • Epoch 1: 0.7209
    • Epoch 2: 0.4441
    • Epoch 3: 0.3683
    • Epoch 4: 0.3358
    • Epoch 5: 0.3145
  • Final Training Loss (Epoch 5): 0.3145

Validation Performance (5 Epochs):

  • Epoch 1:

    • Training Loss: 0.2921
    • Validation Loss: 0.7962
    • Perplexity: 2.22
    • Epoch completed in 321.64 seconds
  • Epoch 2:

    • Training Loss: 0.2872
    • Validation Loss: 0.7672
    • Perplexity: 2.15
    • Epoch completed in 321.91 seconds
  • Epoch 3:

    • Training Loss: 0.2874
    • Validation Loss: 0.7821
    • Perplexity: 2.19
    • Epoch completed in 321.94 seconds
  • Epoch 4:

    • Training Loss: 0.2864
    • Validation Loss: 0.7796
    • Perplexity: 2.18
    • Epoch completed in 322.01 seconds
  • Epoch 5:

    • Training Loss: 0.2831
    • Validation Loss: 0.8017
    • Perplexity: 2.23
    • Epoch completed in 322.01 seconds

Optimizer: AdamW, learning rate: 1e-5
Loss Function: Cross-Entropy Loss, ignoring padding tokens (ignore_index=-100)
Use Case: Conversational AI designed for general, unrestricted conversation, with no filtering on the nature of responses, provided the content is non-violent.


Limitations:

  • Due to the small fine-tuning dataset size (2.7MB), the model may be prone to overfitting and bias.
  • The dataset has been modified to avoid violent language, but the model might still exhibit strong or explicit responses.

Metrics:

  • Loss and perplexity have been tracked, and more conversational metrics (like BLEU, ROUGE, or human evaluation) could be explored.
Downloads last month
10
Safetensors
Model size
1.1B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Collection including datasciguy/TinyLlama-1.1B-Chat-v1.0-Unfiltered