YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

CLIP Sparse Autoencoder Checkpoint

This model is a sparse autoencoder trained on CLIP's internal representations.

Model Details

Architecture

  • Layer: 4
  • Layer Type: hook_resid_post
  • Model: open-clip:laion/CLIP-ViT-B-32-DataComp.XL-s13B-b90K
  • Dictionary Size: 49152
  • Input Dimension: 768
  • Expansion Factor: 64
  • CLS Token Only: False

Training

  • Training Images: 1299988
  • Learning Rate: 0.0014
  • L1 Coefficient: 0.0044
  • Batch Size: 4096
  • Context Size: 50

Performance Metrics

Sparsity

  • L0 (Active Features): 64.0000

  • Dead Features: 0

  • Mean Passes Since Fired: 0.5054

Reconstruction

  • Explained Variance: 0.8013
  • Explained Variance Std: 0.0466
  • MSE Loss: 0.0016
  • L1 Loss: 0
  • Overall Loss: 0.0016

Training Details

  • Training Duration: 4772 seconds
  • Final Learning Rate: 0.0000
  • Warm Up Steps: 200
  • Gradient Clipping: 1

Additional Information

Downloads last month
15
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.