Cached SAE/transcoder acts stored in CSR format. Not especially optimized for others' use/fleshed out. It's mostly just a space for me to place and download activations.

If you want to use them, do

import torch
from huggingface_hub import hf_hub_download

def load_feat_acts(fname, only_active_docs=False):
  local_path = hf_hub_download(repo_id="noanabeshima/tiny_model_cached_acts", filename=fname)
  csr_kwargs = torch.load(local_path)

  # The matrices might be stored in space-efficient formats that're incompatible with torch's sparse csr tensor.
  # Convert them back before constructing the matrix.
  csr_kwargs['crow_indices'] = csr_kwargs['crow_indices'].int()
  csr_kwargs['col_indices'] = csr_kwargs['col_indices'].int()
  values = csr_kwargs['values'].float()
  csr_kwargs['values'] = values/values.max()
  
  feat_acts = torch.sparse_csr_tensor(**csr_kwargs)
  return feat_acts

feat_acts = load_feat_acts(f"mlp_map_test/M2_S-2_R1_P0/{300}.pt").to_dense()

The activations are from 13% of the train split of https://huggingface.co/datasets/noanabeshima/TinyModelTokIds. I gather all the activations per feature and then take the smallest activations prefix so that each feature has at least 3K documents on which it activates and 12K token-activations.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.