SAEs for use with the SAELens library

This repository contains the following SAEs for layers 0 through 11 (inclusive):

  • transformer.h.LAYER
  • blocks.LAYER.hook_resid_pre
  • blocks.LAYER.hook_resid_post

Load these SAEs using SAELens as below:

from sae_lens import SAE

sae, cfg_dict, sparsity = SAE.from_pretrained("austindavis/chess-gpt2-uci-12x12x768-saes", "<sae_id>")
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support