norabelrose commited on
Commit
b6c77b8
·
verified ·
1 Parent(s): 7d9848e

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. layers.0.attention/cfg.json +1 -0
  2. layers.0.attention/sae.safetensors +3 -0
  3. layers.0.mlp/cfg.json +1 -0
  4. layers.0.mlp/sae.safetensors +3 -0
  5. layers.0/cfg.json +1 -0
  6. layers.0/sae.safetensors +3 -0
  7. layers.1.attention/cfg.json +1 -0
  8. layers.1.attention/sae.safetensors +3 -0
  9. layers.1.mlp/cfg.json +1 -0
  10. layers.1.mlp/sae.safetensors +3 -0
  11. layers.1/cfg.json +1 -0
  12. layers.1/sae.safetensors +3 -0
  13. layers.10.attention/cfg.json +1 -0
  14. layers.10.attention/sae.safetensors +3 -0
  15. layers.10.mlp/cfg.json +1 -0
  16. layers.10.mlp/sae.safetensors +3 -0
  17. layers.10/cfg.json +1 -0
  18. layers.10/sae.safetensors +3 -0
  19. layers.11.attention/cfg.json +1 -0
  20. layers.11.attention/sae.safetensors +3 -0
  21. layers.11.mlp/cfg.json +1 -0
  22. layers.11.mlp/sae.safetensors +3 -0
  23. layers.11/cfg.json +1 -0
  24. layers.11/sae.safetensors +3 -0
  25. layers.2.attention/cfg.json +1 -0
  26. layers.2.attention/sae.safetensors +3 -0
  27. layers.2.mlp/cfg.json +1 -0
  28. layers.2.mlp/sae.safetensors +3 -0
  29. layers.2/cfg.json +1 -0
  30. layers.2/sae.safetensors +3 -0
  31. layers.3.attention/cfg.json +1 -0
  32. layers.3.attention/sae.safetensors +3 -0
  33. layers.3.mlp/cfg.json +1 -0
  34. layers.3.mlp/sae.safetensors +3 -0
  35. layers.3/cfg.json +1 -0
  36. layers.3/sae.safetensors +3 -0
  37. layers.4.attention/cfg.json +1 -0
  38. layers.4.attention/sae.safetensors +3 -0
  39. layers.4.mlp/cfg.json +1 -0
  40. layers.4.mlp/sae.safetensors +3 -0
  41. layers.4/cfg.json +1 -0
  42. layers.4/sae.safetensors +3 -0
  43. layers.5.attention/cfg.json +1 -0
  44. layers.5.attention/sae.safetensors +3 -0
  45. layers.5.mlp/cfg.json +1 -0
  46. layers.5.mlp/sae.safetensors +3 -0
  47. layers.5/cfg.json +1 -0
  48. layers.5/sae.safetensors +3 -0
  49. layers.6.attention/cfg.json +1 -0
  50. layers.6.attention/sae.safetensors +3 -0
layers.0.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.0.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8b885fe5e73b25395abff40bc2cb75f201dfbb741e15ec8fb6c399deab8040d
3
+ size 402918736
layers.0.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.0.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd5b565784c9a50e3bad3d653d5ea4e280f12c67146169f182daba5f499bc0fb
3
+ size 402918736
layers.0/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.0/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7695c02bc9908be6881bfe56e3a0beccaf7a36a8a39955336fbd33c0b97964fa
3
+ size 402918736
layers.1.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.1.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ebd16c9b65f5612c6c373b1ffca552fde5210d237dbfc9e74b24a4991fe1390
3
+ size 402918736
layers.1.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.1.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e687f9a04cf563b96664178a0174166feaa2276788edf1dfca6c2b00ea52b20
3
+ size 402918736
layers.1/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.1/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f72dfc33d726abada56eb7129800cbab3705486a55c5a7a90f16bb0e7061c016
3
+ size 402918736
layers.10.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.10.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f8659d9cd1bfbb6b90fc1c61d9b23429895f5f9cde9e10cae42e19b11ad0be2
3
+ size 402918736
layers.10.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.10.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a17e6492c1fd8d1e5e68bd64c5b454cd8fe95989c8288f8c361e3950a0265021
3
+ size 402918736
layers.10/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.10/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a36aca11add76409cc2f2db343405781300ca3fa7bf726b6a46c09ef819f9ff
3
+ size 402918736
layers.11.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.11.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4eb38c18773c440abad7670bcd7809e3fb66d60a4bf45b936061f68fcd9cd6d4
3
+ size 402918736
layers.11.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.11.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37a709b096425be047cf98c5cbfaa9caa56f0b29f4ac59cb63dd8e4aa6848705
3
+ size 402918736
layers.11/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.11/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:234156ef6e6ff0e7582119040581bc3ffe95ae98ef813055c6a97b7e4d026036
3
+ size 402918736
layers.2.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.2.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79c6341dea910f49cc701124e46b507beba08b03a71e1236d041aabed940bdf0
3
+ size 402918736
layers.2.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.2.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de60aa95a93a552c51d271114d82d3d902db17e465bd68cac07dbecbb60bdef0
3
+ size 402918736
layers.2/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.2/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eda1f4bf3af3f5b6ca6cd109e4211afdce29d475bd49507721dd894a31a14bb0
3
+ size 402918736
layers.3.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.3.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d54e455941af1a7357ffacdf910bb5b694f730a3dd44813275df5c6ccaf47700
3
+ size 402918736
layers.3.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.3.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1d72624a4d726584ac8470eab0a8e0ea991f4b6a104f90d87361c73ff7b2958
3
+ size 402918736
layers.3/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.3/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccda6df31c48b1d2961d6a4aeb663a21f064e180112dfa3ddc5805d110acbf2d
3
+ size 402918736
layers.4.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.4.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a76eb054f7b4c83e94a626ce3f967dae3c7b429cb5ba78efe9b0df8190202e16
3
+ size 12865536
layers.4.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.4.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dad4f38db46dfd5ded8deaed837a3b26bc100665e21ee004f7d18d639bda39b7
3
+ size 402918736
layers.4/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.4/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f1aa1b52352dba45336b55312e747bc6214c8bcf2933626671b4ec1af6c6652
3
+ size 402918736
layers.5.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.5.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d76fa053a3611b6b4330cb56ed0924c038beb23120a7ce16490bf7b28b0e845d
3
+ size 402918736
layers.5.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.5.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d1dab915a7d3d9e179c73c36611ac9046687dc40dede474eb35f7213ca0078a
3
+ size 402918736
layers.5/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.5/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0758cd882e2c2e9dd020c2912d8f3022c1de7bb0ad7ff23d581959534235d58c
3
+ size 402918736
layers.6.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 65536, "k": 32, "signed": false, "d_in": 768}
layers.6.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:251758e5609045ce28c79c293af8a1703d429371b4bda26efac246244adc7385
3
+ size 402918736