darpanaswal commited on
Commit
ea510c2
·
verified ·
1 Parent(s): 47a55ec

Upload SAE l0_100_0

Browse files
l0_100_0/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"d_in": 2048, "d_sae": 8192, "dtype": "float32", "device": "cpu", "apply_b_dec_to_input": false, "normalize_activations": "none", "reshape_activations": "none", "metadata": {"sae_lens_version": "6.21.0", "sae_lens_training_version": "6.21.0", "dataset_path": "darpanaswal/LLMSymGuard", "hook_name": "blocks.7.hook_attn_out", "model_name": "meta-llama/Llama-3.2-1B-Instruct", "model_class_name": "HookedTransformer", "hook_head_index": null, "context_size": 128, "seqpos_slice": [null], "model_from_pretrained_kwargs": {"center_writing_weights": false}, "prepend_bos": true, "exclude_special_tokens": false, "sequence_separator_token": "bos", "disable_concat_sequences": false}, "architecture": "jumprelu"}
l0_100_0/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80b82c08ad667afd87e34dbd0f9adcf9937d6679dc989b2fcf986ac39c7f7a9c
3
+ size 134291856