Upload folder using huggingface_hub
Browse files- config.json +1 -0
- layers.10.mlp/cfg.json +1 -0
- layers.10.mlp/sae.safetensors +3 -0
- layers.4.mlp/cfg.json +1 -0
- layers.4.mlp/sae.safetensors +3 -0
- layers.6.mlp/cfg.json +1 -0
- layers.6.mlp/sae.safetensors +3 -0
- layers.8.mlp/cfg.json +1 -0
- layers.8.mlp/sae.safetensors +3 -0
- lr_scheduler.pt +3 -0
- optimizer.pt +3 -0
- state.pt +3 -0
config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"sae": {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 768, "k": 64, "multi_topk": false, "skip_connection": false}, "batch_size": 8, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["layers.4.mlp", "layers.6.mlp", "layers.8.mlp", "layers.10.mlp"], "init_seeds": [0], "layers": [], "layer_stride": 1, "transcode": true, "distribute_modules": false, "save_every": 1000, "log_to_wandb": true, "run_name": "k64-trans-768", "wandb_log_frequency": 1, "model": "EleutherAI/pythia-160m", "dataset": "/mnt/ssd-1/pile_preshuffled/standard/document.bin", "split": "train", "ctx_len": 2049, "hf_token": null, "revision": null, "load_in_8bit": false, "max_examples": 4000000, "resume": false, "text_column": "text", "finetune": null, "shuffle_seed": 42, "data_preprocessing_num_proc": 24}
|
layers.10.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 768, "k": 64, "multi_topk": false, "skip_connection": false, "d_in": 768}
|
layers.10.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cd0d62bfa550ba57852ab766f9366ef9eb8278bc3b2a0d896bb56af8461f91ce
|
3 |
+
size 4725048
|
layers.4.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 768, "k": 64, "multi_topk": false, "skip_connection": false, "d_in": 768}
|
layers.4.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9aa7c8436308917411c4a2defb3d4ae92f70f9fd0e7ca7df7184f878b4fd261c
|
3 |
+
size 4725048
|
layers.6.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 768, "k": 64, "multi_topk": false, "skip_connection": false, "d_in": 768}
|
layers.6.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:964df4fe6cfc3cb0268072c34d994ed36936c864ca36130cc194b37ea949f73b
|
3 |
+
size 4725048
|
layers.8.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 768, "k": 64, "multi_topk": false, "skip_connection": false, "d_in": 768}
|
layers.8.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8212573a98ac7f355912333824edc7faa9dcd63ac5a3ed4c4ca125d2b5d429d4
|
3 |
+
size 4725048
|
lr_scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:76ce317ce7a81776a7034016ebb4dee866ec3026ced3d688395e446fd3cd06e0
|
3 |
+
size 1076
|
optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c9788a30bf39b99e24c8fe5f614c8afa2d88b9c313c6a40f412ba80fa41c7946
|
3 |
+
size 9651244
|
state.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:860727bae580d982997e3deef6115b495b6dd27dc0d7db69cc8ae5d7df6ec4ca
|
3 |
+
size 26368
|