gpt2-ratios / gpt2_expansion_factor_12_20240307175639539342_config.json
jbrinkma's picture
Upload gpt2_expansion_factor_12_20240307175639539342_config.json with huggingface_hub
9755216 verified
{
"seed": 42,
"model_name_or_path": "gpt2",
"hook_point": "transformer.h.3",
"dataset_name_or_path": "Elriggs/openwebtext-100k",
"activation_size": -1,
"add_bos_token": false,
"expansion_factor": 12,
"b_dec_init_method": "",
"n_steps": -1,
"device": "cuda",
"batch_size": 32,
"ctx_length": 256,
"lr": 0.001,
"min_lr": 0.0,
"lr_warmup_steps": 5000,
"sparsity_coefficient": 0.003,
"evaluation_interval": 200,
"beta1": 0.9,
"beta2": 0.999,
"l1_sqrt": false,
"n_tokens_in_feature_cache": 500000.0,
"use_ghost_grads": false,
"output_dir": "outputs",
"cache_dir": "cache",
"checkpoint_interval": 200,
"use_wandb": true,
"wandb_entity": "best_sae",
"wandb_project": "best_sae",
"wandb_name": "gpt2_expansion_factor_12_20240307175639539342",
"wandb_group": "Ratios_GPT2"
}