|
{ |
|
"name": "InterpBench", |
|
"version": "1.0.0", |
|
"description": "A benchmark of transformers with known circuits for evaluating mechanistic interpretability techniques.", |
|
"license": "https://creativecommons.org/licenses/by/4.0/", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench", |
|
"cases": [ |
|
{ |
|
"case_id": "11", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/11", |
|
"task_description": "Counts the number of words in a sequence based on their length.", |
|
"vocab": [ |
|
"J", |
|
"LB", |
|
"TPSI", |
|
"V", |
|
"b", |
|
"no", |
|
"oCLrZaW", |
|
"poiVg" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 12, |
|
"n_ctx": 10, |
|
"d_head": 3, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 48, |
|
"act_fn": "gelu", |
|
"d_vocab": 10, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.1460593486680443, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 5, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 3456, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.01, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 0.4, |
|
"epochs": 500, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 1.0, |
|
"lr_scheduler": "" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "13", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/13", |
|
"task_description": "Analyzes the trend (increasing, decreasing, constant) of numeric tokens.", |
|
"vocab": [ |
|
0, |
|
1, |
|
2 |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 20, |
|
"n_ctx": 10, |
|
"d_head": 5, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 80, |
|
"act_fn": "gelu", |
|
"d_vocab": 5, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "bidirectional", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.1460593486680443, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 3, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 9600, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.01, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 0.4, |
|
"epochs": 500, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 1.0, |
|
"lr_scheduler": "" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "18", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/18", |
|
"task_description": "Classify each token based on its frequency as 'rare', 'common', or 'frequent'.", |
|
"vocab": [ |
|
"a", |
|
"b", |
|
"c", |
|
"d", |
|
"e" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 26, |
|
"n_ctx": 10, |
|
"d_head": 6, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 104, |
|
"act_fn": "gelu", |
|
"d_vocab": 7, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "bidirectional", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.12344267996967354, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 3, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 15808, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.001, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 1.0, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "", |
|
"model_pair": "strict" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "19", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/19", |
|
"task_description": "Removes consecutive duplicate tokens from a sequence.", |
|
"vocab": [ |
|
"a", |
|
"b", |
|
"c" |
|
], |
|
"max_seq_len": 15, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 32, |
|
"n_ctx": 15, |
|
"d_head": 8, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 128, |
|
"act_fn": "gelu", |
|
"d_vocab": 5, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.15689290811054724, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 3, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 24576, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.001, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 0.4, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "20", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/20", |
|
"task_description": "Detect spam messages based on appearance of spam keywords.", |
|
"vocab": [ |
|
"J", |
|
"LB", |
|
"TPSI", |
|
"V", |
|
"b", |
|
"click", |
|
"no", |
|
"now", |
|
"oCLrZaW", |
|
"offer", |
|
"poiVg", |
|
"spam" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 13, |
|
"n_ctx": 10, |
|
"d_head": 3, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 52, |
|
"act_fn": "gelu", |
|
"d_vocab": 14, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.16, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 2, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 3952, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.01, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 0.4, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "", |
|
"model_pair": "strict" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "21", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/21", |
|
"task_description": "Extract unique tokens from a string", |
|
"vocab": [ |
|
"a", |
|
"b", |
|
"c" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 4, |
|
"d_model": 50, |
|
"n_ctx": 10, |
|
"d_head": 12, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 200, |
|
"act_fn": "gelu", |
|
"d_vocab": 5, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.09847319278346618, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 3, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 118400, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.0005, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 0.5, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "", |
|
"model_pair": "strict" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "26", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/26", |
|
"task_description": "Creates a cascading effect by repeating each token in sequence incrementally.", |
|
"vocab": [ |
|
"a", |
|
"b", |
|
"c" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 21, |
|
"n_ctx": 10, |
|
"d_head": 5, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 84, |
|
"act_fn": "gelu", |
|
"d_vocab": 5, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.12344267996967354, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 27, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 10416, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.01, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 0.4, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "", |
|
"model_pair": "strict" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "29", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/29", |
|
"task_description": "Creates abbreviations for each token in the sequence.", |
|
"vocab": [ |
|
"J", |
|
"LB", |
|
"TPSI", |
|
"V", |
|
"b", |
|
"no", |
|
"oCLrZaW", |
|
"poiVg" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 13, |
|
"n_ctx": 10, |
|
"d_head": 3, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 52, |
|
"act_fn": "gelu", |
|
"d_vocab": 10, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.1539600717839002, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 8, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 3952, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.01, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 0.4, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "", |
|
"model_pair": "strict" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "3", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/3", |
|
"task_description": "Returns the fraction of 'x' in the input up to the i-th position for all i.", |
|
"vocab": [ |
|
"a", |
|
"b", |
|
"c", |
|
"x" |
|
], |
|
"max_seq_len": 5, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 12, |
|
"n_ctx": 5, |
|
"d_head": 3, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 48, |
|
"act_fn": "gelu", |
|
"d_vocab": 6, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.22188007849009167, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 1, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 3456, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.01, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 10.0, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "", |
|
"model_pair": "strict", |
|
"same_size": false |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "33", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/33", |
|
"task_description": "Checks if each token's length is odd or even.", |
|
"vocab": [ |
|
"J", |
|
"LB", |
|
"TPSI", |
|
"V", |
|
"b", |
|
"no", |
|
"oCLrZaW", |
|
"poiVg" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 4, |
|
"n_ctx": 10, |
|
"d_head": 1, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 16, |
|
"act_fn": "gelu", |
|
"d_vocab": 10, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.17457431218879393, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 2, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 384, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.001, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 0.4, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "34", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/34", |
|
"task_description": "Calculate the ratio of vowels to consonants in each word.", |
|
"vocab": [ |
|
"J", |
|
"LB", |
|
"TPSI", |
|
"V", |
|
"b", |
|
"no", |
|
"oCLrZaW", |
|
"poiVg" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 16, |
|
"n_ctx": 10, |
|
"d_head": 4, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 64, |
|
"act_fn": "gelu", |
|
"d_vocab": 10, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.16329931618554522, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 5, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 6144, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.01, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 1.0, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "", |
|
"model_pair": "strict" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "35", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/35", |
|
"task_description": "Alternates capitalization of each character in words.", |
|
"vocab": [ |
|
"J", |
|
"LB", |
|
"TPSI", |
|
"V", |
|
"b", |
|
"no", |
|
"oCLrZaW", |
|
"poiVg" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 9, |
|
"n_ctx": 10, |
|
"d_head": 2, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 36, |
|
"act_fn": "gelu", |
|
"d_vocab": 10, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.1539600717839002, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 8, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 1872, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.01, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 1.0, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "", |
|
"model_pair": "strict" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "36", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/36", |
|
"task_description": "Classifies each token as 'positive', 'negative', or 'neutral' based on emojis.", |
|
"vocab": [ |
|
"\ud83d\udcd8", |
|
"\ud83d\ude0a", |
|
"\ud83d\ude22" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 6, |
|
"n_ctx": 10, |
|
"d_head": 1, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 24, |
|
"act_fn": "gelu", |
|
"d_vocab": 5, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.19402850002906638, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 3, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 768, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.01, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 1.0, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "", |
|
"model_pair": "strict" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "37", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/37", |
|
"task_description": "Reverses each word in the sequence except for specified exclusions.", |
|
"vocab": [ |
|
"J", |
|
"LB", |
|
"TPSI", |
|
"V", |
|
"b", |
|
"no", |
|
"oCLrZaW", |
|
"poiVg" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 12, |
|
"n_ctx": 10, |
|
"d_head": 3, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 48, |
|
"act_fn": "gelu", |
|
"d_vocab": 10, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.1539600717839002, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 8, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 3456, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.01, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 1.0, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "", |
|
"model_pair": "strict" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "4", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/4", |
|
"task_description": "Return fraction of previous open tokens minus the fraction of close tokens.", |
|
"vocab": [ |
|
"(", |
|
")", |
|
"a", |
|
"b", |
|
"c" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 20, |
|
"n_ctx": 10, |
|
"d_head": 5, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 80, |
|
"act_fn": "gelu", |
|
"d_vocab": 7, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.17056057308448835, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 1, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 9600, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.001, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 0.4, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "7", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/7", |
|
"task_description": "Returns the number of times each token occurs in the input.", |
|
"vocab": [ |
|
"a", |
|
"b", |
|
"c" |
|
], |
|
"max_seq_len": 10, |
|
"min_seq_len": 4, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/7/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/7/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/7/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/7/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 2, |
|
"d_model": 17, |
|
"n_ctx": 10, |
|
"d_head": 4, |
|
"model_name": "custom", |
|
"n_heads": 4, |
|
"d_mlp": 68, |
|
"act_fn": "gelu", |
|
"d_vocab": 5, |
|
"eps": 1e-05, |
|
"use_attn_result": true, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": true, |
|
"use_hook_mlp_in": true, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": null, |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": null, |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": null, |
|
"n_devices": 1, |
|
"attention_dir": "bidirectional", |
|
"attn_only": false, |
|
"seed": 0, |
|
"initializer_range": 0.15689290811054724, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 10, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 6800, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": null, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/7/ll_model_cfg.pkl", |
|
"training_args": { |
|
"atol": 0.05, |
|
"lr": 0.01, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 0.5, |
|
"epochs": 2000, |
|
"act_fn": "gelu", |
|
"clip_grad_norm": 0.1, |
|
"lr_scheduler": "", |
|
"model_pair": "strict", |
|
"same_size": false, |
|
"seed": 1234, |
|
"batch_size": 256, |
|
"include_mlp": false, |
|
"next_token": false, |
|
"detach_while_caching": true |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/7/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/7/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/7/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "ioi", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/ioi", |
|
"task_description": "Indirect Object Identification (IOI) task.", |
|
"max_seq_len": 16, |
|
"min_seq_len": 16, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 6, |
|
"d_model": 64, |
|
"n_ctx": 1024, |
|
"d_head": 16, |
|
"model_name": "gpt2", |
|
"n_heads": 4, |
|
"d_mlp": 3072, |
|
"act_fn": "gelu_new", |
|
"d_vocab": 50257, |
|
"eps": 1e-05, |
|
"use_attn_result": false, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": false, |
|
"use_hook_mlp_in": false, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": "GPT2LMHeadModel", |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": "gpt2", |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": "LNPre", |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": null, |
|
"initializer_range": 0.02886751345948129, |
|
"init_weights": false, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 50257, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 84934656, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": false, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/ll_model_cfg.pkl", |
|
"training_args": { |
|
"next_token": false, |
|
"non_ioi_thresh": 0.65, |
|
"use_per_token_check": false, |
|
"batch_size": 512, |
|
"num_workers": 0, |
|
"early_stop": true, |
|
"lr_scheduler": null, |
|
"scheduler_val_metric": [ |
|
"val/accuracy", |
|
"val/IIA" |
|
], |
|
"scheduler_mode": "max", |
|
"scheduler_kwargs": {}, |
|
"clip_grad_norm": 1.0, |
|
"seed": 0, |
|
"lr": 0.001, |
|
"detach_while_caching": true, |
|
"optimizer_kwargs": { |
|
"betas": [ |
|
0.9, |
|
0.9 |
|
] |
|
}, |
|
"atol": 0.05, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"val_IIA_sampling": "random", |
|
"use_all_tokens_for_behavior": false, |
|
"strict_weight": 0.4, |
|
"siit_sampling": "individual" |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/edges.pkl" |
|
}, |
|
{ |
|
"case_id": "ioi_next_token", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/ioi_next_token", |
|
"task_description": "Indirect Object Identification (IOI) task, trained using next token prediction.", |
|
"max_seq_len": 16, |
|
"min_seq_len": 16, |
|
"files": [ |
|
{ |
|
"file_name": "edges.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/edges.pkl" |
|
}, |
|
{ |
|
"file_name": "ll_model.pth", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/ll_model.pth" |
|
}, |
|
{ |
|
"file_name": "ll_model_cfg.pkl", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/ll_model_cfg.pkl" |
|
}, |
|
{ |
|
"file_name": "meta.json", |
|
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/meta.json" |
|
} |
|
], |
|
"transformer_cfg": { |
|
"n_layers": 6, |
|
"d_model": 64, |
|
"n_ctx": 1024, |
|
"d_head": 16, |
|
"model_name": "gpt2", |
|
"n_heads": 4, |
|
"d_mlp": 3072, |
|
"act_fn": "gelu_new", |
|
"d_vocab": 50257, |
|
"eps": 1e-05, |
|
"use_attn_result": false, |
|
"use_attn_scale": true, |
|
"use_split_qkv_input": false, |
|
"use_hook_mlp_in": false, |
|
"use_attn_in": false, |
|
"use_local_attn": false, |
|
"original_architecture": "GPT2LMHeadModel", |
|
"from_checkpoint": false, |
|
"checkpoint_index": null, |
|
"checkpoint_label_type": null, |
|
"checkpoint_value": null, |
|
"tokenizer_name": "gpt2", |
|
"window_size": null, |
|
"attn_types": null, |
|
"init_mode": "gpt2", |
|
"normalization_type": "LNPre", |
|
"n_devices": 1, |
|
"attention_dir": "causal", |
|
"attn_only": false, |
|
"seed": null, |
|
"initializer_range": 0.02886751345948129, |
|
"init_weights": true, |
|
"scale_attn_by_inverse_layer_idx": false, |
|
"positional_embedding_type": "standard", |
|
"final_rms": false, |
|
"d_vocab_out": 50257, |
|
"parallel_attn_mlp": false, |
|
"rotary_dim": null, |
|
"n_params": 2457600, |
|
"use_hook_tokens": false, |
|
"gated_mlp": false, |
|
"default_prepend_bos": true, |
|
"dtype": "torch.float32", |
|
"tokenizer_prepends_bos": false, |
|
"n_key_value_heads": null, |
|
"post_embedding_ln": false, |
|
"rotary_base": 10000, |
|
"trust_remote_code": false, |
|
"rotary_adjacent_pairs": false |
|
}, |
|
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/ll_model_cfg.pkl", |
|
"training_args": { |
|
"next_token": true, |
|
"non_ioi_thresh": 0.65, |
|
"use_per_token_check": false, |
|
"batch_size": 256, |
|
"lr": 0.001, |
|
"num_workers": 0, |
|
"early_stop": true, |
|
"lr_scheduler": null, |
|
"scheduler_val_metric": [ |
|
"val/accuracy", |
|
"val/IIA" |
|
], |
|
"scheduler_mode": "max", |
|
"clip_grad_norm": 1.0, |
|
"atol": 0.05, |
|
"use_single_loss": false, |
|
"iit_weight": 1.0, |
|
"behavior_weight": 1.0, |
|
"strict_weight": 0.4 |
|
}, |
|
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/meta.json", |
|
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/ll_model.pth", |
|
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/edges.pkl" |
|
} |
|
] |
|
} |