cybershiptrooper iarcuschin commited on
Commit
9661ddf
1 Parent(s): c843477

Add metadata files (#1)

Browse files

- Add metadata files (a1d74d9edcc5fe196be04281417a40c81dbfae63)


Co-authored-by: Ivan Arcuschin <[email protected]>

benchmark_cases_metadata.csv ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ case_id,task_description,max_seq_len,min_seq_len,training_args.atol,training_args.lr,training_args.use_single_loss,training_args.iit_weight,training_args.behavior_weight,training_args.strict_weight,training_args.epochs,training_args.act_fn,training_args.clip_grad_norm,training_args.lr_scheduler,transformer_cfg.n_layers,transformer_cfg.d_model,transformer_cfg.n_ctx,transformer_cfg.d_head,transformer_cfg.model_name,transformer_cfg.n_heads,transformer_cfg.d_mlp,transformer_cfg.act_fn,transformer_cfg.d_vocab,transformer_cfg.eps,transformer_cfg.use_attn_result,transformer_cfg.use_attn_scale,transformer_cfg.use_split_qkv_input,transformer_cfg.use_hook_mlp_in,transformer_cfg.use_attn_in,transformer_cfg.use_local_attn,transformer_cfg.original_architecture,transformer_cfg.from_checkpoint,transformer_cfg.checkpoint_index,transformer_cfg.checkpoint_label_type,transformer_cfg.checkpoint_value,transformer_cfg.tokenizer_name,transformer_cfg.window_size,transformer_cfg.attn_types,transformer_cfg.init_mode,transformer_cfg.normalization_type,transformer_cfg.device,transformer_cfg.n_devices,transformer_cfg.attention_dir,transformer_cfg.attn_only,transformer_cfg.seed,transformer_cfg.initializer_range,transformer_cfg.init_weights,transformer_cfg.scale_attn_by_inverse_layer_idx,transformer_cfg.positional_embedding_type,transformer_cfg.final_rms,transformer_cfg.d_vocab_out,transformer_cfg.parallel_attn_mlp,transformer_cfg.rotary_dim,transformer_cfg.n_params,transformer_cfg.use_hook_tokens,transformer_cfg.gated_mlp,transformer_cfg.default_prepend_bos,transformer_cfg.dtype,transformer_cfg.tokenizer_prepends_bos,transformer_cfg.n_key_value_heads,transformer_cfg.post_embedding_ln,transformer_cfg.rotary_base,transformer_cfg.trust_remote_code,transformer_cfg.rotary_adjacent_pairs
2
+ 11,Counts the number of words in a sequence based on their length.,10.0,4.0,0.05,0.01,False,1.0,1.0,0.4,500.0,gelu,1.0,,2.0,12.0,10.0,3.0,custom,4.0,48.0,gelu,10.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,causal,False,0.0,0.1460593486680443,True,False,standard,False,5.0,False,,3456.0,False,False,True,torch.float32,,,False,10000.0,False,False
3
+ 13,"Analyzes the trend (increasing, decreasing, constant) of numeric tokens.",10.0,4.0,0.05,0.01,False,1.0,1.0,0.4,500.0,gelu,1.0,,2.0,20.0,10.0,5.0,custom,4.0,80.0,gelu,5.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,bidirectional,False,0.0,0.1460593486680443,True,False,standard,False,3.0,False,,9600.0,False,False,True,torch.float32,,,False,10000.0,False,False
4
+ 18,"Classify each token based on its frequency as 'rare', 'common', or 'frequent'.",10.0,4.0,0.05,0.001,False,1.0,1.0,0.4,2000.0,gelu,0.1,,2.0,12.0,10.0,3.0,custom,4.0,48.0,gelu,7.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,bidirectional,False,0.0,0.12344267996967354,True,False,standard,False,3.0,False,,3456.0,False,False,True,torch.float32,,,False,10000.0,False,False
5
+ 19,Removes consecutive duplicate tokens from a sequence.,15.0,4.0,0.05,0.001,False,1.0,1.0,0.4,2000.0,gelu,0.1,,2.0,32.0,15.0,8.0,custom,4.0,128.0,gelu,5.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,causal,False,0.0,0.15689290811054724,True,False,standard,False,3.0,False,,24576.0,False,False,True,torch.float32,,,False,10000.0,False,False
6
+ 20,Detect spam messages based on appearance of spam keywords.,10.0,4.0,0.05,0.001,False,1.0,1.0,1.0,2000.0,gelu,0.1,,2.0,4.0,10.0,1.0,custom,4.0,16.0,gelu,14.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cuda,1.0,causal,False,0.0,0.16,True,False,standard,False,2.0,False,,384.0,False,False,True,torch.float32,,,False,10000.0,False,False
7
+ 21,Extract unique tokens from a string,10.0,4.0,0.05,0.01,False,1.0,1.0,0.4,500.0,gelu,1.0,,2.0,20.0,10.0,5.0,custom,4.0,80.0,gelu,5.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,causal,False,0.0,0.1885618083164127,True,False,standard,False,3.0,False,,9600.0,False,False,True,torch.float32,,,False,10000.0,False,False
8
+ 24,Identifies the first occurrence of each token in a sequence.,10.0,4.0,0.05,0.01,False,1.0,1.0,0.4,500.0,gelu,1.0,,2.0,20.0,10.0,5.0,custom,4.0,80.0,gelu,5.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,causal,False,0.0,0.1885618083164127,True,False,standard,False,3.0,False,,9600.0,False,False,True,torch.float32,,,False,10000.0,False,False
9
+ 3,Returns the fraction of 'x' in the input up to the i-th position for all i.,5.0,4.0,0.05,0.001,False,1.0,1.0,10.0,2000.0,gelu,0.1,,2.0,12.0,5.0,3.0,custom,4.0,48.0,gelu,6.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,causal,False,0.0,0.22188007849009167,True,False,standard,False,1.0,False,,3456.0,False,False,True,torch.float32,,,False,10000.0,False,False
10
+ 33,Checks if each token's length is odd or even.,10.0,4.0,0.05,0.001,False,1.0,1.0,0.4,2000.0,gelu,0.1,,2.0,4.0,10.0,1.0,custom,4.0,16.0,gelu,10.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,causal,False,0.0,0.17457431218879393,True,False,standard,False,2.0,False,,384.0,False,False,True,torch.float32,,,False,10000.0,False,False
11
+ 34,Calculate the ratio of vowels to consonants in each word.,10.0,4.0,0.05,0.001,False,1.0,1.0,0.4,2000.0,gelu,0.1,,2.0,4.0,10.0,1.0,custom,4.0,16.0,gelu,10.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,causal,False,0.0,0.16329931618554522,True,False,standard,False,5.0,False,,384.0,False,False,True,torch.float32,,,False,10000.0,False,False
12
+ 35,Alternates capitalization of each character in words.,10.0,4.0,0.05,0.001,False,1.0,1.0,0.4,2000.0,gelu,0.1,,2.0,4.0,10.0,1.0,custom,4.0,16.0,gelu,10.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,causal,False,0.0,0.1539600717839002,True,False,standard,False,8.0,False,,384.0,False,False,True,torch.float32,,,False,10000.0,False,False
13
+ 36,"Classifies each token as 'positive', 'negative', or 'neutral' based on emojis.",10.0,4.0,0.05,0.001,False,1.0,1.0,10.0,2000.0,gelu,0.1,,2.0,4.0,10.0,1.0,custom,4.0,16.0,gelu,5.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cuda,1.0,causal,False,0.0,0.19402850002906638,True,False,standard,False,3.0,False,,384.0,False,False,True,torch.float32,,,False,10000.0,False,False
14
+ 37,Reverses each word in the sequence except for specified exclusions.,10.0,4.0,0.05,0.001,False,1.0,1.0,0.4,2000.0,gelu,0.1,,2.0,4.0,10.0,1.0,custom,4.0,16.0,gelu,10.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,causal,False,0.0,0.1539600717839002,True,False,standard,False,8.0,False,,384.0,False,False,True,torch.float32,,,False,10000.0,False,False
15
+ 38,Checks if tokens alternate between two types.,10.0,4.0,0.05,0.001,False,1.0,1.0,0.4,2000.0,gelu,0.1,,2.0,20.0,10.0,5.0,custom,4.0,80.0,gelu,5.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,causal,False,0.0,0.1539600717839002,True,False,standard,False,2.0,False,,9600.0,False,False,True,torch.float32,,,False,10000.0,False,False
16
+ 4,Return fraction of previous open tokens minus the fraction of close tokens.,10.0,4.0,0.05,0.001,False,1.0,1.0,0.4,2000.0,gelu,0.1,,2.0,20.0,10.0,5.0,custom,4.0,80.0,gelu,7.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,causal,False,0.0,0.17056057308448835,True,False,standard,False,1.0,False,,9600.0,False,False,True,torch.float32,,,False,10000.0,False,False
17
+ 8,Fills gaps between tokens with a specified filler.,10.0,4.0,0.05,0.01,False,1.0,1.0,0.4,500.0,gelu,1.0,,2.0,20.0,10.0,5.0,custom,4.0,80.0,gelu,10.0,1e-05,True,True,True,True,False,False,,False,,,,,,,gpt2,,cpu,1.0,causal,False,0.0,0.13333333333333333,True,False,standard,False,8.0,False,,9600.0,False,False,True,torch.float32,,,False,10000.0,False,False
18
+ ioi,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,
19
+ ioi_next_token,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,
benchmark_cases_metadata.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01f473326ab84033784efd049808e3f0395bf577e23311a60f9b05eff32f5ca3
3
+ size 49942
benchmark_metadata.json ADDED
@@ -0,0 +1,1362 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "InterpBench",
3
+ "version": "1.0.0",
4
+ "description": "A benchmark of transformers with known circuits for evaluating mechanistic interpretability techniques.",
5
+ "cases": [
6
+ {
7
+ "case_id": "11",
8
+ "files": [
9
+ "edges.pkl",
10
+ "ll_model_510.pth",
11
+ "ll_model_cfg_510.pkl",
12
+ "meta_510.json"
13
+ ],
14
+ "task_description": "Counts the number of words in a sequence based on their length.",
15
+ "vocab": [
16
+ "J",
17
+ "oCLrZaW",
18
+ "no",
19
+ "poiVg",
20
+ "V",
21
+ "b",
22
+ "LB",
23
+ "TPSI"
24
+ ],
25
+ "max_seq_len": 10,
26
+ "min_seq_len": 4,
27
+ "transformer_cfg": {
28
+ "n_layers": 2,
29
+ "d_model": 12,
30
+ "n_ctx": 10,
31
+ "d_head": 3,
32
+ "model_name": "custom",
33
+ "n_heads": 4,
34
+ "d_mlp": 48,
35
+ "act_fn": "gelu",
36
+ "d_vocab": 10,
37
+ "eps": 1e-05,
38
+ "use_attn_result": true,
39
+ "use_attn_scale": true,
40
+ "use_split_qkv_input": true,
41
+ "use_hook_mlp_in": true,
42
+ "use_attn_in": false,
43
+ "use_local_attn": false,
44
+ "original_architecture": null,
45
+ "from_checkpoint": false,
46
+ "checkpoint_index": null,
47
+ "checkpoint_label_type": null,
48
+ "checkpoint_value": null,
49
+ "tokenizer_name": null,
50
+ "window_size": null,
51
+ "attn_types": null,
52
+ "init_mode": "gpt2",
53
+ "normalization_type": null,
54
+ "device": "cpu",
55
+ "n_devices": 1,
56
+ "attention_dir": "causal",
57
+ "attn_only": false,
58
+ "seed": 0,
59
+ "initializer_range": 0.1460593486680443,
60
+ "init_weights": true,
61
+ "scale_attn_by_inverse_layer_idx": false,
62
+ "positional_embedding_type": "standard",
63
+ "final_rms": false,
64
+ "d_vocab_out": 5,
65
+ "parallel_attn_mlp": false,
66
+ "rotary_dim": null,
67
+ "n_params": 3456,
68
+ "use_hook_tokens": false,
69
+ "gated_mlp": false,
70
+ "default_prepend_bos": true,
71
+ "dtype": "torch.float32",
72
+ "tokenizer_prepends_bos": null,
73
+ "n_key_value_heads": null,
74
+ "post_embedding_ln": false,
75
+ "rotary_base": 10000,
76
+ "trust_remote_code": false,
77
+ "rotary_adjacent_pairs": false
78
+ },
79
+ "training_args": {
80
+ "atol": 0.05,
81
+ "lr": 0.01,
82
+ "use_single_loss": false,
83
+ "iit_weight": 1.0,
84
+ "behavior_weight": 1.0,
85
+ "strict_weight": 0.4,
86
+ "epochs": 500,
87
+ "act_fn": "gelu",
88
+ "clip_grad_norm": 1.0,
89
+ "lr_scheduler": ""
90
+ }
91
+ },
92
+ {
93
+ "case_id": "13",
94
+ "files": [
95
+ "edges.pkl",
96
+ "ll_model_510.pth",
97
+ "ll_model_cfg_510.pkl",
98
+ "meta_510.json"
99
+ ],
100
+ "task_description": "Analyzes the trend (increasing, decreasing, constant) of numeric tokens.",
101
+ "vocab": [
102
+ 0,
103
+ 1,
104
+ 2
105
+ ],
106
+ "max_seq_len": 10,
107
+ "min_seq_len": 4,
108
+ "transformer_cfg": {
109
+ "n_layers": 2,
110
+ "d_model": 20,
111
+ "n_ctx": 10,
112
+ "d_head": 5,
113
+ "model_name": "custom",
114
+ "n_heads": 4,
115
+ "d_mlp": 80,
116
+ "act_fn": "gelu",
117
+ "d_vocab": 5,
118
+ "eps": 1e-05,
119
+ "use_attn_result": true,
120
+ "use_attn_scale": true,
121
+ "use_split_qkv_input": true,
122
+ "use_hook_mlp_in": true,
123
+ "use_attn_in": false,
124
+ "use_local_attn": false,
125
+ "original_architecture": null,
126
+ "from_checkpoint": false,
127
+ "checkpoint_index": null,
128
+ "checkpoint_label_type": null,
129
+ "checkpoint_value": null,
130
+ "tokenizer_name": null,
131
+ "window_size": null,
132
+ "attn_types": null,
133
+ "init_mode": "gpt2",
134
+ "normalization_type": null,
135
+ "device": "cpu",
136
+ "n_devices": 1,
137
+ "attention_dir": "bidirectional",
138
+ "attn_only": false,
139
+ "seed": 0,
140
+ "initializer_range": 0.1460593486680443,
141
+ "init_weights": true,
142
+ "scale_attn_by_inverse_layer_idx": false,
143
+ "positional_embedding_type": "standard",
144
+ "final_rms": false,
145
+ "d_vocab_out": 3,
146
+ "parallel_attn_mlp": false,
147
+ "rotary_dim": null,
148
+ "n_params": 9600,
149
+ "use_hook_tokens": false,
150
+ "gated_mlp": false,
151
+ "default_prepend_bos": true,
152
+ "dtype": "torch.float32",
153
+ "tokenizer_prepends_bos": null,
154
+ "n_key_value_heads": null,
155
+ "post_embedding_ln": false,
156
+ "rotary_base": 10000,
157
+ "trust_remote_code": false,
158
+ "rotary_adjacent_pairs": false
159
+ },
160
+ "training_args": {
161
+ "atol": 0.05,
162
+ "lr": 0.01,
163
+ "use_single_loss": false,
164
+ "iit_weight": 1.0,
165
+ "behavior_weight": 1.0,
166
+ "strict_weight": 0.4,
167
+ "epochs": 500,
168
+ "act_fn": "gelu",
169
+ "clip_grad_norm": 1.0,
170
+ "lr_scheduler": ""
171
+ }
172
+ },
173
+ {
174
+ "case_id": "18",
175
+ "files": [
176
+ "edges.pkl",
177
+ "ll_model_510.pth",
178
+ "ll_model_cfg_510.pkl",
179
+ "meta_510.json"
180
+ ],
181
+ "task_description": "Classify each token based on its frequency as 'rare', 'common', or 'frequent'.",
182
+ "vocab": [
183
+ "c",
184
+ "e",
185
+ "b",
186
+ "d",
187
+ "a"
188
+ ],
189
+ "max_seq_len": 10,
190
+ "min_seq_len": 4,
191
+ "transformer_cfg": {
192
+ "n_layers": 2,
193
+ "d_model": 12,
194
+ "n_ctx": 10,
195
+ "d_head": 3,
196
+ "model_name": "custom",
197
+ "n_heads": 4,
198
+ "d_mlp": 48,
199
+ "act_fn": "gelu",
200
+ "d_vocab": 7,
201
+ "eps": 1e-05,
202
+ "use_attn_result": true,
203
+ "use_attn_scale": true,
204
+ "use_split_qkv_input": true,
205
+ "use_hook_mlp_in": true,
206
+ "use_attn_in": false,
207
+ "use_local_attn": false,
208
+ "original_architecture": null,
209
+ "from_checkpoint": false,
210
+ "checkpoint_index": null,
211
+ "checkpoint_label_type": null,
212
+ "checkpoint_value": null,
213
+ "tokenizer_name": null,
214
+ "window_size": null,
215
+ "attn_types": null,
216
+ "init_mode": "gpt2",
217
+ "normalization_type": null,
218
+ "device": "cpu",
219
+ "n_devices": 1,
220
+ "attention_dir": "bidirectional",
221
+ "attn_only": false,
222
+ "seed": 0,
223
+ "initializer_range": 0.12344267996967354,
224
+ "init_weights": true,
225
+ "scale_attn_by_inverse_layer_idx": false,
226
+ "positional_embedding_type": "standard",
227
+ "final_rms": false,
228
+ "d_vocab_out": 3,
229
+ "parallel_attn_mlp": false,
230
+ "rotary_dim": null,
231
+ "n_params": 3456,
232
+ "use_hook_tokens": false,
233
+ "gated_mlp": false,
234
+ "default_prepend_bos": true,
235
+ "dtype": "torch.float32",
236
+ "tokenizer_prepends_bos": null,
237
+ "n_key_value_heads": null,
238
+ "post_embedding_ln": false,
239
+ "rotary_base": 10000,
240
+ "trust_remote_code": false,
241
+ "rotary_adjacent_pairs": false
242
+ },
243
+ "training_args": {
244
+ "atol": 0.05,
245
+ "lr": 0.001,
246
+ "use_single_loss": false,
247
+ "iit_weight": 1.0,
248
+ "behavior_weight": 1.0,
249
+ "strict_weight": 0.4,
250
+ "epochs": 2000,
251
+ "act_fn": "gelu",
252
+ "clip_grad_norm": 0.1,
253
+ "lr_scheduler": ""
254
+ }
255
+ },
256
+ {
257
+ "case_id": "19",
258
+ "files": [
259
+ "edges.pkl",
260
+ "ll_model_510.pth",
261
+ "ll_model_cfg_510.pkl",
262
+ "meta_510.json"
263
+ ],
264
+ "task_description": "Removes consecutive duplicate tokens from a sequence.",
265
+ "vocab": [
266
+ "b",
267
+ "a",
268
+ "c"
269
+ ],
270
+ "max_seq_len": 15,
271
+ "min_seq_len": 4,
272
+ "transformer_cfg": {
273
+ "n_layers": 2,
274
+ "d_model": 32,
275
+ "n_ctx": 15,
276
+ "d_head": 8,
277
+ "model_name": "custom",
278
+ "n_heads": 4,
279
+ "d_mlp": 128,
280
+ "act_fn": "gelu",
281
+ "d_vocab": 5,
282
+ "eps": 1e-05,
283
+ "use_attn_result": true,
284
+ "use_attn_scale": true,
285
+ "use_split_qkv_input": true,
286
+ "use_hook_mlp_in": true,
287
+ "use_attn_in": false,
288
+ "use_local_attn": false,
289
+ "original_architecture": null,
290
+ "from_checkpoint": false,
291
+ "checkpoint_index": null,
292
+ "checkpoint_label_type": null,
293
+ "checkpoint_value": null,
294
+ "tokenizer_name": null,
295
+ "window_size": null,
296
+ "attn_types": null,
297
+ "init_mode": "gpt2",
298
+ "normalization_type": null,
299
+ "device": "cpu",
300
+ "n_devices": 1,
301
+ "attention_dir": "causal",
302
+ "attn_only": false,
303
+ "seed": 0,
304
+ "initializer_range": 0.15689290811054724,
305
+ "init_weights": true,
306
+ "scale_attn_by_inverse_layer_idx": false,
307
+ "positional_embedding_type": "standard",
308
+ "final_rms": false,
309
+ "d_vocab_out": 3,
310
+ "parallel_attn_mlp": false,
311
+ "rotary_dim": null,
312
+ "n_params": 24576,
313
+ "use_hook_tokens": false,
314
+ "gated_mlp": false,
315
+ "default_prepend_bos": true,
316
+ "dtype": "torch.float32",
317
+ "tokenizer_prepends_bos": null,
318
+ "n_key_value_heads": null,
319
+ "post_embedding_ln": false,
320
+ "rotary_base": 10000,
321
+ "trust_remote_code": false,
322
+ "rotary_adjacent_pairs": false
323
+ },
324
+ "training_args": {
325
+ "atol": 0.05,
326
+ "lr": 0.001,
327
+ "use_single_loss": false,
328
+ "iit_weight": 1.0,
329
+ "behavior_weight": 1.0,
330
+ "strict_weight": 0.4,
331
+ "epochs": 2000,
332
+ "act_fn": "gelu",
333
+ "clip_grad_norm": 0.1,
334
+ "lr_scheduler": ""
335
+ }
336
+ },
337
+ {
338
+ "case_id": "20",
339
+ "files": [
340
+ "edges.pkl",
341
+ "ll_model_1110.pth",
342
+ "ll_model_cfg_1110.pkl",
343
+ "meta_1110.json"
344
+ ],
345
+ "task_description": "Detect spam messages based on appearance of spam keywords.",
346
+ "vocab": [
347
+ "J",
348
+ "spam",
349
+ "offer",
350
+ "click",
351
+ "oCLrZaW",
352
+ "no",
353
+ "poiVg",
354
+ "V",
355
+ "b",
356
+ "LB",
357
+ "now",
358
+ "TPSI"
359
+ ],
360
+ "max_seq_len": 10,
361
+ "min_seq_len": 4,
362
+ "transformer_cfg": {
363
+ "n_layers": 2,
364
+ "d_model": 4,
365
+ "n_ctx": 10,
366
+ "d_head": 1,
367
+ "model_name": "custom",
368
+ "n_heads": 4,
369
+ "d_mlp": 16,
370
+ "act_fn": "gelu",
371
+ "d_vocab": 14,
372
+ "eps": 1e-05,
373
+ "use_attn_result": true,
374
+ "use_attn_scale": true,
375
+ "use_split_qkv_input": true,
376
+ "use_hook_mlp_in": true,
377
+ "use_attn_in": false,
378
+ "use_local_attn": false,
379
+ "original_architecture": null,
380
+ "from_checkpoint": false,
381
+ "checkpoint_index": null,
382
+ "checkpoint_label_type": null,
383
+ "checkpoint_value": null,
384
+ "tokenizer_name": null,
385
+ "window_size": null,
386
+ "attn_types": null,
387
+ "init_mode": "gpt2",
388
+ "normalization_type": null,
389
+ "device": "cuda",
390
+ "n_devices": 1,
391
+ "attention_dir": "causal",
392
+ "attn_only": false,
393
+ "seed": 0,
394
+ "initializer_range": 0.16,
395
+ "init_weights": true,
396
+ "scale_attn_by_inverse_layer_idx": false,
397
+ "positional_embedding_type": "standard",
398
+ "final_rms": false,
399
+ "d_vocab_out": 2,
400
+ "parallel_attn_mlp": false,
401
+ "rotary_dim": null,
402
+ "n_params": 384,
403
+ "use_hook_tokens": false,
404
+ "gated_mlp": false,
405
+ "default_prepend_bos": true,
406
+ "dtype": "torch.float32",
407
+ "tokenizer_prepends_bos": null,
408
+ "n_key_value_heads": null,
409
+ "post_embedding_ln": false,
410
+ "rotary_base": 10000,
411
+ "trust_remote_code": false,
412
+ "rotary_adjacent_pairs": false
413
+ },
414
+ "training_args": {
415
+ "atol": 0.05,
416
+ "lr": 0.001,
417
+ "use_single_loss": false,
418
+ "iit_weight": 1.0,
419
+ "behavior_weight": 1.0,
420
+ "strict_weight": 1.0,
421
+ "epochs": 2000,
422
+ "act_fn": "gelu",
423
+ "clip_grad_norm": 0.1,
424
+ "lr_scheduler": ""
425
+ }
426
+ },
427
+ {
428
+ "case_id": "21",
429
+ "files": [
430
+ "edges.pkl",
431
+ "ll_model_510.pth",
432
+ "ll_model_cfg_510.pkl",
433
+ "meta_510.json"
434
+ ],
435
+ "task_description": "Extract unique tokens from a string",
436
+ "vocab": [
437
+ "b",
438
+ "a",
439
+ "c"
440
+ ],
441
+ "max_seq_len": 10,
442
+ "min_seq_len": 4,
443
+ "transformer_cfg": {
444
+ "n_layers": 2,
445
+ "d_model": 20,
446
+ "n_ctx": 10,
447
+ "d_head": 5,
448
+ "model_name": "custom",
449
+ "n_heads": 4,
450
+ "d_mlp": 80,
451
+ "act_fn": "gelu",
452
+ "d_vocab": 5,
453
+ "eps": 1e-05,
454
+ "use_attn_result": true,
455
+ "use_attn_scale": true,
456
+ "use_split_qkv_input": true,
457
+ "use_hook_mlp_in": true,
458
+ "use_attn_in": false,
459
+ "use_local_attn": false,
460
+ "original_architecture": null,
461
+ "from_checkpoint": false,
462
+ "checkpoint_index": null,
463
+ "checkpoint_label_type": null,
464
+ "checkpoint_value": null,
465
+ "tokenizer_name": null,
466
+ "window_size": null,
467
+ "attn_types": null,
468
+ "init_mode": "gpt2",
469
+ "normalization_type": null,
470
+ "device": "cpu",
471
+ "n_devices": 1,
472
+ "attention_dir": "causal",
473
+ "attn_only": false,
474
+ "seed": 0,
475
+ "initializer_range": 0.1885618083164127,
476
+ "init_weights": true,
477
+ "scale_attn_by_inverse_layer_idx": false,
478
+ "positional_embedding_type": "standard",
479
+ "final_rms": false,
480
+ "d_vocab_out": 3,
481
+ "parallel_attn_mlp": false,
482
+ "rotary_dim": null,
483
+ "n_params": 9600,
484
+ "use_hook_tokens": false,
485
+ "gated_mlp": false,
486
+ "default_prepend_bos": true,
487
+ "dtype": "torch.float32",
488
+ "tokenizer_prepends_bos": null,
489
+ "n_key_value_heads": null,
490
+ "post_embedding_ln": false,
491
+ "rotary_base": 10000,
492
+ "trust_remote_code": false,
493
+ "rotary_adjacent_pairs": false
494
+ },
495
+ "training_args": {
496
+ "atol": 0.05,
497
+ "lr": 0.01,
498
+ "use_single_loss": false,
499
+ "iit_weight": 1.0,
500
+ "behavior_weight": 1.0,
501
+ "strict_weight": 0.4,
502
+ "epochs": 500,
503
+ "act_fn": "gelu",
504
+ "clip_grad_norm": 1.0,
505
+ "lr_scheduler": ""
506
+ }
507
+ },
508
+ {
509
+ "case_id": "24",
510
+ "files": [
511
+ "edges.pkl",
512
+ "ll_model_510.pth",
513
+ "ll_model_cfg_510.pkl",
514
+ "meta_510.json"
515
+ ],
516
+ "task_description": "Identifies the first occurrence of each token in a sequence.",
517
+ "vocab": [
518
+ "b",
519
+ "a",
520
+ "c"
521
+ ],
522
+ "max_seq_len": 10,
523
+ "min_seq_len": 4,
524
+ "transformer_cfg": {
525
+ "n_layers": 2,
526
+ "d_model": 20,
527
+ "n_ctx": 10,
528
+ "d_head": 5,
529
+ "model_name": "custom",
530
+ "n_heads": 4,
531
+ "d_mlp": 80,
532
+ "act_fn": "gelu",
533
+ "d_vocab": 5,
534
+ "eps": 1e-05,
535
+ "use_attn_result": true,
536
+ "use_attn_scale": true,
537
+ "use_split_qkv_input": true,
538
+ "use_hook_mlp_in": true,
539
+ "use_attn_in": false,
540
+ "use_local_attn": false,
541
+ "original_architecture": null,
542
+ "from_checkpoint": false,
543
+ "checkpoint_index": null,
544
+ "checkpoint_label_type": null,
545
+ "checkpoint_value": null,
546
+ "tokenizer_name": null,
547
+ "window_size": null,
548
+ "attn_types": null,
549
+ "init_mode": "gpt2",
550
+ "normalization_type": null,
551
+ "device": "cpu",
552
+ "n_devices": 1,
553
+ "attention_dir": "causal",
554
+ "attn_only": false,
555
+ "seed": 0,
556
+ "initializer_range": 0.1885618083164127,
557
+ "init_weights": true,
558
+ "scale_attn_by_inverse_layer_idx": false,
559
+ "positional_embedding_type": "standard",
560
+ "final_rms": false,
561
+ "d_vocab_out": 3,
562
+ "parallel_attn_mlp": false,
563
+ "rotary_dim": null,
564
+ "n_params": 9600,
565
+ "use_hook_tokens": false,
566
+ "gated_mlp": false,
567
+ "default_prepend_bos": true,
568
+ "dtype": "torch.float32",
569
+ "tokenizer_prepends_bos": null,
570
+ "n_key_value_heads": null,
571
+ "post_embedding_ln": false,
572
+ "rotary_base": 10000,
573
+ "trust_remote_code": false,
574
+ "rotary_adjacent_pairs": false
575
+ },
576
+ "training_args": {
577
+ "atol": 0.05,
578
+ "lr": 0.01,
579
+ "use_single_loss": false,
580
+ "iit_weight": 1.0,
581
+ "behavior_weight": 1.0,
582
+ "strict_weight": 0.4,
583
+ "epochs": 500,
584
+ "act_fn": "gelu",
585
+ "clip_grad_norm": 1.0,
586
+ "lr_scheduler": ""
587
+ }
588
+ },
589
+ {
590
+ "case_id": "3",
591
+ "files": [
592
+ "edges.pkl",
593
+ "ll_model_10110.pth",
594
+ "ll_model_cfg_10110.pkl",
595
+ "meta_10110.json"
596
+ ],
597
+ "task_description": "Returns the fraction of 'x' in the input up to the i-th position for all i.",
598
+ "vocab": [
599
+ "x",
600
+ "b",
601
+ "a",
602
+ "c"
603
+ ],
604
+ "max_seq_len": 5,
605
+ "min_seq_len": 4,
606
+ "transformer_cfg": {
607
+ "n_layers": 2,
608
+ "d_model": 12,
609
+ "n_ctx": 5,
610
+ "d_head": 3,
611
+ "model_name": "custom",
612
+ "n_heads": 4,
613
+ "d_mlp": 48,
614
+ "act_fn": "gelu",
615
+ "d_vocab": 6,
616
+ "eps": 1e-05,
617
+ "use_attn_result": true,
618
+ "use_attn_scale": true,
619
+ "use_split_qkv_input": true,
620
+ "use_hook_mlp_in": true,
621
+ "use_attn_in": false,
622
+ "use_local_attn": false,
623
+ "original_architecture": null,
624
+ "from_checkpoint": false,
625
+ "checkpoint_index": null,
626
+ "checkpoint_label_type": null,
627
+ "checkpoint_value": null,
628
+ "tokenizer_name": null,
629
+ "window_size": null,
630
+ "attn_types": null,
631
+ "init_mode": "gpt2",
632
+ "normalization_type": null,
633
+ "device": "cpu",
634
+ "n_devices": 1,
635
+ "attention_dir": "causal",
636
+ "attn_only": false,
637
+ "seed": 0,
638
+ "initializer_range": 0.22188007849009167,
639
+ "init_weights": true,
640
+ "scale_attn_by_inverse_layer_idx": false,
641
+ "positional_embedding_type": "standard",
642
+ "final_rms": false,
643
+ "d_vocab_out": 1,
644
+ "parallel_attn_mlp": false,
645
+ "rotary_dim": null,
646
+ "n_params": 3456,
647
+ "use_hook_tokens": false,
648
+ "gated_mlp": false,
649
+ "default_prepend_bos": true,
650
+ "dtype": "torch.float32",
651
+ "tokenizer_prepends_bos": null,
652
+ "n_key_value_heads": null,
653
+ "post_embedding_ln": false,
654
+ "rotary_base": 10000,
655
+ "trust_remote_code": false,
656
+ "rotary_adjacent_pairs": false
657
+ },
658
+ "training_args": {
659
+ "atol": 0.05,
660
+ "lr": 0.001,
661
+ "use_single_loss": false,
662
+ "iit_weight": 1.0,
663
+ "behavior_weight": 1.0,
664
+ "strict_weight": 10.0,
665
+ "epochs": 2000,
666
+ "act_fn": "gelu",
667
+ "clip_grad_norm": 0.1,
668
+ "lr_scheduler": ""
669
+ }
670
+ },
671
+ {
672
+ "case_id": "33",
673
+ "files": [
674
+ "edges.pkl",
675
+ "ll_model_510.pth",
676
+ "ll_model_cfg_510.pkl",
677
+ "meta_510.json"
678
+ ],
679
+ "task_description": "Checks if each token's length is odd or even.",
680
+ "vocab": [
681
+ "J",
682
+ "oCLrZaW",
683
+ "no",
684
+ "poiVg",
685
+ "V",
686
+ "b",
687
+ "LB",
688
+ "TPSI"
689
+ ],
690
+ "max_seq_len": 10,
691
+ "min_seq_len": 4,
692
+ "transformer_cfg": {
693
+ "n_layers": 2,
694
+ "d_model": 4,
695
+ "n_ctx": 10,
696
+ "d_head": 1,
697
+ "model_name": "custom",
698
+ "n_heads": 4,
699
+ "d_mlp": 16,
700
+ "act_fn": "gelu",
701
+ "d_vocab": 10,
702
+ "eps": 1e-05,
703
+ "use_attn_result": true,
704
+ "use_attn_scale": true,
705
+ "use_split_qkv_input": true,
706
+ "use_hook_mlp_in": true,
707
+ "use_attn_in": false,
708
+ "use_local_attn": false,
709
+ "original_architecture": null,
710
+ "from_checkpoint": false,
711
+ "checkpoint_index": null,
712
+ "checkpoint_label_type": null,
713
+ "checkpoint_value": null,
714
+ "tokenizer_name": null,
715
+ "window_size": null,
716
+ "attn_types": null,
717
+ "init_mode": "gpt2",
718
+ "normalization_type": null,
719
+ "device": "cpu",
720
+ "n_devices": 1,
721
+ "attention_dir": "causal",
722
+ "attn_only": false,
723
+ "seed": 0,
724
+ "initializer_range": 0.17457431218879393,
725
+ "init_weights": true,
726
+ "scale_attn_by_inverse_layer_idx": false,
727
+ "positional_embedding_type": "standard",
728
+ "final_rms": false,
729
+ "d_vocab_out": 2,
730
+ "parallel_attn_mlp": false,
731
+ "rotary_dim": null,
732
+ "n_params": 384,
733
+ "use_hook_tokens": false,
734
+ "gated_mlp": false,
735
+ "default_prepend_bos": true,
736
+ "dtype": "torch.float32",
737
+ "tokenizer_prepends_bos": null,
738
+ "n_key_value_heads": null,
739
+ "post_embedding_ln": false,
740
+ "rotary_base": 10000,
741
+ "trust_remote_code": false,
742
+ "rotary_adjacent_pairs": false
743
+ },
744
+ "training_args": {
745
+ "atol": 0.05,
746
+ "lr": 0.001,
747
+ "use_single_loss": false,
748
+ "iit_weight": 1.0,
749
+ "behavior_weight": 1.0,
750
+ "strict_weight": 0.4,
751
+ "epochs": 2000,
752
+ "act_fn": "gelu",
753
+ "clip_grad_norm": 0.1,
754
+ "lr_scheduler": ""
755
+ }
756
+ },
757
+ {
758
+ "case_id": "34",
759
+ "files": [
760
+ "edges.pkl",
761
+ "ll_model_510.pth",
762
+ "ll_model_cfg_510.pkl",
763
+ "meta_510.json"
764
+ ],
765
+ "task_description": "Calculate the ratio of vowels to consonants in each word.",
766
+ "vocab": [
767
+ "J",
768
+ "oCLrZaW",
769
+ "no",
770
+ "poiVg",
771
+ "V",
772
+ "b",
773
+ "LB",
774
+ "TPSI"
775
+ ],
776
+ "max_seq_len": 10,
777
+ "min_seq_len": 4,
778
+ "transformer_cfg": {
779
+ "n_layers": 2,
780
+ "d_model": 4,
781
+ "n_ctx": 10,
782
+ "d_head": 1,
783
+ "model_name": "custom",
784
+ "n_heads": 4,
785
+ "d_mlp": 16,
786
+ "act_fn": "gelu",
787
+ "d_vocab": 10,
788
+ "eps": 1e-05,
789
+ "use_attn_result": true,
790
+ "use_attn_scale": true,
791
+ "use_split_qkv_input": true,
792
+ "use_hook_mlp_in": true,
793
+ "use_attn_in": false,
794
+ "use_local_attn": false,
795
+ "original_architecture": null,
796
+ "from_checkpoint": false,
797
+ "checkpoint_index": null,
798
+ "checkpoint_label_type": null,
799
+ "checkpoint_value": null,
800
+ "tokenizer_name": null,
801
+ "window_size": null,
802
+ "attn_types": null,
803
+ "init_mode": "gpt2",
804
+ "normalization_type": null,
805
+ "device": "cpu",
806
+ "n_devices": 1,
807
+ "attention_dir": "causal",
808
+ "attn_only": false,
809
+ "seed": 0,
810
+ "initializer_range": 0.16329931618554522,
811
+ "init_weights": true,
812
+ "scale_attn_by_inverse_layer_idx": false,
813
+ "positional_embedding_type": "standard",
814
+ "final_rms": false,
815
+ "d_vocab_out": 5,
816
+ "parallel_attn_mlp": false,
817
+ "rotary_dim": null,
818
+ "n_params": 384,
819
+ "use_hook_tokens": false,
820
+ "gated_mlp": false,
821
+ "default_prepend_bos": true,
822
+ "dtype": "torch.float32",
823
+ "tokenizer_prepends_bos": null,
824
+ "n_key_value_heads": null,
825
+ "post_embedding_ln": false,
826
+ "rotary_base": 10000,
827
+ "trust_remote_code": false,
828
+ "rotary_adjacent_pairs": false
829
+ },
830
+ "training_args": {
831
+ "atol": 0.05,
832
+ "lr": 0.001,
833
+ "use_single_loss": false,
834
+ "iit_weight": 1.0,
835
+ "behavior_weight": 1.0,
836
+ "strict_weight": 0.4,
837
+ "epochs": 2000,
838
+ "act_fn": "gelu",
839
+ "clip_grad_norm": 0.1,
840
+ "lr_scheduler": ""
841
+ }
842
+ },
843
+ {
844
+ "case_id": "35",
845
+ "files": [
846
+ "edges.pkl",
847
+ "ll_model_510.pth",
848
+ "ll_model_cfg_510.pkl",
849
+ "meta_510.json"
850
+ ],
851
+ "task_description": "Alternates capitalization of each character in words.",
852
+ "vocab": [
853
+ "J",
854
+ "oCLrZaW",
855
+ "no",
856
+ "poiVg",
857
+ "V",
858
+ "b",
859
+ "LB",
860
+ "TPSI"
861
+ ],
862
+ "max_seq_len": 10,
863
+ "min_seq_len": 4,
864
+ "transformer_cfg": {
865
+ "n_layers": 2,
866
+ "d_model": 4,
867
+ "n_ctx": 10,
868
+ "d_head": 1,
869
+ "model_name": "custom",
870
+ "n_heads": 4,
871
+ "d_mlp": 16,
872
+ "act_fn": "gelu",
873
+ "d_vocab": 10,
874
+ "eps": 1e-05,
875
+ "use_attn_result": true,
876
+ "use_attn_scale": true,
877
+ "use_split_qkv_input": true,
878
+ "use_hook_mlp_in": true,
879
+ "use_attn_in": false,
880
+ "use_local_attn": false,
881
+ "original_architecture": null,
882
+ "from_checkpoint": false,
883
+ "checkpoint_index": null,
884
+ "checkpoint_label_type": null,
885
+ "checkpoint_value": null,
886
+ "tokenizer_name": null,
887
+ "window_size": null,
888
+ "attn_types": null,
889
+ "init_mode": "gpt2",
890
+ "normalization_type": null,
891
+ "device": "cpu",
892
+ "n_devices": 1,
893
+ "attention_dir": "causal",
894
+ "attn_only": false,
895
+ "seed": 0,
896
+ "initializer_range": 0.1539600717839002,
897
+ "init_weights": true,
898
+ "scale_attn_by_inverse_layer_idx": false,
899
+ "positional_embedding_type": "standard",
900
+ "final_rms": false,
901
+ "d_vocab_out": 8,
902
+ "parallel_attn_mlp": false,
903
+ "rotary_dim": null,
904
+ "n_params": 384,
905
+ "use_hook_tokens": false,
906
+ "gated_mlp": false,
907
+ "default_prepend_bos": true,
908
+ "dtype": "torch.float32",
909
+ "tokenizer_prepends_bos": null,
910
+ "n_key_value_heads": null,
911
+ "post_embedding_ln": false,
912
+ "rotary_base": 10000,
913
+ "trust_remote_code": false,
914
+ "rotary_adjacent_pairs": false
915
+ },
916
+ "training_args": {
917
+ "atol": 0.05,
918
+ "lr": 0.001,
919
+ "use_single_loss": false,
920
+ "iit_weight": 1.0,
921
+ "behavior_weight": 1.0,
922
+ "strict_weight": 0.4,
923
+ "epochs": 2000,
924
+ "act_fn": "gelu",
925
+ "clip_grad_norm": 0.1,
926
+ "lr_scheduler": ""
927
+ }
928
+ },
929
+ {
930
+ "case_id": "36",
931
+ "files": [
932
+ "edges.pkl",
933
+ "ll_model_10110.pth",
934
+ "ll_model_cfg_10110.pkl",
935
+ "meta_10110.json"
936
+ ],
937
+ "task_description": "Classifies each token as 'positive', 'negative', or 'neutral' based on emojis.",
938
+ "vocab": [
939
+ "\ud83d\ude22",
940
+ "\ud83d\udcd8",
941
+ "\ud83d\ude0a"
942
+ ],
943
+ "max_seq_len": 10,
944
+ "min_seq_len": 4,
945
+ "transformer_cfg": {
946
+ "n_layers": 2,
947
+ "d_model": 4,
948
+ "n_ctx": 10,
949
+ "d_head": 1,
950
+ "model_name": "custom",
951
+ "n_heads": 4,
952
+ "d_mlp": 16,
953
+ "act_fn": "gelu",
954
+ "d_vocab": 5,
955
+ "eps": 1e-05,
956
+ "use_attn_result": true,
957
+ "use_attn_scale": true,
958
+ "use_split_qkv_input": true,
959
+ "use_hook_mlp_in": true,
960
+ "use_attn_in": false,
961
+ "use_local_attn": false,
962
+ "original_architecture": null,
963
+ "from_checkpoint": false,
964
+ "checkpoint_index": null,
965
+ "checkpoint_label_type": null,
966
+ "checkpoint_value": null,
967
+ "tokenizer_name": null,
968
+ "window_size": null,
969
+ "attn_types": null,
970
+ "init_mode": "gpt2",
971
+ "normalization_type": null,
972
+ "device": "cuda",
973
+ "n_devices": 1,
974
+ "attention_dir": "causal",
975
+ "attn_only": false,
976
+ "seed": 0,
977
+ "initializer_range": 0.19402850002906638,
978
+ "init_weights": true,
979
+ "scale_attn_by_inverse_layer_idx": false,
980
+ "positional_embedding_type": "standard",
981
+ "final_rms": false,
982
+ "d_vocab_out": 3,
983
+ "parallel_attn_mlp": false,
984
+ "rotary_dim": null,
985
+ "n_params": 384,
986
+ "use_hook_tokens": false,
987
+ "gated_mlp": false,
988
+ "default_prepend_bos": true,
989
+ "dtype": "torch.float32",
990
+ "tokenizer_prepends_bos": null,
991
+ "n_key_value_heads": null,
992
+ "post_embedding_ln": false,
993
+ "rotary_base": 10000,
994
+ "trust_remote_code": false,
995
+ "rotary_adjacent_pairs": false
996
+ },
997
+ "training_args": {
998
+ "atol": 0.05,
999
+ "lr": 0.001,
1000
+ "use_single_loss": false,
1001
+ "iit_weight": 1.0,
1002
+ "behavior_weight": 1.0,
1003
+ "strict_weight": 10.0,
1004
+ "epochs": 2000,
1005
+ "act_fn": "gelu",
1006
+ "clip_grad_norm": 0.1,
1007
+ "lr_scheduler": ""
1008
+ }
1009
+ },
1010
+ {
1011
+ "case_id": "37",
1012
+ "files": [
1013
+ "edges.pkl",
1014
+ "ll_model_510.pth",
1015
+ "ll_model_cfg_510.pkl",
1016
+ "meta_510.json"
1017
+ ],
1018
+ "task_description": "Reverses each word in the sequence except for specified exclusions.",
1019
+ "vocab": [
1020
+ "J",
1021
+ "oCLrZaW",
1022
+ "no",
1023
+ "poiVg",
1024
+ "V",
1025
+ "b",
1026
+ "LB",
1027
+ "TPSI"
1028
+ ],
1029
+ "max_seq_len": 10,
1030
+ "min_seq_len": 4,
1031
+ "transformer_cfg": {
1032
+ "n_layers": 2,
1033
+ "d_model": 4,
1034
+ "n_ctx": 10,
1035
+ "d_head": 1,
1036
+ "model_name": "custom",
1037
+ "n_heads": 4,
1038
+ "d_mlp": 16,
1039
+ "act_fn": "gelu",
1040
+ "d_vocab": 10,
1041
+ "eps": 1e-05,
1042
+ "use_attn_result": true,
1043
+ "use_attn_scale": true,
1044
+ "use_split_qkv_input": true,
1045
+ "use_hook_mlp_in": true,
1046
+ "use_attn_in": false,
1047
+ "use_local_attn": false,
1048
+ "original_architecture": null,
1049
+ "from_checkpoint": false,
1050
+ "checkpoint_index": null,
1051
+ "checkpoint_label_type": null,
1052
+ "checkpoint_value": null,
1053
+ "tokenizer_name": null,
1054
+ "window_size": null,
1055
+ "attn_types": null,
1056
+ "init_mode": "gpt2",
1057
+ "normalization_type": null,
1058
+ "device": "cpu",
1059
+ "n_devices": 1,
1060
+ "attention_dir": "causal",
1061
+ "attn_only": false,
1062
+ "seed": 0,
1063
+ "initializer_range": 0.1539600717839002,
1064
+ "init_weights": true,
1065
+ "scale_attn_by_inverse_layer_idx": false,
1066
+ "positional_embedding_type": "standard",
1067
+ "final_rms": false,
1068
+ "d_vocab_out": 8,
1069
+ "parallel_attn_mlp": false,
1070
+ "rotary_dim": null,
1071
+ "n_params": 384,
1072
+ "use_hook_tokens": false,
1073
+ "gated_mlp": false,
1074
+ "default_prepend_bos": true,
1075
+ "dtype": "torch.float32",
1076
+ "tokenizer_prepends_bos": null,
1077
+ "n_key_value_heads": null,
1078
+ "post_embedding_ln": false,
1079
+ "rotary_base": 10000,
1080
+ "trust_remote_code": false,
1081
+ "rotary_adjacent_pairs": false
1082
+ },
1083
+ "training_args": {
1084
+ "atol": 0.05,
1085
+ "lr": 0.001,
1086
+ "use_single_loss": false,
1087
+ "iit_weight": 1.0,
1088
+ "behavior_weight": 1.0,
1089
+ "strict_weight": 0.4,
1090
+ "epochs": 2000,
1091
+ "act_fn": "gelu",
1092
+ "clip_grad_norm": 0.1,
1093
+ "lr_scheduler": ""
1094
+ }
1095
+ },
1096
+ {
1097
+ "case_id": "38",
1098
+ "files": [
1099
+ "edges.pkl",
1100
+ "ll_model_510.pth",
1101
+ "ll_model_cfg_510.pkl",
1102
+ "meta_510.json"
1103
+ ],
1104
+ "task_description": "Checks if tokens alternate between two types.",
1105
+ "vocab": [
1106
+ "b",
1107
+ "a",
1108
+ "c"
1109
+ ],
1110
+ "max_seq_len": 10,
1111
+ "min_seq_len": 4,
1112
+ "transformer_cfg": {
1113
+ "n_layers": 2,
1114
+ "d_model": 20,
1115
+ "n_ctx": 10,
1116
+ "d_head": 5,
1117
+ "model_name": "custom",
1118
+ "n_heads": 4,
1119
+ "d_mlp": 80,
1120
+ "act_fn": "gelu",
1121
+ "d_vocab": 5,
1122
+ "eps": 1e-05,
1123
+ "use_attn_result": true,
1124
+ "use_attn_scale": true,
1125
+ "use_split_qkv_input": true,
1126
+ "use_hook_mlp_in": true,
1127
+ "use_attn_in": false,
1128
+ "use_local_attn": false,
1129
+ "original_architecture": null,
1130
+ "from_checkpoint": false,
1131
+ "checkpoint_index": null,
1132
+ "checkpoint_label_type": null,
1133
+ "checkpoint_value": null,
1134
+ "tokenizer_name": null,
1135
+ "window_size": null,
1136
+ "attn_types": null,
1137
+ "init_mode": "gpt2",
1138
+ "normalization_type": null,
1139
+ "device": "cpu",
1140
+ "n_devices": 1,
1141
+ "attention_dir": "causal",
1142
+ "attn_only": false,
1143
+ "seed": 0,
1144
+ "initializer_range": 0.1539600717839002,
1145
+ "init_weights": true,
1146
+ "scale_attn_by_inverse_layer_idx": false,
1147
+ "positional_embedding_type": "standard",
1148
+ "final_rms": false,
1149
+ "d_vocab_out": 2,
1150
+ "parallel_attn_mlp": false,
1151
+ "rotary_dim": null,
1152
+ "n_params": 9600,
1153
+ "use_hook_tokens": false,
1154
+ "gated_mlp": false,
1155
+ "default_prepend_bos": true,
1156
+ "dtype": "torch.float32",
1157
+ "tokenizer_prepends_bos": null,
1158
+ "n_key_value_heads": null,
1159
+ "post_embedding_ln": false,
1160
+ "rotary_base": 10000,
1161
+ "trust_remote_code": false,
1162
+ "rotary_adjacent_pairs": false
1163
+ },
1164
+ "training_args": {
1165
+ "atol": 0.05,
1166
+ "lr": 0.001,
1167
+ "use_single_loss": false,
1168
+ "iit_weight": 1.0,
1169
+ "behavior_weight": 1.0,
1170
+ "strict_weight": 0.4,
1171
+ "epochs": 2000,
1172
+ "act_fn": "gelu",
1173
+ "clip_grad_norm": 0.1,
1174
+ "lr_scheduler": ""
1175
+ }
1176
+ },
1177
+ {
1178
+ "case_id": "4",
1179
+ "files": [
1180
+ "edges.pkl",
1181
+ "ll_model_510.pth",
1182
+ "ll_model_cfg_510.pkl",
1183
+ "meta_510.json"
1184
+ ],
1185
+ "task_description": "Return fraction of previous open tokens minus the fraction of close tokens.",
1186
+ "vocab": [
1187
+ "b",
1188
+ "(",
1189
+ "c",
1190
+ ")",
1191
+ "a"
1192
+ ],
1193
+ "max_seq_len": 10,
1194
+ "min_seq_len": 4,
1195
+ "transformer_cfg": {
1196
+ "n_layers": 2,
1197
+ "d_model": 20,
1198
+ "n_ctx": 10,
1199
+ "d_head": 5,
1200
+ "model_name": "custom",
1201
+ "n_heads": 4,
1202
+ "d_mlp": 80,
1203
+ "act_fn": "gelu",
1204
+ "d_vocab": 7,
1205
+ "eps": 1e-05,
1206
+ "use_attn_result": true,
1207
+ "use_attn_scale": true,
1208
+ "use_split_qkv_input": true,
1209
+ "use_hook_mlp_in": true,
1210
+ "use_attn_in": false,
1211
+ "use_local_attn": false,
1212
+ "original_architecture": null,
1213
+ "from_checkpoint": false,
1214
+ "checkpoint_index": null,
1215
+ "checkpoint_label_type": null,
1216
+ "checkpoint_value": null,
1217
+ "tokenizer_name": null,
1218
+ "window_size": null,
1219
+ "attn_types": null,
1220
+ "init_mode": "gpt2",
1221
+ "normalization_type": null,
1222
+ "device": "cpu",
1223
+ "n_devices": 1,
1224
+ "attention_dir": "causal",
1225
+ "attn_only": false,
1226
+ "seed": 0,
1227
+ "initializer_range": 0.17056057308448835,
1228
+ "init_weights": true,
1229
+ "scale_attn_by_inverse_layer_idx": false,
1230
+ "positional_embedding_type": "standard",
1231
+ "final_rms": false,
1232
+ "d_vocab_out": 1,
1233
+ "parallel_attn_mlp": false,
1234
+ "rotary_dim": null,
1235
+ "n_params": 9600,
1236
+ "use_hook_tokens": false,
1237
+ "gated_mlp": false,
1238
+ "default_prepend_bos": true,
1239
+ "dtype": "torch.float32",
1240
+ "tokenizer_prepends_bos": null,
1241
+ "n_key_value_heads": null,
1242
+ "post_embedding_ln": false,
1243
+ "rotary_base": 10000,
1244
+ "trust_remote_code": false,
1245
+ "rotary_adjacent_pairs": false
1246
+ },
1247
+ "training_args": {
1248
+ "atol": 0.05,
1249
+ "lr": 0.001,
1250
+ "use_single_loss": false,
1251
+ "iit_weight": 1.0,
1252
+ "behavior_weight": 1.0,
1253
+ "strict_weight": 0.4,
1254
+ "epochs": 2000,
1255
+ "act_fn": "gelu",
1256
+ "clip_grad_norm": 0.1,
1257
+ "lr_scheduler": ""
1258
+ }
1259
+ },
1260
+ {
1261
+ "case_id": "8",
1262
+ "files": [
1263
+ "edges.pkl",
1264
+ "ll_model_510.pth",
1265
+ "ll_model_cfg_510.pkl",
1266
+ "meta_510.json"
1267
+ ],
1268
+ "task_description": "Fills gaps between tokens with a specified filler.",
1269
+ "vocab": [
1270
+ "J",
1271
+ "oCLrZaW",
1272
+ "no",
1273
+ "poiVg",
1274
+ "V",
1275
+ "b",
1276
+ "LB",
1277
+ "TPSI"
1278
+ ],
1279
+ "max_seq_len": 10,
1280
+ "min_seq_len": 4,
1281
+ "transformer_cfg": {
1282
+ "n_layers": 2,
1283
+ "d_model": 20,
1284
+ "n_ctx": 10,
1285
+ "d_head": 5,
1286
+ "model_name": "custom",
1287
+ "n_heads": 4,
1288
+ "d_mlp": 80,
1289
+ "act_fn": "gelu",
1290
+ "d_vocab": 10,
1291
+ "eps": 1e-05,
1292
+ "use_attn_result": true,
1293
+ "use_attn_scale": true,
1294
+ "use_split_qkv_input": true,
1295
+ "use_hook_mlp_in": true,
1296
+ "use_attn_in": false,
1297
+ "use_local_attn": false,
1298
+ "original_architecture": null,
1299
+ "from_checkpoint": false,
1300
+ "checkpoint_index": null,
1301
+ "checkpoint_label_type": null,
1302
+ "checkpoint_value": null,
1303
+ "tokenizer_name": null,
1304
+ "window_size": null,
1305
+ "attn_types": null,
1306
+ "init_mode": "gpt2",
1307
+ "normalization_type": null,
1308
+ "device": "cpu",
1309
+ "n_devices": 1,
1310
+ "attention_dir": "causal",
1311
+ "attn_only": false,
1312
+ "seed": 0,
1313
+ "initializer_range": 0.13333333333333333,
1314
+ "init_weights": true,
1315
+ "scale_attn_by_inverse_layer_idx": false,
1316
+ "positional_embedding_type": "standard",
1317
+ "final_rms": false,
1318
+ "d_vocab_out": 8,
1319
+ "parallel_attn_mlp": false,
1320
+ "rotary_dim": null,
1321
+ "n_params": 9600,
1322
+ "use_hook_tokens": false,
1323
+ "gated_mlp": false,
1324
+ "default_prepend_bos": true,
1325
+ "dtype": "torch.float32",
1326
+ "tokenizer_prepends_bos": null,
1327
+ "n_key_value_heads": null,
1328
+ "post_embedding_ln": false,
1329
+ "rotary_base": 10000,
1330
+ "trust_remote_code": false,
1331
+ "rotary_adjacent_pairs": false
1332
+ },
1333
+ "training_args": {
1334
+ "atol": 0.05,
1335
+ "lr": 0.01,
1336
+ "use_single_loss": false,
1337
+ "iit_weight": 1.0,
1338
+ "behavior_weight": 1.0,
1339
+ "strict_weight": 0.4,
1340
+ "epochs": 500,
1341
+ "act_fn": "gelu",
1342
+ "clip_grad_norm": 1.0,
1343
+ "lr_scheduler": ""
1344
+ }
1345
+ },
1346
+ {
1347
+ "case_id": "ioi",
1348
+ "files": [
1349
+ "corr_100_100_40.json",
1350
+ "ll_model_100_100_40.pth"
1351
+ ]
1352
+ },
1353
+ {
1354
+ "case_id": "ioi_next_token",
1355
+ "files": [
1356
+ "corr_100_100_40.json",
1357
+ "ll_model_100_100_40.pth",
1358
+ "training_args.json"
1359
+ ]
1360
+ }
1361
+ ]
1362
+ }