Dataset Viewer
Auto-converted to Parquet Duplicate
model_id
large_stringlengths
13
47
organization
large_stringlengths
4
14
model_type
large_stringlengths
3
16
schema_version
int64
3
3
analyzed_at
large_stringdate
2026-03-29 15:32:25
2026-04-01 23:41:04
config
large_stringlengths
145
207
techniques
large_stringlengths
374
800
provenance
large_stringlengths
356
703
intervention_map
large_stringlengths
11.4k
51.5k
contribution_metadata
large_stringlengths
127
127
access_level
large_stringclasses
1 value
quality_score
float64
0.95
0.95
Qwen/Qwen3-8B
Qwen
qwen3
3
2026-03-29T15:32:25.900493+00:00
{"hidden_size": 4096, "intermediate_size": 12288, "vocab_size": 151936, "max_position_embeddings": 40960, "num_hidden_layers": 36, "rms_norm_eps": 1e-06, "model_type": "qwen3"}
{"rope_type": null, "rope_scaling": {"rope_theta": 1000000, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 40960, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen3RMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"}
{"organization": "Qwen", "first_publish_date": "2025-04-27", "paper_urls": ["https://arxiv.org/abs/2309.00071", "https://arxiv.org/abs/2505.09388"], "base_model": "Qwen/Qwen3-8B-Base", "license": "apache-2.0", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["given inputs.", "switch", "qwen 3t", "qwen 2.5", "qwen 3", "llama", "qwen"], "hub_metadata": {"downloads": 9481229, "likes": 1014, "created_at": "2025-04-27T03:42:21.000Z", "last_modified": "2025-07-26T03:49:13.000Z"}}
{"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}]}
{"contributed_at": "2026-04-01T19:06:09.861598", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1}
public
0.95
Qwen/Qwen2.5-7B
Qwen
qwen2
3
2026-03-29T15:32:29.752121+00:00
{"hidden_size": 3584, "intermediate_size": 18944, "vocab_size": 152064, "max_position_embeddings": 131072, "num_hidden_layers": 28, "rms_norm_eps": 1e-06, "model_type": "qwen2"}
{"rope_type": null, "rope_scaling": {"rope_theta": 1000000.0, "rope_type": "default"}, "attention_bias": null, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 131072, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen2RMSNorm", "linear_bias_present": true, "qkv_projection_style": "separate"}
{"organization": "Qwen", "first_publish_date": "2024-09-15", "paper_urls": [], "base_model": null, "license": "apache-2.0", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["qwen 2.5", "qwen 2", "qwen"], "hub_metadata": {"downloads": 1024285, "likes": 266, "created_at": "2024-09-15T12:17:40.000Z", "last_modified": "2024-09-25T12:32:32.000Z"}}
{"residual_stream_dim": 3584, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}]}
{"contributed_at": "2026-04-01T19:06:09.861726", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1}
public
0.95
Qwen/Qwen3-30B-A3B
Qwen
qwen3_moe
3
2026-03-29T15:32:33.576356+00:00
{"hidden_size": 2048, "intermediate_size": 6144, "vocab_size": 151936, "max_position_embeddings": 40960, "num_hidden_layers": 48, "rms_norm_eps": 1e-06, "model_type": "qwen3_moe"}
{"rope_type": null, "rope_scaling": {"rope_theta": 1000000.0, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 40960, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen3MoeRMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"}
{"organization": "Qwen", "first_publish_date": "2025-04-27", "paper_urls": ["https://arxiv.org/abs/2309.00071", "https://arxiv.org/abs/2505.09388"], "base_model": "Qwen/Qwen3-30B-A3B-Base", "license": "apache-2.0", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["switch", "qwen 2.5", "given inputs.", "llama", "qwen 3", "qwen 3t", "qwen"], "hub_metadata": {"downloads": 1465903, "likes": 869, "created_at": "2025-04-27T03:43:05.000Z", "last_modified": "2025-07-26T03:45:17.000Z"}}
{"residual_stream_dim": 2048, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate", "moe_gate": "layers.0.mlp.gate", "moe_experts": "layers.0.mlp.experts", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate", "moe_gate": "layers.1.mlp.gate", "moe_experts": "layers.1.mlp.experts", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate", "moe_gate": "layers.2.mlp.gate", "moe_experts": "layers.2.mlp.experts", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}]}
{"contributed_at": "2026-04-01T19:06:09.861856", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1}
public
0.95
Qwen/Qwen3-235B-A22B
Qwen
qwen3_moe
3
2026-03-29T15:32:37.467930+00:00
{"hidden_size": 4096, "intermediate_size": 12288, "vocab_size": 151936, "max_position_embeddings": 40960, "num_hidden_layers": 94, "rms_norm_eps": 1e-06, "model_type": "qwen3_moe"}
{"rope_type": null, "rope_scaling": {"rope_theta": 1000000.0, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 40960, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen3MoeRMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"}
{"organization": "Qwen", "first_publish_date": "2025-04-27", "paper_urls": ["https://arxiv.org/abs/2309.00071", "https://arxiv.org/abs/2505.09388"], "base_model": null, "license": "apache-2.0", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["llama", "qwen 3", "switch", "qwen 3t", "qwen", "qwen 2.5", "given inputs."], "hub_metadata": {"downloads": 597542, "likes": 1082, "created_at": "2025-04-27T03:53:46.000Z", "last_modified": "2025-07-26T03:45:13.000Z"}}
{"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate", "moe_gate": "layers.0.mlp.gate", "moe_experts": "layers.0.mlp.experts", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate", "moe_gate": "layers.1.mlp.gate", "moe_experts": "layers.1.mlp.experts", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate", "moe_gate": "layers.2.mlp.gate", "moe_experts": "layers.2.mlp.experts", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_q_proj": "layers.48.self_attn.q_proj", "attn_k_proj": "layers.48.self_attn.k_proj", "attn_v_proj": "layers.48.self_attn.v_proj", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate", "moe_gate": "layers.48.mlp.gate", "moe_experts": "layers.48.mlp.experts", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_q_proj": "layers.49.self_attn.q_proj", "attn_k_proj": "layers.49.self_attn.k_proj", "attn_v_proj": "layers.49.self_attn.v_proj", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate", "moe_gate": "layers.49.mlp.gate", "moe_experts": "layers.49.mlp.experts", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_q_proj": "layers.50.self_attn.q_proj", "attn_k_proj": "layers.50.self_attn.k_proj", "attn_v_proj": "layers.50.self_attn.v_proj", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate", "moe_gate": "layers.50.mlp.gate", "moe_experts": "layers.50.mlp.experts", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate", "moe_gate": "layers.51.mlp.gate", "moe_experts": "layers.51.mlp.experts", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_q_proj": "layers.52.self_attn.q_proj", "attn_k_proj": "layers.52.self_attn.k_proj", "attn_v_proj": "layers.52.self_attn.v_proj", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate", "moe_gate": "layers.52.mlp.gate", "moe_experts": "layers.52.mlp.experts", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_q_proj": "layers.53.self_attn.q_proj", "attn_k_proj": "layers.53.self_attn.k_proj", "attn_v_proj": "layers.53.self_attn.v_proj", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate", "moe_gate": "layers.53.mlp.gate", "moe_experts": "layers.53.mlp.experts", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_q_proj": "layers.54.self_attn.q_proj", "attn_k_proj": "layers.54.self_attn.k_proj", "attn_v_proj": "layers.54.self_attn.v_proj", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate", "moe_gate": "layers.54.mlp.gate", "moe_experts": "layers.54.mlp.experts", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate", "moe_gate": "layers.55.mlp.gate", "moe_experts": "layers.55.mlp.experts", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_q_proj": "layers.56.self_attn.q_proj", "attn_k_proj": "layers.56.self_attn.k_proj", "attn_v_proj": "layers.56.self_attn.v_proj", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate", "moe_gate": "layers.56.mlp.gate", "moe_experts": "layers.56.mlp.experts", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_q_proj": "layers.57.self_attn.q_proj", "attn_k_proj": "layers.57.self_attn.k_proj", "attn_v_proj": "layers.57.self_attn.v_proj", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate", "moe_gate": "layers.57.mlp.gate", "moe_experts": "layers.57.mlp.experts", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_q_proj": "layers.58.self_attn.q_proj", "attn_k_proj": "layers.58.self_attn.k_proj", "attn_v_proj": "layers.58.self_attn.v_proj", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate", "moe_gate": "layers.58.mlp.gate", "moe_experts": "layers.58.mlp.experts", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate", "moe_gate": "layers.59.mlp.gate", "moe_experts": "layers.59.mlp.experts", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_q_proj": "layers.60.self_attn.q_proj", "attn_k_proj": "layers.60.self_attn.k_proj", "attn_v_proj": "layers.60.self_attn.v_proj", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate", "moe_gate": "layers.60.mlp.gate", "moe_experts": "layers.60.mlp.experts", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}, {"index": 61, "intervention_points": {"attn_input": "layers.61.self_attn.", "attn_q_proj": "layers.61.self_attn.q_proj", "attn_k_proj": "layers.61.self_attn.k_proj", "attn_v_proj": "layers.61.self_attn.v_proj", "attn_output_proj": "layers.61.self_attn.o_proj", "mlp_input": "layers.61.mlp.", "mlp_gate_proj": "layers.61.mlp.gate", "moe_gate": "layers.61.mlp.gate", "moe_experts": "layers.61.mlp.experts", "attn_pre_norm": "layers.61.input_layernorm", "attn_post_norm": "layers.61.post_attention_layernorm"}}, {"index": 62, "intervention_points": {"attn_input": "layers.62.self_attn.", "attn_q_proj": "layers.62.self_attn.q_proj", "attn_k_proj": "layers.62.self_attn.k_proj", "attn_v_proj": "layers.62.self_attn.v_proj", "attn_output_proj": "layers.62.self_attn.o_proj", "mlp_input": "layers.62.mlp.", "mlp_gate_proj": "layers.62.mlp.gate", "moe_gate": "layers.62.mlp.gate", "moe_experts": "layers.62.mlp.experts", "attn_pre_norm": "layers.62.input_layernorm", "attn_post_norm": "layers.62.post_attention_layernorm"}}, {"index": 63, "intervention_points": {"attn_input": "layers.63.self_attn.", "attn_q_proj": "layers.63.self_attn.q_proj", "attn_k_proj": "layers.63.self_attn.k_proj", "attn_v_proj": "layers.63.self_attn.v_proj", "attn_output_proj": "layers.63.self_attn.o_proj", "mlp_input": "layers.63.mlp.", "mlp_gate_proj": "layers.63.mlp.gate", "moe_gate": "layers.63.mlp.gate", "moe_experts": "layers.63.mlp.experts", "attn_pre_norm": "layers.63.input_layernorm", "attn_post_norm": "layers.63.post_attention_layernorm"}}, {"index": 64, "intervention_points": {"attn_input": "layers.64.self_attn.", "attn_q_proj": "layers.64.self_attn.q_proj", "attn_k_proj": "layers.64.self_attn.k_proj", "attn_v_proj": "layers.64.self_attn.v_proj", "attn_output_proj": "layers.64.self_attn.o_proj", "mlp_input": "layers.64.mlp.", "mlp_gate_proj": "layers.64.mlp.gate", "moe_gate": "layers.64.mlp.gate", "moe_experts": "layers.64.mlp.experts", "attn_pre_norm": "layers.64.input_layernorm", "attn_post_norm": "layers.64.post_attention_layernorm"}}, {"index": 65, "intervention_points": {"attn_input": "layers.65.self_attn.", "attn_q_proj": "layers.65.self_attn.q_proj", "attn_k_proj": "layers.65.self_attn.k_proj", "attn_v_proj": "layers.65.self_attn.v_proj", "attn_output_proj": "layers.65.self_attn.o_proj", "mlp_input": "layers.65.mlp.", "mlp_gate_proj": "layers.65.mlp.gate", "moe_gate": "layers.65.mlp.gate", "moe_experts": "layers.65.mlp.experts", "attn_pre_norm": "layers.65.input_layernorm", "attn_post_norm": "layers.65.post_attention_layernorm"}}, {"index": 66, "intervention_points": {"attn_input": "layers.66.self_attn.", "attn_q_proj": "layers.66.self_attn.q_proj", "attn_k_proj": "layers.66.self_attn.k_proj", "attn_v_proj": "layers.66.self_attn.v_proj", "attn_output_proj": "layers.66.self_attn.o_proj", "mlp_input": "layers.66.mlp.", "mlp_gate_proj": "layers.66.mlp.gate", "moe_gate": "layers.66.mlp.gate", "moe_experts": "layers.66.mlp.experts", "attn_pre_norm": "layers.66.input_layernorm", "attn_post_norm": "layers.66.post_attention_layernorm"}}, {"index": 67, "intervention_points": {"attn_input": "layers.67.self_attn.", "attn_q_proj": "layers.67.self_attn.q_proj", "attn_k_proj": "layers.67.self_attn.k_proj", "attn_v_proj": "layers.67.self_attn.v_proj", "attn_output_proj": "layers.67.self_attn.o_proj", "mlp_input": "layers.67.mlp.", "mlp_gate_proj": "layers.67.mlp.gate", "moe_gate": "layers.67.mlp.gate", "moe_experts": "layers.67.mlp.experts", "attn_pre_norm": "layers.67.input_layernorm", "attn_post_norm": "layers.67.post_attention_layernorm"}}, {"index": 68, "intervention_points": {"attn_input": "layers.68.self_attn.", "attn_q_proj": "layers.68.self_attn.q_proj", "attn_k_proj": "layers.68.self_attn.k_proj", "attn_v_proj": "layers.68.self_attn.v_proj", "attn_output_proj": "layers.68.self_attn.o_proj", "mlp_input": "layers.68.mlp.", "mlp_gate_proj": "layers.68.mlp.gate", "moe_gate": "layers.68.mlp.gate", "moe_experts": "layers.68.mlp.experts", "attn_pre_norm": "layers.68.input_layernorm", "attn_post_norm": "layers.68.post_attention_layernorm"}}, {"index": 69, "intervention_points": {"attn_input": "layers.69.self_attn.", "attn_q_proj": "layers.69.self_attn.q_proj", "attn_k_proj": "layers.69.self_attn.k_proj", "attn_v_proj": "layers.69.self_attn.v_proj", "attn_output_proj": "layers.69.self_attn.o_proj", "mlp_input": "layers.69.mlp.", "mlp_gate_proj": "layers.69.mlp.gate", "moe_gate": "layers.69.mlp.gate", "moe_experts": "layers.69.mlp.experts", "attn_pre_norm": "layers.69.input_layernorm", "attn_post_norm": "layers.69.post_attention_layernorm"}}, {"index": 70, "intervention_points": {"attn_input": "layers.70.self_attn.", "attn_q_proj": "layers.70.self_attn.q_proj", "attn_k_proj": "layers.70.self_attn.k_proj", "attn_v_proj": "layers.70.self_attn.v_proj", "attn_output_proj": "layers.70.self_attn.o_proj", "mlp_input": "layers.70.mlp.", "mlp_gate_proj": "layers.70.mlp.gate", "moe_gate": "layers.70.mlp.gate", "moe_experts": "layers.70.mlp.experts", "attn_pre_norm": "layers.70.input_layernorm", "attn_post_norm": "layers.70.post_attention_layernorm"}}, {"index": 71, "intervention_points": {"attn_input": "layers.71.self_attn.", "attn_q_proj": "layers.71.self_attn.q_proj", "attn_k_proj": "layers.71.self_attn.k_proj", "attn_v_proj": "layers.71.self_attn.v_proj", "attn_output_proj": "layers.71.self_attn.o_proj", "mlp_input": "layers.71.mlp.", "mlp_gate_proj": "layers.71.mlp.gate", "moe_gate": "layers.71.mlp.gate", "moe_experts": "layers.71.mlp.experts", "attn_pre_norm": "layers.71.input_layernorm", "attn_post_norm": "layers.71.post_attention_layernorm"}}, {"index": 72, "intervention_points": {"attn_input": "layers.72.self_attn.", "attn_q_proj": "layers.72.self_attn.q_proj", "attn_k_proj": "layers.72.self_attn.k_proj", "attn_v_proj": "layers.72.self_attn.v_proj", "attn_output_proj": "layers.72.self_attn.o_proj", "mlp_input": "layers.72.mlp.", "mlp_gate_proj": "layers.72.mlp.gate", "moe_gate": "layers.72.mlp.gate", "moe_experts": "layers.72.mlp.experts", "attn_pre_norm": "layers.72.input_layernorm", "attn_post_norm": "layers.72.post_attention_layernorm"}}, {"index": 73, "intervention_points": {"attn_input": "layers.73.self_attn.", "attn_q_proj": "layers.73.self_attn.q_proj", "attn_k_proj": "layers.73.self_attn.k_proj", "attn_v_proj": "layers.73.self_attn.v_proj", "attn_output_proj": "layers.73.self_attn.o_proj", "mlp_input": "layers.73.mlp.", "mlp_gate_proj": "layers.73.mlp.gate", "moe_gate": "layers.73.mlp.gate", "moe_experts": "layers.73.mlp.experts", "attn_pre_norm": "layers.73.input_layernorm", "attn_post_norm": "layers.73.post_attention_layernorm"}}, {"index": 74, "intervention_points": {"attn_input": "layers.74.self_attn.", "attn_q_proj": "layers.74.self_attn.q_proj", "attn_k_proj": "layers.74.self_attn.k_proj", "attn_v_proj": "layers.74.self_attn.v_proj", "attn_output_proj": "layers.74.self_attn.o_proj", "mlp_input": "layers.74.mlp.", "mlp_gate_proj": "layers.74.mlp.gate", "moe_gate": "layers.74.mlp.gate", "moe_experts": "layers.74.mlp.experts", "attn_pre_norm": "layers.74.input_layernorm", "attn_post_norm": "layers.74.post_attention_layernorm"}}, {"index": 75, "intervention_points": {"attn_input": "layers.75.self_attn.", "attn_q_proj": "layers.75.self_attn.q_proj", "attn_k_proj": "layers.75.self_attn.k_proj", "attn_v_proj": "layers.75.self_attn.v_proj", "attn_output_proj": "layers.75.self_attn.o_proj", "mlp_input": "layers.75.mlp.", "mlp_gate_proj": "layers.75.mlp.gate", "moe_gate": "layers.75.mlp.gate", "moe_experts": "layers.75.mlp.experts", "attn_pre_norm": "layers.75.input_layernorm", "attn_post_norm": "layers.75.post_attention_layernorm"}}, {"index": 76, "intervention_points": {"attn_input": "layers.76.self_attn.", "attn_q_proj": "layers.76.self_attn.q_proj", "attn_k_proj": "layers.76.self_attn.k_proj", "attn_v_proj": "layers.76.self_attn.v_proj", "attn_output_proj": "layers.76.self_attn.o_proj", "mlp_input": "layers.76.mlp.", "mlp_gate_proj": "layers.76.mlp.gate", "moe_gate": "layers.76.mlp.gate", "moe_experts": "layers.76.mlp.experts", "attn_pre_norm": "layers.76.input_layernorm", "attn_post_norm": "layers.76.post_attention_layernorm"}}, {"index": 77, "intervention_points": {"attn_input": "layers.77.self_attn.", "attn_q_proj": "layers.77.self_attn.q_proj", "attn_k_proj": "layers.77.self_attn.k_proj", "attn_v_proj": "layers.77.self_attn.v_proj", "attn_output_proj": "layers.77.self_attn.o_proj", "mlp_input": "layers.77.mlp.", "mlp_gate_proj": "layers.77.mlp.gate", "moe_gate": "layers.77.mlp.gate", "moe_experts": "layers.77.mlp.experts", "attn_pre_norm": "layers.77.input_layernorm", "attn_post_norm": "layers.77.post_attention_layernorm"}}, {"index": 78, "intervention_points": {"attn_input": "layers.78.self_attn.", "attn_q_proj": "layers.78.self_attn.q_proj", "attn_k_proj": "layers.78.self_attn.k_proj", "attn_v_proj": "layers.78.self_attn.v_proj", "attn_output_proj": "layers.78.self_attn.o_proj", "mlp_input": "layers.78.mlp.", "mlp_gate_proj": "layers.78.mlp.gate", "moe_gate": "layers.78.mlp.gate", "moe_experts": "layers.78.mlp.experts", "attn_pre_norm": "layers.78.input_layernorm", "attn_post_norm": "layers.78.post_attention_layernorm"}}, {"index": 79, "intervention_points": {"attn_input": "layers.79.self_attn.", "attn_q_proj": "layers.79.self_attn.q_proj", "attn_k_proj": "layers.79.self_attn.k_proj", "attn_v_proj": "layers.79.self_attn.v_proj", "attn_output_proj": "layers.79.self_attn.o_proj", "mlp_input": "layers.79.mlp.", "mlp_gate_proj": "layers.79.mlp.gate", "moe_gate": "layers.79.mlp.gate", "moe_experts": "layers.79.mlp.experts", "attn_pre_norm": "layers.79.input_layernorm", "attn_post_norm": "layers.79.post_attention_layernorm"}}, {"index": 80, "intervention_points": {"attn_input": "layers.80.self_attn.", "attn_q_proj": "layers.80.self_attn.q_proj", "attn_k_proj": "layers.80.self_attn.k_proj", "attn_v_proj": "layers.80.self_attn.v_proj", "attn_output_proj": "layers.80.self_attn.o_proj", "mlp_input": "layers.80.mlp.", "mlp_gate_proj": "layers.80.mlp.gate", "moe_gate": "layers.80.mlp.gate", "moe_experts": "layers.80.mlp.experts", "attn_pre_norm": "layers.80.input_layernorm", "attn_post_norm": "layers.80.post_attention_layernorm"}}, {"index": 81, "intervention_points": {"attn_input": "layers.81.self_attn.", "attn_q_proj": "layers.81.self_attn.q_proj", "attn_k_proj": "layers.81.self_attn.k_proj", "attn_v_proj": "layers.81.self_attn.v_proj", "attn_output_proj": "layers.81.self_attn.o_proj", "mlp_input": "layers.81.mlp.", "mlp_gate_proj": "layers.81.mlp.gate", "moe_gate": "layers.81.mlp.gate", "moe_experts": "layers.81.mlp.experts", "attn_pre_norm": "layers.81.input_layernorm", "attn_post_norm": "layers.81.post_attention_layernorm"}}, {"index": 82, "intervention_points": {"attn_input": "layers.82.self_attn.", "attn_q_proj": "layers.82.self_attn.q_proj", "attn_k_proj": "layers.82.self_attn.k_proj", "attn_v_proj": "layers.82.self_attn.v_proj", "attn_output_proj": "layers.82.self_attn.o_proj", "mlp_input": "layers.82.mlp.", "mlp_gate_proj": "layers.82.mlp.gate", "moe_gate": "layers.82.mlp.gate", "moe_experts": "layers.82.mlp.experts", "attn_pre_norm": "layers.82.input_layernorm", "attn_post_norm": "layers.82.post_attention_layernorm"}}, {"index": 83, "intervention_points": {"attn_input": "layers.83.self_attn.", "attn_q_proj": "layers.83.self_attn.q_proj", "attn_k_proj": "layers.83.self_attn.k_proj", "attn_v_proj": "layers.83.self_attn.v_proj", "attn_output_proj": "layers.83.self_attn.o_proj", "mlp_input": "layers.83.mlp.", "mlp_gate_proj": "layers.83.mlp.gate", "moe_gate": "layers.83.mlp.gate", "moe_experts": "layers.83.mlp.experts", "attn_pre_norm": "layers.83.input_layernorm", "attn_post_norm": "layers.83.post_attention_layernorm"}}, {"index": 84, "intervention_points": {"attn_input": "layers.84.self_attn.", "attn_q_proj": "layers.84.self_attn.q_proj", "attn_k_proj": "layers.84.self_attn.k_proj", "attn_v_proj": "layers.84.self_attn.v_proj", "attn_output_proj": "layers.84.self_attn.o_proj", "mlp_input": "layers.84.mlp.", "mlp_gate_proj": "layers.84.mlp.gate", "moe_gate": "layers.84.mlp.gate", "moe_experts": "layers.84.mlp.experts", "attn_pre_norm": "layers.84.input_layernorm", "attn_post_norm": "layers.84.post_attention_layernorm"}}, {"index": 85, "intervention_points": {"attn_input": "layers.85.self_attn.", "attn_q_proj": "layers.85.self_attn.q_proj", "attn_k_proj": "layers.85.self_attn.k_proj", "attn_v_proj": "layers.85.self_attn.v_proj", "attn_output_proj": "layers.85.self_attn.o_proj", "mlp_input": "layers.85.mlp.", "mlp_gate_proj": "layers.85.mlp.gate", "moe_gate": "layers.85.mlp.gate", "moe_experts": "layers.85.mlp.experts", "attn_pre_norm": "layers.85.input_layernorm", "attn_post_norm": "layers.85.post_attention_layernorm"}}, {"index": 86, "intervention_points": {"attn_input": "layers.86.self_attn.", "attn_q_proj": "layers.86.self_attn.q_proj", "attn_k_proj": "layers.86.self_attn.k_proj", "attn_v_proj": "layers.86.self_attn.v_proj", "attn_output_proj": "layers.86.self_attn.o_proj", "mlp_input": "layers.86.mlp.", "mlp_gate_proj": "layers.86.mlp.gate", "moe_gate": "layers.86.mlp.gate", "moe_experts": "layers.86.mlp.experts", "attn_pre_norm": "layers.86.input_layernorm", "attn_post_norm": "layers.86.post_attention_layernorm"}}, {"index": 87, "intervention_points": {"attn_input": "layers.87.self_attn.", "attn_q_proj": "layers.87.self_attn.q_proj", "attn_k_proj": "layers.87.self_attn.k_proj", "attn_v_proj": "layers.87.self_attn.v_proj", "attn_output_proj": "layers.87.self_attn.o_proj", "mlp_input": "layers.87.mlp.", "mlp_gate_proj": "layers.87.mlp.gate", "moe_gate": "layers.87.mlp.gate", "moe_experts": "layers.87.mlp.experts", "attn_pre_norm": "layers.87.input_layernorm", "attn_post_norm": "layers.87.post_attention_layernorm"}}, {"index": 88, "intervention_points": {"attn_input": "layers.88.self_attn.", "attn_q_proj": "layers.88.self_attn.q_proj", "attn_k_proj": "layers.88.self_attn.k_proj", "attn_v_proj": "layers.88.self_attn.v_proj", "attn_output_proj": "layers.88.self_attn.o_proj", "mlp_input": "layers.88.mlp.", "mlp_gate_proj": "layers.88.mlp.gate", "moe_gate": "layers.88.mlp.gate", "moe_experts": "layers.88.mlp.experts", "attn_pre_norm": "layers.88.input_layernorm", "attn_post_norm": "layers.88.post_attention_layernorm"}}, {"index": 89, "intervention_points": {"attn_input": "layers.89.self_attn.", "attn_q_proj": "layers.89.self_attn.q_proj", "attn_k_proj": "layers.89.self_attn.k_proj", "attn_v_proj": "layers.89.self_attn.v_proj", "attn_output_proj": "layers.89.self_attn.o_proj", "mlp_input": "layers.89.mlp.", "mlp_gate_proj": "layers.89.mlp.gate", "moe_gate": "layers.89.mlp.gate", "moe_experts": "layers.89.mlp.experts", "attn_pre_norm": "layers.89.input_layernorm", "attn_post_norm": "layers.89.post_attention_layernorm"}}, {"index": 90, "intervention_points": {"attn_input": "layers.90.self_attn.", "attn_q_proj": "layers.90.self_attn.q_proj", "attn_k_proj": "layers.90.self_attn.k_proj", "attn_v_proj": "layers.90.self_attn.v_proj", "attn_output_proj": "layers.90.self_attn.o_proj", "mlp_input": "layers.90.mlp.", "mlp_gate_proj": "layers.90.mlp.gate", "moe_gate": "layers.90.mlp.gate", "moe_experts": "layers.90.mlp.experts", "attn_pre_norm": "layers.90.input_layernorm", "attn_post_norm": "layers.90.post_attention_layernorm"}}, {"index": 91, "intervention_points": {"attn_input": "layers.91.self_attn.", "attn_q_proj": "layers.91.self_attn.q_proj", "attn_k_proj": "layers.91.self_attn.k_proj", "attn_v_proj": "layers.91.self_attn.v_proj", "attn_output_proj": "layers.91.self_attn.o_proj", "mlp_input": "layers.91.mlp.", "mlp_gate_proj": "layers.91.mlp.gate", "moe_gate": "layers.91.mlp.gate", "moe_experts": "layers.91.mlp.experts", "attn_pre_norm": "layers.91.input_layernorm", "attn_post_norm": "layers.91.post_attention_layernorm"}}, {"index": 92, "intervention_points": {"attn_input": "layers.92.self_attn.", "attn_q_proj": "layers.92.self_attn.q_proj", "attn_k_proj": "layers.92.self_attn.k_proj", "attn_v_proj": "layers.92.self_attn.v_proj", "attn_output_proj": "layers.92.self_attn.o_proj", "mlp_input": "layers.92.mlp.", "mlp_gate_proj": "layers.92.mlp.gate", "moe_gate": "layers.92.mlp.gate", "moe_experts": "layers.92.mlp.experts", "attn_pre_norm": "layers.92.input_layernorm", "attn_post_norm": "layers.92.post_attention_layernorm"}}, {"index": 93, "intervention_points": {"attn_input": "layers.93.self_attn.", "attn_q_proj": "layers.93.self_attn.q_proj", "attn_k_proj": "layers.93.self_attn.k_proj", "attn_v_proj": "layers.93.self_attn.v_proj", "attn_output_proj": "layers.93.self_attn.o_proj", "mlp_input": "layers.93.mlp.", "mlp_gate_proj": "layers.93.mlp.gate", "moe_gate": "layers.93.mlp.gate", "moe_experts": "layers.93.mlp.experts", "attn_pre_norm": "layers.93.input_layernorm", "attn_post_norm": "layers.93.post_attention_layernorm"}}]}
{"contributed_at": "2026-04-01T19:06:09.862043", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1}
public
0.95
deepseek-ai/DeepSeek-R1-Distill-Qwen-7B
deepseek-ai
qwen2
3
2026-03-29T15:32:46.354569+00:00
{"hidden_size": 3584, "intermediate_size": 18944, "vocab_size": 152064, "max_position_embeddings": 131072, "num_hidden_layers": 28, "rms_norm_eps": 1e-06, "model_type": "qwen2"}
{"rope_type": null, "rope_scaling": {"rope_theta": 10000, "rope_type": "default"}, "attention_bias": null, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 131072, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen2RMSNorm", "linear_bias_present": true, "qkv_projection_style": "separate"}
{"organization": "deepseek-ai", "first_publish_date": "2025-01-20", "paper_urls": ["https://arxiv.org/abs/2501.12948"], "base_model": null, "license": "mit", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["switch", "gpt 4", "llama 70b", "gpt", "qwen 7b", "claude", "gpt 4o", "llama 3", "llama 3.3", "qwen 1.5b", "llama 8b", "llama 3.1", "qwen 32b", "llama3.1-8b-base and is originally licensed under", "qwen 14b", "qwen2.5 and llama3 series to the community.", "open-source models", "qwen 2.5", "qwen", "claude 3.5", "llama"], "hub_metadata": {"downloads": 620266, "likes": 799, "created_at": "2025-01-20T09:16:14.000Z", "last_modified": "2025-02-24T03:32:20.000Z"}}
{"residual_stream_dim": 3584, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}]}
{"contributed_at": "2026-04-01T19:06:09.862159", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1}
public
0.95
deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
deepseek-ai
qwen2
3
2026-03-29T15:32:50.200935+00:00
{"hidden_size": 5120, "intermediate_size": 27648, "vocab_size": 152064, "max_position_embeddings": 131072, "num_hidden_layers": 64, "rms_norm_eps": 1e-05, "model_type": "qwen2"}
{"rope_type": null, "rope_scaling": {"rope_theta": 1000000.0, "rope_type": "default"}, "attention_bias": null, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 131072, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-05, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen2RMSNorm", "linear_bias_present": true, "qkv_projection_style": "separate"}
{"organization": "deepseek-ai", "first_publish_date": "2025-01-20", "paper_urls": ["https://arxiv.org/abs/2501.12948"], "base_model": null, "license": "mit", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["qwen2.5 and llama3 series to the community.", "llama 70b", "gpt", "claude 3.5", "open-source models", "qwen 14b", "llama 3.3", "llama", "switch", "gpt 4o", "qwen 7b", "llama 3", "llama 8b", "qwen 1.5b", "llama 3.1", "llama3.1-8b-base and is originally licensed under", "claude", "qwen", "gpt 4", "qwen 2.5", "qwen 32b"], "hub_metadata": {"downloads": 965513, "likes": 1526, "created_at": "2025-01-20T09:19:00.000Z", "last_modified": "2025-02-24T03:31:29.000Z"}}
{"residual_stream_dim": 5120, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate_proj", "mlp_up_proj": "layers.36.mlp.up_proj", "mlp_down_proj": "layers.36.mlp.down_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate_proj", "mlp_up_proj": "layers.37.mlp.up_proj", "mlp_down_proj": "layers.37.mlp.down_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate_proj", "mlp_up_proj": "layers.38.mlp.up_proj", "mlp_down_proj": "layers.38.mlp.down_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate_proj", "mlp_up_proj": "layers.39.mlp.up_proj", "mlp_down_proj": "layers.39.mlp.down_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate_proj", "mlp_up_proj": "layers.40.mlp.up_proj", "mlp_down_proj": "layers.40.mlp.down_proj", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate_proj", "mlp_up_proj": "layers.41.mlp.up_proj", "mlp_down_proj": "layers.41.mlp.down_proj", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate_proj", "mlp_up_proj": "layers.42.mlp.up_proj", "mlp_down_proj": "layers.42.mlp.down_proj", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate_proj", "mlp_up_proj": "layers.43.mlp.up_proj", "mlp_down_proj": "layers.43.mlp.down_proj", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate_proj", "mlp_up_proj": "layers.44.mlp.up_proj", "mlp_down_proj": "layers.44.mlp.down_proj", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate_proj", "mlp_up_proj": "layers.45.mlp.up_proj", "mlp_down_proj": "layers.45.mlp.down_proj", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate_proj", "mlp_up_proj": "layers.46.mlp.up_proj", "mlp_down_proj": "layers.46.mlp.down_proj", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate_proj", "mlp_up_proj": "layers.47.mlp.up_proj", "mlp_down_proj": "layers.47.mlp.down_proj", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_q_proj": "layers.48.self_attn.q_proj", "attn_k_proj": "layers.48.self_attn.k_proj", "attn_v_proj": "layers.48.self_attn.v_proj", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate_proj", "mlp_up_proj": "layers.48.mlp.up_proj", "mlp_down_proj": "layers.48.mlp.down_proj", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_q_proj": "layers.49.self_attn.q_proj", "attn_k_proj": "layers.49.self_attn.k_proj", "attn_v_proj": "layers.49.self_attn.v_proj", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate_proj", "mlp_up_proj": "layers.49.mlp.up_proj", "mlp_down_proj": "layers.49.mlp.down_proj", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_q_proj": "layers.50.self_attn.q_proj", "attn_k_proj": "layers.50.self_attn.k_proj", "attn_v_proj": "layers.50.self_attn.v_proj", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate_proj", "mlp_up_proj": "layers.50.mlp.up_proj", "mlp_down_proj": "layers.50.mlp.down_proj", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate_proj", "mlp_up_proj": "layers.51.mlp.up_proj", "mlp_down_proj": "layers.51.mlp.down_proj", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_q_proj": "layers.52.self_attn.q_proj", "attn_k_proj": "layers.52.self_attn.k_proj", "attn_v_proj": "layers.52.self_attn.v_proj", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate_proj", "mlp_up_proj": "layers.52.mlp.up_proj", "mlp_down_proj": "layers.52.mlp.down_proj", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_q_proj": "layers.53.self_attn.q_proj", "attn_k_proj": "layers.53.self_attn.k_proj", "attn_v_proj": "layers.53.self_attn.v_proj", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate_proj", "mlp_up_proj": "layers.53.mlp.up_proj", "mlp_down_proj": "layers.53.mlp.down_proj", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_q_proj": "layers.54.self_attn.q_proj", "attn_k_proj": "layers.54.self_attn.k_proj", "attn_v_proj": "layers.54.self_attn.v_proj", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate_proj", "mlp_up_proj": "layers.54.mlp.up_proj", "mlp_down_proj": "layers.54.mlp.down_proj", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate_proj", "mlp_up_proj": "layers.55.mlp.up_proj", "mlp_down_proj": "layers.55.mlp.down_proj", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_q_proj": "layers.56.self_attn.q_proj", "attn_k_proj": "layers.56.self_attn.k_proj", "attn_v_proj": "layers.56.self_attn.v_proj", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate_proj", "mlp_up_proj": "layers.56.mlp.up_proj", "mlp_down_proj": "layers.56.mlp.down_proj", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_q_proj": "layers.57.self_attn.q_proj", "attn_k_proj": "layers.57.self_attn.k_proj", "attn_v_proj": "layers.57.self_attn.v_proj", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate_proj", "mlp_up_proj": "layers.57.mlp.up_proj", "mlp_down_proj": "layers.57.mlp.down_proj", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_q_proj": "layers.58.self_attn.q_proj", "attn_k_proj": "layers.58.self_attn.k_proj", "attn_v_proj": "layers.58.self_attn.v_proj", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate_proj", "mlp_up_proj": "layers.58.mlp.up_proj", "mlp_down_proj": "layers.58.mlp.down_proj", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate_proj", "mlp_up_proj": "layers.59.mlp.up_proj", "mlp_down_proj": "layers.59.mlp.down_proj", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_q_proj": "layers.60.self_attn.q_proj", "attn_k_proj": "layers.60.self_attn.k_proj", "attn_v_proj": "layers.60.self_attn.v_proj", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate_proj", "mlp_up_proj": "layers.60.mlp.up_proj", "mlp_down_proj": "layers.60.mlp.down_proj", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}, {"index": 61, "intervention_points": {"attn_input": "layers.61.self_attn.", "attn_q_proj": "layers.61.self_attn.q_proj", "attn_k_proj": "layers.61.self_attn.k_proj", "attn_v_proj": "layers.61.self_attn.v_proj", "attn_output_proj": "layers.61.self_attn.o_proj", "mlp_input": "layers.61.mlp.", "mlp_gate_proj": "layers.61.mlp.gate_proj", "mlp_up_proj": "layers.61.mlp.up_proj", "mlp_down_proj": "layers.61.mlp.down_proj", "attn_pre_norm": "layers.61.input_layernorm", "attn_post_norm": "layers.61.post_attention_layernorm"}}, {"index": 62, "intervention_points": {"attn_input": "layers.62.self_attn.", "attn_q_proj": "layers.62.self_attn.q_proj", "attn_k_proj": "layers.62.self_attn.k_proj", "attn_v_proj": "layers.62.self_attn.v_proj", "attn_output_proj": "layers.62.self_attn.o_proj", "mlp_input": "layers.62.mlp.", "mlp_gate_proj": "layers.62.mlp.gate_proj", "mlp_up_proj": "layers.62.mlp.up_proj", "mlp_down_proj": "layers.62.mlp.down_proj", "attn_pre_norm": "layers.62.input_layernorm", "attn_post_norm": "layers.62.post_attention_layernorm"}}, {"index": 63, "intervention_points": {"attn_input": "layers.63.self_attn.", "attn_q_proj": "layers.63.self_attn.q_proj", "attn_k_proj": "layers.63.self_attn.k_proj", "attn_v_proj": "layers.63.self_attn.v_proj", "attn_output_proj": "layers.63.self_attn.o_proj", "mlp_input": "layers.63.mlp.", "mlp_gate_proj": "layers.63.mlp.gate_proj", "mlp_up_proj": "layers.63.mlp.up_proj", "mlp_down_proj": "layers.63.mlp.down_proj", "attn_pre_norm": "layers.63.input_layernorm", "attn_post_norm": "layers.63.post_attention_layernorm"}}]}
{"contributed_at": "2026-04-01T19:06:09.862302", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1}
public
0.95
deepseek-ai/DeepSeek-V2.5
deepseek-ai
deepseek_v2
3
2026-03-29T15:32:56.833668+00:00
"{\"hidden_size\": 5120, \"intermediate_size\": 12288, \"vocab_size\": 102400, \"max_position_embedd(...TRUNCATED)
"{\"rope_type\": null, \"rope_scaling\": {\"beta_fast\": 32, \"beta_slow\": 1, \"factor\": 40, \"msc(...TRUNCATED)
"{\"organization\": \"deepseek-ai\", \"first_publish_date\": \"2024-09-05\", \"paper_urls\": [\"http(...TRUNCATED)
"{\"residual_stream_dim\": 5120, \"layers\": [{\"index\": 0, \"intervention_points\": {\"attn_input\(...TRUNCATED)
"{\"contributed_at\": \"2026-04-01T19:06:09.862432\", \"contributor\": \"trohrbaugh\", \"attribution(...TRUNCATED)
public
0.95
deepseek-ai/DeepSeek-V3-Base
deepseek-ai
deepseek_v3
3
2026-03-29T15:33:04.667941+00:00
"{\"hidden_size\": 7168, \"intermediate_size\": 18432, \"vocab_size\": 129280, \"max_position_embedd(...TRUNCATED)
"{\"rope_type\": null, \"rope_scaling\": {\"beta_fast\": 32, \"beta_slow\": 1, \"factor\": 40, \"msc(...TRUNCATED)
"{\"organization\": \"deepseek-ai\", \"first_publish_date\": \"2024-12-25\", \"paper_urls\": [\"http(...TRUNCATED)
"{\"residual_stream_dim\": 7168, \"layers\": [{\"index\": 0, \"intervention_points\": {\"attn_input\(...TRUNCATED)
"{\"contributed_at\": \"2026-04-01T19:06:09.862572\", \"contributor\": \"trohrbaugh\", \"attribution(...TRUNCATED)
public
0.95
XiaomiMiMo/MiMo-7B-RL
XiaomiMiMo
mimo
3
2026-03-29T15:33:14.094468+00:00
"{\"hidden_size\": 4096, \"intermediate_size\": 11008, \"vocab_size\": 151680, \"max_position_embedd(...TRUNCATED)
"{\"rope_type\": null, \"rope_scaling\": {\"rope_theta\": 640000, \"rope_type\": \"default\"}, \"att(...TRUNCATED)
"{\"organization\": \"XiaomiMiMo\", \"first_publish_date\": \"2025-04-29\", \"paper_urls\": [\"https(...TRUNCATED)
"{\"residual_stream_dim\": 4096, \"layers\": [{\"index\": 0, \"intervention_points\": {\"attn_input\(...TRUNCATED)
"{\"contributed_at\": \"2026-04-01T19:06:09.862690\", \"contributor\": \"trohrbaugh\", \"attribution(...TRUNCATED)
public
0.95
THUDM/glm-4-9b-chat-hf
THUDM
glm
3
2026-03-29T15:33:18.543209+00:00
"{\"hidden_size\": 4096, \"intermediate_size\": 13696, \"vocab_size\": 151552, \"max_position_embedd(...TRUNCATED)
"{\"rope_type\": null, \"rope_scaling\": {\"rope_theta\": 10000.0, \"partial_rotary_factor\": 0.5, \(...TRUNCATED)
"{\"organization\": \"THUDM\", \"first_publish_date\": \"2024-10-23\", \"paper_urls\": [], \"base_mo(...TRUNCATED)
"{\"residual_stream_dim\": 4096, \"layers\": [{\"index\": 0, \"intervention_points\": {\"attn_input\(...TRUNCATED)
"{\"contributed_at\": \"2026-04-01T19:06:09.862794\", \"contributor\": \"trohrbaugh\", \"attribution(...TRUNCATED)
public
0.95
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
35

Space using RadicalNotionAI/community-analyses 1