model_id large_stringlengths 13 47 | organization large_stringlengths 4 14 | model_type large_stringlengths 3 16 | schema_version int64 3 3 | analyzed_at large_stringdate 2026-03-29 15:32:25 2026-04-01 23:41:04 | config large_stringlengths 145 207 | techniques large_stringlengths 374 800 | provenance large_stringlengths 356 703 | intervention_map large_stringlengths 11.4k 51.5k | contribution_metadata large_stringlengths 127 127 | access_level large_stringclasses 1
value | quality_score float64 0.95 0.95 |
|---|---|---|---|---|---|---|---|---|---|---|---|
Qwen/Qwen3-8B | Qwen | qwen3 | 3 | 2026-03-29T15:32:25.900493+00:00 | {"hidden_size": 4096, "intermediate_size": 12288, "vocab_size": 151936, "max_position_embeddings": 40960, "num_hidden_layers": 36, "rms_norm_eps": 1e-06, "model_type": "qwen3"} | {"rope_type": null, "rope_scaling": {"rope_theta": 1000000, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 40960, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen3RMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "Qwen", "first_publish_date": "2025-04-27", "paper_urls": ["https://arxiv.org/abs/2309.00071", "https://arxiv.org/abs/2505.09388"], "base_model": "Qwen/Qwen3-8B-Base", "license": "apache-2.0", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["given inputs.", "switch", "qwen 3t", "qwen 2.5", "qwen 3", "llama", "qwen"], "hub_metadata": {"downloads": 9481229, "likes": 1014, "created_at": "2025-04-27T03:42:21.000Z", "last_modified": "2025-07-26T03:49:13.000Z"}} | {"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.861598", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
Qwen/Qwen2.5-7B | Qwen | qwen2 | 3 | 2026-03-29T15:32:29.752121+00:00 | {"hidden_size": 3584, "intermediate_size": 18944, "vocab_size": 152064, "max_position_embeddings": 131072, "num_hidden_layers": 28, "rms_norm_eps": 1e-06, "model_type": "qwen2"} | {"rope_type": null, "rope_scaling": {"rope_theta": 1000000.0, "rope_type": "default"}, "attention_bias": null, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 131072, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen2RMSNorm", "linear_bias_present": true, "qkv_projection_style": "separate"} | {"organization": "Qwen", "first_publish_date": "2024-09-15", "paper_urls": [], "base_model": null, "license": "apache-2.0", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["qwen 2.5", "qwen 2", "qwen"], "hub_metadata": {"downloads": 1024285, "likes": 266, "created_at": "2024-09-15T12:17:40.000Z", "last_modified": "2024-09-25T12:32:32.000Z"}} | {"residual_stream_dim": 3584, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.861726", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
Qwen/Qwen3-30B-A3B | Qwen | qwen3_moe | 3 | 2026-03-29T15:32:33.576356+00:00 | {"hidden_size": 2048, "intermediate_size": 6144, "vocab_size": 151936, "max_position_embeddings": 40960, "num_hidden_layers": 48, "rms_norm_eps": 1e-06, "model_type": "qwen3_moe"} | {"rope_type": null, "rope_scaling": {"rope_theta": 1000000.0, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 40960, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen3MoeRMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "Qwen", "first_publish_date": "2025-04-27", "paper_urls": ["https://arxiv.org/abs/2309.00071", "https://arxiv.org/abs/2505.09388"], "base_model": "Qwen/Qwen3-30B-A3B-Base", "license": "apache-2.0", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["switch", "qwen 2.5", "given inputs.", "llama", "qwen 3", "qwen 3t", "qwen"], "hub_metadata": {"downloads": 1465903, "likes": 869, "created_at": "2025-04-27T03:43:05.000Z", "last_modified": "2025-07-26T03:45:17.000Z"}} | {"residual_stream_dim": 2048, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate", "moe_gate": "layers.0.mlp.gate", "moe_experts": "layers.0.mlp.experts", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate", "moe_gate": "layers.1.mlp.gate", "moe_experts": "layers.1.mlp.experts", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate", "moe_gate": "layers.2.mlp.gate", "moe_experts": "layers.2.mlp.experts", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.861856", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
Qwen/Qwen3-235B-A22B | Qwen | qwen3_moe | 3 | 2026-03-29T15:32:37.467930+00:00 | {"hidden_size": 4096, "intermediate_size": 12288, "vocab_size": 151936, "max_position_embeddings": 40960, "num_hidden_layers": 94, "rms_norm_eps": 1e-06, "model_type": "qwen3_moe"} | {"rope_type": null, "rope_scaling": {"rope_theta": 1000000.0, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 40960, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen3MoeRMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "Qwen", "first_publish_date": "2025-04-27", "paper_urls": ["https://arxiv.org/abs/2309.00071", "https://arxiv.org/abs/2505.09388"], "base_model": null, "license": "apache-2.0", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["llama", "qwen 3", "switch", "qwen 3t", "qwen", "qwen 2.5", "given inputs."], "hub_metadata": {"downloads": 597542, "likes": 1082, "created_at": "2025-04-27T03:53:46.000Z", "last_modified": "2025-07-26T03:45:13.000Z"}} | {"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate", "moe_gate": "layers.0.mlp.gate", "moe_experts": "layers.0.mlp.experts", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate", "moe_gate": "layers.1.mlp.gate", "moe_experts": "layers.1.mlp.experts", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate", "moe_gate": "layers.2.mlp.gate", "moe_experts": "layers.2.mlp.experts", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_q_proj": "layers.48.self_attn.q_proj", "attn_k_proj": "layers.48.self_attn.k_proj", "attn_v_proj": "layers.48.self_attn.v_proj", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate", "moe_gate": "layers.48.mlp.gate", "moe_experts": "layers.48.mlp.experts", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_q_proj": "layers.49.self_attn.q_proj", "attn_k_proj": "layers.49.self_attn.k_proj", "attn_v_proj": "layers.49.self_attn.v_proj", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate", "moe_gate": "layers.49.mlp.gate", "moe_experts": "layers.49.mlp.experts", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_q_proj": "layers.50.self_attn.q_proj", "attn_k_proj": "layers.50.self_attn.k_proj", "attn_v_proj": "layers.50.self_attn.v_proj", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate", "moe_gate": "layers.50.mlp.gate", "moe_experts": "layers.50.mlp.experts", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate", "moe_gate": "layers.51.mlp.gate", "moe_experts": "layers.51.mlp.experts", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_q_proj": "layers.52.self_attn.q_proj", "attn_k_proj": "layers.52.self_attn.k_proj", "attn_v_proj": "layers.52.self_attn.v_proj", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate", "moe_gate": "layers.52.mlp.gate", "moe_experts": "layers.52.mlp.experts", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_q_proj": "layers.53.self_attn.q_proj", "attn_k_proj": "layers.53.self_attn.k_proj", "attn_v_proj": "layers.53.self_attn.v_proj", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate", "moe_gate": "layers.53.mlp.gate", "moe_experts": "layers.53.mlp.experts", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_q_proj": "layers.54.self_attn.q_proj", "attn_k_proj": "layers.54.self_attn.k_proj", "attn_v_proj": "layers.54.self_attn.v_proj", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate", "moe_gate": "layers.54.mlp.gate", "moe_experts": "layers.54.mlp.experts", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate", "moe_gate": "layers.55.mlp.gate", "moe_experts": "layers.55.mlp.experts", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_q_proj": "layers.56.self_attn.q_proj", "attn_k_proj": "layers.56.self_attn.k_proj", "attn_v_proj": "layers.56.self_attn.v_proj", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate", "moe_gate": "layers.56.mlp.gate", "moe_experts": "layers.56.mlp.experts", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_q_proj": "layers.57.self_attn.q_proj", "attn_k_proj": "layers.57.self_attn.k_proj", "attn_v_proj": "layers.57.self_attn.v_proj", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate", "moe_gate": "layers.57.mlp.gate", "moe_experts": "layers.57.mlp.experts", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_q_proj": "layers.58.self_attn.q_proj", "attn_k_proj": "layers.58.self_attn.k_proj", "attn_v_proj": "layers.58.self_attn.v_proj", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate", "moe_gate": "layers.58.mlp.gate", "moe_experts": "layers.58.mlp.experts", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate", "moe_gate": "layers.59.mlp.gate", "moe_experts": "layers.59.mlp.experts", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_q_proj": "layers.60.self_attn.q_proj", "attn_k_proj": "layers.60.self_attn.k_proj", "attn_v_proj": "layers.60.self_attn.v_proj", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate", "moe_gate": "layers.60.mlp.gate", "moe_experts": "layers.60.mlp.experts", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}, {"index": 61, "intervention_points": {"attn_input": "layers.61.self_attn.", "attn_q_proj": "layers.61.self_attn.q_proj", "attn_k_proj": "layers.61.self_attn.k_proj", "attn_v_proj": "layers.61.self_attn.v_proj", "attn_output_proj": "layers.61.self_attn.o_proj", "mlp_input": "layers.61.mlp.", "mlp_gate_proj": "layers.61.mlp.gate", "moe_gate": "layers.61.mlp.gate", "moe_experts": "layers.61.mlp.experts", "attn_pre_norm": "layers.61.input_layernorm", "attn_post_norm": "layers.61.post_attention_layernorm"}}, {"index": 62, "intervention_points": {"attn_input": "layers.62.self_attn.", "attn_q_proj": "layers.62.self_attn.q_proj", "attn_k_proj": "layers.62.self_attn.k_proj", "attn_v_proj": "layers.62.self_attn.v_proj", "attn_output_proj": "layers.62.self_attn.o_proj", "mlp_input": "layers.62.mlp.", "mlp_gate_proj": "layers.62.mlp.gate", "moe_gate": "layers.62.mlp.gate", "moe_experts": "layers.62.mlp.experts", "attn_pre_norm": "layers.62.input_layernorm", "attn_post_norm": "layers.62.post_attention_layernorm"}}, {"index": 63, "intervention_points": {"attn_input": "layers.63.self_attn.", "attn_q_proj": "layers.63.self_attn.q_proj", "attn_k_proj": "layers.63.self_attn.k_proj", "attn_v_proj": "layers.63.self_attn.v_proj", "attn_output_proj": "layers.63.self_attn.o_proj", "mlp_input": "layers.63.mlp.", "mlp_gate_proj": "layers.63.mlp.gate", "moe_gate": "layers.63.mlp.gate", "moe_experts": "layers.63.mlp.experts", "attn_pre_norm": "layers.63.input_layernorm", "attn_post_norm": "layers.63.post_attention_layernorm"}}, {"index": 64, "intervention_points": {"attn_input": "layers.64.self_attn.", "attn_q_proj": "layers.64.self_attn.q_proj", "attn_k_proj": "layers.64.self_attn.k_proj", "attn_v_proj": "layers.64.self_attn.v_proj", "attn_output_proj": "layers.64.self_attn.o_proj", "mlp_input": "layers.64.mlp.", "mlp_gate_proj": "layers.64.mlp.gate", "moe_gate": "layers.64.mlp.gate", "moe_experts": "layers.64.mlp.experts", "attn_pre_norm": "layers.64.input_layernorm", "attn_post_norm": "layers.64.post_attention_layernorm"}}, {"index": 65, "intervention_points": {"attn_input": "layers.65.self_attn.", "attn_q_proj": "layers.65.self_attn.q_proj", "attn_k_proj": "layers.65.self_attn.k_proj", "attn_v_proj": "layers.65.self_attn.v_proj", "attn_output_proj": "layers.65.self_attn.o_proj", "mlp_input": "layers.65.mlp.", "mlp_gate_proj": "layers.65.mlp.gate", "moe_gate": "layers.65.mlp.gate", "moe_experts": "layers.65.mlp.experts", "attn_pre_norm": "layers.65.input_layernorm", "attn_post_norm": "layers.65.post_attention_layernorm"}}, {"index": 66, "intervention_points": {"attn_input": "layers.66.self_attn.", "attn_q_proj": "layers.66.self_attn.q_proj", "attn_k_proj": "layers.66.self_attn.k_proj", "attn_v_proj": "layers.66.self_attn.v_proj", "attn_output_proj": "layers.66.self_attn.o_proj", "mlp_input": "layers.66.mlp.", "mlp_gate_proj": "layers.66.mlp.gate", "moe_gate": "layers.66.mlp.gate", "moe_experts": "layers.66.mlp.experts", "attn_pre_norm": "layers.66.input_layernorm", "attn_post_norm": "layers.66.post_attention_layernorm"}}, {"index": 67, "intervention_points": {"attn_input": "layers.67.self_attn.", "attn_q_proj": "layers.67.self_attn.q_proj", "attn_k_proj": "layers.67.self_attn.k_proj", "attn_v_proj": "layers.67.self_attn.v_proj", "attn_output_proj": "layers.67.self_attn.o_proj", "mlp_input": "layers.67.mlp.", "mlp_gate_proj": "layers.67.mlp.gate", "moe_gate": "layers.67.mlp.gate", "moe_experts": "layers.67.mlp.experts", "attn_pre_norm": "layers.67.input_layernorm", "attn_post_norm": "layers.67.post_attention_layernorm"}}, {"index": 68, "intervention_points": {"attn_input": "layers.68.self_attn.", "attn_q_proj": "layers.68.self_attn.q_proj", "attn_k_proj": "layers.68.self_attn.k_proj", "attn_v_proj": "layers.68.self_attn.v_proj", "attn_output_proj": "layers.68.self_attn.o_proj", "mlp_input": "layers.68.mlp.", "mlp_gate_proj": "layers.68.mlp.gate", "moe_gate": "layers.68.mlp.gate", "moe_experts": "layers.68.mlp.experts", "attn_pre_norm": "layers.68.input_layernorm", "attn_post_norm": "layers.68.post_attention_layernorm"}}, {"index": 69, "intervention_points": {"attn_input": "layers.69.self_attn.", "attn_q_proj": "layers.69.self_attn.q_proj", "attn_k_proj": "layers.69.self_attn.k_proj", "attn_v_proj": "layers.69.self_attn.v_proj", "attn_output_proj": "layers.69.self_attn.o_proj", "mlp_input": "layers.69.mlp.", "mlp_gate_proj": "layers.69.mlp.gate", "moe_gate": "layers.69.mlp.gate", "moe_experts": "layers.69.mlp.experts", "attn_pre_norm": "layers.69.input_layernorm", "attn_post_norm": "layers.69.post_attention_layernorm"}}, {"index": 70, "intervention_points": {"attn_input": "layers.70.self_attn.", "attn_q_proj": "layers.70.self_attn.q_proj", "attn_k_proj": "layers.70.self_attn.k_proj", "attn_v_proj": "layers.70.self_attn.v_proj", "attn_output_proj": "layers.70.self_attn.o_proj", "mlp_input": "layers.70.mlp.", "mlp_gate_proj": "layers.70.mlp.gate", "moe_gate": "layers.70.mlp.gate", "moe_experts": "layers.70.mlp.experts", "attn_pre_norm": "layers.70.input_layernorm", "attn_post_norm": "layers.70.post_attention_layernorm"}}, {"index": 71, "intervention_points": {"attn_input": "layers.71.self_attn.", "attn_q_proj": "layers.71.self_attn.q_proj", "attn_k_proj": "layers.71.self_attn.k_proj", "attn_v_proj": "layers.71.self_attn.v_proj", "attn_output_proj": "layers.71.self_attn.o_proj", "mlp_input": "layers.71.mlp.", "mlp_gate_proj": "layers.71.mlp.gate", "moe_gate": "layers.71.mlp.gate", "moe_experts": "layers.71.mlp.experts", "attn_pre_norm": "layers.71.input_layernorm", "attn_post_norm": "layers.71.post_attention_layernorm"}}, {"index": 72, "intervention_points": {"attn_input": "layers.72.self_attn.", "attn_q_proj": "layers.72.self_attn.q_proj", "attn_k_proj": "layers.72.self_attn.k_proj", "attn_v_proj": "layers.72.self_attn.v_proj", "attn_output_proj": "layers.72.self_attn.o_proj", "mlp_input": "layers.72.mlp.", "mlp_gate_proj": "layers.72.mlp.gate", "moe_gate": "layers.72.mlp.gate", "moe_experts": "layers.72.mlp.experts", "attn_pre_norm": "layers.72.input_layernorm", "attn_post_norm": "layers.72.post_attention_layernorm"}}, {"index": 73, "intervention_points": {"attn_input": "layers.73.self_attn.", "attn_q_proj": "layers.73.self_attn.q_proj", "attn_k_proj": "layers.73.self_attn.k_proj", "attn_v_proj": "layers.73.self_attn.v_proj", "attn_output_proj": "layers.73.self_attn.o_proj", "mlp_input": "layers.73.mlp.", "mlp_gate_proj": "layers.73.mlp.gate", "moe_gate": "layers.73.mlp.gate", "moe_experts": "layers.73.mlp.experts", "attn_pre_norm": "layers.73.input_layernorm", "attn_post_norm": "layers.73.post_attention_layernorm"}}, {"index": 74, "intervention_points": {"attn_input": "layers.74.self_attn.", "attn_q_proj": "layers.74.self_attn.q_proj", "attn_k_proj": "layers.74.self_attn.k_proj", "attn_v_proj": "layers.74.self_attn.v_proj", "attn_output_proj": "layers.74.self_attn.o_proj", "mlp_input": "layers.74.mlp.", "mlp_gate_proj": "layers.74.mlp.gate", "moe_gate": "layers.74.mlp.gate", "moe_experts": "layers.74.mlp.experts", "attn_pre_norm": "layers.74.input_layernorm", "attn_post_norm": "layers.74.post_attention_layernorm"}}, {"index": 75, "intervention_points": {"attn_input": "layers.75.self_attn.", "attn_q_proj": "layers.75.self_attn.q_proj", "attn_k_proj": "layers.75.self_attn.k_proj", "attn_v_proj": "layers.75.self_attn.v_proj", "attn_output_proj": "layers.75.self_attn.o_proj", "mlp_input": "layers.75.mlp.", "mlp_gate_proj": "layers.75.mlp.gate", "moe_gate": "layers.75.mlp.gate", "moe_experts": "layers.75.mlp.experts", "attn_pre_norm": "layers.75.input_layernorm", "attn_post_norm": "layers.75.post_attention_layernorm"}}, {"index": 76, "intervention_points": {"attn_input": "layers.76.self_attn.", "attn_q_proj": "layers.76.self_attn.q_proj", "attn_k_proj": "layers.76.self_attn.k_proj", "attn_v_proj": "layers.76.self_attn.v_proj", "attn_output_proj": "layers.76.self_attn.o_proj", "mlp_input": "layers.76.mlp.", "mlp_gate_proj": "layers.76.mlp.gate", "moe_gate": "layers.76.mlp.gate", "moe_experts": "layers.76.mlp.experts", "attn_pre_norm": "layers.76.input_layernorm", "attn_post_norm": "layers.76.post_attention_layernorm"}}, {"index": 77, "intervention_points": {"attn_input": "layers.77.self_attn.", "attn_q_proj": "layers.77.self_attn.q_proj", "attn_k_proj": "layers.77.self_attn.k_proj", "attn_v_proj": "layers.77.self_attn.v_proj", "attn_output_proj": "layers.77.self_attn.o_proj", "mlp_input": "layers.77.mlp.", "mlp_gate_proj": "layers.77.mlp.gate", "moe_gate": "layers.77.mlp.gate", "moe_experts": "layers.77.mlp.experts", "attn_pre_norm": "layers.77.input_layernorm", "attn_post_norm": "layers.77.post_attention_layernorm"}}, {"index": 78, "intervention_points": {"attn_input": "layers.78.self_attn.", "attn_q_proj": "layers.78.self_attn.q_proj", "attn_k_proj": "layers.78.self_attn.k_proj", "attn_v_proj": "layers.78.self_attn.v_proj", "attn_output_proj": "layers.78.self_attn.o_proj", "mlp_input": "layers.78.mlp.", "mlp_gate_proj": "layers.78.mlp.gate", "moe_gate": "layers.78.mlp.gate", "moe_experts": "layers.78.mlp.experts", "attn_pre_norm": "layers.78.input_layernorm", "attn_post_norm": "layers.78.post_attention_layernorm"}}, {"index": 79, "intervention_points": {"attn_input": "layers.79.self_attn.", "attn_q_proj": "layers.79.self_attn.q_proj", "attn_k_proj": "layers.79.self_attn.k_proj", "attn_v_proj": "layers.79.self_attn.v_proj", "attn_output_proj": "layers.79.self_attn.o_proj", "mlp_input": "layers.79.mlp.", "mlp_gate_proj": "layers.79.mlp.gate", "moe_gate": "layers.79.mlp.gate", "moe_experts": "layers.79.mlp.experts", "attn_pre_norm": "layers.79.input_layernorm", "attn_post_norm": "layers.79.post_attention_layernorm"}}, {"index": 80, "intervention_points": {"attn_input": "layers.80.self_attn.", "attn_q_proj": "layers.80.self_attn.q_proj", "attn_k_proj": "layers.80.self_attn.k_proj", "attn_v_proj": "layers.80.self_attn.v_proj", "attn_output_proj": "layers.80.self_attn.o_proj", "mlp_input": "layers.80.mlp.", "mlp_gate_proj": "layers.80.mlp.gate", "moe_gate": "layers.80.mlp.gate", "moe_experts": "layers.80.mlp.experts", "attn_pre_norm": "layers.80.input_layernorm", "attn_post_norm": "layers.80.post_attention_layernorm"}}, {"index": 81, "intervention_points": {"attn_input": "layers.81.self_attn.", "attn_q_proj": "layers.81.self_attn.q_proj", "attn_k_proj": "layers.81.self_attn.k_proj", "attn_v_proj": "layers.81.self_attn.v_proj", "attn_output_proj": "layers.81.self_attn.o_proj", "mlp_input": "layers.81.mlp.", "mlp_gate_proj": "layers.81.mlp.gate", "moe_gate": "layers.81.mlp.gate", "moe_experts": "layers.81.mlp.experts", "attn_pre_norm": "layers.81.input_layernorm", "attn_post_norm": "layers.81.post_attention_layernorm"}}, {"index": 82, "intervention_points": {"attn_input": "layers.82.self_attn.", "attn_q_proj": "layers.82.self_attn.q_proj", "attn_k_proj": "layers.82.self_attn.k_proj", "attn_v_proj": "layers.82.self_attn.v_proj", "attn_output_proj": "layers.82.self_attn.o_proj", "mlp_input": "layers.82.mlp.", "mlp_gate_proj": "layers.82.mlp.gate", "moe_gate": "layers.82.mlp.gate", "moe_experts": "layers.82.mlp.experts", "attn_pre_norm": "layers.82.input_layernorm", "attn_post_norm": "layers.82.post_attention_layernorm"}}, {"index": 83, "intervention_points": {"attn_input": "layers.83.self_attn.", "attn_q_proj": "layers.83.self_attn.q_proj", "attn_k_proj": "layers.83.self_attn.k_proj", "attn_v_proj": "layers.83.self_attn.v_proj", "attn_output_proj": "layers.83.self_attn.o_proj", "mlp_input": "layers.83.mlp.", "mlp_gate_proj": "layers.83.mlp.gate", "moe_gate": "layers.83.mlp.gate", "moe_experts": "layers.83.mlp.experts", "attn_pre_norm": "layers.83.input_layernorm", "attn_post_norm": "layers.83.post_attention_layernorm"}}, {"index": 84, "intervention_points": {"attn_input": "layers.84.self_attn.", "attn_q_proj": "layers.84.self_attn.q_proj", "attn_k_proj": "layers.84.self_attn.k_proj", "attn_v_proj": "layers.84.self_attn.v_proj", "attn_output_proj": "layers.84.self_attn.o_proj", "mlp_input": "layers.84.mlp.", "mlp_gate_proj": "layers.84.mlp.gate", "moe_gate": "layers.84.mlp.gate", "moe_experts": "layers.84.mlp.experts", "attn_pre_norm": "layers.84.input_layernorm", "attn_post_norm": "layers.84.post_attention_layernorm"}}, {"index": 85, "intervention_points": {"attn_input": "layers.85.self_attn.", "attn_q_proj": "layers.85.self_attn.q_proj", "attn_k_proj": "layers.85.self_attn.k_proj", "attn_v_proj": "layers.85.self_attn.v_proj", "attn_output_proj": "layers.85.self_attn.o_proj", "mlp_input": "layers.85.mlp.", "mlp_gate_proj": "layers.85.mlp.gate", "moe_gate": "layers.85.mlp.gate", "moe_experts": "layers.85.mlp.experts", "attn_pre_norm": "layers.85.input_layernorm", "attn_post_norm": "layers.85.post_attention_layernorm"}}, {"index": 86, "intervention_points": {"attn_input": "layers.86.self_attn.", "attn_q_proj": "layers.86.self_attn.q_proj", "attn_k_proj": "layers.86.self_attn.k_proj", "attn_v_proj": "layers.86.self_attn.v_proj", "attn_output_proj": "layers.86.self_attn.o_proj", "mlp_input": "layers.86.mlp.", "mlp_gate_proj": "layers.86.mlp.gate", "moe_gate": "layers.86.mlp.gate", "moe_experts": "layers.86.mlp.experts", "attn_pre_norm": "layers.86.input_layernorm", "attn_post_norm": "layers.86.post_attention_layernorm"}}, {"index": 87, "intervention_points": {"attn_input": "layers.87.self_attn.", "attn_q_proj": "layers.87.self_attn.q_proj", "attn_k_proj": "layers.87.self_attn.k_proj", "attn_v_proj": "layers.87.self_attn.v_proj", "attn_output_proj": "layers.87.self_attn.o_proj", "mlp_input": "layers.87.mlp.", "mlp_gate_proj": "layers.87.mlp.gate", "moe_gate": "layers.87.mlp.gate", "moe_experts": "layers.87.mlp.experts", "attn_pre_norm": "layers.87.input_layernorm", "attn_post_norm": "layers.87.post_attention_layernorm"}}, {"index": 88, "intervention_points": {"attn_input": "layers.88.self_attn.", "attn_q_proj": "layers.88.self_attn.q_proj", "attn_k_proj": "layers.88.self_attn.k_proj", "attn_v_proj": "layers.88.self_attn.v_proj", "attn_output_proj": "layers.88.self_attn.o_proj", "mlp_input": "layers.88.mlp.", "mlp_gate_proj": "layers.88.mlp.gate", "moe_gate": "layers.88.mlp.gate", "moe_experts": "layers.88.mlp.experts", "attn_pre_norm": "layers.88.input_layernorm", "attn_post_norm": "layers.88.post_attention_layernorm"}}, {"index": 89, "intervention_points": {"attn_input": "layers.89.self_attn.", "attn_q_proj": "layers.89.self_attn.q_proj", "attn_k_proj": "layers.89.self_attn.k_proj", "attn_v_proj": "layers.89.self_attn.v_proj", "attn_output_proj": "layers.89.self_attn.o_proj", "mlp_input": "layers.89.mlp.", "mlp_gate_proj": "layers.89.mlp.gate", "moe_gate": "layers.89.mlp.gate", "moe_experts": "layers.89.mlp.experts", "attn_pre_norm": "layers.89.input_layernorm", "attn_post_norm": "layers.89.post_attention_layernorm"}}, {"index": 90, "intervention_points": {"attn_input": "layers.90.self_attn.", "attn_q_proj": "layers.90.self_attn.q_proj", "attn_k_proj": "layers.90.self_attn.k_proj", "attn_v_proj": "layers.90.self_attn.v_proj", "attn_output_proj": "layers.90.self_attn.o_proj", "mlp_input": "layers.90.mlp.", "mlp_gate_proj": "layers.90.mlp.gate", "moe_gate": "layers.90.mlp.gate", "moe_experts": "layers.90.mlp.experts", "attn_pre_norm": "layers.90.input_layernorm", "attn_post_norm": "layers.90.post_attention_layernorm"}}, {"index": 91, "intervention_points": {"attn_input": "layers.91.self_attn.", "attn_q_proj": "layers.91.self_attn.q_proj", "attn_k_proj": "layers.91.self_attn.k_proj", "attn_v_proj": "layers.91.self_attn.v_proj", "attn_output_proj": "layers.91.self_attn.o_proj", "mlp_input": "layers.91.mlp.", "mlp_gate_proj": "layers.91.mlp.gate", "moe_gate": "layers.91.mlp.gate", "moe_experts": "layers.91.mlp.experts", "attn_pre_norm": "layers.91.input_layernorm", "attn_post_norm": "layers.91.post_attention_layernorm"}}, {"index": 92, "intervention_points": {"attn_input": "layers.92.self_attn.", "attn_q_proj": "layers.92.self_attn.q_proj", "attn_k_proj": "layers.92.self_attn.k_proj", "attn_v_proj": "layers.92.self_attn.v_proj", "attn_output_proj": "layers.92.self_attn.o_proj", "mlp_input": "layers.92.mlp.", "mlp_gate_proj": "layers.92.mlp.gate", "moe_gate": "layers.92.mlp.gate", "moe_experts": "layers.92.mlp.experts", "attn_pre_norm": "layers.92.input_layernorm", "attn_post_norm": "layers.92.post_attention_layernorm"}}, {"index": 93, "intervention_points": {"attn_input": "layers.93.self_attn.", "attn_q_proj": "layers.93.self_attn.q_proj", "attn_k_proj": "layers.93.self_attn.k_proj", "attn_v_proj": "layers.93.self_attn.v_proj", "attn_output_proj": "layers.93.self_attn.o_proj", "mlp_input": "layers.93.mlp.", "mlp_gate_proj": "layers.93.mlp.gate", "moe_gate": "layers.93.mlp.gate", "moe_experts": "layers.93.mlp.experts", "attn_pre_norm": "layers.93.input_layernorm", "attn_post_norm": "layers.93.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.862043", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
deepseek-ai/DeepSeek-R1-Distill-Qwen-7B | deepseek-ai | qwen2 | 3 | 2026-03-29T15:32:46.354569+00:00 | {"hidden_size": 3584, "intermediate_size": 18944, "vocab_size": 152064, "max_position_embeddings": 131072, "num_hidden_layers": 28, "rms_norm_eps": 1e-06, "model_type": "qwen2"} | {"rope_type": null, "rope_scaling": {"rope_theta": 10000, "rope_type": "default"}, "attention_bias": null, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 131072, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen2RMSNorm", "linear_bias_present": true, "qkv_projection_style": "separate"} | {"organization": "deepseek-ai", "first_publish_date": "2025-01-20", "paper_urls": ["https://arxiv.org/abs/2501.12948"], "base_model": null, "license": "mit", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["switch", "gpt 4", "llama 70b", "gpt", "qwen 7b", "claude", "gpt 4o", "llama 3", "llama 3.3", "qwen 1.5b", "llama 8b", "llama 3.1", "qwen 32b", "llama3.1-8b-base and is originally licensed under", "qwen 14b", "qwen2.5 and llama3 series to the community.", "open-source models", "qwen 2.5", "qwen", "claude 3.5", "llama"], "hub_metadata": {"downloads": 620266, "likes": 799, "created_at": "2025-01-20T09:16:14.000Z", "last_modified": "2025-02-24T03:32:20.000Z"}} | {"residual_stream_dim": 3584, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.862159", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
deepseek-ai/DeepSeek-R1-Distill-Qwen-32B | deepseek-ai | qwen2 | 3 | 2026-03-29T15:32:50.200935+00:00 | {"hidden_size": 5120, "intermediate_size": 27648, "vocab_size": 152064, "max_position_embeddings": 131072, "num_hidden_layers": 64, "rms_norm_eps": 1e-05, "model_type": "qwen2"} | {"rope_type": null, "rope_scaling": {"rope_theta": 1000000.0, "rope_type": "default"}, "attention_bias": null, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 131072, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-05, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen2RMSNorm", "linear_bias_present": true, "qkv_projection_style": "separate"} | {"organization": "deepseek-ai", "first_publish_date": "2025-01-20", "paper_urls": ["https://arxiv.org/abs/2501.12948"], "base_model": null, "license": "mit", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["qwen2.5 and llama3 series to the community.", "llama 70b", "gpt", "claude 3.5", "open-source models", "qwen 14b", "llama 3.3", "llama", "switch", "gpt 4o", "qwen 7b", "llama 3", "llama 8b", "qwen 1.5b", "llama 3.1", "llama3.1-8b-base and is originally licensed under", "claude", "qwen", "gpt 4", "qwen 2.5", "qwen 32b"], "hub_metadata": {"downloads": 965513, "likes": 1526, "created_at": "2025-01-20T09:19:00.000Z", "last_modified": "2025-02-24T03:31:29.000Z"}} | {"residual_stream_dim": 5120, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate_proj", "mlp_up_proj": "layers.36.mlp.up_proj", "mlp_down_proj": "layers.36.mlp.down_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate_proj", "mlp_up_proj": "layers.37.mlp.up_proj", "mlp_down_proj": "layers.37.mlp.down_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate_proj", "mlp_up_proj": "layers.38.mlp.up_proj", "mlp_down_proj": "layers.38.mlp.down_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate_proj", "mlp_up_proj": "layers.39.mlp.up_proj", "mlp_down_proj": "layers.39.mlp.down_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate_proj", "mlp_up_proj": "layers.40.mlp.up_proj", "mlp_down_proj": "layers.40.mlp.down_proj", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate_proj", "mlp_up_proj": "layers.41.mlp.up_proj", "mlp_down_proj": "layers.41.mlp.down_proj", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate_proj", "mlp_up_proj": "layers.42.mlp.up_proj", "mlp_down_proj": "layers.42.mlp.down_proj", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate_proj", "mlp_up_proj": "layers.43.mlp.up_proj", "mlp_down_proj": "layers.43.mlp.down_proj", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate_proj", "mlp_up_proj": "layers.44.mlp.up_proj", "mlp_down_proj": "layers.44.mlp.down_proj", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate_proj", "mlp_up_proj": "layers.45.mlp.up_proj", "mlp_down_proj": "layers.45.mlp.down_proj", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate_proj", "mlp_up_proj": "layers.46.mlp.up_proj", "mlp_down_proj": "layers.46.mlp.down_proj", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate_proj", "mlp_up_proj": "layers.47.mlp.up_proj", "mlp_down_proj": "layers.47.mlp.down_proj", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_q_proj": "layers.48.self_attn.q_proj", "attn_k_proj": "layers.48.self_attn.k_proj", "attn_v_proj": "layers.48.self_attn.v_proj", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate_proj", "mlp_up_proj": "layers.48.mlp.up_proj", "mlp_down_proj": "layers.48.mlp.down_proj", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_q_proj": "layers.49.self_attn.q_proj", "attn_k_proj": "layers.49.self_attn.k_proj", "attn_v_proj": "layers.49.self_attn.v_proj", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate_proj", "mlp_up_proj": "layers.49.mlp.up_proj", "mlp_down_proj": "layers.49.mlp.down_proj", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_q_proj": "layers.50.self_attn.q_proj", "attn_k_proj": "layers.50.self_attn.k_proj", "attn_v_proj": "layers.50.self_attn.v_proj", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate_proj", "mlp_up_proj": "layers.50.mlp.up_proj", "mlp_down_proj": "layers.50.mlp.down_proj", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate_proj", "mlp_up_proj": "layers.51.mlp.up_proj", "mlp_down_proj": "layers.51.mlp.down_proj", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_q_proj": "layers.52.self_attn.q_proj", "attn_k_proj": "layers.52.self_attn.k_proj", "attn_v_proj": "layers.52.self_attn.v_proj", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate_proj", "mlp_up_proj": "layers.52.mlp.up_proj", "mlp_down_proj": "layers.52.mlp.down_proj", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_q_proj": "layers.53.self_attn.q_proj", "attn_k_proj": "layers.53.self_attn.k_proj", "attn_v_proj": "layers.53.self_attn.v_proj", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate_proj", "mlp_up_proj": "layers.53.mlp.up_proj", "mlp_down_proj": "layers.53.mlp.down_proj", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_q_proj": "layers.54.self_attn.q_proj", "attn_k_proj": "layers.54.self_attn.k_proj", "attn_v_proj": "layers.54.self_attn.v_proj", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate_proj", "mlp_up_proj": "layers.54.mlp.up_proj", "mlp_down_proj": "layers.54.mlp.down_proj", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate_proj", "mlp_up_proj": "layers.55.mlp.up_proj", "mlp_down_proj": "layers.55.mlp.down_proj", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_q_proj": "layers.56.self_attn.q_proj", "attn_k_proj": "layers.56.self_attn.k_proj", "attn_v_proj": "layers.56.self_attn.v_proj", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate_proj", "mlp_up_proj": "layers.56.mlp.up_proj", "mlp_down_proj": "layers.56.mlp.down_proj", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_q_proj": "layers.57.self_attn.q_proj", "attn_k_proj": "layers.57.self_attn.k_proj", "attn_v_proj": "layers.57.self_attn.v_proj", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate_proj", "mlp_up_proj": "layers.57.mlp.up_proj", "mlp_down_proj": "layers.57.mlp.down_proj", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_q_proj": "layers.58.self_attn.q_proj", "attn_k_proj": "layers.58.self_attn.k_proj", "attn_v_proj": "layers.58.self_attn.v_proj", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate_proj", "mlp_up_proj": "layers.58.mlp.up_proj", "mlp_down_proj": "layers.58.mlp.down_proj", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate_proj", "mlp_up_proj": "layers.59.mlp.up_proj", "mlp_down_proj": "layers.59.mlp.down_proj", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_q_proj": "layers.60.self_attn.q_proj", "attn_k_proj": "layers.60.self_attn.k_proj", "attn_v_proj": "layers.60.self_attn.v_proj", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate_proj", "mlp_up_proj": "layers.60.mlp.up_proj", "mlp_down_proj": "layers.60.mlp.down_proj", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}, {"index": 61, "intervention_points": {"attn_input": "layers.61.self_attn.", "attn_q_proj": "layers.61.self_attn.q_proj", "attn_k_proj": "layers.61.self_attn.k_proj", "attn_v_proj": "layers.61.self_attn.v_proj", "attn_output_proj": "layers.61.self_attn.o_proj", "mlp_input": "layers.61.mlp.", "mlp_gate_proj": "layers.61.mlp.gate_proj", "mlp_up_proj": "layers.61.mlp.up_proj", "mlp_down_proj": "layers.61.mlp.down_proj", "attn_pre_norm": "layers.61.input_layernorm", "attn_post_norm": "layers.61.post_attention_layernorm"}}, {"index": 62, "intervention_points": {"attn_input": "layers.62.self_attn.", "attn_q_proj": "layers.62.self_attn.q_proj", "attn_k_proj": "layers.62.self_attn.k_proj", "attn_v_proj": "layers.62.self_attn.v_proj", "attn_output_proj": "layers.62.self_attn.o_proj", "mlp_input": "layers.62.mlp.", "mlp_gate_proj": "layers.62.mlp.gate_proj", "mlp_up_proj": "layers.62.mlp.up_proj", "mlp_down_proj": "layers.62.mlp.down_proj", "attn_pre_norm": "layers.62.input_layernorm", "attn_post_norm": "layers.62.post_attention_layernorm"}}, {"index": 63, "intervention_points": {"attn_input": "layers.63.self_attn.", "attn_q_proj": "layers.63.self_attn.q_proj", "attn_k_proj": "layers.63.self_attn.k_proj", "attn_v_proj": "layers.63.self_attn.v_proj", "attn_output_proj": "layers.63.self_attn.o_proj", "mlp_input": "layers.63.mlp.", "mlp_gate_proj": "layers.63.mlp.gate_proj", "mlp_up_proj": "layers.63.mlp.up_proj", "mlp_down_proj": "layers.63.mlp.down_proj", "attn_pre_norm": "layers.63.input_layernorm", "attn_post_norm": "layers.63.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.862302", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
deepseek-ai/DeepSeek-V2.5 | deepseek-ai | deepseek_v2 | 3 | 2026-03-29T15:32:56.833668+00:00 | {"hidden_size": 5120, "intermediate_size": 12288, "vocab_size": 102400, "max_position_embeddings": 163840, "num_hidden_layers": 60, "rope_theta": 10000, "rms_norm_eps": 1e-06, "model_type": "deepseek_v2"} | {"rope_type": null, "rope_scaling": {"beta_fast": 32, "beta_slow": 1, "factor": 40, "mscale": 1.0, "mscale_all_dim": 1.0, "original_max_position_embeddings": 4096, "type": "yarn", "rope_theta": 10000, "rope_type": "yarn"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "eager", "attention_architecture": "mla", "mla_q_lora_rank": 1536, "mla_kv_lora_rank": 512, "mla_qk_nope_head_dim": 128, "mla_qk_rope_head_dim": 64, "flash_attention_capable": true, "position_embedding_type": null, "max_position_embeddings": 163840, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "DeepseekV2RMSNorm", "linear_bias_present": false} | {"organization": "deepseek-ai", "first_publish_date": "2024-09-05", "paper_urls": ["https://arxiv.org/abs/2405.04434"], "base_model": null, "license": "other", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["your devices\nmax"], "hub_metadata": {"downloads": 5837, "likes": 733, "created_at": "2024-09-05T09:30:06.000Z", "last_modified": "2024-12-11T12:08:06.000Z"}} | {"residual_stream_dim": 5120, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate", "moe_gate": "layers.1.mlp.gate", "moe_experts": "layers.1.mlp.experts", "moe_shared_expert": "layers.1.mlp.shared_experts", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate", "moe_gate": "layers.2.mlp.gate", "moe_experts": "layers.2.mlp.experts", "moe_shared_expert": "layers.2.mlp.shared_experts", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "moe_shared_expert": "layers.3.mlp.shared_experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "moe_shared_expert": "layers.4.mlp.shared_experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "moe_shared_expert": "layers.5.mlp.shared_experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "moe_shared_expert": "layers.6.mlp.shared_experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "moe_shared_expert": "layers.7.mlp.shared_experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "moe_shared_expert": "layers.8.mlp.shared_experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "moe_shared_expert": "layers.9.mlp.shared_experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "moe_shared_expert": "layers.10.mlp.shared_experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "moe_shared_expert": "layers.11.mlp.shared_experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "moe_shared_expert": "layers.12.mlp.shared_experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "moe_shared_expert": "layers.13.mlp.shared_experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "moe_shared_expert": "layers.14.mlp.shared_experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "moe_shared_expert": "layers.15.mlp.shared_experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "moe_shared_expert": "layers.16.mlp.shared_experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "moe_shared_expert": "layers.17.mlp.shared_experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "moe_shared_expert": "layers.18.mlp.shared_experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "moe_shared_expert": "layers.19.mlp.shared_experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "moe_shared_expert": "layers.20.mlp.shared_experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "moe_shared_expert": "layers.21.mlp.shared_experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "moe_shared_expert": "layers.22.mlp.shared_experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "moe_shared_expert": "layers.23.mlp.shared_experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "moe_shared_expert": "layers.24.mlp.shared_experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "moe_shared_expert": "layers.25.mlp.shared_experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "moe_shared_expert": "layers.26.mlp.shared_experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "moe_shared_expert": "layers.27.mlp.shared_experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "moe_shared_expert": "layers.28.mlp.shared_experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "moe_shared_expert": "layers.29.mlp.shared_experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "moe_shared_expert": "layers.30.mlp.shared_experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "moe_shared_expert": "layers.31.mlp.shared_experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "moe_shared_expert": "layers.32.mlp.shared_experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "moe_shared_expert": "layers.33.mlp.shared_experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "moe_shared_expert": "layers.34.mlp.shared_experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "moe_shared_expert": "layers.35.mlp.shared_experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "moe_shared_expert": "layers.36.mlp.shared_experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "moe_shared_expert": "layers.37.mlp.shared_experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "moe_shared_expert": "layers.38.mlp.shared_experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "moe_shared_expert": "layers.39.mlp.shared_experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "moe_shared_expert": "layers.40.mlp.shared_experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "moe_shared_expert": "layers.41.mlp.shared_experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "moe_shared_expert": "layers.42.mlp.shared_experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "moe_shared_expert": "layers.43.mlp.shared_experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "moe_shared_expert": "layers.44.mlp.shared_experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "moe_shared_expert": "layers.45.mlp.shared_experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "moe_shared_expert": "layers.46.mlp.shared_experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "moe_shared_expert": "layers.47.mlp.shared_experts", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate", "moe_gate": "layers.48.mlp.gate", "moe_experts": "layers.48.mlp.experts", "moe_shared_expert": "layers.48.mlp.shared_experts", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate", "moe_gate": "layers.49.mlp.gate", "moe_experts": "layers.49.mlp.experts", "moe_shared_expert": "layers.49.mlp.shared_experts", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate", "moe_gate": "layers.50.mlp.gate", "moe_experts": "layers.50.mlp.experts", "moe_shared_expert": "layers.50.mlp.shared_experts", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate", "moe_gate": "layers.51.mlp.gate", "moe_experts": "layers.51.mlp.experts", "moe_shared_expert": "layers.51.mlp.shared_experts", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate", "moe_gate": "layers.52.mlp.gate", "moe_experts": "layers.52.mlp.experts", "moe_shared_expert": "layers.52.mlp.shared_experts", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate", "moe_gate": "layers.53.mlp.gate", "moe_experts": "layers.53.mlp.experts", "moe_shared_expert": "layers.53.mlp.shared_experts", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate", "moe_gate": "layers.54.mlp.gate", "moe_experts": "layers.54.mlp.experts", "moe_shared_expert": "layers.54.mlp.shared_experts", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate", "moe_gate": "layers.55.mlp.gate", "moe_experts": "layers.55.mlp.experts", "moe_shared_expert": "layers.55.mlp.shared_experts", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate", "moe_gate": "layers.56.mlp.gate", "moe_experts": "layers.56.mlp.experts", "moe_shared_expert": "layers.56.mlp.shared_experts", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate", "moe_gate": "layers.57.mlp.gate", "moe_experts": "layers.57.mlp.experts", "moe_shared_expert": "layers.57.mlp.shared_experts", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate", "moe_gate": "layers.58.mlp.gate", "moe_experts": "layers.58.mlp.experts", "moe_shared_expert": "layers.58.mlp.shared_experts", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate", "moe_gate": "layers.59.mlp.gate", "moe_experts": "layers.59.mlp.experts", "moe_shared_expert": "layers.59.mlp.shared_experts", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.862432", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
deepseek-ai/DeepSeek-V3-Base | deepseek-ai | deepseek_v3 | 3 | 2026-03-29T15:33:04.667941+00:00 | {"hidden_size": 7168, "intermediate_size": 18432, "vocab_size": 129280, "max_position_embeddings": 163840, "num_hidden_layers": 61, "rope_theta": 10000, "rms_norm_eps": 1e-06, "model_type": "deepseek_v3"} | {"rope_type": null, "rope_scaling": {"beta_fast": 32, "beta_slow": 1, "factor": 40, "mscale": 1.0, "mscale_all_dim": 1.0, "original_max_position_embeddings": 4096, "type": "yarn", "rope_theta": 10000, "rope_type": "yarn"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "eager", "attention_architecture": "mla", "mla_q_lora_rank": 1536, "mla_kv_lora_rank": 512, "mla_qk_nope_head_dim": 128, "mla_qk_rope_head_dim": 64, "flash_attention_capable": true, "position_embedding_type": null, "max_position_embeddings": 163840, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "DeepseekV3RMSNorm", "linear_bias_present": false} | {"organization": "deepseek-ai", "first_publish_date": "2024-12-25", "paper_urls": ["https://arxiv.org/abs/2412.19437"], "base_model": null, "license": null, "architecture_tags": [], "model_card_lineage": ["llama", "gpt", "claude 3.5", "gpt 4o", "internlm", "qwen 2.5", "llama 3.1", "claude"], "hub_metadata": {"downloads": 18178, "likes": 1684, "created_at": "2024-12-25T12:52:06.000Z", "last_modified": "2025-03-27T04:00:09.000Z"}} | {"residual_stream_dim": 7168, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "moe_shared_expert": "layers.3.mlp.shared_experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "moe_shared_expert": "layers.4.mlp.shared_experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "moe_shared_expert": "layers.5.mlp.shared_experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "moe_shared_expert": "layers.6.mlp.shared_experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "moe_shared_expert": "layers.7.mlp.shared_experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "moe_shared_expert": "layers.8.mlp.shared_experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "moe_shared_expert": "layers.9.mlp.shared_experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "moe_shared_expert": "layers.10.mlp.shared_experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "moe_shared_expert": "layers.11.mlp.shared_experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "moe_shared_expert": "layers.12.mlp.shared_experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "moe_shared_expert": "layers.13.mlp.shared_experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "moe_shared_expert": "layers.14.mlp.shared_experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "moe_shared_expert": "layers.15.mlp.shared_experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "moe_shared_expert": "layers.16.mlp.shared_experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "moe_shared_expert": "layers.17.mlp.shared_experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "moe_shared_expert": "layers.18.mlp.shared_experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "moe_shared_expert": "layers.19.mlp.shared_experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "moe_shared_expert": "layers.20.mlp.shared_experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "moe_shared_expert": "layers.21.mlp.shared_experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "moe_shared_expert": "layers.22.mlp.shared_experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "moe_shared_expert": "layers.23.mlp.shared_experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "moe_shared_expert": "layers.24.mlp.shared_experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "moe_shared_expert": "layers.25.mlp.shared_experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "moe_shared_expert": "layers.26.mlp.shared_experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "moe_shared_expert": "layers.27.mlp.shared_experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "moe_shared_expert": "layers.28.mlp.shared_experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "moe_shared_expert": "layers.29.mlp.shared_experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "moe_shared_expert": "layers.30.mlp.shared_experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "moe_shared_expert": "layers.31.mlp.shared_experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "moe_shared_expert": "layers.32.mlp.shared_experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "moe_shared_expert": "layers.33.mlp.shared_experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "moe_shared_expert": "layers.34.mlp.shared_experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "moe_shared_expert": "layers.35.mlp.shared_experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "moe_shared_expert": "layers.36.mlp.shared_experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "moe_shared_expert": "layers.37.mlp.shared_experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "moe_shared_expert": "layers.38.mlp.shared_experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "moe_shared_expert": "layers.39.mlp.shared_experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "moe_shared_expert": "layers.40.mlp.shared_experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "moe_shared_expert": "layers.41.mlp.shared_experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "moe_shared_expert": "layers.42.mlp.shared_experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "moe_shared_expert": "layers.43.mlp.shared_experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "moe_shared_expert": "layers.44.mlp.shared_experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "moe_shared_expert": "layers.45.mlp.shared_experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "moe_shared_expert": "layers.46.mlp.shared_experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "moe_shared_expert": "layers.47.mlp.shared_experts", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate", "moe_gate": "layers.48.mlp.gate", "moe_experts": "layers.48.mlp.experts", "moe_shared_expert": "layers.48.mlp.shared_experts", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate", "moe_gate": "layers.49.mlp.gate", "moe_experts": "layers.49.mlp.experts", "moe_shared_expert": "layers.49.mlp.shared_experts", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate", "moe_gate": "layers.50.mlp.gate", "moe_experts": "layers.50.mlp.experts", "moe_shared_expert": "layers.50.mlp.shared_experts", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate", "moe_gate": "layers.51.mlp.gate", "moe_experts": "layers.51.mlp.experts", "moe_shared_expert": "layers.51.mlp.shared_experts", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate", "moe_gate": "layers.52.mlp.gate", "moe_experts": "layers.52.mlp.experts", "moe_shared_expert": "layers.52.mlp.shared_experts", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate", "moe_gate": "layers.53.mlp.gate", "moe_experts": "layers.53.mlp.experts", "moe_shared_expert": "layers.53.mlp.shared_experts", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate", "moe_gate": "layers.54.mlp.gate", "moe_experts": "layers.54.mlp.experts", "moe_shared_expert": "layers.54.mlp.shared_experts", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate", "moe_gate": "layers.55.mlp.gate", "moe_experts": "layers.55.mlp.experts", "moe_shared_expert": "layers.55.mlp.shared_experts", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate", "moe_gate": "layers.56.mlp.gate", "moe_experts": "layers.56.mlp.experts", "moe_shared_expert": "layers.56.mlp.shared_experts", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate", "moe_gate": "layers.57.mlp.gate", "moe_experts": "layers.57.mlp.experts", "moe_shared_expert": "layers.57.mlp.shared_experts", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate", "moe_gate": "layers.58.mlp.gate", "moe_experts": "layers.58.mlp.experts", "moe_shared_expert": "layers.58.mlp.shared_experts", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate", "moe_gate": "layers.59.mlp.gate", "moe_experts": "layers.59.mlp.experts", "moe_shared_expert": "layers.59.mlp.shared_experts", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate", "moe_gate": "layers.60.mlp.gate", "moe_experts": "layers.60.mlp.experts", "moe_shared_expert": "layers.60.mlp.shared_experts", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.862572", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
XiaomiMiMo/MiMo-7B-RL | XiaomiMiMo | mimo | 3 | 2026-03-29T15:33:14.094468+00:00 | {"hidden_size": 4096, "intermediate_size": 11008, "vocab_size": 151680, "max_position_embeddings": 32768, "num_hidden_layers": 36, "rms_norm_eps": 1e-05, "model_type": "mimo"} | {"rope_type": null, "rope_scaling": {"rope_theta": 640000, "rope_type": "default"}, "attention_bias": true, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 32768, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-05, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen2RMSNorm", "linear_bias_present": true, "qkv_projection_style": "separate"} | {"organization": "XiaomiMiMo", "first_publish_date": "2025-04-29", "paper_urls": ["https://arxiv.org/abs/2505.07608"], "base_model": null, "license": "mit", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["claude 3.5", "qwen", "gpt 4.1", "claude", "qwen 14b", "qwen 7b", "gpt", "the model definition in the huggingface repo", "gpt 4o"], "hub_metadata": {"downloads": 44084, "likes": 276, "created_at": "2025-04-29T23:48:55.000Z", "last_modified": "2025-06-05T15:55:53.000Z"}} | {"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.862690", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
THUDM/glm-4-9b-chat-hf | THUDM | glm | 3 | 2026-03-29T15:33:18.543209+00:00 | {"hidden_size": 4096, "intermediate_size": 13696, "vocab_size": 151552, "max_position_embeddings": 131072, "num_hidden_layers": 40, "rms_norm_eps": 1.5625e-07, "model_type": "glm"} | {"rope_type": null, "rope_scaling": {"rope_theta": 10000.0, "partial_rotary_factor": 0.5, "rope_type": "default"}, "attention_bias": true, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 131072, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1.5625e-07, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "GlmRMSNorm", "linear_bias_present": true, "qkv_projection_style": "separate"} | {"organization": "THUDM", "first_publish_date": "2024-10-23", "paper_urls": [], "base_model": "THUDM/glm-4-9b-chat", "license": "other", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["chatglm 3", "glm-4-9b.", "llama", "claude", "qwen", "gpt 4", "glm 2024c", "chatglm", "claude 3", "gpt", "glm 4v", "glm", "glm 3", "glm 130b", "glm 4", "llama 3"], "hub_metadata": {"downloads": 37657, "likes": 24, "created_at": "2024-10-23T05:22:52.000Z", "last_modified": "2025-01-27T05:26:17.000Z"}} | {"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_down_proj": "layers.36.mlp.down_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_down_proj": "layers.37.mlp.down_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_down_proj": "layers.38.mlp.down_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_down_proj": "layers.39.mlp.down_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.862794", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
THUDM/glm-4-9b-hf | THUDM | glm | 3 | 2026-03-29T15:33:22.734695+00:00 | {"hidden_size": 4096, "intermediate_size": 13696, "vocab_size": 151552, "max_position_embeddings": 8192, "num_hidden_layers": 40, "rms_norm_eps": 1.5625e-07, "model_type": "glm"} | {"rope_type": null, "rope_scaling": {"rope_theta": 10000.0, "partial_rotary_factor": 0.5, "rope_type": "default"}, "attention_bias": true, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 8192, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1.5625e-07, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "GlmRMSNorm", "linear_bias_present": true, "qkv_projection_style": "separate"} | {"organization": "THUDM", "first_publish_date": "2025-01-16", "paper_urls": [], "base_model": null, "license": "other", "architecture_tags": ["text-generation"], "model_card_lineage": ["gpt 4", "glm 4v", "gpt", "glm 2024c", "glm-4-9b.", "chatglm 3", "llama", "glm 4", "glm", "qwen", "claude 3", "glm 130b", "chatglm", "glm 3", "claude", "llama 3"], "hub_metadata": {"downloads": 1380, "likes": 8, "created_at": "2025-01-16T06:02:41.000Z", "last_modified": "2025-01-27T06:33:22.000Z"}} | {"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_down_proj": "layers.36.mlp.down_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_down_proj": "layers.37.mlp.down_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_down_proj": "layers.38.mlp.down_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_down_proj": "layers.39.mlp.down_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.862891", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
zai-org/GLM-4-32B-0414 | zai-org | glm4 | 3 | 2026-03-29T15:33:26.619721+00:00 | {"hidden_size": 6144, "intermediate_size": 23040, "vocab_size": 151552, "max_position_embeddings": 32768, "num_hidden_layers": 61, "rms_norm_eps": 1e-05, "model_type": "glm4"} | {"rope_type": null, "rope_scaling": {"rope_theta": 10000.0, "partial_rotary_factor": 0.5, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 32768, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-05, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Glm4RMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "zai-org", "first_publish_date": "2025-04-07", "paper_urls": [], "base_model": null, "license": "mit", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["glm", "switch", "gpt 4o", "qwen 2.5", "search results", "glm 4", "pairwise ranking feedback", "the list of available tools.", "gpt", "academic research", "glm-4-32b-0414 through cold start"], "hub_metadata": {"downloads": 1724, "likes": 483, "created_at": "2025-04-07T12:53:41.000Z", "last_modified": "2025-05-01T15:42:37.000Z"}} | {"residual_stream_dim": 6144, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_down_proj": "layers.36.mlp.down_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_down_proj": "layers.37.mlp.down_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_down_proj": "layers.38.mlp.down_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_down_proj": "layers.39.mlp.down_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_down_proj": "layers.40.mlp.down_proj", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_down_proj": "layers.41.mlp.down_proj", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_down_proj": "layers.42.mlp.down_proj", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_down_proj": "layers.43.mlp.down_proj", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_down_proj": "layers.44.mlp.down_proj", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_down_proj": "layers.45.mlp.down_proj", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_down_proj": "layers.46.mlp.down_proj", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_down_proj": "layers.47.mlp.down_proj", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_q_proj": "layers.48.self_attn.q_proj", "attn_k_proj": "layers.48.self_attn.k_proj", "attn_v_proj": "layers.48.self_attn.v_proj", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_down_proj": "layers.48.mlp.down_proj", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_q_proj": "layers.49.self_attn.q_proj", "attn_k_proj": "layers.49.self_attn.k_proj", "attn_v_proj": "layers.49.self_attn.v_proj", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_down_proj": "layers.49.mlp.down_proj", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_q_proj": "layers.50.self_attn.q_proj", "attn_k_proj": "layers.50.self_attn.k_proj", "attn_v_proj": "layers.50.self_attn.v_proj", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_down_proj": "layers.50.mlp.down_proj", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_down_proj": "layers.51.mlp.down_proj", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_q_proj": "layers.52.self_attn.q_proj", "attn_k_proj": "layers.52.self_attn.k_proj", "attn_v_proj": "layers.52.self_attn.v_proj", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_down_proj": "layers.52.mlp.down_proj", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_q_proj": "layers.53.self_attn.q_proj", "attn_k_proj": "layers.53.self_attn.k_proj", "attn_v_proj": "layers.53.self_attn.v_proj", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_down_proj": "layers.53.mlp.down_proj", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_q_proj": "layers.54.self_attn.q_proj", "attn_k_proj": "layers.54.self_attn.k_proj", "attn_v_proj": "layers.54.self_attn.v_proj", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_down_proj": "layers.54.mlp.down_proj", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_down_proj": "layers.55.mlp.down_proj", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_q_proj": "layers.56.self_attn.q_proj", "attn_k_proj": "layers.56.self_attn.k_proj", "attn_v_proj": "layers.56.self_attn.v_proj", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_down_proj": "layers.56.mlp.down_proj", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_q_proj": "layers.57.self_attn.q_proj", "attn_k_proj": "layers.57.self_attn.k_proj", "attn_v_proj": "layers.57.self_attn.v_proj", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_down_proj": "layers.57.mlp.down_proj", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_q_proj": "layers.58.self_attn.q_proj", "attn_k_proj": "layers.58.self_attn.k_proj", "attn_v_proj": "layers.58.self_attn.v_proj", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_down_proj": "layers.58.mlp.down_proj", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_down_proj": "layers.59.mlp.down_proj", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_q_proj": "layers.60.self_attn.q_proj", "attn_k_proj": "layers.60.self_attn.k_proj", "attn_v_proj": "layers.60.self_attn.v_proj", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_down_proj": "layers.60.mlp.down_proj", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.863014", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
01-ai/Yi-1.5-9B-Chat | 01-ai | llama | 3 | 2026-03-29T15:33:30.543328+00:00 | {"hidden_size": 4096, "intermediate_size": 11008, "vocab_size": 64000, "max_position_embeddings": 4096, "num_hidden_layers": 48, "rms_norm_eps": 1e-06, "model_type": "llama"} | {"rope_type": null, "rope_scaling": {"rope_theta": 5000000.0, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 4096, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "LlamaRMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "01-ai", "first_publish_date": "2024-05-10", "paper_urls": ["https://arxiv.org/abs/2403.04652"], "base_model": null, "license": "apache-2.0", "architecture_tags": ["llama", "text-generation", "conversational"], "model_card_lineage": ["yi 1.5", "yi 15", "yi"], "hub_metadata": {"downloads": 18917, "likes": 148, "created_at": "2024-05-10T05:38:24.000Z", "last_modified": "2024-06-26T10:41:03.000Z"}} | {"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate_proj", "mlp_up_proj": "layers.36.mlp.up_proj", "mlp_down_proj": "layers.36.mlp.down_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate_proj", "mlp_up_proj": "layers.37.mlp.up_proj", "mlp_down_proj": "layers.37.mlp.down_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate_proj", "mlp_up_proj": "layers.38.mlp.up_proj", "mlp_down_proj": "layers.38.mlp.down_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate_proj", "mlp_up_proj": "layers.39.mlp.up_proj", "mlp_down_proj": "layers.39.mlp.down_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate_proj", "mlp_up_proj": "layers.40.mlp.up_proj", "mlp_down_proj": "layers.40.mlp.down_proj", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate_proj", "mlp_up_proj": "layers.41.mlp.up_proj", "mlp_down_proj": "layers.41.mlp.down_proj", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate_proj", "mlp_up_proj": "layers.42.mlp.up_proj", "mlp_down_proj": "layers.42.mlp.down_proj", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate_proj", "mlp_up_proj": "layers.43.mlp.up_proj", "mlp_down_proj": "layers.43.mlp.down_proj", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate_proj", "mlp_up_proj": "layers.44.mlp.up_proj", "mlp_down_proj": "layers.44.mlp.down_proj", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate_proj", "mlp_up_proj": "layers.45.mlp.up_proj", "mlp_down_proj": "layers.45.mlp.down_proj", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate_proj", "mlp_up_proj": "layers.46.mlp.up_proj", "mlp_down_proj": "layers.46.mlp.down_proj", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate_proj", "mlp_up_proj": "layers.47.mlp.up_proj", "mlp_down_proj": "layers.47.mlp.down_proj", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.863130", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
baichuan-inc/Baichuan2-13B-Chat | baichuan-inc | baichuan | 3 | 2026-03-29T15:33:34.877086+00:00 | {"hidden_size": 5120, "intermediate_size": 13696, "vocab_size": 125696, "num_hidden_layers": 40, "rms_norm_eps": 1e-06, "model_type": "baichuan"} | {"rope_type": null, "attention_bias": null, "sliding_window_size": null, "attention_implementation": "eager", "position_embedding_type": null, "max_position_embeddings": null, "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": null, "actual_norm_class": "RMSNorm", "linear_bias_present": false} | {"organization": "baichuan-inc", "first_publish_date": "2023-08-29", "paper_urls": [], "base_model": null, "license": "other", "architecture_tags": ["text-generation"], "model_card_lineage": ["llama", "baichuan 13b", "glm 2", "llama 13b", "gpt 4", "chatglm 2", "baichuan 7b", "baichuan 2", "gpt", "baichuan", "baichuan 2 models", "llama 2", "llama 7b", "gpt 3.5"], "hub_metadata": {"downloads": 11672, "likes": 432, "created_at": "2023-08-29T02:30:01.000Z", "last_modified": "2024-02-26T08:58:32.000Z"}} | {"residual_stream_dim": 5120, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate_proj", "mlp_up_proj": "layers.36.mlp.up_proj", "mlp_down_proj": "layers.36.mlp.down_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate_proj", "mlp_up_proj": "layers.37.mlp.up_proj", "mlp_down_proj": "layers.37.mlp.down_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate_proj", "mlp_up_proj": "layers.38.mlp.up_proj", "mlp_down_proj": "layers.38.mlp.down_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate_proj", "mlp_up_proj": "layers.39.mlp.up_proj", "mlp_down_proj": "layers.39.mlp.down_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.863225", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
ByteDance-Seed/Seed-Coder-8B-Instruct | ByteDance-Seed | llama | 3 | 2026-03-29T15:33:38.656812+00:00 | {"hidden_size": 4096, "intermediate_size": 14336, "vocab_size": 155136, "max_position_embeddings": 32768, "num_hidden_layers": 32, "rms_norm_eps": 1e-06, "model_type": "llama"} | {"rope_type": null, "rope_scaling": {"rope_theta": 500000.0, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 32768, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": 0.1, "attention_dropout": 0.1, "actual_norm_class": "LlamaRMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "ByteDance-Seed", "first_publish_date": "2025-04-27", "paper_urls": ["https://arxiv.org/abs/2506.03524"], "base_model": "ByteDance-Seed/Seed-Coder-8B-Base", "license": "mit", "architecture_tags": ["llama", "text-generation", "conversational"], "model_card_lineage": ["yi", "llama", "qwen 1.5", "codellama", "llama 3.1", "qwen 3", "llama 7b", "codellama 7b", "qwen 2.5"], "hub_metadata": {"downloads": 7704, "likes": 110, "created_at": "2025-04-27T07:52:37.000Z", "last_modified": "2025-06-06T02:18:40.000Z"}} | {"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.863320", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
moonshotai/Kimi-K2-Instruct | moonshotai | kimi_k2 | 3 | 2026-03-29T15:57:11.098748+00:00 | {"hidden_size": 7168, "intermediate_size": 18432, "vocab_size": 163840, "max_position_embeddings": 131072, "num_hidden_layers": 61, "rope_theta": 50000.0, "rms_norm_eps": 1e-06, "model_type": "kimi_k2"} | {"rope_type": null, "rope_scaling": {"beta_fast": 1.0, "beta_slow": 1.0, "factor": 32.0, "mscale": 1.0, "mscale_all_dim": 1.0, "original_max_position_embeddings": 4096, "type": "yarn", "rope_theta": 50000.0, "rope_type": "yarn"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "eager", "attention_architecture": "mla", "mla_q_lora_rank": 1536, "mla_kv_lora_rank": 512, "mla_qk_nope_head_dim": 128, "mla_qk_rope_head_dim": 64, "flash_attention_capable": true, "position_embedding_type": null, "max_position_embeddings": 131072, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "DeepseekV3RMSNorm", "linear_bias_present": false} | {"organization": "moonshotai", "first_publish_date": "2025-07-11", "paper_urls": [], "base_model": null, "license": "other", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["qwen 3", "llama 4", "claude", "gpt", "llama", "qwen 2.5", "gpt 4.1"], "hub_metadata": {"downloads": 113628, "likes": 2330, "created_at": "2025-07-11T00:55:12.000Z", "last_modified": "2026-01-30T04:50:07.000Z"}} | {"residual_stream_dim": 7168, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate", "moe_gate": "layers.1.mlp.gate", "moe_experts": "layers.1.mlp.experts", "moe_shared_expert": "layers.1.mlp.shared_experts", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate", "moe_gate": "layers.2.mlp.gate", "moe_experts": "layers.2.mlp.experts", "moe_shared_expert": "layers.2.mlp.shared_experts", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "moe_shared_expert": "layers.3.mlp.shared_experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "moe_shared_expert": "layers.4.mlp.shared_experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "moe_shared_expert": "layers.5.mlp.shared_experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "moe_shared_expert": "layers.6.mlp.shared_experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "moe_shared_expert": "layers.7.mlp.shared_experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "moe_shared_expert": "layers.8.mlp.shared_experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "moe_shared_expert": "layers.9.mlp.shared_experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "moe_shared_expert": "layers.10.mlp.shared_experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "moe_shared_expert": "layers.11.mlp.shared_experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "moe_shared_expert": "layers.12.mlp.shared_experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "moe_shared_expert": "layers.13.mlp.shared_experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "moe_shared_expert": "layers.14.mlp.shared_experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "moe_shared_expert": "layers.15.mlp.shared_experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "moe_shared_expert": "layers.16.mlp.shared_experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "moe_shared_expert": "layers.17.mlp.shared_experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "moe_shared_expert": "layers.18.mlp.shared_experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "moe_shared_expert": "layers.19.mlp.shared_experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "moe_shared_expert": "layers.20.mlp.shared_experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "moe_shared_expert": "layers.21.mlp.shared_experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "moe_shared_expert": "layers.22.mlp.shared_experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "moe_shared_expert": "layers.23.mlp.shared_experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "moe_shared_expert": "layers.24.mlp.shared_experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "moe_shared_expert": "layers.25.mlp.shared_experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "moe_shared_expert": "layers.26.mlp.shared_experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "moe_shared_expert": "layers.27.mlp.shared_experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "moe_shared_expert": "layers.28.mlp.shared_experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "moe_shared_expert": "layers.29.mlp.shared_experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "moe_shared_expert": "layers.30.mlp.shared_experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "moe_shared_expert": "layers.31.mlp.shared_experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "moe_shared_expert": "layers.32.mlp.shared_experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "moe_shared_expert": "layers.33.mlp.shared_experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "moe_shared_expert": "layers.34.mlp.shared_experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "moe_shared_expert": "layers.35.mlp.shared_experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "moe_shared_expert": "layers.36.mlp.shared_experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "moe_shared_expert": "layers.37.mlp.shared_experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "moe_shared_expert": "layers.38.mlp.shared_experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "moe_shared_expert": "layers.39.mlp.shared_experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "moe_shared_expert": "layers.40.mlp.shared_experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "moe_shared_expert": "layers.41.mlp.shared_experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "moe_shared_expert": "layers.42.mlp.shared_experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "moe_shared_expert": "layers.43.mlp.shared_experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "moe_shared_expert": "layers.44.mlp.shared_experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "moe_shared_expert": "layers.45.mlp.shared_experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "moe_shared_expert": "layers.46.mlp.shared_experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "moe_shared_expert": "layers.47.mlp.shared_experts", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate", "moe_gate": "layers.48.mlp.gate", "moe_experts": "layers.48.mlp.experts", "moe_shared_expert": "layers.48.mlp.shared_experts", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate", "moe_gate": "layers.49.mlp.gate", "moe_experts": "layers.49.mlp.experts", "moe_shared_expert": "layers.49.mlp.shared_experts", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate", "moe_gate": "layers.50.mlp.gate", "moe_experts": "layers.50.mlp.experts", "moe_shared_expert": "layers.50.mlp.shared_experts", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate", "moe_gate": "layers.51.mlp.gate", "moe_experts": "layers.51.mlp.experts", "moe_shared_expert": "layers.51.mlp.shared_experts", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate", "moe_gate": "layers.52.mlp.gate", "moe_experts": "layers.52.mlp.experts", "moe_shared_expert": "layers.52.mlp.shared_experts", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate", "moe_gate": "layers.53.mlp.gate", "moe_experts": "layers.53.mlp.experts", "moe_shared_expert": "layers.53.mlp.shared_experts", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate", "moe_gate": "layers.54.mlp.gate", "moe_experts": "layers.54.mlp.experts", "moe_shared_expert": "layers.54.mlp.shared_experts", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate", "moe_gate": "layers.55.mlp.gate", "moe_experts": "layers.55.mlp.experts", "moe_shared_expert": "layers.55.mlp.shared_experts", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate", "moe_gate": "layers.56.mlp.gate", "moe_experts": "layers.56.mlp.experts", "moe_shared_expert": "layers.56.mlp.shared_experts", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate", "moe_gate": "layers.57.mlp.gate", "moe_experts": "layers.57.mlp.experts", "moe_shared_expert": "layers.57.mlp.shared_experts", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate", "moe_gate": "layers.58.mlp.gate", "moe_experts": "layers.58.mlp.experts", "moe_shared_expert": "layers.58.mlp.shared_experts", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate", "moe_gate": "layers.59.mlp.gate", "moe_experts": "layers.59.mlp.experts", "moe_shared_expert": "layers.59.mlp.shared_experts", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate", "moe_gate": "layers.60.mlp.gate", "moe_experts": "layers.60.mlp.experts", "moe_shared_expert": "layers.60.mlp.shared_experts", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.863446", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
MiniMaxAI/MiniMax-M2.1 | MiniMaxAI | minimax_m2 | 3 | 2026-03-29T15:33:55.295821+00:00 | {"hidden_size": 3072, "intermediate_size": 1536, "vocab_size": 200064, "max_position_embeddings": 196608, "num_hidden_layers": 62, "rope_theta": 5000000, "rms_norm_eps": 1e-06, "model_type": "minimax_m2"} | {"rope_type": null, "attention_bias": null, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 196608, "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "MiniMaxM2RMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "MiniMaxAI", "first_publish_date": "2025-12-20", "paper_urls": ["https://arxiv.org/abs/2509.06501"], "base_model": null, "license": "other", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["gpt 5.2", "minimax-m2.1", "gpt", "claude"], "hub_metadata": {"downloads": 41782, "likes": 1271, "created_at": "2025-12-20T05:45:05.000Z", "last_modified": "2026-02-13T09:24:21.000Z"}} | {"residual_stream_dim": 3072, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_q_proj": "layers.48.self_attn.q_proj", "attn_k_proj": "layers.48.self_attn.k_proj", "attn_v_proj": "layers.48.self_attn.v_proj", "attn_output_proj": "layers.48.self_attn.o_proj", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_q_proj": "layers.49.self_attn.q_proj", "attn_k_proj": "layers.49.self_attn.k_proj", "attn_v_proj": "layers.49.self_attn.v_proj", "attn_output_proj": "layers.49.self_attn.o_proj", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_q_proj": "layers.50.self_attn.q_proj", "attn_k_proj": "layers.50.self_attn.k_proj", "attn_v_proj": "layers.50.self_attn.v_proj", "attn_output_proj": "layers.50.self_attn.o_proj", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_q_proj": "layers.52.self_attn.q_proj", "attn_k_proj": "layers.52.self_attn.k_proj", "attn_v_proj": "layers.52.self_attn.v_proj", "attn_output_proj": "layers.52.self_attn.o_proj", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_q_proj": "layers.53.self_attn.q_proj", "attn_k_proj": "layers.53.self_attn.k_proj", "attn_v_proj": "layers.53.self_attn.v_proj", "attn_output_proj": "layers.53.self_attn.o_proj", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_q_proj": "layers.54.self_attn.q_proj", "attn_k_proj": "layers.54.self_attn.k_proj", "attn_v_proj": "layers.54.self_attn.v_proj", "attn_output_proj": "layers.54.self_attn.o_proj", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_q_proj": "layers.56.self_attn.q_proj", "attn_k_proj": "layers.56.self_attn.k_proj", "attn_v_proj": "layers.56.self_attn.v_proj", "attn_output_proj": "layers.56.self_attn.o_proj", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_q_proj": "layers.57.self_attn.q_proj", "attn_k_proj": "layers.57.self_attn.k_proj", "attn_v_proj": "layers.57.self_attn.v_proj", "attn_output_proj": "layers.57.self_attn.o_proj", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_q_proj": "layers.58.self_attn.q_proj", "attn_k_proj": "layers.58.self_attn.k_proj", "attn_v_proj": "layers.58.self_attn.v_proj", "attn_output_proj": "layers.58.self_attn.o_proj", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_q_proj": "layers.60.self_attn.q_proj", "attn_k_proj": "layers.60.self_attn.k_proj", "attn_v_proj": "layers.60.self_attn.v_proj", "attn_output_proj": "layers.60.self_attn.o_proj", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}, {"index": 61, "intervention_points": {"attn_input": "layers.61.self_attn.", "attn_q_proj": "layers.61.self_attn.q_proj", "attn_k_proj": "layers.61.self_attn.k_proj", "attn_v_proj": "layers.61.self_attn.v_proj", "attn_output_proj": "layers.61.self_attn.o_proj", "attn_pre_norm": "layers.61.input_layernorm", "attn_post_norm": "layers.61.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.863582", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
Qwen/Qwen3.5-9B | Qwen | qwen3_5_text | 3 | 2026-03-29T15:43:11.356013+00:00 | {"hidden_size": 4096, "intermediate_size": 12288, "vocab_size": 248320, "max_position_embeddings": 262144, "num_hidden_layers": 32, "rms_norm_eps": 1e-06, "model_type": "qwen3_5_text"} | {"rope_type": null, "rope_scaling": {"mrope_interleaved": true, "mrope_section": [11, 11, 10], "rope_type": "default", "rope_theta": 10000000, "partial_rotary_factor": 0.25}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 262144, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "['Qwen3_5RMSNorm', 'Qwen3_5RMSNormGated']", "linear_bias_present": false, "qkv_projection_style": "fused"} | {"organization": "Qwen", "first_publish_date": "2026-02-27", "paper_urls": [], "base_model": "Qwen/Qwen3.5-9B-Base", "license": "apache-2.0", "architecture_tags": ["conversational"], "model_card_lineage": ["qwen 3", "claude", "gpt", "qwen 3.5", "qwen", "switch", "gpt 5"], "hub_metadata": {"downloads": 4286464, "likes": 1083, "created_at": "2026-02-27T12:58:26.000Z", "last_modified": "2026-03-02T00:51:43.000Z"}} | {"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.863670", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
Qwen/Qwen3.5-35B-A3B | Qwen | qwen3_5_moe_text | 3 | 2026-03-29T15:43:22.406730+00:00 | {"hidden_size": 2048, "vocab_size": 248320, "max_position_embeddings": 262144, "num_hidden_layers": 40, "rms_norm_eps": 1e-06, "model_type": "qwen3_5_moe_text"} | {"rope_type": null, "rope_scaling": {"mrope_interleaved": true, "mrope_section": [11, 11, 10], "rope_type": "default", "rope_theta": 10000000, "partial_rotary_factor": 0.25}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 262144, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "['Qwen3_5MoeRMSNorm', 'Qwen3_5MoeRMSNormGated']", "linear_bias_present": false, "qkv_projection_style": "fused"} | {"organization": "Qwen", "first_publish_date": "2026-02-24", "paper_urls": [], "base_model": "Qwen/Qwen3.5-35B-A3B-Base", "license": "apache-2.0", "architecture_tags": ["conversational"], "model_card_lineage": ["qwen", "gpt 5", "claude", "gpt", "our model adopt a simple context-folding strategy", "qwen 3", "qwen 3.5", "switch"], "hub_metadata": {"downloads": 3031012, "likes": 1285, "created_at": "2026-02-24T09:39:25.000Z", "last_modified": "2026-02-27T09:48:22.000Z"}} | {"residual_stream_dim": 2048, "layers": [{"index": 0, "intervention_points": {"mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate", "moe_gate": "layers.0.mlp.gate", "moe_experts": "layers.0.mlp.experts", "moe_shared_expert": "layers.0.mlp.shared_expert", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate", "moe_gate": "layers.1.mlp.gate", "moe_experts": "layers.1.mlp.experts", "moe_shared_expert": "layers.1.mlp.shared_expert", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate", "moe_gate": "layers.2.mlp.gate", "moe_experts": "layers.2.mlp.experts", "moe_shared_expert": "layers.2.mlp.shared_expert", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "moe_shared_expert": "layers.3.mlp.shared_expert", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "moe_shared_expert": "layers.4.mlp.shared_expert", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "moe_shared_expert": "layers.5.mlp.shared_expert", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "moe_shared_expert": "layers.6.mlp.shared_expert", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "moe_shared_expert": "layers.7.mlp.shared_expert", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "moe_shared_expert": "layers.8.mlp.shared_expert", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "moe_shared_expert": "layers.9.mlp.shared_expert", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "moe_shared_expert": "layers.10.mlp.shared_expert", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "moe_shared_expert": "layers.11.mlp.shared_expert", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "moe_shared_expert": "layers.12.mlp.shared_expert", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "moe_shared_expert": "layers.13.mlp.shared_expert", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "moe_shared_expert": "layers.14.mlp.shared_expert", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "moe_shared_expert": "layers.15.mlp.shared_expert", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "moe_shared_expert": "layers.16.mlp.shared_expert", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "moe_shared_expert": "layers.17.mlp.shared_expert", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "moe_shared_expert": "layers.18.mlp.shared_expert", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "moe_shared_expert": "layers.19.mlp.shared_expert", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "moe_shared_expert": "layers.20.mlp.shared_expert", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "moe_shared_expert": "layers.21.mlp.shared_expert", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "moe_shared_expert": "layers.22.mlp.shared_expert", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "moe_shared_expert": "layers.23.mlp.shared_expert", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "moe_shared_expert": "layers.24.mlp.shared_expert", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "moe_shared_expert": "layers.25.mlp.shared_expert", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "moe_shared_expert": "layers.26.mlp.shared_expert", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "moe_shared_expert": "layers.27.mlp.shared_expert", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "moe_shared_expert": "layers.28.mlp.shared_expert", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "moe_shared_expert": "layers.29.mlp.shared_expert", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "moe_shared_expert": "layers.30.mlp.shared_expert", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "moe_shared_expert": "layers.31.mlp.shared_expert", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "moe_shared_expert": "layers.32.mlp.shared_expert", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "moe_shared_expert": "layers.33.mlp.shared_expert", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "moe_shared_expert": "layers.34.mlp.shared_expert", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "moe_shared_expert": "layers.35.mlp.shared_expert", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "moe_shared_expert": "layers.36.mlp.shared_expert", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "moe_shared_expert": "layers.37.mlp.shared_expert", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "moe_shared_expert": "layers.38.mlp.shared_expert", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "moe_shared_expert": "layers.39.mlp.shared_expert", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.863769", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
Qwen/Qwen3.5-397B-A17B | Qwen | qwen3_5_moe_text | 3 | 2026-03-29T15:43:32.479492+00:00 | {"hidden_size": 4096, "vocab_size": 248320, "max_position_embeddings": 262144, "num_hidden_layers": 60, "rms_norm_eps": 1e-06, "model_type": "qwen3_5_moe_text"} | {"rope_type": null, "rope_scaling": {"mrope_interleaved": true, "mrope_section": [11, 11, 10], "rope_type": "default", "rope_theta": 10000000, "partial_rotary_factor": 0.25}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 262144, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "['Qwen3_5MoeRMSNorm', 'Qwen3_5MoeRMSNormGated']", "linear_bias_present": false, "qkv_projection_style": "fused"} | {"organization": "Qwen", "first_publish_date": "2026-02-16", "paper_urls": [], "base_model": null, "license": "apache-2.0", "architecture_tags": ["conversational"], "model_card_lineage": ["qwen 3", "switch", "qwen", "gpt 5.2", "claude 4.5", "our model adopt a simple context-folding strategy", "claude", "qwen 3.5"], "hub_metadata": {"downloads": 1381182, "likes": 1387, "created_at": "2026-02-16T04:55:12.000Z", "last_modified": "2026-03-15T07:47:35.000Z"}} | {"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate", "moe_gate": "layers.0.mlp.gate", "moe_experts": "layers.0.mlp.experts", "moe_shared_expert": "layers.0.mlp.shared_expert", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate", "moe_gate": "layers.1.mlp.gate", "moe_experts": "layers.1.mlp.experts", "moe_shared_expert": "layers.1.mlp.shared_expert", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate", "moe_gate": "layers.2.mlp.gate", "moe_experts": "layers.2.mlp.experts", "moe_shared_expert": "layers.2.mlp.shared_expert", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "moe_shared_expert": "layers.3.mlp.shared_expert", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "moe_shared_expert": "layers.4.mlp.shared_expert", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "moe_shared_expert": "layers.5.mlp.shared_expert", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "moe_shared_expert": "layers.6.mlp.shared_expert", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "moe_shared_expert": "layers.7.mlp.shared_expert", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "moe_shared_expert": "layers.8.mlp.shared_expert", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "moe_shared_expert": "layers.9.mlp.shared_expert", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "moe_shared_expert": "layers.10.mlp.shared_expert", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "moe_shared_expert": "layers.11.mlp.shared_expert", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "moe_shared_expert": "layers.12.mlp.shared_expert", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "moe_shared_expert": "layers.13.mlp.shared_expert", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "moe_shared_expert": "layers.14.mlp.shared_expert", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "moe_shared_expert": "layers.15.mlp.shared_expert", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "moe_shared_expert": "layers.16.mlp.shared_expert", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "moe_shared_expert": "layers.17.mlp.shared_expert", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "moe_shared_expert": "layers.18.mlp.shared_expert", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "moe_shared_expert": "layers.19.mlp.shared_expert", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "moe_shared_expert": "layers.20.mlp.shared_expert", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "moe_shared_expert": "layers.21.mlp.shared_expert", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "moe_shared_expert": "layers.22.mlp.shared_expert", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "moe_shared_expert": "layers.23.mlp.shared_expert", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "moe_shared_expert": "layers.24.mlp.shared_expert", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "moe_shared_expert": "layers.25.mlp.shared_expert", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "moe_shared_expert": "layers.26.mlp.shared_expert", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "moe_shared_expert": "layers.27.mlp.shared_expert", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "moe_shared_expert": "layers.28.mlp.shared_expert", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "moe_shared_expert": "layers.29.mlp.shared_expert", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "moe_shared_expert": "layers.30.mlp.shared_expert", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "moe_shared_expert": "layers.31.mlp.shared_expert", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "moe_shared_expert": "layers.32.mlp.shared_expert", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "moe_shared_expert": "layers.33.mlp.shared_expert", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "moe_shared_expert": "layers.34.mlp.shared_expert", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "moe_shared_expert": "layers.35.mlp.shared_expert", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "moe_shared_expert": "layers.36.mlp.shared_expert", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "moe_shared_expert": "layers.37.mlp.shared_expert", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "moe_shared_expert": "layers.38.mlp.shared_expert", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "moe_shared_expert": "layers.39.mlp.shared_expert", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "moe_shared_expert": "layers.40.mlp.shared_expert", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "moe_shared_expert": "layers.41.mlp.shared_expert", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "moe_shared_expert": "layers.42.mlp.shared_expert", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "moe_shared_expert": "layers.43.mlp.shared_expert", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "moe_shared_expert": "layers.44.mlp.shared_expert", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "moe_shared_expert": "layers.45.mlp.shared_expert", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "moe_shared_expert": "layers.46.mlp.shared_expert", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "moe_shared_expert": "layers.47.mlp.shared_expert", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate", "moe_gate": "layers.48.mlp.gate", "moe_experts": "layers.48.mlp.experts", "moe_shared_expert": "layers.48.mlp.shared_expert", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate", "moe_gate": "layers.49.mlp.gate", "moe_experts": "layers.49.mlp.experts", "moe_shared_expert": "layers.49.mlp.shared_expert", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate", "moe_gate": "layers.50.mlp.gate", "moe_experts": "layers.50.mlp.experts", "moe_shared_expert": "layers.50.mlp.shared_expert", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate", "moe_gate": "layers.51.mlp.gate", "moe_experts": "layers.51.mlp.experts", "moe_shared_expert": "layers.51.mlp.shared_expert", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate", "moe_gate": "layers.52.mlp.gate", "moe_experts": "layers.52.mlp.experts", "moe_shared_expert": "layers.52.mlp.shared_expert", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate", "moe_gate": "layers.53.mlp.gate", "moe_experts": "layers.53.mlp.experts", "moe_shared_expert": "layers.53.mlp.shared_expert", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate", "moe_gate": "layers.54.mlp.gate", "moe_experts": "layers.54.mlp.experts", "moe_shared_expert": "layers.54.mlp.shared_expert", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate", "moe_gate": "layers.55.mlp.gate", "moe_experts": "layers.55.mlp.experts", "moe_shared_expert": "layers.55.mlp.shared_expert", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate", "moe_gate": "layers.56.mlp.gate", "moe_experts": "layers.56.mlp.experts", "moe_shared_expert": "layers.56.mlp.shared_expert", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate", "moe_gate": "layers.57.mlp.gate", "moe_experts": "layers.57.mlp.experts", "moe_shared_expert": "layers.57.mlp.shared_expert", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate", "moe_gate": "layers.58.mlp.gate", "moe_experts": "layers.58.mlp.experts", "moe_shared_expert": "layers.58.mlp.shared_expert", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate", "moe_gate": "layers.59.mlp.gate", "moe_experts": "layers.59.mlp.experts", "moe_shared_expert": "layers.59.mlp.shared_expert", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.863891", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
Qwen/Qwen3.5-27B | Qwen | qwen3_5_text | 3 | 2026-03-29T15:43:42.293743+00:00 | {"hidden_size": 5120, "intermediate_size": 17408, "vocab_size": 248320, "max_position_embeddings": 262144, "num_hidden_layers": 64, "rms_norm_eps": 1e-06, "model_type": "qwen3_5_text"} | {"rope_type": null, "rope_scaling": {"mrope_interleaved": true, "mrope_section": [11, 11, 10], "rope_type": "default", "rope_theta": 10000000, "partial_rotary_factor": 0.25}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 262144, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "['Qwen3_5RMSNorm', 'Qwen3_5RMSNormGated']", "linear_bias_present": false, "qkv_projection_style": "fused"} | {"organization": "Qwen", "first_publish_date": "2026-02-24", "paper_urls": [], "base_model": null, "license": "apache-2.0", "architecture_tags": ["conversational"], "model_card_lineage": ["claude", "qwen 3.5", "gpt", "qwen", "switch", "our model adopt a simple context-folding strategy", "qwen 3", "gpt 5"], "hub_metadata": {"downloads": 2651267, "likes": 797, "created_at": "2026-02-24T09:41:56.000Z", "last_modified": "2026-02-25T02:43:25.000Z"}} | {"residual_stream_dim": 5120, "layers": [{"index": 0, "intervention_points": {"mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate_proj", "mlp_up_proj": "layers.36.mlp.up_proj", "mlp_down_proj": "layers.36.mlp.down_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate_proj", "mlp_up_proj": "layers.37.mlp.up_proj", "mlp_down_proj": "layers.37.mlp.down_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate_proj", "mlp_up_proj": "layers.38.mlp.up_proj", "mlp_down_proj": "layers.38.mlp.down_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate_proj", "mlp_up_proj": "layers.39.mlp.up_proj", "mlp_down_proj": "layers.39.mlp.down_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate_proj", "mlp_up_proj": "layers.40.mlp.up_proj", "mlp_down_proj": "layers.40.mlp.down_proj", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate_proj", "mlp_up_proj": "layers.41.mlp.up_proj", "mlp_down_proj": "layers.41.mlp.down_proj", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate_proj", "mlp_up_proj": "layers.42.mlp.up_proj", "mlp_down_proj": "layers.42.mlp.down_proj", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate_proj", "mlp_up_proj": "layers.43.mlp.up_proj", "mlp_down_proj": "layers.43.mlp.down_proj", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate_proj", "mlp_up_proj": "layers.44.mlp.up_proj", "mlp_down_proj": "layers.44.mlp.down_proj", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate_proj", "mlp_up_proj": "layers.45.mlp.up_proj", "mlp_down_proj": "layers.45.mlp.down_proj", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate_proj", "mlp_up_proj": "layers.46.mlp.up_proj", "mlp_down_proj": "layers.46.mlp.down_proj", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate_proj", "mlp_up_proj": "layers.47.mlp.up_proj", "mlp_down_proj": "layers.47.mlp.down_proj", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate_proj", "mlp_up_proj": "layers.48.mlp.up_proj", "mlp_down_proj": "layers.48.mlp.down_proj", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate_proj", "mlp_up_proj": "layers.49.mlp.up_proj", "mlp_down_proj": "layers.49.mlp.down_proj", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate_proj", "mlp_up_proj": "layers.50.mlp.up_proj", "mlp_down_proj": "layers.50.mlp.down_proj", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate_proj", "mlp_up_proj": "layers.51.mlp.up_proj", "mlp_down_proj": "layers.51.mlp.down_proj", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate_proj", "mlp_up_proj": "layers.52.mlp.up_proj", "mlp_down_proj": "layers.52.mlp.down_proj", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate_proj", "mlp_up_proj": "layers.53.mlp.up_proj", "mlp_down_proj": "layers.53.mlp.down_proj", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate_proj", "mlp_up_proj": "layers.54.mlp.up_proj", "mlp_down_proj": "layers.54.mlp.down_proj", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate_proj", "mlp_up_proj": "layers.55.mlp.up_proj", "mlp_down_proj": "layers.55.mlp.down_proj", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate_proj", "mlp_up_proj": "layers.56.mlp.up_proj", "mlp_down_proj": "layers.56.mlp.down_proj", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate_proj", "mlp_up_proj": "layers.57.mlp.up_proj", "mlp_down_proj": "layers.57.mlp.down_proj", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate_proj", "mlp_up_proj": "layers.58.mlp.up_proj", "mlp_down_proj": "layers.58.mlp.down_proj", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate_proj", "mlp_up_proj": "layers.59.mlp.up_proj", "mlp_down_proj": "layers.59.mlp.down_proj", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate_proj", "mlp_up_proj": "layers.60.mlp.up_proj", "mlp_down_proj": "layers.60.mlp.down_proj", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}, {"index": 61, "intervention_points": {"mlp_input": "layers.61.mlp.", "mlp_gate_proj": "layers.61.mlp.gate_proj", "mlp_up_proj": "layers.61.mlp.up_proj", "mlp_down_proj": "layers.61.mlp.down_proj", "attn_pre_norm": "layers.61.input_layernorm", "attn_post_norm": "layers.61.post_attention_layernorm"}}, {"index": 62, "intervention_points": {"mlp_input": "layers.62.mlp.", "mlp_gate_proj": "layers.62.mlp.gate_proj", "mlp_up_proj": "layers.62.mlp.up_proj", "mlp_down_proj": "layers.62.mlp.down_proj", "attn_pre_norm": "layers.62.input_layernorm", "attn_post_norm": "layers.62.post_attention_layernorm"}}, {"index": 63, "intervention_points": {"attn_input": "layers.63.self_attn.", "attn_q_proj": "layers.63.self_attn.q_proj", "attn_k_proj": "layers.63.self_attn.k_proj", "attn_v_proj": "layers.63.self_attn.v_proj", "attn_output_proj": "layers.63.self_attn.o_proj", "mlp_input": "layers.63.mlp.", "mlp_gate_proj": "layers.63.mlp.gate_proj", "mlp_up_proj": "layers.63.mlp.up_proj", "mlp_down_proj": "layers.63.mlp.down_proj", "attn_pre_norm": "layers.63.input_layernorm", "attn_post_norm": "layers.63.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.864010", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
deepseek-ai/DeepSeek-V3.1 | deepseek-ai | deepseek_v3 | 3 | 2026-03-29T15:43:57.325330+00:00 | {"hidden_size": 7168, "intermediate_size": 18432, "vocab_size": 129280, "max_position_embeddings": 163840, "num_hidden_layers": 61, "rope_theta": 10000, "rms_norm_eps": 1e-06, "model_type": "deepseek_v3"} | {"rope_type": null, "rope_scaling": {"beta_fast": 32, "beta_slow": 1, "factor": 40, "mscale": 1.0, "mscale_all_dim": 1.0, "original_max_position_embeddings": 4096, "type": "yarn", "rope_theta": 10000, "rope_type": "yarn"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "eager", "attention_architecture": "mla", "mla_q_lora_rank": 1536, "mla_kv_lora_rank": 512, "mla_qk_nope_head_dim": 128, "mla_qk_rope_head_dim": 64, "flash_attention_capable": true, "position_embedding_type": null, "max_position_embeddings": 163840, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "DeepseekV3RMSNorm", "linear_bias_present": false} | {"organization": "deepseek-ai", "first_publish_date": "2025-08-21", "paper_urls": ["https://arxiv.org/abs/2412.19437"], "base_model": "deepseek-ai/DeepSeek-V3.1-Base", "license": "mit", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["deepseek-r1."], "hub_metadata": {"downloads": 153389, "likes": 819, "created_at": "2025-08-21T02:37:52.000Z", "last_modified": "2025-09-05T11:30:15.000Z"}} | {"residual_stream_dim": 7168, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "moe_shared_expert": "layers.3.mlp.shared_experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "moe_shared_expert": "layers.4.mlp.shared_experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "moe_shared_expert": "layers.5.mlp.shared_experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "moe_shared_expert": "layers.6.mlp.shared_experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "moe_shared_expert": "layers.7.mlp.shared_experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "moe_shared_expert": "layers.8.mlp.shared_experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "moe_shared_expert": "layers.9.mlp.shared_experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "moe_shared_expert": "layers.10.mlp.shared_experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "moe_shared_expert": "layers.11.mlp.shared_experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "moe_shared_expert": "layers.12.mlp.shared_experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "moe_shared_expert": "layers.13.mlp.shared_experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "moe_shared_expert": "layers.14.mlp.shared_experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "moe_shared_expert": "layers.15.mlp.shared_experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "moe_shared_expert": "layers.16.mlp.shared_experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "moe_shared_expert": "layers.17.mlp.shared_experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "moe_shared_expert": "layers.18.mlp.shared_experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "moe_shared_expert": "layers.19.mlp.shared_experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "moe_shared_expert": "layers.20.mlp.shared_experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "moe_shared_expert": "layers.21.mlp.shared_experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "moe_shared_expert": "layers.22.mlp.shared_experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "moe_shared_expert": "layers.23.mlp.shared_experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "moe_shared_expert": "layers.24.mlp.shared_experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "moe_shared_expert": "layers.25.mlp.shared_experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "moe_shared_expert": "layers.26.mlp.shared_experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "moe_shared_expert": "layers.27.mlp.shared_experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "moe_shared_expert": "layers.28.mlp.shared_experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "moe_shared_expert": "layers.29.mlp.shared_experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "moe_shared_expert": "layers.30.mlp.shared_experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "moe_shared_expert": "layers.31.mlp.shared_experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "moe_shared_expert": "layers.32.mlp.shared_experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "moe_shared_expert": "layers.33.mlp.shared_experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "moe_shared_expert": "layers.34.mlp.shared_experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "moe_shared_expert": "layers.35.mlp.shared_experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "moe_shared_expert": "layers.36.mlp.shared_experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "moe_shared_expert": "layers.37.mlp.shared_experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "moe_shared_expert": "layers.38.mlp.shared_experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "moe_shared_expert": "layers.39.mlp.shared_experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "moe_shared_expert": "layers.40.mlp.shared_experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "moe_shared_expert": "layers.41.mlp.shared_experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "moe_shared_expert": "layers.42.mlp.shared_experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "moe_shared_expert": "layers.43.mlp.shared_experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "moe_shared_expert": "layers.44.mlp.shared_experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "moe_shared_expert": "layers.45.mlp.shared_experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "moe_shared_expert": "layers.46.mlp.shared_experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "moe_shared_expert": "layers.47.mlp.shared_experts", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate", "moe_gate": "layers.48.mlp.gate", "moe_experts": "layers.48.mlp.experts", "moe_shared_expert": "layers.48.mlp.shared_experts", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate", "moe_gate": "layers.49.mlp.gate", "moe_experts": "layers.49.mlp.experts", "moe_shared_expert": "layers.49.mlp.shared_experts", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate", "moe_gate": "layers.50.mlp.gate", "moe_experts": "layers.50.mlp.experts", "moe_shared_expert": "layers.50.mlp.shared_experts", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate", "moe_gate": "layers.51.mlp.gate", "moe_experts": "layers.51.mlp.experts", "moe_shared_expert": "layers.51.mlp.shared_experts", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate", "moe_gate": "layers.52.mlp.gate", "moe_experts": "layers.52.mlp.experts", "moe_shared_expert": "layers.52.mlp.shared_experts", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate", "moe_gate": "layers.53.mlp.gate", "moe_experts": "layers.53.mlp.experts", "moe_shared_expert": "layers.53.mlp.shared_experts", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate", "moe_gate": "layers.54.mlp.gate", "moe_experts": "layers.54.mlp.experts", "moe_shared_expert": "layers.54.mlp.shared_experts", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate", "moe_gate": "layers.55.mlp.gate", "moe_experts": "layers.55.mlp.experts", "moe_shared_expert": "layers.55.mlp.shared_experts", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate", "moe_gate": "layers.56.mlp.gate", "moe_experts": "layers.56.mlp.experts", "moe_shared_expert": "layers.56.mlp.shared_experts", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate", "moe_gate": "layers.57.mlp.gate", "moe_experts": "layers.57.mlp.experts", "moe_shared_expert": "layers.57.mlp.shared_experts", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate", "moe_gate": "layers.58.mlp.gate", "moe_experts": "layers.58.mlp.experts", "moe_shared_expert": "layers.58.mlp.shared_experts", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate", "moe_gate": "layers.59.mlp.gate", "moe_experts": "layers.59.mlp.experts", "moe_shared_expert": "layers.59.mlp.shared_experts", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate", "moe_gate": "layers.60.mlp.gate", "moe_experts": "layers.60.mlp.experts", "moe_shared_expert": "layers.60.mlp.shared_experts", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.864137", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
deepseek-ai/DeepSeek-R1-0528-Qwen3-8B | deepseek-ai | qwen3 | 3 | 2026-03-29T15:44:03.764306+00:00 | {"hidden_size": 4096, "intermediate_size": 12288, "vocab_size": 151936, "max_position_embeddings": 131072, "num_hidden_layers": 36, "rms_norm_eps": 1e-06, "model_type": "qwen3"} | {"rope_type": null, "rope_scaling": {"rope_type": "yarn", "factor": 4.0, "original_max_position_embeddings": 32768, "attn_factor": 0.8782488562869419, "rope_theta": 1000000}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 131072, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen3RMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "deepseek-ai", "first_publish_date": "2025-05-29", "paper_urls": ["https://arxiv.org/abs/2501.12948"], "base_model": null, "license": "mit", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["phi", "the user", "gpt 4.1", "qwen 3", "gpt", "the question.\r\n- for listing-type questions", "phi 4", "switch", "multiple contexts"], "hub_metadata": {"downloads": 149988, "likes": 1042, "created_at": "2025-05-29T11:07:47.000Z", "last_modified": "2025-05-29T13:13:34.000Z"}} | {"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.864254", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
zai-org/GLM-4.7 | zai-org | glm4_moe | 3 | 2026-03-29T15:44:49.559148+00:00 | {"hidden_size": 5120, "intermediate_size": 12288, "vocab_size": 151552, "max_position_embeddings": 202752, "num_hidden_layers": 92, "rms_norm_eps": 1e-05, "model_type": "glm4_moe"} | {"rope_type": null, "rope_scaling": {"rope_theta": 1000000, "partial_rotary_factor": 0.5, "rope_type": "default"}, "attention_bias": true, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 202752, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-05, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Glm4MoeRMSNorm", "linear_bias_present": true, "qkv_projection_style": "separate"} | {"organization": "zai-org", "first_publish_date": "2025-12-22", "paper_urls": ["https://arxiv.org/abs/2508.06471"], "base_model": null, "license": "mit", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["gpt 5", "switch", "claude", "glm", "gpt", "glm 4.6", "glm 45a", "yi", "glm 4.5", "glm 4.7", "glm 45", "glm 47", "gpt 5.1"], "hub_metadata": {"downloads": 142541, "likes": 1947, "created_at": "2025-12-22T07:45:52.000Z", "last_modified": "2026-01-29T08:05:58.000Z"}} | {"residual_stream_dim": 5120, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "moe_shared_expert": "layers.3.mlp.shared_experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "moe_shared_expert": "layers.4.mlp.shared_experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "moe_shared_expert": "layers.5.mlp.shared_experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "moe_shared_expert": "layers.6.mlp.shared_experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "moe_shared_expert": "layers.7.mlp.shared_experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "moe_shared_expert": "layers.8.mlp.shared_experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "moe_shared_expert": "layers.9.mlp.shared_experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "moe_shared_expert": "layers.10.mlp.shared_experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "moe_shared_expert": "layers.11.mlp.shared_experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "moe_shared_expert": "layers.12.mlp.shared_experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "moe_shared_expert": "layers.13.mlp.shared_experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "moe_shared_expert": "layers.14.mlp.shared_experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "moe_shared_expert": "layers.15.mlp.shared_experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "moe_shared_expert": "layers.16.mlp.shared_experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "moe_shared_expert": "layers.17.mlp.shared_experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "moe_shared_expert": "layers.18.mlp.shared_experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "moe_shared_expert": "layers.19.mlp.shared_experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "moe_shared_expert": "layers.20.mlp.shared_experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "moe_shared_expert": "layers.21.mlp.shared_experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "moe_shared_expert": "layers.22.mlp.shared_experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "moe_shared_expert": "layers.23.mlp.shared_experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "moe_shared_expert": "layers.24.mlp.shared_experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "moe_shared_expert": "layers.25.mlp.shared_experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "moe_shared_expert": "layers.26.mlp.shared_experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "moe_shared_expert": "layers.27.mlp.shared_experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "moe_shared_expert": "layers.28.mlp.shared_experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "moe_shared_expert": "layers.29.mlp.shared_experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "moe_shared_expert": "layers.30.mlp.shared_experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "moe_shared_expert": "layers.31.mlp.shared_experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "moe_shared_expert": "layers.32.mlp.shared_experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "moe_shared_expert": "layers.33.mlp.shared_experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "moe_shared_expert": "layers.34.mlp.shared_experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "moe_shared_expert": "layers.35.mlp.shared_experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "moe_shared_expert": "layers.36.mlp.shared_experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "moe_shared_expert": "layers.37.mlp.shared_experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "moe_shared_expert": "layers.38.mlp.shared_experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "moe_shared_expert": "layers.39.mlp.shared_experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "moe_shared_expert": "layers.40.mlp.shared_experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "moe_shared_expert": "layers.41.mlp.shared_experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "moe_shared_expert": "layers.42.mlp.shared_experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "moe_shared_expert": "layers.43.mlp.shared_experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "moe_shared_expert": "layers.44.mlp.shared_experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "moe_shared_expert": "layers.45.mlp.shared_experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "moe_shared_expert": "layers.46.mlp.shared_experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "moe_shared_expert": "layers.47.mlp.shared_experts", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_q_proj": "layers.48.self_attn.q_proj", "attn_k_proj": "layers.48.self_attn.k_proj", "attn_v_proj": "layers.48.self_attn.v_proj", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate", "moe_gate": "layers.48.mlp.gate", "moe_experts": "layers.48.mlp.experts", "moe_shared_expert": "layers.48.mlp.shared_experts", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_q_proj": "layers.49.self_attn.q_proj", "attn_k_proj": "layers.49.self_attn.k_proj", "attn_v_proj": "layers.49.self_attn.v_proj", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate", "moe_gate": "layers.49.mlp.gate", "moe_experts": "layers.49.mlp.experts", "moe_shared_expert": "layers.49.mlp.shared_experts", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_q_proj": "layers.50.self_attn.q_proj", "attn_k_proj": "layers.50.self_attn.k_proj", "attn_v_proj": "layers.50.self_attn.v_proj", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate", "moe_gate": "layers.50.mlp.gate", "moe_experts": "layers.50.mlp.experts", "moe_shared_expert": "layers.50.mlp.shared_experts", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate", "moe_gate": "layers.51.mlp.gate", "moe_experts": "layers.51.mlp.experts", "moe_shared_expert": "layers.51.mlp.shared_experts", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_q_proj": "layers.52.self_attn.q_proj", "attn_k_proj": "layers.52.self_attn.k_proj", "attn_v_proj": "layers.52.self_attn.v_proj", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate", "moe_gate": "layers.52.mlp.gate", "moe_experts": "layers.52.mlp.experts", "moe_shared_expert": "layers.52.mlp.shared_experts", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_q_proj": "layers.53.self_attn.q_proj", "attn_k_proj": "layers.53.self_attn.k_proj", "attn_v_proj": "layers.53.self_attn.v_proj", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate", "moe_gate": "layers.53.mlp.gate", "moe_experts": "layers.53.mlp.experts", "moe_shared_expert": "layers.53.mlp.shared_experts", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_q_proj": "layers.54.self_attn.q_proj", "attn_k_proj": "layers.54.self_attn.k_proj", "attn_v_proj": "layers.54.self_attn.v_proj", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate", "moe_gate": "layers.54.mlp.gate", "moe_experts": "layers.54.mlp.experts", "moe_shared_expert": "layers.54.mlp.shared_experts", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate", "moe_gate": "layers.55.mlp.gate", "moe_experts": "layers.55.mlp.experts", "moe_shared_expert": "layers.55.mlp.shared_experts", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_q_proj": "layers.56.self_attn.q_proj", "attn_k_proj": "layers.56.self_attn.k_proj", "attn_v_proj": "layers.56.self_attn.v_proj", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate", "moe_gate": "layers.56.mlp.gate", "moe_experts": "layers.56.mlp.experts", "moe_shared_expert": "layers.56.mlp.shared_experts", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_q_proj": "layers.57.self_attn.q_proj", "attn_k_proj": "layers.57.self_attn.k_proj", "attn_v_proj": "layers.57.self_attn.v_proj", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate", "moe_gate": "layers.57.mlp.gate", "moe_experts": "layers.57.mlp.experts", "moe_shared_expert": "layers.57.mlp.shared_experts", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_q_proj": "layers.58.self_attn.q_proj", "attn_k_proj": "layers.58.self_attn.k_proj", "attn_v_proj": "layers.58.self_attn.v_proj", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate", "moe_gate": "layers.58.mlp.gate", "moe_experts": "layers.58.mlp.experts", "moe_shared_expert": "layers.58.mlp.shared_experts", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate", "moe_gate": "layers.59.mlp.gate", "moe_experts": "layers.59.mlp.experts", "moe_shared_expert": "layers.59.mlp.shared_experts", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_q_proj": "layers.60.self_attn.q_proj", "attn_k_proj": "layers.60.self_attn.k_proj", "attn_v_proj": "layers.60.self_attn.v_proj", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate", "moe_gate": "layers.60.mlp.gate", "moe_experts": "layers.60.mlp.experts", "moe_shared_expert": "layers.60.mlp.shared_experts", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}, {"index": 61, "intervention_points": {"attn_input": "layers.61.self_attn.", "attn_q_proj": "layers.61.self_attn.q_proj", "attn_k_proj": "layers.61.self_attn.k_proj", "attn_v_proj": "layers.61.self_attn.v_proj", "attn_output_proj": "layers.61.self_attn.o_proj", "mlp_input": "layers.61.mlp.", "mlp_gate_proj": "layers.61.mlp.gate", "moe_gate": "layers.61.mlp.gate", "moe_experts": "layers.61.mlp.experts", "moe_shared_expert": "layers.61.mlp.shared_experts", "attn_pre_norm": "layers.61.input_layernorm", "attn_post_norm": "layers.61.post_attention_layernorm"}}, {"index": 62, "intervention_points": {"attn_input": "layers.62.self_attn.", "attn_q_proj": "layers.62.self_attn.q_proj", "attn_k_proj": "layers.62.self_attn.k_proj", "attn_v_proj": "layers.62.self_attn.v_proj", "attn_output_proj": "layers.62.self_attn.o_proj", "mlp_input": "layers.62.mlp.", "mlp_gate_proj": "layers.62.mlp.gate", "moe_gate": "layers.62.mlp.gate", "moe_experts": "layers.62.mlp.experts", "moe_shared_expert": "layers.62.mlp.shared_experts", "attn_pre_norm": "layers.62.input_layernorm", "attn_post_norm": "layers.62.post_attention_layernorm"}}, {"index": 63, "intervention_points": {"attn_input": "layers.63.self_attn.", "attn_q_proj": "layers.63.self_attn.q_proj", "attn_k_proj": "layers.63.self_attn.k_proj", "attn_v_proj": "layers.63.self_attn.v_proj", "attn_output_proj": "layers.63.self_attn.o_proj", "mlp_input": "layers.63.mlp.", "mlp_gate_proj": "layers.63.mlp.gate", "moe_gate": "layers.63.mlp.gate", "moe_experts": "layers.63.mlp.experts", "moe_shared_expert": "layers.63.mlp.shared_experts", "attn_pre_norm": "layers.63.input_layernorm", "attn_post_norm": "layers.63.post_attention_layernorm"}}, {"index": 64, "intervention_points": {"attn_input": "layers.64.self_attn.", "attn_q_proj": "layers.64.self_attn.q_proj", "attn_k_proj": "layers.64.self_attn.k_proj", "attn_v_proj": "layers.64.self_attn.v_proj", "attn_output_proj": "layers.64.self_attn.o_proj", "mlp_input": "layers.64.mlp.", "mlp_gate_proj": "layers.64.mlp.gate", "moe_gate": "layers.64.mlp.gate", "moe_experts": "layers.64.mlp.experts", "moe_shared_expert": "layers.64.mlp.shared_experts", "attn_pre_norm": "layers.64.input_layernorm", "attn_post_norm": "layers.64.post_attention_layernorm"}}, {"index": 65, "intervention_points": {"attn_input": "layers.65.self_attn.", "attn_q_proj": "layers.65.self_attn.q_proj", "attn_k_proj": "layers.65.self_attn.k_proj", "attn_v_proj": "layers.65.self_attn.v_proj", "attn_output_proj": "layers.65.self_attn.o_proj", "mlp_input": "layers.65.mlp.", "mlp_gate_proj": "layers.65.mlp.gate", "moe_gate": "layers.65.mlp.gate", "moe_experts": "layers.65.mlp.experts", "moe_shared_expert": "layers.65.mlp.shared_experts", "attn_pre_norm": "layers.65.input_layernorm", "attn_post_norm": "layers.65.post_attention_layernorm"}}, {"index": 66, "intervention_points": {"attn_input": "layers.66.self_attn.", "attn_q_proj": "layers.66.self_attn.q_proj", "attn_k_proj": "layers.66.self_attn.k_proj", "attn_v_proj": "layers.66.self_attn.v_proj", "attn_output_proj": "layers.66.self_attn.o_proj", "mlp_input": "layers.66.mlp.", "mlp_gate_proj": "layers.66.mlp.gate", "moe_gate": "layers.66.mlp.gate", "moe_experts": "layers.66.mlp.experts", "moe_shared_expert": "layers.66.mlp.shared_experts", "attn_pre_norm": "layers.66.input_layernorm", "attn_post_norm": "layers.66.post_attention_layernorm"}}, {"index": 67, "intervention_points": {"attn_input": "layers.67.self_attn.", "attn_q_proj": "layers.67.self_attn.q_proj", "attn_k_proj": "layers.67.self_attn.k_proj", "attn_v_proj": "layers.67.self_attn.v_proj", "attn_output_proj": "layers.67.self_attn.o_proj", "mlp_input": "layers.67.mlp.", "mlp_gate_proj": "layers.67.mlp.gate", "moe_gate": "layers.67.mlp.gate", "moe_experts": "layers.67.mlp.experts", "moe_shared_expert": "layers.67.mlp.shared_experts", "attn_pre_norm": "layers.67.input_layernorm", "attn_post_norm": "layers.67.post_attention_layernorm"}}, {"index": 68, "intervention_points": {"attn_input": "layers.68.self_attn.", "attn_q_proj": "layers.68.self_attn.q_proj", "attn_k_proj": "layers.68.self_attn.k_proj", "attn_v_proj": "layers.68.self_attn.v_proj", "attn_output_proj": "layers.68.self_attn.o_proj", "mlp_input": "layers.68.mlp.", "mlp_gate_proj": "layers.68.mlp.gate", "moe_gate": "layers.68.mlp.gate", "moe_experts": "layers.68.mlp.experts", "moe_shared_expert": "layers.68.mlp.shared_experts", "attn_pre_norm": "layers.68.input_layernorm", "attn_post_norm": "layers.68.post_attention_layernorm"}}, {"index": 69, "intervention_points": {"attn_input": "layers.69.self_attn.", "attn_q_proj": "layers.69.self_attn.q_proj", "attn_k_proj": "layers.69.self_attn.k_proj", "attn_v_proj": "layers.69.self_attn.v_proj", "attn_output_proj": "layers.69.self_attn.o_proj", "mlp_input": "layers.69.mlp.", "mlp_gate_proj": "layers.69.mlp.gate", "moe_gate": "layers.69.mlp.gate", "moe_experts": "layers.69.mlp.experts", "moe_shared_expert": "layers.69.mlp.shared_experts", "attn_pre_norm": "layers.69.input_layernorm", "attn_post_norm": "layers.69.post_attention_layernorm"}}, {"index": 70, "intervention_points": {"attn_input": "layers.70.self_attn.", "attn_q_proj": "layers.70.self_attn.q_proj", "attn_k_proj": "layers.70.self_attn.k_proj", "attn_v_proj": "layers.70.self_attn.v_proj", "attn_output_proj": "layers.70.self_attn.o_proj", "mlp_input": "layers.70.mlp.", "mlp_gate_proj": "layers.70.mlp.gate", "moe_gate": "layers.70.mlp.gate", "moe_experts": "layers.70.mlp.experts", "moe_shared_expert": "layers.70.mlp.shared_experts", "attn_pre_norm": "layers.70.input_layernorm", "attn_post_norm": "layers.70.post_attention_layernorm"}}, {"index": 71, "intervention_points": {"attn_input": "layers.71.self_attn.", "attn_q_proj": "layers.71.self_attn.q_proj", "attn_k_proj": "layers.71.self_attn.k_proj", "attn_v_proj": "layers.71.self_attn.v_proj", "attn_output_proj": "layers.71.self_attn.o_proj", "mlp_input": "layers.71.mlp.", "mlp_gate_proj": "layers.71.mlp.gate", "moe_gate": "layers.71.mlp.gate", "moe_experts": "layers.71.mlp.experts", "moe_shared_expert": "layers.71.mlp.shared_experts", "attn_pre_norm": "layers.71.input_layernorm", "attn_post_norm": "layers.71.post_attention_layernorm"}}, {"index": 72, "intervention_points": {"attn_input": "layers.72.self_attn.", "attn_q_proj": "layers.72.self_attn.q_proj", "attn_k_proj": "layers.72.self_attn.k_proj", "attn_v_proj": "layers.72.self_attn.v_proj", "attn_output_proj": "layers.72.self_attn.o_proj", "mlp_input": "layers.72.mlp.", "mlp_gate_proj": "layers.72.mlp.gate", "moe_gate": "layers.72.mlp.gate", "moe_experts": "layers.72.mlp.experts", "moe_shared_expert": "layers.72.mlp.shared_experts", "attn_pre_norm": "layers.72.input_layernorm", "attn_post_norm": "layers.72.post_attention_layernorm"}}, {"index": 73, "intervention_points": {"attn_input": "layers.73.self_attn.", "attn_q_proj": "layers.73.self_attn.q_proj", "attn_k_proj": "layers.73.self_attn.k_proj", "attn_v_proj": "layers.73.self_attn.v_proj", "attn_output_proj": "layers.73.self_attn.o_proj", "mlp_input": "layers.73.mlp.", "mlp_gate_proj": "layers.73.mlp.gate", "moe_gate": "layers.73.mlp.gate", "moe_experts": "layers.73.mlp.experts", "moe_shared_expert": "layers.73.mlp.shared_experts", "attn_pre_norm": "layers.73.input_layernorm", "attn_post_norm": "layers.73.post_attention_layernorm"}}, {"index": 74, "intervention_points": {"attn_input": "layers.74.self_attn.", "attn_q_proj": "layers.74.self_attn.q_proj", "attn_k_proj": "layers.74.self_attn.k_proj", "attn_v_proj": "layers.74.self_attn.v_proj", "attn_output_proj": "layers.74.self_attn.o_proj", "mlp_input": "layers.74.mlp.", "mlp_gate_proj": "layers.74.mlp.gate", "moe_gate": "layers.74.mlp.gate", "moe_experts": "layers.74.mlp.experts", "moe_shared_expert": "layers.74.mlp.shared_experts", "attn_pre_norm": "layers.74.input_layernorm", "attn_post_norm": "layers.74.post_attention_layernorm"}}, {"index": 75, "intervention_points": {"attn_input": "layers.75.self_attn.", "attn_q_proj": "layers.75.self_attn.q_proj", "attn_k_proj": "layers.75.self_attn.k_proj", "attn_v_proj": "layers.75.self_attn.v_proj", "attn_output_proj": "layers.75.self_attn.o_proj", "mlp_input": "layers.75.mlp.", "mlp_gate_proj": "layers.75.mlp.gate", "moe_gate": "layers.75.mlp.gate", "moe_experts": "layers.75.mlp.experts", "moe_shared_expert": "layers.75.mlp.shared_experts", "attn_pre_norm": "layers.75.input_layernorm", "attn_post_norm": "layers.75.post_attention_layernorm"}}, {"index": 76, "intervention_points": {"attn_input": "layers.76.self_attn.", "attn_q_proj": "layers.76.self_attn.q_proj", "attn_k_proj": "layers.76.self_attn.k_proj", "attn_v_proj": "layers.76.self_attn.v_proj", "attn_output_proj": "layers.76.self_attn.o_proj", "mlp_input": "layers.76.mlp.", "mlp_gate_proj": "layers.76.mlp.gate", "moe_gate": "layers.76.mlp.gate", "moe_experts": "layers.76.mlp.experts", "moe_shared_expert": "layers.76.mlp.shared_experts", "attn_pre_norm": "layers.76.input_layernorm", "attn_post_norm": "layers.76.post_attention_layernorm"}}, {"index": 77, "intervention_points": {"attn_input": "layers.77.self_attn.", "attn_q_proj": "layers.77.self_attn.q_proj", "attn_k_proj": "layers.77.self_attn.k_proj", "attn_v_proj": "layers.77.self_attn.v_proj", "attn_output_proj": "layers.77.self_attn.o_proj", "mlp_input": "layers.77.mlp.", "mlp_gate_proj": "layers.77.mlp.gate", "moe_gate": "layers.77.mlp.gate", "moe_experts": "layers.77.mlp.experts", "moe_shared_expert": "layers.77.mlp.shared_experts", "attn_pre_norm": "layers.77.input_layernorm", "attn_post_norm": "layers.77.post_attention_layernorm"}}, {"index": 78, "intervention_points": {"attn_input": "layers.78.self_attn.", "attn_q_proj": "layers.78.self_attn.q_proj", "attn_k_proj": "layers.78.self_attn.k_proj", "attn_v_proj": "layers.78.self_attn.v_proj", "attn_output_proj": "layers.78.self_attn.o_proj", "mlp_input": "layers.78.mlp.", "mlp_gate_proj": "layers.78.mlp.gate", "moe_gate": "layers.78.mlp.gate", "moe_experts": "layers.78.mlp.experts", "moe_shared_expert": "layers.78.mlp.shared_experts", "attn_pre_norm": "layers.78.input_layernorm", "attn_post_norm": "layers.78.post_attention_layernorm"}}, {"index": 79, "intervention_points": {"attn_input": "layers.79.self_attn.", "attn_q_proj": "layers.79.self_attn.q_proj", "attn_k_proj": "layers.79.self_attn.k_proj", "attn_v_proj": "layers.79.self_attn.v_proj", "attn_output_proj": "layers.79.self_attn.o_proj", "mlp_input": "layers.79.mlp.", "mlp_gate_proj": "layers.79.mlp.gate", "moe_gate": "layers.79.mlp.gate", "moe_experts": "layers.79.mlp.experts", "moe_shared_expert": "layers.79.mlp.shared_experts", "attn_pre_norm": "layers.79.input_layernorm", "attn_post_norm": "layers.79.post_attention_layernorm"}}, {"index": 80, "intervention_points": {"attn_input": "layers.80.self_attn.", "attn_q_proj": "layers.80.self_attn.q_proj", "attn_k_proj": "layers.80.self_attn.k_proj", "attn_v_proj": "layers.80.self_attn.v_proj", "attn_output_proj": "layers.80.self_attn.o_proj", "mlp_input": "layers.80.mlp.", "mlp_gate_proj": "layers.80.mlp.gate", "moe_gate": "layers.80.mlp.gate", "moe_experts": "layers.80.mlp.experts", "moe_shared_expert": "layers.80.mlp.shared_experts", "attn_pre_norm": "layers.80.input_layernorm", "attn_post_norm": "layers.80.post_attention_layernorm"}}, {"index": 81, "intervention_points": {"attn_input": "layers.81.self_attn.", "attn_q_proj": "layers.81.self_attn.q_proj", "attn_k_proj": "layers.81.self_attn.k_proj", "attn_v_proj": "layers.81.self_attn.v_proj", "attn_output_proj": "layers.81.self_attn.o_proj", "mlp_input": "layers.81.mlp.", "mlp_gate_proj": "layers.81.mlp.gate", "moe_gate": "layers.81.mlp.gate", "moe_experts": "layers.81.mlp.experts", "moe_shared_expert": "layers.81.mlp.shared_experts", "attn_pre_norm": "layers.81.input_layernorm", "attn_post_norm": "layers.81.post_attention_layernorm"}}, {"index": 82, "intervention_points": {"attn_input": "layers.82.self_attn.", "attn_q_proj": "layers.82.self_attn.q_proj", "attn_k_proj": "layers.82.self_attn.k_proj", "attn_v_proj": "layers.82.self_attn.v_proj", "attn_output_proj": "layers.82.self_attn.o_proj", "mlp_input": "layers.82.mlp.", "mlp_gate_proj": "layers.82.mlp.gate", "moe_gate": "layers.82.mlp.gate", "moe_experts": "layers.82.mlp.experts", "moe_shared_expert": "layers.82.mlp.shared_experts", "attn_pre_norm": "layers.82.input_layernorm", "attn_post_norm": "layers.82.post_attention_layernorm"}}, {"index": 83, "intervention_points": {"attn_input": "layers.83.self_attn.", "attn_q_proj": "layers.83.self_attn.q_proj", "attn_k_proj": "layers.83.self_attn.k_proj", "attn_v_proj": "layers.83.self_attn.v_proj", "attn_output_proj": "layers.83.self_attn.o_proj", "mlp_input": "layers.83.mlp.", "mlp_gate_proj": "layers.83.mlp.gate", "moe_gate": "layers.83.mlp.gate", "moe_experts": "layers.83.mlp.experts", "moe_shared_expert": "layers.83.mlp.shared_experts", "attn_pre_norm": "layers.83.input_layernorm", "attn_post_norm": "layers.83.post_attention_layernorm"}}, {"index": 84, "intervention_points": {"attn_input": "layers.84.self_attn.", "attn_q_proj": "layers.84.self_attn.q_proj", "attn_k_proj": "layers.84.self_attn.k_proj", "attn_v_proj": "layers.84.self_attn.v_proj", "attn_output_proj": "layers.84.self_attn.o_proj", "mlp_input": "layers.84.mlp.", "mlp_gate_proj": "layers.84.mlp.gate", "moe_gate": "layers.84.mlp.gate", "moe_experts": "layers.84.mlp.experts", "moe_shared_expert": "layers.84.mlp.shared_experts", "attn_pre_norm": "layers.84.input_layernorm", "attn_post_norm": "layers.84.post_attention_layernorm"}}, {"index": 85, "intervention_points": {"attn_input": "layers.85.self_attn.", "attn_q_proj": "layers.85.self_attn.q_proj", "attn_k_proj": "layers.85.self_attn.k_proj", "attn_v_proj": "layers.85.self_attn.v_proj", "attn_output_proj": "layers.85.self_attn.o_proj", "mlp_input": "layers.85.mlp.", "mlp_gate_proj": "layers.85.mlp.gate", "moe_gate": "layers.85.mlp.gate", "moe_experts": "layers.85.mlp.experts", "moe_shared_expert": "layers.85.mlp.shared_experts", "attn_pre_norm": "layers.85.input_layernorm", "attn_post_norm": "layers.85.post_attention_layernorm"}}, {"index": 86, "intervention_points": {"attn_input": "layers.86.self_attn.", "attn_q_proj": "layers.86.self_attn.q_proj", "attn_k_proj": "layers.86.self_attn.k_proj", "attn_v_proj": "layers.86.self_attn.v_proj", "attn_output_proj": "layers.86.self_attn.o_proj", "mlp_input": "layers.86.mlp.", "mlp_gate_proj": "layers.86.mlp.gate", "moe_gate": "layers.86.mlp.gate", "moe_experts": "layers.86.mlp.experts", "moe_shared_expert": "layers.86.mlp.shared_experts", "attn_pre_norm": "layers.86.input_layernorm", "attn_post_norm": "layers.86.post_attention_layernorm"}}, {"index": 87, "intervention_points": {"attn_input": "layers.87.self_attn.", "attn_q_proj": "layers.87.self_attn.q_proj", "attn_k_proj": "layers.87.self_attn.k_proj", "attn_v_proj": "layers.87.self_attn.v_proj", "attn_output_proj": "layers.87.self_attn.o_proj", "mlp_input": "layers.87.mlp.", "mlp_gate_proj": "layers.87.mlp.gate", "moe_gate": "layers.87.mlp.gate", "moe_experts": "layers.87.mlp.experts", "moe_shared_expert": "layers.87.mlp.shared_experts", "attn_pre_norm": "layers.87.input_layernorm", "attn_post_norm": "layers.87.post_attention_layernorm"}}, {"index": 88, "intervention_points": {"attn_input": "layers.88.self_attn.", "attn_q_proj": "layers.88.self_attn.q_proj", "attn_k_proj": "layers.88.self_attn.k_proj", "attn_v_proj": "layers.88.self_attn.v_proj", "attn_output_proj": "layers.88.self_attn.o_proj", "mlp_input": "layers.88.mlp.", "mlp_gate_proj": "layers.88.mlp.gate", "moe_gate": "layers.88.mlp.gate", "moe_experts": "layers.88.mlp.experts", "moe_shared_expert": "layers.88.mlp.shared_experts", "attn_pre_norm": "layers.88.input_layernorm", "attn_post_norm": "layers.88.post_attention_layernorm"}}, {"index": 89, "intervention_points": {"attn_input": "layers.89.self_attn.", "attn_q_proj": "layers.89.self_attn.q_proj", "attn_k_proj": "layers.89.self_attn.k_proj", "attn_v_proj": "layers.89.self_attn.v_proj", "attn_output_proj": "layers.89.self_attn.o_proj", "mlp_input": "layers.89.mlp.", "mlp_gate_proj": "layers.89.mlp.gate", "moe_gate": "layers.89.mlp.gate", "moe_experts": "layers.89.mlp.experts", "moe_shared_expert": "layers.89.mlp.shared_experts", "attn_pre_norm": "layers.89.input_layernorm", "attn_post_norm": "layers.89.post_attention_layernorm"}}, {"index": 90, "intervention_points": {"attn_input": "layers.90.self_attn.", "attn_q_proj": "layers.90.self_attn.q_proj", "attn_k_proj": "layers.90.self_attn.k_proj", "attn_v_proj": "layers.90.self_attn.v_proj", "attn_output_proj": "layers.90.self_attn.o_proj", "mlp_input": "layers.90.mlp.", "mlp_gate_proj": "layers.90.mlp.gate", "moe_gate": "layers.90.mlp.gate", "moe_experts": "layers.90.mlp.experts", "moe_shared_expert": "layers.90.mlp.shared_experts", "attn_pre_norm": "layers.90.input_layernorm", "attn_post_norm": "layers.90.post_attention_layernorm"}}, {"index": 91, "intervention_points": {"attn_input": "layers.91.self_attn.", "attn_q_proj": "layers.91.self_attn.q_proj", "attn_k_proj": "layers.91.self_attn.k_proj", "attn_v_proj": "layers.91.self_attn.v_proj", "attn_output_proj": "layers.91.self_attn.o_proj", "mlp_input": "layers.91.mlp.", "mlp_gate_proj": "layers.91.mlp.gate", "moe_gate": "layers.91.mlp.gate", "moe_experts": "layers.91.mlp.experts", "moe_shared_expert": "layers.91.mlp.shared_experts", "attn_pre_norm": "layers.91.input_layernorm", "attn_post_norm": "layers.91.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.864459", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
zai-org/GLM-4.7-Flash | zai-org | glm4_moe_lite | 3 | 2026-03-29T15:44:53.562389+00:00 | {"hidden_size": 2048, "intermediate_size": 10240, "vocab_size": 154880, "max_position_embeddings": 202752, "num_hidden_layers": 47, "rms_norm_eps": 1e-05, "model_type": "glm4_moe_lite"} | {"rope_type": null, "rope_scaling": {"rope_theta": 1000000, "partial_rotary_factor": 1.0, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "attention_architecture": "mla", "mla_q_lora_rank": 768, "mla_kv_lora_rank": 512, "mla_qk_nope_head_dim": 192, "mla_qk_rope_head_dim": 64, "position_embedding_type": null, "max_position_embeddings": 202752, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-05, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Glm4MoeLiteRMSNorm", "linear_bias_present": false} | {"organization": "zai-org", "first_publish_date": "2026-01-19", "paper_urls": ["https://arxiv.org/abs/2508.06471"], "base_model": null, "license": "mit", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["glm 4.5", "glm", "glm 47", "qwen 3", "glm 45a", "claude", "glm 45", "yi", "gpt", "glm 4.7"], "hub_metadata": {"downloads": 1395480, "likes": 1632, "created_at": "2026-01-19T06:28:10.000Z", "last_modified": "2026-01-29T08:06:19.000Z"}} | {"residual_stream_dim": 2048, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate", "moe_gate": "layers.1.mlp.gate", "moe_experts": "layers.1.mlp.experts", "moe_shared_expert": "layers.1.mlp.shared_experts", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate", "moe_gate": "layers.2.mlp.gate", "moe_experts": "layers.2.mlp.experts", "moe_shared_expert": "layers.2.mlp.shared_experts", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "moe_shared_expert": "layers.3.mlp.shared_experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "moe_shared_expert": "layers.4.mlp.shared_experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "moe_shared_expert": "layers.5.mlp.shared_experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "moe_shared_expert": "layers.6.mlp.shared_experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "moe_shared_expert": "layers.7.mlp.shared_experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "moe_shared_expert": "layers.8.mlp.shared_experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "moe_shared_expert": "layers.9.mlp.shared_experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "moe_shared_expert": "layers.10.mlp.shared_experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "moe_shared_expert": "layers.11.mlp.shared_experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "moe_shared_expert": "layers.12.mlp.shared_experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "moe_shared_expert": "layers.13.mlp.shared_experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "moe_shared_expert": "layers.14.mlp.shared_experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "moe_shared_expert": "layers.15.mlp.shared_experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "moe_shared_expert": "layers.16.mlp.shared_experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "moe_shared_expert": "layers.17.mlp.shared_experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "moe_shared_expert": "layers.18.mlp.shared_experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "moe_shared_expert": "layers.19.mlp.shared_experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "moe_shared_expert": "layers.20.mlp.shared_experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "moe_shared_expert": "layers.21.mlp.shared_experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "moe_shared_expert": "layers.22.mlp.shared_experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "moe_shared_expert": "layers.23.mlp.shared_experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "moe_shared_expert": "layers.24.mlp.shared_experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "moe_shared_expert": "layers.25.mlp.shared_experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "moe_shared_expert": "layers.26.mlp.shared_experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "moe_shared_expert": "layers.27.mlp.shared_experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "moe_shared_expert": "layers.28.mlp.shared_experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "moe_shared_expert": "layers.29.mlp.shared_experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "moe_shared_expert": "layers.30.mlp.shared_experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "moe_shared_expert": "layers.31.mlp.shared_experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "moe_shared_expert": "layers.32.mlp.shared_experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "moe_shared_expert": "layers.33.mlp.shared_experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "moe_shared_expert": "layers.34.mlp.shared_experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "moe_shared_expert": "layers.35.mlp.shared_experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "moe_shared_expert": "layers.36.mlp.shared_experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "moe_shared_expert": "layers.37.mlp.shared_experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "moe_shared_expert": "layers.38.mlp.shared_experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "moe_shared_expert": "layers.39.mlp.shared_experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "moe_shared_expert": "layers.40.mlp.shared_experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "moe_shared_expert": "layers.41.mlp.shared_experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "moe_shared_expert": "layers.42.mlp.shared_experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "moe_shared_expert": "layers.43.mlp.shared_experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "moe_shared_expert": "layers.44.mlp.shared_experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "moe_shared_expert": "layers.45.mlp.shared_experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "moe_shared_expert": "layers.46.mlp.shared_experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.864569", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
XiaomiMiMo/MiMo-V2-Flash | XiaomiMiMo | mimo_v2_flash | 3 | 2026-03-29T15:45:04.990961+00:00 | {"hidden_size": 4096, "intermediate_size": 16384, "vocab_size": 152576, "max_position_embeddings": 262144, "num_hidden_layers": 48, "rope_theta": 10000, "model_type": "mimo_v2_flash"} | {"rope_type": null, "rope_scaling": {"rope_type": "default", "rope_theta": 5000000, "partial_rotary_factor": 0.334}, "attention_bias": false, "sliding_window_size": 128, "attention_implementation": "eager", "flash_attention_capable": true, "position_embedding_type": null, "max_position_embeddings": 262144, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": null, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "MiMoV2RMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "XiaomiMiMo", "first_publish_date": "2025-12-16", "paper_urls": [], "base_model": null, "license": "mit", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["top of sglang. with", "gpt 5", "gpt", "qwen 3", "claude"], "hub_metadata": {"downloads": 91678, "likes": 677, "created_at": "2025-12-16T08:47:02.000Z", "last_modified": "2026-02-27T16:45:45.000Z"}} | {"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate", "moe_gate": "layers.1.mlp.gate", "moe_experts": "layers.1.mlp.experts", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate", "moe_gate": "layers.2.mlp.gate", "moe_experts": "layers.2.mlp.experts", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.864690", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
Qwen/Qwen3-Coder-30B-A3B-Instruct | Qwen | qwen3_moe | 3 | 2026-03-29T15:45:38.648112+00:00 | {"hidden_size": 2048, "intermediate_size": 6144, "vocab_size": 151936, "max_position_embeddings": 262144, "num_hidden_layers": 48, "rms_norm_eps": 1e-06, "model_type": "qwen3_moe"} | {"rope_type": null, "rope_scaling": {"rope_theta": 10000000, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "sdpa", "position_embedding_type": null, "max_position_embeddings": 262144, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen3MoeRMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "Qwen", "first_publish_date": "2025-07-31", "paper_urls": ["https://arxiv.org/abs/2505.09388"], "base_model": null, "license": "apache-2.0", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["qwen 3t", "qwen 3", "given inputs.", "qwen", "llama"], "hub_metadata": {"downloads": 1134092, "likes": 985, "created_at": "2025-07-31T07:04:55.000Z", "last_modified": "2025-12-03T08:05:17.000Z"}} | {"residual_stream_dim": 2048, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate", "moe_gate": "layers.0.mlp.gate", "moe_experts": "layers.0.mlp.experts", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate", "moe_gate": "layers.1.mlp.gate", "moe_experts": "layers.1.mlp.experts", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate", "moe_gate": "layers.2.mlp.gate", "moe_experts": "layers.2.mlp.experts", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.864808", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
Skywork/MindLink-72B-0801 | Skywork | qwen2 | 3 | 2026-03-29T16:18:40.753724+00:00 | {"hidden_size": 8192, "intermediate_size": 29568, "vocab_size": 152064, "max_position_embeddings": 32768, "num_hidden_layers": 80, "rms_norm_eps": 1e-06, "model_type": "qwen2"} | {"rope_type": null, "rope_scaling": {"rope_theta": 1000000.0, "rope_type": "default"}, "attention_bias": null, "sliding_window_size": null, "attention_implementation": "eager", "position_embedding_type": null, "max_position_embeddings": 32768, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-06, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "Qwen2RMSNorm", "linear_bias_present": true, "qkv_projection_style": "separate"} | {"organization": "Skywork", "first_publish_date": "2025-08-01", "paper_urls": [], "base_model": "Qwen/Qwen2.5-72B-Instruct", "license": "apache-2.0", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["qwen", "task complexity", "improvements from", "qwen 2.5"], "hub_metadata": {"downloads": 12, "likes": 32, "created_at": "2025-08-01T12:52:49.000Z", "last_modified": "2025-08-02T15:17:25.000Z"}} | {"residual_stream_dim": 8192, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate_proj", "mlp_up_proj": "layers.36.mlp.up_proj", "mlp_down_proj": "layers.36.mlp.down_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate_proj", "mlp_up_proj": "layers.37.mlp.up_proj", "mlp_down_proj": "layers.37.mlp.down_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate_proj", "mlp_up_proj": "layers.38.mlp.up_proj", "mlp_down_proj": "layers.38.mlp.down_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate_proj", "mlp_up_proj": "layers.39.mlp.up_proj", "mlp_down_proj": "layers.39.mlp.down_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate_proj", "mlp_up_proj": "layers.40.mlp.up_proj", "mlp_down_proj": "layers.40.mlp.down_proj", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate_proj", "mlp_up_proj": "layers.41.mlp.up_proj", "mlp_down_proj": "layers.41.mlp.down_proj", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate_proj", "mlp_up_proj": "layers.42.mlp.up_proj", "mlp_down_proj": "layers.42.mlp.down_proj", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate_proj", "mlp_up_proj": "layers.43.mlp.up_proj", "mlp_down_proj": "layers.43.mlp.down_proj", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate_proj", "mlp_up_proj": "layers.44.mlp.up_proj", "mlp_down_proj": "layers.44.mlp.down_proj", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate_proj", "mlp_up_proj": "layers.45.mlp.up_proj", "mlp_down_proj": "layers.45.mlp.down_proj", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate_proj", "mlp_up_proj": "layers.46.mlp.up_proj", "mlp_down_proj": "layers.46.mlp.down_proj", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate_proj", "mlp_up_proj": "layers.47.mlp.up_proj", "mlp_down_proj": "layers.47.mlp.down_proj", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_q_proj": "layers.48.self_attn.q_proj", "attn_k_proj": "layers.48.self_attn.k_proj", "attn_v_proj": "layers.48.self_attn.v_proj", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate_proj", "mlp_up_proj": "layers.48.mlp.up_proj", "mlp_down_proj": "layers.48.mlp.down_proj", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_q_proj": "layers.49.self_attn.q_proj", "attn_k_proj": "layers.49.self_attn.k_proj", "attn_v_proj": "layers.49.self_attn.v_proj", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate_proj", "mlp_up_proj": "layers.49.mlp.up_proj", "mlp_down_proj": "layers.49.mlp.down_proj", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_q_proj": "layers.50.self_attn.q_proj", "attn_k_proj": "layers.50.self_attn.k_proj", "attn_v_proj": "layers.50.self_attn.v_proj", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate_proj", "mlp_up_proj": "layers.50.mlp.up_proj", "mlp_down_proj": "layers.50.mlp.down_proj", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate_proj", "mlp_up_proj": "layers.51.mlp.up_proj", "mlp_down_proj": "layers.51.mlp.down_proj", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_q_proj": "layers.52.self_attn.q_proj", "attn_k_proj": "layers.52.self_attn.k_proj", "attn_v_proj": "layers.52.self_attn.v_proj", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate_proj", "mlp_up_proj": "layers.52.mlp.up_proj", "mlp_down_proj": "layers.52.mlp.down_proj", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_q_proj": "layers.53.self_attn.q_proj", "attn_k_proj": "layers.53.self_attn.k_proj", "attn_v_proj": "layers.53.self_attn.v_proj", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate_proj", "mlp_up_proj": "layers.53.mlp.up_proj", "mlp_down_proj": "layers.53.mlp.down_proj", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_q_proj": "layers.54.self_attn.q_proj", "attn_k_proj": "layers.54.self_attn.k_proj", "attn_v_proj": "layers.54.self_attn.v_proj", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate_proj", "mlp_up_proj": "layers.54.mlp.up_proj", "mlp_down_proj": "layers.54.mlp.down_proj", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate_proj", "mlp_up_proj": "layers.55.mlp.up_proj", "mlp_down_proj": "layers.55.mlp.down_proj", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_q_proj": "layers.56.self_attn.q_proj", "attn_k_proj": "layers.56.self_attn.k_proj", "attn_v_proj": "layers.56.self_attn.v_proj", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate_proj", "mlp_up_proj": "layers.56.mlp.up_proj", "mlp_down_proj": "layers.56.mlp.down_proj", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_q_proj": "layers.57.self_attn.q_proj", "attn_k_proj": "layers.57.self_attn.k_proj", "attn_v_proj": "layers.57.self_attn.v_proj", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate_proj", "mlp_up_proj": "layers.57.mlp.up_proj", "mlp_down_proj": "layers.57.mlp.down_proj", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_q_proj": "layers.58.self_attn.q_proj", "attn_k_proj": "layers.58.self_attn.k_proj", "attn_v_proj": "layers.58.self_attn.v_proj", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate_proj", "mlp_up_proj": "layers.58.mlp.up_proj", "mlp_down_proj": "layers.58.mlp.down_proj", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate_proj", "mlp_up_proj": "layers.59.mlp.up_proj", "mlp_down_proj": "layers.59.mlp.down_proj", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_q_proj": "layers.60.self_attn.q_proj", "attn_k_proj": "layers.60.self_attn.k_proj", "attn_v_proj": "layers.60.self_attn.v_proj", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate_proj", "mlp_up_proj": "layers.60.mlp.up_proj", "mlp_down_proj": "layers.60.mlp.down_proj", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}, {"index": 61, "intervention_points": {"attn_input": "layers.61.self_attn.", "attn_q_proj": "layers.61.self_attn.q_proj", "attn_k_proj": "layers.61.self_attn.k_proj", "attn_v_proj": "layers.61.self_attn.v_proj", "attn_output_proj": "layers.61.self_attn.o_proj", "mlp_input": "layers.61.mlp.", "mlp_gate_proj": "layers.61.mlp.gate_proj", "mlp_up_proj": "layers.61.mlp.up_proj", "mlp_down_proj": "layers.61.mlp.down_proj", "attn_pre_norm": "layers.61.input_layernorm", "attn_post_norm": "layers.61.post_attention_layernorm"}}, {"index": 62, "intervention_points": {"attn_input": "layers.62.self_attn.", "attn_q_proj": "layers.62.self_attn.q_proj", "attn_k_proj": "layers.62.self_attn.k_proj", "attn_v_proj": "layers.62.self_attn.v_proj", "attn_output_proj": "layers.62.self_attn.o_proj", "mlp_input": "layers.62.mlp.", "mlp_gate_proj": "layers.62.mlp.gate_proj", "mlp_up_proj": "layers.62.mlp.up_proj", "mlp_down_proj": "layers.62.mlp.down_proj", "attn_pre_norm": "layers.62.input_layernorm", "attn_post_norm": "layers.62.post_attention_layernorm"}}, {"index": 63, "intervention_points": {"attn_input": "layers.63.self_attn.", "attn_q_proj": "layers.63.self_attn.q_proj", "attn_k_proj": "layers.63.self_attn.k_proj", "attn_v_proj": "layers.63.self_attn.v_proj", "attn_output_proj": "layers.63.self_attn.o_proj", "mlp_input": "layers.63.mlp.", "mlp_gate_proj": "layers.63.mlp.gate_proj", "mlp_up_proj": "layers.63.mlp.up_proj", "mlp_down_proj": "layers.63.mlp.down_proj", "attn_pre_norm": "layers.63.input_layernorm", "attn_post_norm": "layers.63.post_attention_layernorm"}}, {"index": 64, "intervention_points": {"attn_input": "layers.64.self_attn.", "attn_q_proj": "layers.64.self_attn.q_proj", "attn_k_proj": "layers.64.self_attn.k_proj", "attn_v_proj": "layers.64.self_attn.v_proj", "attn_output_proj": "layers.64.self_attn.o_proj", "mlp_input": "layers.64.mlp.", "mlp_gate_proj": "layers.64.mlp.gate_proj", "mlp_up_proj": "layers.64.mlp.up_proj", "mlp_down_proj": "layers.64.mlp.down_proj", "attn_pre_norm": "layers.64.input_layernorm", "attn_post_norm": "layers.64.post_attention_layernorm"}}, {"index": 65, "intervention_points": {"attn_input": "layers.65.self_attn.", "attn_q_proj": "layers.65.self_attn.q_proj", "attn_k_proj": "layers.65.self_attn.k_proj", "attn_v_proj": "layers.65.self_attn.v_proj", "attn_output_proj": "layers.65.self_attn.o_proj", "mlp_input": "layers.65.mlp.", "mlp_gate_proj": "layers.65.mlp.gate_proj", "mlp_up_proj": "layers.65.mlp.up_proj", "mlp_down_proj": "layers.65.mlp.down_proj", "attn_pre_norm": "layers.65.input_layernorm", "attn_post_norm": "layers.65.post_attention_layernorm"}}, {"index": 66, "intervention_points": {"attn_input": "layers.66.self_attn.", "attn_q_proj": "layers.66.self_attn.q_proj", "attn_k_proj": "layers.66.self_attn.k_proj", "attn_v_proj": "layers.66.self_attn.v_proj", "attn_output_proj": "layers.66.self_attn.o_proj", "mlp_input": "layers.66.mlp.", "mlp_gate_proj": "layers.66.mlp.gate_proj", "mlp_up_proj": "layers.66.mlp.up_proj", "mlp_down_proj": "layers.66.mlp.down_proj", "attn_pre_norm": "layers.66.input_layernorm", "attn_post_norm": "layers.66.post_attention_layernorm"}}, {"index": 67, "intervention_points": {"attn_input": "layers.67.self_attn.", "attn_q_proj": "layers.67.self_attn.q_proj", "attn_k_proj": "layers.67.self_attn.k_proj", "attn_v_proj": "layers.67.self_attn.v_proj", "attn_output_proj": "layers.67.self_attn.o_proj", "mlp_input": "layers.67.mlp.", "mlp_gate_proj": "layers.67.mlp.gate_proj", "mlp_up_proj": "layers.67.mlp.up_proj", "mlp_down_proj": "layers.67.mlp.down_proj", "attn_pre_norm": "layers.67.input_layernorm", "attn_post_norm": "layers.67.post_attention_layernorm"}}, {"index": 68, "intervention_points": {"attn_input": "layers.68.self_attn.", "attn_q_proj": "layers.68.self_attn.q_proj", "attn_k_proj": "layers.68.self_attn.k_proj", "attn_v_proj": "layers.68.self_attn.v_proj", "attn_output_proj": "layers.68.self_attn.o_proj", "mlp_input": "layers.68.mlp.", "mlp_gate_proj": "layers.68.mlp.gate_proj", "mlp_up_proj": "layers.68.mlp.up_proj", "mlp_down_proj": "layers.68.mlp.down_proj", "attn_pre_norm": "layers.68.input_layernorm", "attn_post_norm": "layers.68.post_attention_layernorm"}}, {"index": 69, "intervention_points": {"attn_input": "layers.69.self_attn.", "attn_q_proj": "layers.69.self_attn.q_proj", "attn_k_proj": "layers.69.self_attn.k_proj", "attn_v_proj": "layers.69.self_attn.v_proj", "attn_output_proj": "layers.69.self_attn.o_proj", "mlp_input": "layers.69.mlp.", "mlp_gate_proj": "layers.69.mlp.gate_proj", "mlp_up_proj": "layers.69.mlp.up_proj", "mlp_down_proj": "layers.69.mlp.down_proj", "attn_pre_norm": "layers.69.input_layernorm", "attn_post_norm": "layers.69.post_attention_layernorm"}}, {"index": 70, "intervention_points": {"attn_input": "layers.70.self_attn.", "attn_q_proj": "layers.70.self_attn.q_proj", "attn_k_proj": "layers.70.self_attn.k_proj", "attn_v_proj": "layers.70.self_attn.v_proj", "attn_output_proj": "layers.70.self_attn.o_proj", "mlp_input": "layers.70.mlp.", "mlp_gate_proj": "layers.70.mlp.gate_proj", "mlp_up_proj": "layers.70.mlp.up_proj", "mlp_down_proj": "layers.70.mlp.down_proj", "attn_pre_norm": "layers.70.input_layernorm", "attn_post_norm": "layers.70.post_attention_layernorm"}}, {"index": 71, "intervention_points": {"attn_input": "layers.71.self_attn.", "attn_q_proj": "layers.71.self_attn.q_proj", "attn_k_proj": "layers.71.self_attn.k_proj", "attn_v_proj": "layers.71.self_attn.v_proj", "attn_output_proj": "layers.71.self_attn.o_proj", "mlp_input": "layers.71.mlp.", "mlp_gate_proj": "layers.71.mlp.gate_proj", "mlp_up_proj": "layers.71.mlp.up_proj", "mlp_down_proj": "layers.71.mlp.down_proj", "attn_pre_norm": "layers.71.input_layernorm", "attn_post_norm": "layers.71.post_attention_layernorm"}}, {"index": 72, "intervention_points": {"attn_input": "layers.72.self_attn.", "attn_q_proj": "layers.72.self_attn.q_proj", "attn_k_proj": "layers.72.self_attn.k_proj", "attn_v_proj": "layers.72.self_attn.v_proj", "attn_output_proj": "layers.72.self_attn.o_proj", "mlp_input": "layers.72.mlp.", "mlp_gate_proj": "layers.72.mlp.gate_proj", "mlp_up_proj": "layers.72.mlp.up_proj", "mlp_down_proj": "layers.72.mlp.down_proj", "attn_pre_norm": "layers.72.input_layernorm", "attn_post_norm": "layers.72.post_attention_layernorm"}}, {"index": 73, "intervention_points": {"attn_input": "layers.73.self_attn.", "attn_q_proj": "layers.73.self_attn.q_proj", "attn_k_proj": "layers.73.self_attn.k_proj", "attn_v_proj": "layers.73.self_attn.v_proj", "attn_output_proj": "layers.73.self_attn.o_proj", "mlp_input": "layers.73.mlp.", "mlp_gate_proj": "layers.73.mlp.gate_proj", "mlp_up_proj": "layers.73.mlp.up_proj", "mlp_down_proj": "layers.73.mlp.down_proj", "attn_pre_norm": "layers.73.input_layernorm", "attn_post_norm": "layers.73.post_attention_layernorm"}}, {"index": 74, "intervention_points": {"attn_input": "layers.74.self_attn.", "attn_q_proj": "layers.74.self_attn.q_proj", "attn_k_proj": "layers.74.self_attn.k_proj", "attn_v_proj": "layers.74.self_attn.v_proj", "attn_output_proj": "layers.74.self_attn.o_proj", "mlp_input": "layers.74.mlp.", "mlp_gate_proj": "layers.74.mlp.gate_proj", "mlp_up_proj": "layers.74.mlp.up_proj", "mlp_down_proj": "layers.74.mlp.down_proj", "attn_pre_norm": "layers.74.input_layernorm", "attn_post_norm": "layers.74.post_attention_layernorm"}}, {"index": 75, "intervention_points": {"attn_input": "layers.75.self_attn.", "attn_q_proj": "layers.75.self_attn.q_proj", "attn_k_proj": "layers.75.self_attn.k_proj", "attn_v_proj": "layers.75.self_attn.v_proj", "attn_output_proj": "layers.75.self_attn.o_proj", "mlp_input": "layers.75.mlp.", "mlp_gate_proj": "layers.75.mlp.gate_proj", "mlp_up_proj": "layers.75.mlp.up_proj", "mlp_down_proj": "layers.75.mlp.down_proj", "attn_pre_norm": "layers.75.input_layernorm", "attn_post_norm": "layers.75.post_attention_layernorm"}}, {"index": 76, "intervention_points": {"attn_input": "layers.76.self_attn.", "attn_q_proj": "layers.76.self_attn.q_proj", "attn_k_proj": "layers.76.self_attn.k_proj", "attn_v_proj": "layers.76.self_attn.v_proj", "attn_output_proj": "layers.76.self_attn.o_proj", "mlp_input": "layers.76.mlp.", "mlp_gate_proj": "layers.76.mlp.gate_proj", "mlp_up_proj": "layers.76.mlp.up_proj", "mlp_down_proj": "layers.76.mlp.down_proj", "attn_pre_norm": "layers.76.input_layernorm", "attn_post_norm": "layers.76.post_attention_layernorm"}}, {"index": 77, "intervention_points": {"attn_input": "layers.77.self_attn.", "attn_q_proj": "layers.77.self_attn.q_proj", "attn_k_proj": "layers.77.self_attn.k_proj", "attn_v_proj": "layers.77.self_attn.v_proj", "attn_output_proj": "layers.77.self_attn.o_proj", "mlp_input": "layers.77.mlp.", "mlp_gate_proj": "layers.77.mlp.gate_proj", "mlp_up_proj": "layers.77.mlp.up_proj", "mlp_down_proj": "layers.77.mlp.down_proj", "attn_pre_norm": "layers.77.input_layernorm", "attn_post_norm": "layers.77.post_attention_layernorm"}}, {"index": 78, "intervention_points": {"attn_input": "layers.78.self_attn.", "attn_q_proj": "layers.78.self_attn.q_proj", "attn_k_proj": "layers.78.self_attn.k_proj", "attn_v_proj": "layers.78.self_attn.v_proj", "attn_output_proj": "layers.78.self_attn.o_proj", "mlp_input": "layers.78.mlp.", "mlp_gate_proj": "layers.78.mlp.gate_proj", "mlp_up_proj": "layers.78.mlp.up_proj", "mlp_down_proj": "layers.78.mlp.down_proj", "attn_pre_norm": "layers.78.input_layernorm", "attn_post_norm": "layers.78.post_attention_layernorm"}}, {"index": 79, "intervention_points": {"attn_input": "layers.79.self_attn.", "attn_q_proj": "layers.79.self_attn.q_proj", "attn_k_proj": "layers.79.self_attn.k_proj", "attn_v_proj": "layers.79.self_attn.v_proj", "attn_output_proj": "layers.79.self_attn.o_proj", "mlp_input": "layers.79.mlp.", "mlp_gate_proj": "layers.79.mlp.gate_proj", "mlp_up_proj": "layers.79.mlp.up_proj", "mlp_down_proj": "layers.79.mlp.down_proj", "attn_pre_norm": "layers.79.input_layernorm", "attn_post_norm": "layers.79.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.864995", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
tencent/Hunyuan-A13B-Instruct | tencent | hunyuan_v1_moe | 3 | 2026-03-29T16:20:27.238698+00:00 | {"hidden_size": 4096, "intermediate_size": 3072, "vocab_size": 128167, "max_position_embeddings": 32768, "num_hidden_layers": 32, "rope_theta": 10000.0, "rms_norm_eps": 1e-05, "model_type": "hunyuan_v1_moe"} | {"rope_type": null, "rope_scaling": {"alpha": 1000.0, "beta_fast": 32, "beta_slow": 1, "factor": 1.0, "mscale": 1.0, "mscale_all_dim": 1.0, "type": "dynamic", "rope_theta": 10000.0, "rope_type": "dynamic"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "eager", "flash_attention_capable": true, "position_embedding_type": null, "max_position_embeddings": 32768, "positional_encoding": "rope", "tie_word_embeddings": true, "norm_epsilon": 1e-05, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.1, "actual_norm_class": "HunYuanRMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "tencent", "first_publish_date": "2025-06-25", "paper_urls": [], "base_model": null, "license": "other", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["qwen 2.5", "qwen 3", "switch", "a fine-grained mixture-of-experts", "opt", "the latest version of sglang.\n\nto get started"], "hub_metadata": {"downloads": 22754, "likes": 616, "created_at": "2025-06-25T12:39:52.000Z", "last_modified": "2025-08-21T08:36:52.000Z"}} | {"residual_stream_dim": 4096, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate", "moe_gate": "layers.0.mlp.gate", "moe_experts": "layers.0.mlp.experts", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate", "moe_gate": "layers.1.mlp.gate", "moe_experts": "layers.1.mlp.experts", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate", "moe_gate": "layers.2.mlp.gate", "moe_experts": "layers.2.mlp.experts", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.865095", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
zai-org/GLM-5 | zai-org | glm_moe_dsa | 3 | 2026-03-29T16:30:24.700922+00:00 | {"hidden_size": 6144, "intermediate_size": 12288, "vocab_size": 154880, "max_position_embeddings": 202752, "num_hidden_layers": 78, "rms_norm_eps": 1e-05, "model_type": "glm_moe_dsa"} | {"rope_type": null, "rope_scaling": {"rope_theta": 1000000, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "eager", "attention_architecture": "mla", "mla_q_lora_rank": 2048, "mla_kv_lora_rank": 512, "mla_qk_nope_head_dim": 192, "mla_qk_rope_head_dim": 64, "position_embedding_type": null, "max_position_embeddings": 202752, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-05, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "['GlmMoeDsaRMSNorm', 'LayerNorm']", "linear_bias_present": false} | {"organization": "zai-org", "first_publish_date": "2026-02-11", "paper_urls": [], "base_model": null, "license": "mit", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["glm 47", "glm", "glm 5t", "glm 45", "gpt", "glm 5", "glm 4.7", "glm 4.5", "claude", "gpt 5.2"], "hub_metadata": {"downloads": 214293, "likes": 1885, "created_at": "2026-02-11T04:55:46.000Z", "last_modified": "2026-03-24T06:45:55.000Z"}} | {"residual_stream_dim": 6144, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate", "moe_gate": "layers.3.mlp.gate", "moe_experts": "layers.3.mlp.experts", "moe_shared_expert": "layers.3.mlp.shared_experts", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate", "moe_gate": "layers.4.mlp.gate", "moe_experts": "layers.4.mlp.experts", "moe_shared_expert": "layers.4.mlp.shared_experts", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate", "moe_gate": "layers.5.mlp.gate", "moe_experts": "layers.5.mlp.experts", "moe_shared_expert": "layers.5.mlp.shared_experts", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate", "moe_gate": "layers.6.mlp.gate", "moe_experts": "layers.6.mlp.experts", "moe_shared_expert": "layers.6.mlp.shared_experts", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate", "moe_gate": "layers.7.mlp.gate", "moe_experts": "layers.7.mlp.experts", "moe_shared_expert": "layers.7.mlp.shared_experts", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate", "moe_gate": "layers.8.mlp.gate", "moe_experts": "layers.8.mlp.experts", "moe_shared_expert": "layers.8.mlp.shared_experts", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate", "moe_gate": "layers.9.mlp.gate", "moe_experts": "layers.9.mlp.experts", "moe_shared_expert": "layers.9.mlp.shared_experts", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate", "moe_gate": "layers.10.mlp.gate", "moe_experts": "layers.10.mlp.experts", "moe_shared_expert": "layers.10.mlp.shared_experts", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate", "moe_gate": "layers.11.mlp.gate", "moe_experts": "layers.11.mlp.experts", "moe_shared_expert": "layers.11.mlp.shared_experts", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate", "moe_gate": "layers.12.mlp.gate", "moe_experts": "layers.12.mlp.experts", "moe_shared_expert": "layers.12.mlp.shared_experts", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate", "moe_gate": "layers.13.mlp.gate", "moe_experts": "layers.13.mlp.experts", "moe_shared_expert": "layers.13.mlp.shared_experts", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate", "moe_gate": "layers.14.mlp.gate", "moe_experts": "layers.14.mlp.experts", "moe_shared_expert": "layers.14.mlp.shared_experts", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate", "moe_gate": "layers.15.mlp.gate", "moe_experts": "layers.15.mlp.experts", "moe_shared_expert": "layers.15.mlp.shared_experts", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate", "moe_gate": "layers.16.mlp.gate", "moe_experts": "layers.16.mlp.experts", "moe_shared_expert": "layers.16.mlp.shared_experts", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate", "moe_gate": "layers.17.mlp.gate", "moe_experts": "layers.17.mlp.experts", "moe_shared_expert": "layers.17.mlp.shared_experts", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate", "moe_gate": "layers.18.mlp.gate", "moe_experts": "layers.18.mlp.experts", "moe_shared_expert": "layers.18.mlp.shared_experts", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate", "moe_gate": "layers.19.mlp.gate", "moe_experts": "layers.19.mlp.experts", "moe_shared_expert": "layers.19.mlp.shared_experts", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate", "moe_gate": "layers.20.mlp.gate", "moe_experts": "layers.20.mlp.experts", "moe_shared_expert": "layers.20.mlp.shared_experts", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate", "moe_gate": "layers.21.mlp.gate", "moe_experts": "layers.21.mlp.experts", "moe_shared_expert": "layers.21.mlp.shared_experts", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate", "moe_gate": "layers.22.mlp.gate", "moe_experts": "layers.22.mlp.experts", "moe_shared_expert": "layers.22.mlp.shared_experts", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate", "moe_gate": "layers.23.mlp.gate", "moe_experts": "layers.23.mlp.experts", "moe_shared_expert": "layers.23.mlp.shared_experts", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate", "moe_gate": "layers.24.mlp.gate", "moe_experts": "layers.24.mlp.experts", "moe_shared_expert": "layers.24.mlp.shared_experts", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate", "moe_gate": "layers.25.mlp.gate", "moe_experts": "layers.25.mlp.experts", "moe_shared_expert": "layers.25.mlp.shared_experts", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate", "moe_gate": "layers.26.mlp.gate", "moe_experts": "layers.26.mlp.experts", "moe_shared_expert": "layers.26.mlp.shared_experts", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate", "moe_gate": "layers.27.mlp.gate", "moe_experts": "layers.27.mlp.experts", "moe_shared_expert": "layers.27.mlp.shared_experts", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate", "moe_gate": "layers.28.mlp.gate", "moe_experts": "layers.28.mlp.experts", "moe_shared_expert": "layers.28.mlp.shared_experts", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate", "moe_gate": "layers.29.mlp.gate", "moe_experts": "layers.29.mlp.experts", "moe_shared_expert": "layers.29.mlp.shared_experts", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate", "moe_gate": "layers.30.mlp.gate", "moe_experts": "layers.30.mlp.experts", "moe_shared_expert": "layers.30.mlp.shared_experts", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate", "moe_gate": "layers.31.mlp.gate", "moe_experts": "layers.31.mlp.experts", "moe_shared_expert": "layers.31.mlp.shared_experts", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate", "moe_gate": "layers.32.mlp.gate", "moe_experts": "layers.32.mlp.experts", "moe_shared_expert": "layers.32.mlp.shared_experts", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate", "moe_gate": "layers.33.mlp.gate", "moe_experts": "layers.33.mlp.experts", "moe_shared_expert": "layers.33.mlp.shared_experts", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate", "moe_gate": "layers.34.mlp.gate", "moe_experts": "layers.34.mlp.experts", "moe_shared_expert": "layers.34.mlp.shared_experts", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate", "moe_gate": "layers.35.mlp.gate", "moe_experts": "layers.35.mlp.experts", "moe_shared_expert": "layers.35.mlp.shared_experts", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate", "moe_gate": "layers.36.mlp.gate", "moe_experts": "layers.36.mlp.experts", "moe_shared_expert": "layers.36.mlp.shared_experts", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate", "moe_gate": "layers.37.mlp.gate", "moe_experts": "layers.37.mlp.experts", "moe_shared_expert": "layers.37.mlp.shared_experts", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate", "moe_gate": "layers.38.mlp.gate", "moe_experts": "layers.38.mlp.experts", "moe_shared_expert": "layers.38.mlp.shared_experts", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate", "moe_gate": "layers.39.mlp.gate", "moe_experts": "layers.39.mlp.experts", "moe_shared_expert": "layers.39.mlp.shared_experts", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate", "moe_gate": "layers.40.mlp.gate", "moe_experts": "layers.40.mlp.experts", "moe_shared_expert": "layers.40.mlp.shared_experts", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate", "moe_gate": "layers.41.mlp.gate", "moe_experts": "layers.41.mlp.experts", "moe_shared_expert": "layers.41.mlp.shared_experts", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate", "moe_gate": "layers.42.mlp.gate", "moe_experts": "layers.42.mlp.experts", "moe_shared_expert": "layers.42.mlp.shared_experts", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate", "moe_gate": "layers.43.mlp.gate", "moe_experts": "layers.43.mlp.experts", "moe_shared_expert": "layers.43.mlp.shared_experts", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate", "moe_gate": "layers.44.mlp.gate", "moe_experts": "layers.44.mlp.experts", "moe_shared_expert": "layers.44.mlp.shared_experts", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate", "moe_gate": "layers.45.mlp.gate", "moe_experts": "layers.45.mlp.experts", "moe_shared_expert": "layers.45.mlp.shared_experts", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate", "moe_gate": "layers.46.mlp.gate", "moe_experts": "layers.46.mlp.experts", "moe_shared_expert": "layers.46.mlp.shared_experts", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate", "moe_gate": "layers.47.mlp.gate", "moe_experts": "layers.47.mlp.experts", "moe_shared_expert": "layers.47.mlp.shared_experts", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate", "moe_gate": "layers.48.mlp.gate", "moe_experts": "layers.48.mlp.experts", "moe_shared_expert": "layers.48.mlp.shared_experts", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate", "moe_gate": "layers.49.mlp.gate", "moe_experts": "layers.49.mlp.experts", "moe_shared_expert": "layers.49.mlp.shared_experts", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate", "moe_gate": "layers.50.mlp.gate", "moe_experts": "layers.50.mlp.experts", "moe_shared_expert": "layers.50.mlp.shared_experts", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate", "moe_gate": "layers.51.mlp.gate", "moe_experts": "layers.51.mlp.experts", "moe_shared_expert": "layers.51.mlp.shared_experts", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate", "moe_gate": "layers.52.mlp.gate", "moe_experts": "layers.52.mlp.experts", "moe_shared_expert": "layers.52.mlp.shared_experts", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate", "moe_gate": "layers.53.mlp.gate", "moe_experts": "layers.53.mlp.experts", "moe_shared_expert": "layers.53.mlp.shared_experts", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate", "moe_gate": "layers.54.mlp.gate", "moe_experts": "layers.54.mlp.experts", "moe_shared_expert": "layers.54.mlp.shared_experts", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate", "moe_gate": "layers.55.mlp.gate", "moe_experts": "layers.55.mlp.experts", "moe_shared_expert": "layers.55.mlp.shared_experts", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate", "moe_gate": "layers.56.mlp.gate", "moe_experts": "layers.56.mlp.experts", "moe_shared_expert": "layers.56.mlp.shared_experts", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate", "moe_gate": "layers.57.mlp.gate", "moe_experts": "layers.57.mlp.experts", "moe_shared_expert": "layers.57.mlp.shared_experts", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate", "moe_gate": "layers.58.mlp.gate", "moe_experts": "layers.58.mlp.experts", "moe_shared_expert": "layers.58.mlp.shared_experts", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate", "moe_gate": "layers.59.mlp.gate", "moe_experts": "layers.59.mlp.experts", "moe_shared_expert": "layers.59.mlp.shared_experts", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate", "moe_gate": "layers.60.mlp.gate", "moe_experts": "layers.60.mlp.experts", "moe_shared_expert": "layers.60.mlp.shared_experts", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}, {"index": 61, "intervention_points": {"attn_input": "layers.61.self_attn.", "attn_output_proj": "layers.61.self_attn.o_proj", "mlp_input": "layers.61.mlp.", "mlp_gate_proj": "layers.61.mlp.gate", "moe_gate": "layers.61.mlp.gate", "moe_experts": "layers.61.mlp.experts", "moe_shared_expert": "layers.61.mlp.shared_experts", "attn_pre_norm": "layers.61.input_layernorm", "attn_post_norm": "layers.61.post_attention_layernorm"}}, {"index": 62, "intervention_points": {"attn_input": "layers.62.self_attn.", "attn_output_proj": "layers.62.self_attn.o_proj", "mlp_input": "layers.62.mlp.", "mlp_gate_proj": "layers.62.mlp.gate", "moe_gate": "layers.62.mlp.gate", "moe_experts": "layers.62.mlp.experts", "moe_shared_expert": "layers.62.mlp.shared_experts", "attn_pre_norm": "layers.62.input_layernorm", "attn_post_norm": "layers.62.post_attention_layernorm"}}, {"index": 63, "intervention_points": {"attn_input": "layers.63.self_attn.", "attn_output_proj": "layers.63.self_attn.o_proj", "mlp_input": "layers.63.mlp.", "mlp_gate_proj": "layers.63.mlp.gate", "moe_gate": "layers.63.mlp.gate", "moe_experts": "layers.63.mlp.experts", "moe_shared_expert": "layers.63.mlp.shared_experts", "attn_pre_norm": "layers.63.input_layernorm", "attn_post_norm": "layers.63.post_attention_layernorm"}}, {"index": 64, "intervention_points": {"attn_input": "layers.64.self_attn.", "attn_output_proj": "layers.64.self_attn.o_proj", "mlp_input": "layers.64.mlp.", "mlp_gate_proj": "layers.64.mlp.gate", "moe_gate": "layers.64.mlp.gate", "moe_experts": "layers.64.mlp.experts", "moe_shared_expert": "layers.64.mlp.shared_experts", "attn_pre_norm": "layers.64.input_layernorm", "attn_post_norm": "layers.64.post_attention_layernorm"}}, {"index": 65, "intervention_points": {"attn_input": "layers.65.self_attn.", "attn_output_proj": "layers.65.self_attn.o_proj", "mlp_input": "layers.65.mlp.", "mlp_gate_proj": "layers.65.mlp.gate", "moe_gate": "layers.65.mlp.gate", "moe_experts": "layers.65.mlp.experts", "moe_shared_expert": "layers.65.mlp.shared_experts", "attn_pre_norm": "layers.65.input_layernorm", "attn_post_norm": "layers.65.post_attention_layernorm"}}, {"index": 66, "intervention_points": {"attn_input": "layers.66.self_attn.", "attn_output_proj": "layers.66.self_attn.o_proj", "mlp_input": "layers.66.mlp.", "mlp_gate_proj": "layers.66.mlp.gate", "moe_gate": "layers.66.mlp.gate", "moe_experts": "layers.66.mlp.experts", "moe_shared_expert": "layers.66.mlp.shared_experts", "attn_pre_norm": "layers.66.input_layernorm", "attn_post_norm": "layers.66.post_attention_layernorm"}}, {"index": 67, "intervention_points": {"attn_input": "layers.67.self_attn.", "attn_output_proj": "layers.67.self_attn.o_proj", "mlp_input": "layers.67.mlp.", "mlp_gate_proj": "layers.67.mlp.gate", "moe_gate": "layers.67.mlp.gate", "moe_experts": "layers.67.mlp.experts", "moe_shared_expert": "layers.67.mlp.shared_experts", "attn_pre_norm": "layers.67.input_layernorm", "attn_post_norm": "layers.67.post_attention_layernorm"}}, {"index": 68, "intervention_points": {"attn_input": "layers.68.self_attn.", "attn_output_proj": "layers.68.self_attn.o_proj", "mlp_input": "layers.68.mlp.", "mlp_gate_proj": "layers.68.mlp.gate", "moe_gate": "layers.68.mlp.gate", "moe_experts": "layers.68.mlp.experts", "moe_shared_expert": "layers.68.mlp.shared_experts", "attn_pre_norm": "layers.68.input_layernorm", "attn_post_norm": "layers.68.post_attention_layernorm"}}, {"index": 69, "intervention_points": {"attn_input": "layers.69.self_attn.", "attn_output_proj": "layers.69.self_attn.o_proj", "mlp_input": "layers.69.mlp.", "mlp_gate_proj": "layers.69.mlp.gate", "moe_gate": "layers.69.mlp.gate", "moe_experts": "layers.69.mlp.experts", "moe_shared_expert": "layers.69.mlp.shared_experts", "attn_pre_norm": "layers.69.input_layernorm", "attn_post_norm": "layers.69.post_attention_layernorm"}}, {"index": 70, "intervention_points": {"attn_input": "layers.70.self_attn.", "attn_output_proj": "layers.70.self_attn.o_proj", "mlp_input": "layers.70.mlp.", "mlp_gate_proj": "layers.70.mlp.gate", "moe_gate": "layers.70.mlp.gate", "moe_experts": "layers.70.mlp.experts", "moe_shared_expert": "layers.70.mlp.shared_experts", "attn_pre_norm": "layers.70.input_layernorm", "attn_post_norm": "layers.70.post_attention_layernorm"}}, {"index": 71, "intervention_points": {"attn_input": "layers.71.self_attn.", "attn_output_proj": "layers.71.self_attn.o_proj", "mlp_input": "layers.71.mlp.", "mlp_gate_proj": "layers.71.mlp.gate", "moe_gate": "layers.71.mlp.gate", "moe_experts": "layers.71.mlp.experts", "moe_shared_expert": "layers.71.mlp.shared_experts", "attn_pre_norm": "layers.71.input_layernorm", "attn_post_norm": "layers.71.post_attention_layernorm"}}, {"index": 72, "intervention_points": {"attn_input": "layers.72.self_attn.", "attn_output_proj": "layers.72.self_attn.o_proj", "mlp_input": "layers.72.mlp.", "mlp_gate_proj": "layers.72.mlp.gate", "moe_gate": "layers.72.mlp.gate", "moe_experts": "layers.72.mlp.experts", "moe_shared_expert": "layers.72.mlp.shared_experts", "attn_pre_norm": "layers.72.input_layernorm", "attn_post_norm": "layers.72.post_attention_layernorm"}}, {"index": 73, "intervention_points": {"attn_input": "layers.73.self_attn.", "attn_output_proj": "layers.73.self_attn.o_proj", "mlp_input": "layers.73.mlp.", "mlp_gate_proj": "layers.73.mlp.gate", "moe_gate": "layers.73.mlp.gate", "moe_experts": "layers.73.mlp.experts", "moe_shared_expert": "layers.73.mlp.shared_experts", "attn_pre_norm": "layers.73.input_layernorm", "attn_post_norm": "layers.73.post_attention_layernorm"}}, {"index": 74, "intervention_points": {"attn_input": "layers.74.self_attn.", "attn_output_proj": "layers.74.self_attn.o_proj", "mlp_input": "layers.74.mlp.", "mlp_gate_proj": "layers.74.mlp.gate", "moe_gate": "layers.74.mlp.gate", "moe_experts": "layers.74.mlp.experts", "moe_shared_expert": "layers.74.mlp.shared_experts", "attn_pre_norm": "layers.74.input_layernorm", "attn_post_norm": "layers.74.post_attention_layernorm"}}, {"index": 75, "intervention_points": {"attn_input": "layers.75.self_attn.", "attn_output_proj": "layers.75.self_attn.o_proj", "mlp_input": "layers.75.mlp.", "mlp_gate_proj": "layers.75.mlp.gate", "moe_gate": "layers.75.mlp.gate", "moe_experts": "layers.75.mlp.experts", "moe_shared_expert": "layers.75.mlp.shared_experts", "attn_pre_norm": "layers.75.input_layernorm", "attn_post_norm": "layers.75.post_attention_layernorm"}}, {"index": 76, "intervention_points": {"attn_input": "layers.76.self_attn.", "attn_output_proj": "layers.76.self_attn.o_proj", "mlp_input": "layers.76.mlp.", "mlp_gate_proj": "layers.76.mlp.gate", "moe_gate": "layers.76.mlp.gate", "moe_experts": "layers.76.mlp.experts", "moe_shared_expert": "layers.76.mlp.shared_experts", "attn_pre_norm": "layers.76.input_layernorm", "attn_post_norm": "layers.76.post_attention_layernorm"}}, {"index": 77, "intervention_points": {"attn_input": "layers.77.self_attn.", "attn_output_proj": "layers.77.self_attn.o_proj", "mlp_input": "layers.77.mlp.", "mlp_gate_proj": "layers.77.mlp.gate", "moe_gate": "layers.77.mlp.gate", "moe_experts": "layers.77.mlp.experts", "moe_shared_expert": "layers.77.mlp.shared_experts", "attn_pre_norm": "layers.77.input_layernorm", "attn_post_norm": "layers.77.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:06:09.865240", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
IQuestLab/IQuest-Coder-V1-40B-Instruct | IQuestLab | iquestcoder | 3 | 2026-04-01T23:40:19.538892+00:00 | {"hidden_size": 5120, "intermediate_size": 27648, "vocab_size": 76800, "max_position_embeddings": 131072, "num_hidden_layers": 80, "rope_theta": 500000.0, "rms_norm_eps": 1e-05, "model_type": "iquestcoder"} | {"rope_type": null, "rope_scaling": {"rope_theta": 500000.0, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "eager", "flash_attention_capable": true, "position_embedding_type": null, "max_position_embeddings": 131072, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-05, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "IQuestCoderRMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "IQuestLab", "first_publish_date": "2025-12-30", "paper_urls": [], "base_model": null, "license": "other", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["qwen 3"], "hub_metadata": {"downloads": 12216, "likes": 289, "created_at": "2025-12-30T15:30:10.000Z", "last_modified": "2026-03-04T12:15:39.000Z"}} | {"residual_stream_dim": 5120, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate_proj", "mlp_up_proj": "layers.36.mlp.up_proj", "mlp_down_proj": "layers.36.mlp.down_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate_proj", "mlp_up_proj": "layers.37.mlp.up_proj", "mlp_down_proj": "layers.37.mlp.down_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate_proj", "mlp_up_proj": "layers.38.mlp.up_proj", "mlp_down_proj": "layers.38.mlp.down_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate_proj", "mlp_up_proj": "layers.39.mlp.up_proj", "mlp_down_proj": "layers.39.mlp.down_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate_proj", "mlp_up_proj": "layers.40.mlp.up_proj", "mlp_down_proj": "layers.40.mlp.down_proj", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate_proj", "mlp_up_proj": "layers.41.mlp.up_proj", "mlp_down_proj": "layers.41.mlp.down_proj", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate_proj", "mlp_up_proj": "layers.42.mlp.up_proj", "mlp_down_proj": "layers.42.mlp.down_proj", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate_proj", "mlp_up_proj": "layers.43.mlp.up_proj", "mlp_down_proj": "layers.43.mlp.down_proj", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate_proj", "mlp_up_proj": "layers.44.mlp.up_proj", "mlp_down_proj": "layers.44.mlp.down_proj", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate_proj", "mlp_up_proj": "layers.45.mlp.up_proj", "mlp_down_proj": "layers.45.mlp.down_proj", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate_proj", "mlp_up_proj": "layers.46.mlp.up_proj", "mlp_down_proj": "layers.46.mlp.down_proj", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate_proj", "mlp_up_proj": "layers.47.mlp.up_proj", "mlp_down_proj": "layers.47.mlp.down_proj", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_q_proj": "layers.48.self_attn.q_proj", "attn_k_proj": "layers.48.self_attn.k_proj", "attn_v_proj": "layers.48.self_attn.v_proj", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate_proj", "mlp_up_proj": "layers.48.mlp.up_proj", "mlp_down_proj": "layers.48.mlp.down_proj", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_q_proj": "layers.49.self_attn.q_proj", "attn_k_proj": "layers.49.self_attn.k_proj", "attn_v_proj": "layers.49.self_attn.v_proj", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate_proj", "mlp_up_proj": "layers.49.mlp.up_proj", "mlp_down_proj": "layers.49.mlp.down_proj", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_q_proj": "layers.50.self_attn.q_proj", "attn_k_proj": "layers.50.self_attn.k_proj", "attn_v_proj": "layers.50.self_attn.v_proj", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate_proj", "mlp_up_proj": "layers.50.mlp.up_proj", "mlp_down_proj": "layers.50.mlp.down_proj", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate_proj", "mlp_up_proj": "layers.51.mlp.up_proj", "mlp_down_proj": "layers.51.mlp.down_proj", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_q_proj": "layers.52.self_attn.q_proj", "attn_k_proj": "layers.52.self_attn.k_proj", "attn_v_proj": "layers.52.self_attn.v_proj", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate_proj", "mlp_up_proj": "layers.52.mlp.up_proj", "mlp_down_proj": "layers.52.mlp.down_proj", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_q_proj": "layers.53.self_attn.q_proj", "attn_k_proj": "layers.53.self_attn.k_proj", "attn_v_proj": "layers.53.self_attn.v_proj", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate_proj", "mlp_up_proj": "layers.53.mlp.up_proj", "mlp_down_proj": "layers.53.mlp.down_proj", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_q_proj": "layers.54.self_attn.q_proj", "attn_k_proj": "layers.54.self_attn.k_proj", "attn_v_proj": "layers.54.self_attn.v_proj", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate_proj", "mlp_up_proj": "layers.54.mlp.up_proj", "mlp_down_proj": "layers.54.mlp.down_proj", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate_proj", "mlp_up_proj": "layers.55.mlp.up_proj", "mlp_down_proj": "layers.55.mlp.down_proj", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_q_proj": "layers.56.self_attn.q_proj", "attn_k_proj": "layers.56.self_attn.k_proj", "attn_v_proj": "layers.56.self_attn.v_proj", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate_proj", "mlp_up_proj": "layers.56.mlp.up_proj", "mlp_down_proj": "layers.56.mlp.down_proj", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_q_proj": "layers.57.self_attn.q_proj", "attn_k_proj": "layers.57.self_attn.k_proj", "attn_v_proj": "layers.57.self_attn.v_proj", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate_proj", "mlp_up_proj": "layers.57.mlp.up_proj", "mlp_down_proj": "layers.57.mlp.down_proj", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_q_proj": "layers.58.self_attn.q_proj", "attn_k_proj": "layers.58.self_attn.k_proj", "attn_v_proj": "layers.58.self_attn.v_proj", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate_proj", "mlp_up_proj": "layers.58.mlp.up_proj", "mlp_down_proj": "layers.58.mlp.down_proj", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate_proj", "mlp_up_proj": "layers.59.mlp.up_proj", "mlp_down_proj": "layers.59.mlp.down_proj", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_q_proj": "layers.60.self_attn.q_proj", "attn_k_proj": "layers.60.self_attn.k_proj", "attn_v_proj": "layers.60.self_attn.v_proj", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate_proj", "mlp_up_proj": "layers.60.mlp.up_proj", "mlp_down_proj": "layers.60.mlp.down_proj", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}, {"index": 61, "intervention_points": {"attn_input": "layers.61.self_attn.", "attn_q_proj": "layers.61.self_attn.q_proj", "attn_k_proj": "layers.61.self_attn.k_proj", "attn_v_proj": "layers.61.self_attn.v_proj", "attn_output_proj": "layers.61.self_attn.o_proj", "mlp_input": "layers.61.mlp.", "mlp_gate_proj": "layers.61.mlp.gate_proj", "mlp_up_proj": "layers.61.mlp.up_proj", "mlp_down_proj": "layers.61.mlp.down_proj", "attn_pre_norm": "layers.61.input_layernorm", "attn_post_norm": "layers.61.post_attention_layernorm"}}, {"index": 62, "intervention_points": {"attn_input": "layers.62.self_attn.", "attn_q_proj": "layers.62.self_attn.q_proj", "attn_k_proj": "layers.62.self_attn.k_proj", "attn_v_proj": "layers.62.self_attn.v_proj", "attn_output_proj": "layers.62.self_attn.o_proj", "mlp_input": "layers.62.mlp.", "mlp_gate_proj": "layers.62.mlp.gate_proj", "mlp_up_proj": "layers.62.mlp.up_proj", "mlp_down_proj": "layers.62.mlp.down_proj", "attn_pre_norm": "layers.62.input_layernorm", "attn_post_norm": "layers.62.post_attention_layernorm"}}, {"index": 63, "intervention_points": {"attn_input": "layers.63.self_attn.", "attn_q_proj": "layers.63.self_attn.q_proj", "attn_k_proj": "layers.63.self_attn.k_proj", "attn_v_proj": "layers.63.self_attn.v_proj", "attn_output_proj": "layers.63.self_attn.o_proj", "mlp_input": "layers.63.mlp.", "mlp_gate_proj": "layers.63.mlp.gate_proj", "mlp_up_proj": "layers.63.mlp.up_proj", "mlp_down_proj": "layers.63.mlp.down_proj", "attn_pre_norm": "layers.63.input_layernorm", "attn_post_norm": "layers.63.post_attention_layernorm"}}, {"index": 64, "intervention_points": {"attn_input": "layers.64.self_attn.", "attn_q_proj": "layers.64.self_attn.q_proj", "attn_k_proj": "layers.64.self_attn.k_proj", "attn_v_proj": "layers.64.self_attn.v_proj", "attn_output_proj": "layers.64.self_attn.o_proj", "mlp_input": "layers.64.mlp.", "mlp_gate_proj": "layers.64.mlp.gate_proj", "mlp_up_proj": "layers.64.mlp.up_proj", "mlp_down_proj": "layers.64.mlp.down_proj", "attn_pre_norm": "layers.64.input_layernorm", "attn_post_norm": "layers.64.post_attention_layernorm"}}, {"index": 65, "intervention_points": {"attn_input": "layers.65.self_attn.", "attn_q_proj": "layers.65.self_attn.q_proj", "attn_k_proj": "layers.65.self_attn.k_proj", "attn_v_proj": "layers.65.self_attn.v_proj", "attn_output_proj": "layers.65.self_attn.o_proj", "mlp_input": "layers.65.mlp.", "mlp_gate_proj": "layers.65.mlp.gate_proj", "mlp_up_proj": "layers.65.mlp.up_proj", "mlp_down_proj": "layers.65.mlp.down_proj", "attn_pre_norm": "layers.65.input_layernorm", "attn_post_norm": "layers.65.post_attention_layernorm"}}, {"index": 66, "intervention_points": {"attn_input": "layers.66.self_attn.", "attn_q_proj": "layers.66.self_attn.q_proj", "attn_k_proj": "layers.66.self_attn.k_proj", "attn_v_proj": "layers.66.self_attn.v_proj", "attn_output_proj": "layers.66.self_attn.o_proj", "mlp_input": "layers.66.mlp.", "mlp_gate_proj": "layers.66.mlp.gate_proj", "mlp_up_proj": "layers.66.mlp.up_proj", "mlp_down_proj": "layers.66.mlp.down_proj", "attn_pre_norm": "layers.66.input_layernorm", "attn_post_norm": "layers.66.post_attention_layernorm"}}, {"index": 67, "intervention_points": {"attn_input": "layers.67.self_attn.", "attn_q_proj": "layers.67.self_attn.q_proj", "attn_k_proj": "layers.67.self_attn.k_proj", "attn_v_proj": "layers.67.self_attn.v_proj", "attn_output_proj": "layers.67.self_attn.o_proj", "mlp_input": "layers.67.mlp.", "mlp_gate_proj": "layers.67.mlp.gate_proj", "mlp_up_proj": "layers.67.mlp.up_proj", "mlp_down_proj": "layers.67.mlp.down_proj", "attn_pre_norm": "layers.67.input_layernorm", "attn_post_norm": "layers.67.post_attention_layernorm"}}, {"index": 68, "intervention_points": {"attn_input": "layers.68.self_attn.", "attn_q_proj": "layers.68.self_attn.q_proj", "attn_k_proj": "layers.68.self_attn.k_proj", "attn_v_proj": "layers.68.self_attn.v_proj", "attn_output_proj": "layers.68.self_attn.o_proj", "mlp_input": "layers.68.mlp.", "mlp_gate_proj": "layers.68.mlp.gate_proj", "mlp_up_proj": "layers.68.mlp.up_proj", "mlp_down_proj": "layers.68.mlp.down_proj", "attn_pre_norm": "layers.68.input_layernorm", "attn_post_norm": "layers.68.post_attention_layernorm"}}, {"index": 69, "intervention_points": {"attn_input": "layers.69.self_attn.", "attn_q_proj": "layers.69.self_attn.q_proj", "attn_k_proj": "layers.69.self_attn.k_proj", "attn_v_proj": "layers.69.self_attn.v_proj", "attn_output_proj": "layers.69.self_attn.o_proj", "mlp_input": "layers.69.mlp.", "mlp_gate_proj": "layers.69.mlp.gate_proj", "mlp_up_proj": "layers.69.mlp.up_proj", "mlp_down_proj": "layers.69.mlp.down_proj", "attn_pre_norm": "layers.69.input_layernorm", "attn_post_norm": "layers.69.post_attention_layernorm"}}, {"index": 70, "intervention_points": {"attn_input": "layers.70.self_attn.", "attn_q_proj": "layers.70.self_attn.q_proj", "attn_k_proj": "layers.70.self_attn.k_proj", "attn_v_proj": "layers.70.self_attn.v_proj", "attn_output_proj": "layers.70.self_attn.o_proj", "mlp_input": "layers.70.mlp.", "mlp_gate_proj": "layers.70.mlp.gate_proj", "mlp_up_proj": "layers.70.mlp.up_proj", "mlp_down_proj": "layers.70.mlp.down_proj", "attn_pre_norm": "layers.70.input_layernorm", "attn_post_norm": "layers.70.post_attention_layernorm"}}, {"index": 71, "intervention_points": {"attn_input": "layers.71.self_attn.", "attn_q_proj": "layers.71.self_attn.q_proj", "attn_k_proj": "layers.71.self_attn.k_proj", "attn_v_proj": "layers.71.self_attn.v_proj", "attn_output_proj": "layers.71.self_attn.o_proj", "mlp_input": "layers.71.mlp.", "mlp_gate_proj": "layers.71.mlp.gate_proj", "mlp_up_proj": "layers.71.mlp.up_proj", "mlp_down_proj": "layers.71.mlp.down_proj", "attn_pre_norm": "layers.71.input_layernorm", "attn_post_norm": "layers.71.post_attention_layernorm"}}, {"index": 72, "intervention_points": {"attn_input": "layers.72.self_attn.", "attn_q_proj": "layers.72.self_attn.q_proj", "attn_k_proj": "layers.72.self_attn.k_proj", "attn_v_proj": "layers.72.self_attn.v_proj", "attn_output_proj": "layers.72.self_attn.o_proj", "mlp_input": "layers.72.mlp.", "mlp_gate_proj": "layers.72.mlp.gate_proj", "mlp_up_proj": "layers.72.mlp.up_proj", "mlp_down_proj": "layers.72.mlp.down_proj", "attn_pre_norm": "layers.72.input_layernorm", "attn_post_norm": "layers.72.post_attention_layernorm"}}, {"index": 73, "intervention_points": {"attn_input": "layers.73.self_attn.", "attn_q_proj": "layers.73.self_attn.q_proj", "attn_k_proj": "layers.73.self_attn.k_proj", "attn_v_proj": "layers.73.self_attn.v_proj", "attn_output_proj": "layers.73.self_attn.o_proj", "mlp_input": "layers.73.mlp.", "mlp_gate_proj": "layers.73.mlp.gate_proj", "mlp_up_proj": "layers.73.mlp.up_proj", "mlp_down_proj": "layers.73.mlp.down_proj", "attn_pre_norm": "layers.73.input_layernorm", "attn_post_norm": "layers.73.post_attention_layernorm"}}, {"index": 74, "intervention_points": {"attn_input": "layers.74.self_attn.", "attn_q_proj": "layers.74.self_attn.q_proj", "attn_k_proj": "layers.74.self_attn.k_proj", "attn_v_proj": "layers.74.self_attn.v_proj", "attn_output_proj": "layers.74.self_attn.o_proj", "mlp_input": "layers.74.mlp.", "mlp_gate_proj": "layers.74.mlp.gate_proj", "mlp_up_proj": "layers.74.mlp.up_proj", "mlp_down_proj": "layers.74.mlp.down_proj", "attn_pre_norm": "layers.74.input_layernorm", "attn_post_norm": "layers.74.post_attention_layernorm"}}, {"index": 75, "intervention_points": {"attn_input": "layers.75.self_attn.", "attn_q_proj": "layers.75.self_attn.q_proj", "attn_k_proj": "layers.75.self_attn.k_proj", "attn_v_proj": "layers.75.self_attn.v_proj", "attn_output_proj": "layers.75.self_attn.o_proj", "mlp_input": "layers.75.mlp.", "mlp_gate_proj": "layers.75.mlp.gate_proj", "mlp_up_proj": "layers.75.mlp.up_proj", "mlp_down_proj": "layers.75.mlp.down_proj", "attn_pre_norm": "layers.75.input_layernorm", "attn_post_norm": "layers.75.post_attention_layernorm"}}, {"index": 76, "intervention_points": {"attn_input": "layers.76.self_attn.", "attn_q_proj": "layers.76.self_attn.q_proj", "attn_k_proj": "layers.76.self_attn.k_proj", "attn_v_proj": "layers.76.self_attn.v_proj", "attn_output_proj": "layers.76.self_attn.o_proj", "mlp_input": "layers.76.mlp.", "mlp_gate_proj": "layers.76.mlp.gate_proj", "mlp_up_proj": "layers.76.mlp.up_proj", "mlp_down_proj": "layers.76.mlp.down_proj", "attn_pre_norm": "layers.76.input_layernorm", "attn_post_norm": "layers.76.post_attention_layernorm"}}, {"index": 77, "intervention_points": {"attn_input": "layers.77.self_attn.", "attn_q_proj": "layers.77.self_attn.q_proj", "attn_k_proj": "layers.77.self_attn.k_proj", "attn_v_proj": "layers.77.self_attn.v_proj", "attn_output_proj": "layers.77.self_attn.o_proj", "mlp_input": "layers.77.mlp.", "mlp_gate_proj": "layers.77.mlp.gate_proj", "mlp_up_proj": "layers.77.mlp.up_proj", "mlp_down_proj": "layers.77.mlp.down_proj", "attn_pre_norm": "layers.77.input_layernorm", "attn_post_norm": "layers.77.post_attention_layernorm"}}, {"index": 78, "intervention_points": {"attn_input": "layers.78.self_attn.", "attn_q_proj": "layers.78.self_attn.q_proj", "attn_k_proj": "layers.78.self_attn.k_proj", "attn_v_proj": "layers.78.self_attn.v_proj", "attn_output_proj": "layers.78.self_attn.o_proj", "mlp_input": "layers.78.mlp.", "mlp_gate_proj": "layers.78.mlp.gate_proj", "mlp_up_proj": "layers.78.mlp.up_proj", "mlp_down_proj": "layers.78.mlp.down_proj", "attn_pre_norm": "layers.78.input_layernorm", "attn_post_norm": "layers.78.post_attention_layernorm"}}, {"index": 79, "intervention_points": {"attn_input": "layers.79.self_attn.", "attn_q_proj": "layers.79.self_attn.q_proj", "attn_k_proj": "layers.79.self_attn.k_proj", "attn_v_proj": "layers.79.self_attn.v_proj", "attn_output_proj": "layers.79.self_attn.o_proj", "mlp_input": "layers.79.mlp.", "mlp_gate_proj": "layers.79.mlp.gate_proj", "mlp_up_proj": "layers.79.mlp.up_proj", "mlp_down_proj": "layers.79.mlp.down_proj", "attn_pre_norm": "layers.79.input_layernorm", "attn_post_norm": "layers.79.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:41:14.509541", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
trohrbaugh/IQuest-Coder-V1-40B-Thinking-heretic | trohrbaugh | iquestcoder | 3 | 2026-04-01T23:40:47.306445+00:00 | {"hidden_size": 5120, "intermediate_size": 27648, "vocab_size": 76800, "max_position_embeddings": 131072, "num_hidden_layers": 80, "rope_theta": 500000.0, "rms_norm_eps": 1e-05, "model_type": "iquestcoder"} | {"rope_type": null, "rope_scaling": {"rope_theta": 500000.0, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "eager", "flash_attention_capable": true, "position_embedding_type": null, "max_position_embeddings": 131072, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-05, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "IQuestCoderRMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "trohrbaugh", "first_publish_date": "2026-04-01", "paper_urls": [], "base_model": null, "license": "other", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["qwen 3", "claude"], "hub_metadata": {"downloads": 0, "likes": 0, "created_at": "2026-04-01T18:23:14.000Z", "last_modified": "2026-04-01T18:38:13.000Z"}} | {"residual_stream_dim": 5120, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate_proj", "mlp_up_proj": "layers.36.mlp.up_proj", "mlp_down_proj": "layers.36.mlp.down_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate_proj", "mlp_up_proj": "layers.37.mlp.up_proj", "mlp_down_proj": "layers.37.mlp.down_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate_proj", "mlp_up_proj": "layers.38.mlp.up_proj", "mlp_down_proj": "layers.38.mlp.down_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate_proj", "mlp_up_proj": "layers.39.mlp.up_proj", "mlp_down_proj": "layers.39.mlp.down_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate_proj", "mlp_up_proj": "layers.40.mlp.up_proj", "mlp_down_proj": "layers.40.mlp.down_proj", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate_proj", "mlp_up_proj": "layers.41.mlp.up_proj", "mlp_down_proj": "layers.41.mlp.down_proj", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate_proj", "mlp_up_proj": "layers.42.mlp.up_proj", "mlp_down_proj": "layers.42.mlp.down_proj", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate_proj", "mlp_up_proj": "layers.43.mlp.up_proj", "mlp_down_proj": "layers.43.mlp.down_proj", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate_proj", "mlp_up_proj": "layers.44.mlp.up_proj", "mlp_down_proj": "layers.44.mlp.down_proj", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate_proj", "mlp_up_proj": "layers.45.mlp.up_proj", "mlp_down_proj": "layers.45.mlp.down_proj", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate_proj", "mlp_up_proj": "layers.46.mlp.up_proj", "mlp_down_proj": "layers.46.mlp.down_proj", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate_proj", "mlp_up_proj": "layers.47.mlp.up_proj", "mlp_down_proj": "layers.47.mlp.down_proj", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_q_proj": "layers.48.self_attn.q_proj", "attn_k_proj": "layers.48.self_attn.k_proj", "attn_v_proj": "layers.48.self_attn.v_proj", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate_proj", "mlp_up_proj": "layers.48.mlp.up_proj", "mlp_down_proj": "layers.48.mlp.down_proj", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_q_proj": "layers.49.self_attn.q_proj", "attn_k_proj": "layers.49.self_attn.k_proj", "attn_v_proj": "layers.49.self_attn.v_proj", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate_proj", "mlp_up_proj": "layers.49.mlp.up_proj", "mlp_down_proj": "layers.49.mlp.down_proj", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_q_proj": "layers.50.self_attn.q_proj", "attn_k_proj": "layers.50.self_attn.k_proj", "attn_v_proj": "layers.50.self_attn.v_proj", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate_proj", "mlp_up_proj": "layers.50.mlp.up_proj", "mlp_down_proj": "layers.50.mlp.down_proj", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate_proj", "mlp_up_proj": "layers.51.mlp.up_proj", "mlp_down_proj": "layers.51.mlp.down_proj", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_q_proj": "layers.52.self_attn.q_proj", "attn_k_proj": "layers.52.self_attn.k_proj", "attn_v_proj": "layers.52.self_attn.v_proj", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate_proj", "mlp_up_proj": "layers.52.mlp.up_proj", "mlp_down_proj": "layers.52.mlp.down_proj", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_q_proj": "layers.53.self_attn.q_proj", "attn_k_proj": "layers.53.self_attn.k_proj", "attn_v_proj": "layers.53.self_attn.v_proj", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate_proj", "mlp_up_proj": "layers.53.mlp.up_proj", "mlp_down_proj": "layers.53.mlp.down_proj", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_q_proj": "layers.54.self_attn.q_proj", "attn_k_proj": "layers.54.self_attn.k_proj", "attn_v_proj": "layers.54.self_attn.v_proj", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate_proj", "mlp_up_proj": "layers.54.mlp.up_proj", "mlp_down_proj": "layers.54.mlp.down_proj", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate_proj", "mlp_up_proj": "layers.55.mlp.up_proj", "mlp_down_proj": "layers.55.mlp.down_proj", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_q_proj": "layers.56.self_attn.q_proj", "attn_k_proj": "layers.56.self_attn.k_proj", "attn_v_proj": "layers.56.self_attn.v_proj", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate_proj", "mlp_up_proj": "layers.56.mlp.up_proj", "mlp_down_proj": "layers.56.mlp.down_proj", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_q_proj": "layers.57.self_attn.q_proj", "attn_k_proj": "layers.57.self_attn.k_proj", "attn_v_proj": "layers.57.self_attn.v_proj", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate_proj", "mlp_up_proj": "layers.57.mlp.up_proj", "mlp_down_proj": "layers.57.mlp.down_proj", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_q_proj": "layers.58.self_attn.q_proj", "attn_k_proj": "layers.58.self_attn.k_proj", "attn_v_proj": "layers.58.self_attn.v_proj", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate_proj", "mlp_up_proj": "layers.58.mlp.up_proj", "mlp_down_proj": "layers.58.mlp.down_proj", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate_proj", "mlp_up_proj": "layers.59.mlp.up_proj", "mlp_down_proj": "layers.59.mlp.down_proj", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_q_proj": "layers.60.self_attn.q_proj", "attn_k_proj": "layers.60.self_attn.k_proj", "attn_v_proj": "layers.60.self_attn.v_proj", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate_proj", "mlp_up_proj": "layers.60.mlp.up_proj", "mlp_down_proj": "layers.60.mlp.down_proj", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}, {"index": 61, "intervention_points": {"attn_input": "layers.61.self_attn.", "attn_q_proj": "layers.61.self_attn.q_proj", "attn_k_proj": "layers.61.self_attn.k_proj", "attn_v_proj": "layers.61.self_attn.v_proj", "attn_output_proj": "layers.61.self_attn.o_proj", "mlp_input": "layers.61.mlp.", "mlp_gate_proj": "layers.61.mlp.gate_proj", "mlp_up_proj": "layers.61.mlp.up_proj", "mlp_down_proj": "layers.61.mlp.down_proj", "attn_pre_norm": "layers.61.input_layernorm", "attn_post_norm": "layers.61.post_attention_layernorm"}}, {"index": 62, "intervention_points": {"attn_input": "layers.62.self_attn.", "attn_q_proj": "layers.62.self_attn.q_proj", "attn_k_proj": "layers.62.self_attn.k_proj", "attn_v_proj": "layers.62.self_attn.v_proj", "attn_output_proj": "layers.62.self_attn.o_proj", "mlp_input": "layers.62.mlp.", "mlp_gate_proj": "layers.62.mlp.gate_proj", "mlp_up_proj": "layers.62.mlp.up_proj", "mlp_down_proj": "layers.62.mlp.down_proj", "attn_pre_norm": "layers.62.input_layernorm", "attn_post_norm": "layers.62.post_attention_layernorm"}}, {"index": 63, "intervention_points": {"attn_input": "layers.63.self_attn.", "attn_q_proj": "layers.63.self_attn.q_proj", "attn_k_proj": "layers.63.self_attn.k_proj", "attn_v_proj": "layers.63.self_attn.v_proj", "attn_output_proj": "layers.63.self_attn.o_proj", "mlp_input": "layers.63.mlp.", "mlp_gate_proj": "layers.63.mlp.gate_proj", "mlp_up_proj": "layers.63.mlp.up_proj", "mlp_down_proj": "layers.63.mlp.down_proj", "attn_pre_norm": "layers.63.input_layernorm", "attn_post_norm": "layers.63.post_attention_layernorm"}}, {"index": 64, "intervention_points": {"attn_input": "layers.64.self_attn.", "attn_q_proj": "layers.64.self_attn.q_proj", "attn_k_proj": "layers.64.self_attn.k_proj", "attn_v_proj": "layers.64.self_attn.v_proj", "attn_output_proj": "layers.64.self_attn.o_proj", "mlp_input": "layers.64.mlp.", "mlp_gate_proj": "layers.64.mlp.gate_proj", "mlp_up_proj": "layers.64.mlp.up_proj", "mlp_down_proj": "layers.64.mlp.down_proj", "attn_pre_norm": "layers.64.input_layernorm", "attn_post_norm": "layers.64.post_attention_layernorm"}}, {"index": 65, "intervention_points": {"attn_input": "layers.65.self_attn.", "attn_q_proj": "layers.65.self_attn.q_proj", "attn_k_proj": "layers.65.self_attn.k_proj", "attn_v_proj": "layers.65.self_attn.v_proj", "attn_output_proj": "layers.65.self_attn.o_proj", "mlp_input": "layers.65.mlp.", "mlp_gate_proj": "layers.65.mlp.gate_proj", "mlp_up_proj": "layers.65.mlp.up_proj", "mlp_down_proj": "layers.65.mlp.down_proj", "attn_pre_norm": "layers.65.input_layernorm", "attn_post_norm": "layers.65.post_attention_layernorm"}}, {"index": 66, "intervention_points": {"attn_input": "layers.66.self_attn.", "attn_q_proj": "layers.66.self_attn.q_proj", "attn_k_proj": "layers.66.self_attn.k_proj", "attn_v_proj": "layers.66.self_attn.v_proj", "attn_output_proj": "layers.66.self_attn.o_proj", "mlp_input": "layers.66.mlp.", "mlp_gate_proj": "layers.66.mlp.gate_proj", "mlp_up_proj": "layers.66.mlp.up_proj", "mlp_down_proj": "layers.66.mlp.down_proj", "attn_pre_norm": "layers.66.input_layernorm", "attn_post_norm": "layers.66.post_attention_layernorm"}}, {"index": 67, "intervention_points": {"attn_input": "layers.67.self_attn.", "attn_q_proj": "layers.67.self_attn.q_proj", "attn_k_proj": "layers.67.self_attn.k_proj", "attn_v_proj": "layers.67.self_attn.v_proj", "attn_output_proj": "layers.67.self_attn.o_proj", "mlp_input": "layers.67.mlp.", "mlp_gate_proj": "layers.67.mlp.gate_proj", "mlp_up_proj": "layers.67.mlp.up_proj", "mlp_down_proj": "layers.67.mlp.down_proj", "attn_pre_norm": "layers.67.input_layernorm", "attn_post_norm": "layers.67.post_attention_layernorm"}}, {"index": 68, "intervention_points": {"attn_input": "layers.68.self_attn.", "attn_q_proj": "layers.68.self_attn.q_proj", "attn_k_proj": "layers.68.self_attn.k_proj", "attn_v_proj": "layers.68.self_attn.v_proj", "attn_output_proj": "layers.68.self_attn.o_proj", "mlp_input": "layers.68.mlp.", "mlp_gate_proj": "layers.68.mlp.gate_proj", "mlp_up_proj": "layers.68.mlp.up_proj", "mlp_down_proj": "layers.68.mlp.down_proj", "attn_pre_norm": "layers.68.input_layernorm", "attn_post_norm": "layers.68.post_attention_layernorm"}}, {"index": 69, "intervention_points": {"attn_input": "layers.69.self_attn.", "attn_q_proj": "layers.69.self_attn.q_proj", "attn_k_proj": "layers.69.self_attn.k_proj", "attn_v_proj": "layers.69.self_attn.v_proj", "attn_output_proj": "layers.69.self_attn.o_proj", "mlp_input": "layers.69.mlp.", "mlp_gate_proj": "layers.69.mlp.gate_proj", "mlp_up_proj": "layers.69.mlp.up_proj", "mlp_down_proj": "layers.69.mlp.down_proj", "attn_pre_norm": "layers.69.input_layernorm", "attn_post_norm": "layers.69.post_attention_layernorm"}}, {"index": 70, "intervention_points": {"attn_input": "layers.70.self_attn.", "attn_q_proj": "layers.70.self_attn.q_proj", "attn_k_proj": "layers.70.self_attn.k_proj", "attn_v_proj": "layers.70.self_attn.v_proj", "attn_output_proj": "layers.70.self_attn.o_proj", "mlp_input": "layers.70.mlp.", "mlp_gate_proj": "layers.70.mlp.gate_proj", "mlp_up_proj": "layers.70.mlp.up_proj", "mlp_down_proj": "layers.70.mlp.down_proj", "attn_pre_norm": "layers.70.input_layernorm", "attn_post_norm": "layers.70.post_attention_layernorm"}}, {"index": 71, "intervention_points": {"attn_input": "layers.71.self_attn.", "attn_q_proj": "layers.71.self_attn.q_proj", "attn_k_proj": "layers.71.self_attn.k_proj", "attn_v_proj": "layers.71.self_attn.v_proj", "attn_output_proj": "layers.71.self_attn.o_proj", "mlp_input": "layers.71.mlp.", "mlp_gate_proj": "layers.71.mlp.gate_proj", "mlp_up_proj": "layers.71.mlp.up_proj", "mlp_down_proj": "layers.71.mlp.down_proj", "attn_pre_norm": "layers.71.input_layernorm", "attn_post_norm": "layers.71.post_attention_layernorm"}}, {"index": 72, "intervention_points": {"attn_input": "layers.72.self_attn.", "attn_q_proj": "layers.72.self_attn.q_proj", "attn_k_proj": "layers.72.self_attn.k_proj", "attn_v_proj": "layers.72.self_attn.v_proj", "attn_output_proj": "layers.72.self_attn.o_proj", "mlp_input": "layers.72.mlp.", "mlp_gate_proj": "layers.72.mlp.gate_proj", "mlp_up_proj": "layers.72.mlp.up_proj", "mlp_down_proj": "layers.72.mlp.down_proj", "attn_pre_norm": "layers.72.input_layernorm", "attn_post_norm": "layers.72.post_attention_layernorm"}}, {"index": 73, "intervention_points": {"attn_input": "layers.73.self_attn.", "attn_q_proj": "layers.73.self_attn.q_proj", "attn_k_proj": "layers.73.self_attn.k_proj", "attn_v_proj": "layers.73.self_attn.v_proj", "attn_output_proj": "layers.73.self_attn.o_proj", "mlp_input": "layers.73.mlp.", "mlp_gate_proj": "layers.73.mlp.gate_proj", "mlp_up_proj": "layers.73.mlp.up_proj", "mlp_down_proj": "layers.73.mlp.down_proj", "attn_pre_norm": "layers.73.input_layernorm", "attn_post_norm": "layers.73.post_attention_layernorm"}}, {"index": 74, "intervention_points": {"attn_input": "layers.74.self_attn.", "attn_q_proj": "layers.74.self_attn.q_proj", "attn_k_proj": "layers.74.self_attn.k_proj", "attn_v_proj": "layers.74.self_attn.v_proj", "attn_output_proj": "layers.74.self_attn.o_proj", "mlp_input": "layers.74.mlp.", "mlp_gate_proj": "layers.74.mlp.gate_proj", "mlp_up_proj": "layers.74.mlp.up_proj", "mlp_down_proj": "layers.74.mlp.down_proj", "attn_pre_norm": "layers.74.input_layernorm", "attn_post_norm": "layers.74.post_attention_layernorm"}}, {"index": 75, "intervention_points": {"attn_input": "layers.75.self_attn.", "attn_q_proj": "layers.75.self_attn.q_proj", "attn_k_proj": "layers.75.self_attn.k_proj", "attn_v_proj": "layers.75.self_attn.v_proj", "attn_output_proj": "layers.75.self_attn.o_proj", "mlp_input": "layers.75.mlp.", "mlp_gate_proj": "layers.75.mlp.gate_proj", "mlp_up_proj": "layers.75.mlp.up_proj", "mlp_down_proj": "layers.75.mlp.down_proj", "attn_pre_norm": "layers.75.input_layernorm", "attn_post_norm": "layers.75.post_attention_layernorm"}}, {"index": 76, "intervention_points": {"attn_input": "layers.76.self_attn.", "attn_q_proj": "layers.76.self_attn.q_proj", "attn_k_proj": "layers.76.self_attn.k_proj", "attn_v_proj": "layers.76.self_attn.v_proj", "attn_output_proj": "layers.76.self_attn.o_proj", "mlp_input": "layers.76.mlp.", "mlp_gate_proj": "layers.76.mlp.gate_proj", "mlp_up_proj": "layers.76.mlp.up_proj", "mlp_down_proj": "layers.76.mlp.down_proj", "attn_pre_norm": "layers.76.input_layernorm", "attn_post_norm": "layers.76.post_attention_layernorm"}}, {"index": 77, "intervention_points": {"attn_input": "layers.77.self_attn.", "attn_q_proj": "layers.77.self_attn.q_proj", "attn_k_proj": "layers.77.self_attn.k_proj", "attn_v_proj": "layers.77.self_attn.v_proj", "attn_output_proj": "layers.77.self_attn.o_proj", "mlp_input": "layers.77.mlp.", "mlp_gate_proj": "layers.77.mlp.gate_proj", "mlp_up_proj": "layers.77.mlp.up_proj", "mlp_down_proj": "layers.77.mlp.down_proj", "attn_pre_norm": "layers.77.input_layernorm", "attn_post_norm": "layers.77.post_attention_layernorm"}}, {"index": 78, "intervention_points": {"attn_input": "layers.78.self_attn.", "attn_q_proj": "layers.78.self_attn.q_proj", "attn_k_proj": "layers.78.self_attn.k_proj", "attn_v_proj": "layers.78.self_attn.v_proj", "attn_output_proj": "layers.78.self_attn.o_proj", "mlp_input": "layers.78.mlp.", "mlp_gate_proj": "layers.78.mlp.gate_proj", "mlp_up_proj": "layers.78.mlp.up_proj", "mlp_down_proj": "layers.78.mlp.down_proj", "attn_pre_norm": "layers.78.input_layernorm", "attn_post_norm": "layers.78.post_attention_layernorm"}}, {"index": 79, "intervention_points": {"attn_input": "layers.79.self_attn.", "attn_q_proj": "layers.79.self_attn.q_proj", "attn_k_proj": "layers.79.self_attn.k_proj", "attn_v_proj": "layers.79.self_attn.v_proj", "attn_output_proj": "layers.79.self_attn.o_proj", "mlp_input": "layers.79.mlp.", "mlp_gate_proj": "layers.79.mlp.gate_proj", "mlp_up_proj": "layers.79.mlp.up_proj", "mlp_down_proj": "layers.79.mlp.down_proj", "attn_pre_norm": "layers.79.input_layernorm", "attn_post_norm": "layers.79.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:41:14.509889", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
IQuestLab/IQuest-Coder-V1-40B-Thinking | IQuestLab | iquestcoder | 3 | 2026-04-01T23:41:05.000325+00:00 | {"hidden_size": 5120, "intermediate_size": 27648, "vocab_size": 76800, "max_position_embeddings": 131072, "num_hidden_layers": 80, "rope_theta": 500000.0, "rms_norm_eps": 1e-05, "model_type": "iquestcoder"} | {"rope_type": null, "rope_scaling": {"rope_theta": 500000.0, "rope_type": "default"}, "attention_bias": false, "sliding_window_size": null, "attention_implementation": "eager", "flash_attention_capable": true, "position_embedding_type": null, "max_position_embeddings": 131072, "positional_encoding": "rope", "tie_word_embeddings": false, "norm_epsilon": 1e-05, "activation_function": "silu", "residual_dropout": null, "attention_dropout": 0.0, "actual_norm_class": "IQuestCoderRMSNorm", "linear_bias_present": false, "qkv_projection_style": "separate"} | {"organization": "IQuestLab", "first_publish_date": "2025-12-30", "paper_urls": [], "base_model": null, "license": "other", "architecture_tags": ["text-generation", "conversational"], "model_card_lineage": ["qwen 3", "claude"], "hub_metadata": {"downloads": 330, "likes": 16, "created_at": "2025-12-30T15:32:25.000Z", "last_modified": "2026-03-04T12:45:22.000Z"}} | {"residual_stream_dim": 5120, "layers": [{"index": 0, "intervention_points": {"attn_input": "layers.0.self_attn.", "attn_q_proj": "layers.0.self_attn.q_proj", "attn_k_proj": "layers.0.self_attn.k_proj", "attn_v_proj": "layers.0.self_attn.v_proj", "attn_output_proj": "layers.0.self_attn.o_proj", "mlp_input": "layers.0.mlp.", "mlp_gate_proj": "layers.0.mlp.gate_proj", "mlp_up_proj": "layers.0.mlp.up_proj", "mlp_down_proj": "layers.0.mlp.down_proj", "attn_pre_norm": "layers.0.input_layernorm", "attn_post_norm": "layers.0.post_attention_layernorm"}}, {"index": 1, "intervention_points": {"attn_input": "layers.1.self_attn.", "attn_q_proj": "layers.1.self_attn.q_proj", "attn_k_proj": "layers.1.self_attn.k_proj", "attn_v_proj": "layers.1.self_attn.v_proj", "attn_output_proj": "layers.1.self_attn.o_proj", "mlp_input": "layers.1.mlp.", "mlp_gate_proj": "layers.1.mlp.gate_proj", "mlp_up_proj": "layers.1.mlp.up_proj", "mlp_down_proj": "layers.1.mlp.down_proj", "attn_pre_norm": "layers.1.input_layernorm", "attn_post_norm": "layers.1.post_attention_layernorm"}}, {"index": 2, "intervention_points": {"attn_input": "layers.2.self_attn.", "attn_q_proj": "layers.2.self_attn.q_proj", "attn_k_proj": "layers.2.self_attn.k_proj", "attn_v_proj": "layers.2.self_attn.v_proj", "attn_output_proj": "layers.2.self_attn.o_proj", "mlp_input": "layers.2.mlp.", "mlp_gate_proj": "layers.2.mlp.gate_proj", "mlp_up_proj": "layers.2.mlp.up_proj", "mlp_down_proj": "layers.2.mlp.down_proj", "attn_pre_norm": "layers.2.input_layernorm", "attn_post_norm": "layers.2.post_attention_layernorm"}}, {"index": 3, "intervention_points": {"attn_input": "layers.3.self_attn.", "attn_q_proj": "layers.3.self_attn.q_proj", "attn_k_proj": "layers.3.self_attn.k_proj", "attn_v_proj": "layers.3.self_attn.v_proj", "attn_output_proj": "layers.3.self_attn.o_proj", "mlp_input": "layers.3.mlp.", "mlp_gate_proj": "layers.3.mlp.gate_proj", "mlp_up_proj": "layers.3.mlp.up_proj", "mlp_down_proj": "layers.3.mlp.down_proj", "attn_pre_norm": "layers.3.input_layernorm", "attn_post_norm": "layers.3.post_attention_layernorm"}}, {"index": 4, "intervention_points": {"attn_input": "layers.4.self_attn.", "attn_q_proj": "layers.4.self_attn.q_proj", "attn_k_proj": "layers.4.self_attn.k_proj", "attn_v_proj": "layers.4.self_attn.v_proj", "attn_output_proj": "layers.4.self_attn.o_proj", "mlp_input": "layers.4.mlp.", "mlp_gate_proj": "layers.4.mlp.gate_proj", "mlp_up_proj": "layers.4.mlp.up_proj", "mlp_down_proj": "layers.4.mlp.down_proj", "attn_pre_norm": "layers.4.input_layernorm", "attn_post_norm": "layers.4.post_attention_layernorm"}}, {"index": 5, "intervention_points": {"attn_input": "layers.5.self_attn.", "attn_q_proj": "layers.5.self_attn.q_proj", "attn_k_proj": "layers.5.self_attn.k_proj", "attn_v_proj": "layers.5.self_attn.v_proj", "attn_output_proj": "layers.5.self_attn.o_proj", "mlp_input": "layers.5.mlp.", "mlp_gate_proj": "layers.5.mlp.gate_proj", "mlp_up_proj": "layers.5.mlp.up_proj", "mlp_down_proj": "layers.5.mlp.down_proj", "attn_pre_norm": "layers.5.input_layernorm", "attn_post_norm": "layers.5.post_attention_layernorm"}}, {"index": 6, "intervention_points": {"attn_input": "layers.6.self_attn.", "attn_q_proj": "layers.6.self_attn.q_proj", "attn_k_proj": "layers.6.self_attn.k_proj", "attn_v_proj": "layers.6.self_attn.v_proj", "attn_output_proj": "layers.6.self_attn.o_proj", "mlp_input": "layers.6.mlp.", "mlp_gate_proj": "layers.6.mlp.gate_proj", "mlp_up_proj": "layers.6.mlp.up_proj", "mlp_down_proj": "layers.6.mlp.down_proj", "attn_pre_norm": "layers.6.input_layernorm", "attn_post_norm": "layers.6.post_attention_layernorm"}}, {"index": 7, "intervention_points": {"attn_input": "layers.7.self_attn.", "attn_q_proj": "layers.7.self_attn.q_proj", "attn_k_proj": "layers.7.self_attn.k_proj", "attn_v_proj": "layers.7.self_attn.v_proj", "attn_output_proj": "layers.7.self_attn.o_proj", "mlp_input": "layers.7.mlp.", "mlp_gate_proj": "layers.7.mlp.gate_proj", "mlp_up_proj": "layers.7.mlp.up_proj", "mlp_down_proj": "layers.7.mlp.down_proj", "attn_pre_norm": "layers.7.input_layernorm", "attn_post_norm": "layers.7.post_attention_layernorm"}}, {"index": 8, "intervention_points": {"attn_input": "layers.8.self_attn.", "attn_q_proj": "layers.8.self_attn.q_proj", "attn_k_proj": "layers.8.self_attn.k_proj", "attn_v_proj": "layers.8.self_attn.v_proj", "attn_output_proj": "layers.8.self_attn.o_proj", "mlp_input": "layers.8.mlp.", "mlp_gate_proj": "layers.8.mlp.gate_proj", "mlp_up_proj": "layers.8.mlp.up_proj", "mlp_down_proj": "layers.8.mlp.down_proj", "attn_pre_norm": "layers.8.input_layernorm", "attn_post_norm": "layers.8.post_attention_layernorm"}}, {"index": 9, "intervention_points": {"attn_input": "layers.9.self_attn.", "attn_q_proj": "layers.9.self_attn.q_proj", "attn_k_proj": "layers.9.self_attn.k_proj", "attn_v_proj": "layers.9.self_attn.v_proj", "attn_output_proj": "layers.9.self_attn.o_proj", "mlp_input": "layers.9.mlp.", "mlp_gate_proj": "layers.9.mlp.gate_proj", "mlp_up_proj": "layers.9.mlp.up_proj", "mlp_down_proj": "layers.9.mlp.down_proj", "attn_pre_norm": "layers.9.input_layernorm", "attn_post_norm": "layers.9.post_attention_layernorm"}}, {"index": 10, "intervention_points": {"attn_input": "layers.10.self_attn.", "attn_q_proj": "layers.10.self_attn.q_proj", "attn_k_proj": "layers.10.self_attn.k_proj", "attn_v_proj": "layers.10.self_attn.v_proj", "attn_output_proj": "layers.10.self_attn.o_proj", "mlp_input": "layers.10.mlp.", "mlp_gate_proj": "layers.10.mlp.gate_proj", "mlp_up_proj": "layers.10.mlp.up_proj", "mlp_down_proj": "layers.10.mlp.down_proj", "attn_pre_norm": "layers.10.input_layernorm", "attn_post_norm": "layers.10.post_attention_layernorm"}}, {"index": 11, "intervention_points": {"attn_input": "layers.11.self_attn.", "attn_q_proj": "layers.11.self_attn.q_proj", "attn_k_proj": "layers.11.self_attn.k_proj", "attn_v_proj": "layers.11.self_attn.v_proj", "attn_output_proj": "layers.11.self_attn.o_proj", "mlp_input": "layers.11.mlp.", "mlp_gate_proj": "layers.11.mlp.gate_proj", "mlp_up_proj": "layers.11.mlp.up_proj", "mlp_down_proj": "layers.11.mlp.down_proj", "attn_pre_norm": "layers.11.input_layernorm", "attn_post_norm": "layers.11.post_attention_layernorm"}}, {"index": 12, "intervention_points": {"attn_input": "layers.12.self_attn.", "attn_q_proj": "layers.12.self_attn.q_proj", "attn_k_proj": "layers.12.self_attn.k_proj", "attn_v_proj": "layers.12.self_attn.v_proj", "attn_output_proj": "layers.12.self_attn.o_proj", "mlp_input": "layers.12.mlp.", "mlp_gate_proj": "layers.12.mlp.gate_proj", "mlp_up_proj": "layers.12.mlp.up_proj", "mlp_down_proj": "layers.12.mlp.down_proj", "attn_pre_norm": "layers.12.input_layernorm", "attn_post_norm": "layers.12.post_attention_layernorm"}}, {"index": 13, "intervention_points": {"attn_input": "layers.13.self_attn.", "attn_q_proj": "layers.13.self_attn.q_proj", "attn_k_proj": "layers.13.self_attn.k_proj", "attn_v_proj": "layers.13.self_attn.v_proj", "attn_output_proj": "layers.13.self_attn.o_proj", "mlp_input": "layers.13.mlp.", "mlp_gate_proj": "layers.13.mlp.gate_proj", "mlp_up_proj": "layers.13.mlp.up_proj", "mlp_down_proj": "layers.13.mlp.down_proj", "attn_pre_norm": "layers.13.input_layernorm", "attn_post_norm": "layers.13.post_attention_layernorm"}}, {"index": 14, "intervention_points": {"attn_input": "layers.14.self_attn.", "attn_q_proj": "layers.14.self_attn.q_proj", "attn_k_proj": "layers.14.self_attn.k_proj", "attn_v_proj": "layers.14.self_attn.v_proj", "attn_output_proj": "layers.14.self_attn.o_proj", "mlp_input": "layers.14.mlp.", "mlp_gate_proj": "layers.14.mlp.gate_proj", "mlp_up_proj": "layers.14.mlp.up_proj", "mlp_down_proj": "layers.14.mlp.down_proj", "attn_pre_norm": "layers.14.input_layernorm", "attn_post_norm": "layers.14.post_attention_layernorm"}}, {"index": 15, "intervention_points": {"attn_input": "layers.15.self_attn.", "attn_q_proj": "layers.15.self_attn.q_proj", "attn_k_proj": "layers.15.self_attn.k_proj", "attn_v_proj": "layers.15.self_attn.v_proj", "attn_output_proj": "layers.15.self_attn.o_proj", "mlp_input": "layers.15.mlp.", "mlp_gate_proj": "layers.15.mlp.gate_proj", "mlp_up_proj": "layers.15.mlp.up_proj", "mlp_down_proj": "layers.15.mlp.down_proj", "attn_pre_norm": "layers.15.input_layernorm", "attn_post_norm": "layers.15.post_attention_layernorm"}}, {"index": 16, "intervention_points": {"attn_input": "layers.16.self_attn.", "attn_q_proj": "layers.16.self_attn.q_proj", "attn_k_proj": "layers.16.self_attn.k_proj", "attn_v_proj": "layers.16.self_attn.v_proj", "attn_output_proj": "layers.16.self_attn.o_proj", "mlp_input": "layers.16.mlp.", "mlp_gate_proj": "layers.16.mlp.gate_proj", "mlp_up_proj": "layers.16.mlp.up_proj", "mlp_down_proj": "layers.16.mlp.down_proj", "attn_pre_norm": "layers.16.input_layernorm", "attn_post_norm": "layers.16.post_attention_layernorm"}}, {"index": 17, "intervention_points": {"attn_input": "layers.17.self_attn.", "attn_q_proj": "layers.17.self_attn.q_proj", "attn_k_proj": "layers.17.self_attn.k_proj", "attn_v_proj": "layers.17.self_attn.v_proj", "attn_output_proj": "layers.17.self_attn.o_proj", "mlp_input": "layers.17.mlp.", "mlp_gate_proj": "layers.17.mlp.gate_proj", "mlp_up_proj": "layers.17.mlp.up_proj", "mlp_down_proj": "layers.17.mlp.down_proj", "attn_pre_norm": "layers.17.input_layernorm", "attn_post_norm": "layers.17.post_attention_layernorm"}}, {"index": 18, "intervention_points": {"attn_input": "layers.18.self_attn.", "attn_q_proj": "layers.18.self_attn.q_proj", "attn_k_proj": "layers.18.self_attn.k_proj", "attn_v_proj": "layers.18.self_attn.v_proj", "attn_output_proj": "layers.18.self_attn.o_proj", "mlp_input": "layers.18.mlp.", "mlp_gate_proj": "layers.18.mlp.gate_proj", "mlp_up_proj": "layers.18.mlp.up_proj", "mlp_down_proj": "layers.18.mlp.down_proj", "attn_pre_norm": "layers.18.input_layernorm", "attn_post_norm": "layers.18.post_attention_layernorm"}}, {"index": 19, "intervention_points": {"attn_input": "layers.19.self_attn.", "attn_q_proj": "layers.19.self_attn.q_proj", "attn_k_proj": "layers.19.self_attn.k_proj", "attn_v_proj": "layers.19.self_attn.v_proj", "attn_output_proj": "layers.19.self_attn.o_proj", "mlp_input": "layers.19.mlp.", "mlp_gate_proj": "layers.19.mlp.gate_proj", "mlp_up_proj": "layers.19.mlp.up_proj", "mlp_down_proj": "layers.19.mlp.down_proj", "attn_pre_norm": "layers.19.input_layernorm", "attn_post_norm": "layers.19.post_attention_layernorm"}}, {"index": 20, "intervention_points": {"attn_input": "layers.20.self_attn.", "attn_q_proj": "layers.20.self_attn.q_proj", "attn_k_proj": "layers.20.self_attn.k_proj", "attn_v_proj": "layers.20.self_attn.v_proj", "attn_output_proj": "layers.20.self_attn.o_proj", "mlp_input": "layers.20.mlp.", "mlp_gate_proj": "layers.20.mlp.gate_proj", "mlp_up_proj": "layers.20.mlp.up_proj", "mlp_down_proj": "layers.20.mlp.down_proj", "attn_pre_norm": "layers.20.input_layernorm", "attn_post_norm": "layers.20.post_attention_layernorm"}}, {"index": 21, "intervention_points": {"attn_input": "layers.21.self_attn.", "attn_q_proj": "layers.21.self_attn.q_proj", "attn_k_proj": "layers.21.self_attn.k_proj", "attn_v_proj": "layers.21.self_attn.v_proj", "attn_output_proj": "layers.21.self_attn.o_proj", "mlp_input": "layers.21.mlp.", "mlp_gate_proj": "layers.21.mlp.gate_proj", "mlp_up_proj": "layers.21.mlp.up_proj", "mlp_down_proj": "layers.21.mlp.down_proj", "attn_pre_norm": "layers.21.input_layernorm", "attn_post_norm": "layers.21.post_attention_layernorm"}}, {"index": 22, "intervention_points": {"attn_input": "layers.22.self_attn.", "attn_q_proj": "layers.22.self_attn.q_proj", "attn_k_proj": "layers.22.self_attn.k_proj", "attn_v_proj": "layers.22.self_attn.v_proj", "attn_output_proj": "layers.22.self_attn.o_proj", "mlp_input": "layers.22.mlp.", "mlp_gate_proj": "layers.22.mlp.gate_proj", "mlp_up_proj": "layers.22.mlp.up_proj", "mlp_down_proj": "layers.22.mlp.down_proj", "attn_pre_norm": "layers.22.input_layernorm", "attn_post_norm": "layers.22.post_attention_layernorm"}}, {"index": 23, "intervention_points": {"attn_input": "layers.23.self_attn.", "attn_q_proj": "layers.23.self_attn.q_proj", "attn_k_proj": "layers.23.self_attn.k_proj", "attn_v_proj": "layers.23.self_attn.v_proj", "attn_output_proj": "layers.23.self_attn.o_proj", "mlp_input": "layers.23.mlp.", "mlp_gate_proj": "layers.23.mlp.gate_proj", "mlp_up_proj": "layers.23.mlp.up_proj", "mlp_down_proj": "layers.23.mlp.down_proj", "attn_pre_norm": "layers.23.input_layernorm", "attn_post_norm": "layers.23.post_attention_layernorm"}}, {"index": 24, "intervention_points": {"attn_input": "layers.24.self_attn.", "attn_q_proj": "layers.24.self_attn.q_proj", "attn_k_proj": "layers.24.self_attn.k_proj", "attn_v_proj": "layers.24.self_attn.v_proj", "attn_output_proj": "layers.24.self_attn.o_proj", "mlp_input": "layers.24.mlp.", "mlp_gate_proj": "layers.24.mlp.gate_proj", "mlp_up_proj": "layers.24.mlp.up_proj", "mlp_down_proj": "layers.24.mlp.down_proj", "attn_pre_norm": "layers.24.input_layernorm", "attn_post_norm": "layers.24.post_attention_layernorm"}}, {"index": 25, "intervention_points": {"attn_input": "layers.25.self_attn.", "attn_q_proj": "layers.25.self_attn.q_proj", "attn_k_proj": "layers.25.self_attn.k_proj", "attn_v_proj": "layers.25.self_attn.v_proj", "attn_output_proj": "layers.25.self_attn.o_proj", "mlp_input": "layers.25.mlp.", "mlp_gate_proj": "layers.25.mlp.gate_proj", "mlp_up_proj": "layers.25.mlp.up_proj", "mlp_down_proj": "layers.25.mlp.down_proj", "attn_pre_norm": "layers.25.input_layernorm", "attn_post_norm": "layers.25.post_attention_layernorm"}}, {"index": 26, "intervention_points": {"attn_input": "layers.26.self_attn.", "attn_q_proj": "layers.26.self_attn.q_proj", "attn_k_proj": "layers.26.self_attn.k_proj", "attn_v_proj": "layers.26.self_attn.v_proj", "attn_output_proj": "layers.26.self_attn.o_proj", "mlp_input": "layers.26.mlp.", "mlp_gate_proj": "layers.26.mlp.gate_proj", "mlp_up_proj": "layers.26.mlp.up_proj", "mlp_down_proj": "layers.26.mlp.down_proj", "attn_pre_norm": "layers.26.input_layernorm", "attn_post_norm": "layers.26.post_attention_layernorm"}}, {"index": 27, "intervention_points": {"attn_input": "layers.27.self_attn.", "attn_q_proj": "layers.27.self_attn.q_proj", "attn_k_proj": "layers.27.self_attn.k_proj", "attn_v_proj": "layers.27.self_attn.v_proj", "attn_output_proj": "layers.27.self_attn.o_proj", "mlp_input": "layers.27.mlp.", "mlp_gate_proj": "layers.27.mlp.gate_proj", "mlp_up_proj": "layers.27.mlp.up_proj", "mlp_down_proj": "layers.27.mlp.down_proj", "attn_pre_norm": "layers.27.input_layernorm", "attn_post_norm": "layers.27.post_attention_layernorm"}}, {"index": 28, "intervention_points": {"attn_input": "layers.28.self_attn.", "attn_q_proj": "layers.28.self_attn.q_proj", "attn_k_proj": "layers.28.self_attn.k_proj", "attn_v_proj": "layers.28.self_attn.v_proj", "attn_output_proj": "layers.28.self_attn.o_proj", "mlp_input": "layers.28.mlp.", "mlp_gate_proj": "layers.28.mlp.gate_proj", "mlp_up_proj": "layers.28.mlp.up_proj", "mlp_down_proj": "layers.28.mlp.down_proj", "attn_pre_norm": "layers.28.input_layernorm", "attn_post_norm": "layers.28.post_attention_layernorm"}}, {"index": 29, "intervention_points": {"attn_input": "layers.29.self_attn.", "attn_q_proj": "layers.29.self_attn.q_proj", "attn_k_proj": "layers.29.self_attn.k_proj", "attn_v_proj": "layers.29.self_attn.v_proj", "attn_output_proj": "layers.29.self_attn.o_proj", "mlp_input": "layers.29.mlp.", "mlp_gate_proj": "layers.29.mlp.gate_proj", "mlp_up_proj": "layers.29.mlp.up_proj", "mlp_down_proj": "layers.29.mlp.down_proj", "attn_pre_norm": "layers.29.input_layernorm", "attn_post_norm": "layers.29.post_attention_layernorm"}}, {"index": 30, "intervention_points": {"attn_input": "layers.30.self_attn.", "attn_q_proj": "layers.30.self_attn.q_proj", "attn_k_proj": "layers.30.self_attn.k_proj", "attn_v_proj": "layers.30.self_attn.v_proj", "attn_output_proj": "layers.30.self_attn.o_proj", "mlp_input": "layers.30.mlp.", "mlp_gate_proj": "layers.30.mlp.gate_proj", "mlp_up_proj": "layers.30.mlp.up_proj", "mlp_down_proj": "layers.30.mlp.down_proj", "attn_pre_norm": "layers.30.input_layernorm", "attn_post_norm": "layers.30.post_attention_layernorm"}}, {"index": 31, "intervention_points": {"attn_input": "layers.31.self_attn.", "attn_q_proj": "layers.31.self_attn.q_proj", "attn_k_proj": "layers.31.self_attn.k_proj", "attn_v_proj": "layers.31.self_attn.v_proj", "attn_output_proj": "layers.31.self_attn.o_proj", "mlp_input": "layers.31.mlp.", "mlp_gate_proj": "layers.31.mlp.gate_proj", "mlp_up_proj": "layers.31.mlp.up_proj", "mlp_down_proj": "layers.31.mlp.down_proj", "attn_pre_norm": "layers.31.input_layernorm", "attn_post_norm": "layers.31.post_attention_layernorm"}}, {"index": 32, "intervention_points": {"attn_input": "layers.32.self_attn.", "attn_q_proj": "layers.32.self_attn.q_proj", "attn_k_proj": "layers.32.self_attn.k_proj", "attn_v_proj": "layers.32.self_attn.v_proj", "attn_output_proj": "layers.32.self_attn.o_proj", "mlp_input": "layers.32.mlp.", "mlp_gate_proj": "layers.32.mlp.gate_proj", "mlp_up_proj": "layers.32.mlp.up_proj", "mlp_down_proj": "layers.32.mlp.down_proj", "attn_pre_norm": "layers.32.input_layernorm", "attn_post_norm": "layers.32.post_attention_layernorm"}}, {"index": 33, "intervention_points": {"attn_input": "layers.33.self_attn.", "attn_q_proj": "layers.33.self_attn.q_proj", "attn_k_proj": "layers.33.self_attn.k_proj", "attn_v_proj": "layers.33.self_attn.v_proj", "attn_output_proj": "layers.33.self_attn.o_proj", "mlp_input": "layers.33.mlp.", "mlp_gate_proj": "layers.33.mlp.gate_proj", "mlp_up_proj": "layers.33.mlp.up_proj", "mlp_down_proj": "layers.33.mlp.down_proj", "attn_pre_norm": "layers.33.input_layernorm", "attn_post_norm": "layers.33.post_attention_layernorm"}}, {"index": 34, "intervention_points": {"attn_input": "layers.34.self_attn.", "attn_q_proj": "layers.34.self_attn.q_proj", "attn_k_proj": "layers.34.self_attn.k_proj", "attn_v_proj": "layers.34.self_attn.v_proj", "attn_output_proj": "layers.34.self_attn.o_proj", "mlp_input": "layers.34.mlp.", "mlp_gate_proj": "layers.34.mlp.gate_proj", "mlp_up_proj": "layers.34.mlp.up_proj", "mlp_down_proj": "layers.34.mlp.down_proj", "attn_pre_norm": "layers.34.input_layernorm", "attn_post_norm": "layers.34.post_attention_layernorm"}}, {"index": 35, "intervention_points": {"attn_input": "layers.35.self_attn.", "attn_q_proj": "layers.35.self_attn.q_proj", "attn_k_proj": "layers.35.self_attn.k_proj", "attn_v_proj": "layers.35.self_attn.v_proj", "attn_output_proj": "layers.35.self_attn.o_proj", "mlp_input": "layers.35.mlp.", "mlp_gate_proj": "layers.35.mlp.gate_proj", "mlp_up_proj": "layers.35.mlp.up_proj", "mlp_down_proj": "layers.35.mlp.down_proj", "attn_pre_norm": "layers.35.input_layernorm", "attn_post_norm": "layers.35.post_attention_layernorm"}}, {"index": 36, "intervention_points": {"attn_input": "layers.36.self_attn.", "attn_q_proj": "layers.36.self_attn.q_proj", "attn_k_proj": "layers.36.self_attn.k_proj", "attn_v_proj": "layers.36.self_attn.v_proj", "attn_output_proj": "layers.36.self_attn.o_proj", "mlp_input": "layers.36.mlp.", "mlp_gate_proj": "layers.36.mlp.gate_proj", "mlp_up_proj": "layers.36.mlp.up_proj", "mlp_down_proj": "layers.36.mlp.down_proj", "attn_pre_norm": "layers.36.input_layernorm", "attn_post_norm": "layers.36.post_attention_layernorm"}}, {"index": 37, "intervention_points": {"attn_input": "layers.37.self_attn.", "attn_q_proj": "layers.37.self_attn.q_proj", "attn_k_proj": "layers.37.self_attn.k_proj", "attn_v_proj": "layers.37.self_attn.v_proj", "attn_output_proj": "layers.37.self_attn.o_proj", "mlp_input": "layers.37.mlp.", "mlp_gate_proj": "layers.37.mlp.gate_proj", "mlp_up_proj": "layers.37.mlp.up_proj", "mlp_down_proj": "layers.37.mlp.down_proj", "attn_pre_norm": "layers.37.input_layernorm", "attn_post_norm": "layers.37.post_attention_layernorm"}}, {"index": 38, "intervention_points": {"attn_input": "layers.38.self_attn.", "attn_q_proj": "layers.38.self_attn.q_proj", "attn_k_proj": "layers.38.self_attn.k_proj", "attn_v_proj": "layers.38.self_attn.v_proj", "attn_output_proj": "layers.38.self_attn.o_proj", "mlp_input": "layers.38.mlp.", "mlp_gate_proj": "layers.38.mlp.gate_proj", "mlp_up_proj": "layers.38.mlp.up_proj", "mlp_down_proj": "layers.38.mlp.down_proj", "attn_pre_norm": "layers.38.input_layernorm", "attn_post_norm": "layers.38.post_attention_layernorm"}}, {"index": 39, "intervention_points": {"attn_input": "layers.39.self_attn.", "attn_q_proj": "layers.39.self_attn.q_proj", "attn_k_proj": "layers.39.self_attn.k_proj", "attn_v_proj": "layers.39.self_attn.v_proj", "attn_output_proj": "layers.39.self_attn.o_proj", "mlp_input": "layers.39.mlp.", "mlp_gate_proj": "layers.39.mlp.gate_proj", "mlp_up_proj": "layers.39.mlp.up_proj", "mlp_down_proj": "layers.39.mlp.down_proj", "attn_pre_norm": "layers.39.input_layernorm", "attn_post_norm": "layers.39.post_attention_layernorm"}}, {"index": 40, "intervention_points": {"attn_input": "layers.40.self_attn.", "attn_q_proj": "layers.40.self_attn.q_proj", "attn_k_proj": "layers.40.self_attn.k_proj", "attn_v_proj": "layers.40.self_attn.v_proj", "attn_output_proj": "layers.40.self_attn.o_proj", "mlp_input": "layers.40.mlp.", "mlp_gate_proj": "layers.40.mlp.gate_proj", "mlp_up_proj": "layers.40.mlp.up_proj", "mlp_down_proj": "layers.40.mlp.down_proj", "attn_pre_norm": "layers.40.input_layernorm", "attn_post_norm": "layers.40.post_attention_layernorm"}}, {"index": 41, "intervention_points": {"attn_input": "layers.41.self_attn.", "attn_q_proj": "layers.41.self_attn.q_proj", "attn_k_proj": "layers.41.self_attn.k_proj", "attn_v_proj": "layers.41.self_attn.v_proj", "attn_output_proj": "layers.41.self_attn.o_proj", "mlp_input": "layers.41.mlp.", "mlp_gate_proj": "layers.41.mlp.gate_proj", "mlp_up_proj": "layers.41.mlp.up_proj", "mlp_down_proj": "layers.41.mlp.down_proj", "attn_pre_norm": "layers.41.input_layernorm", "attn_post_norm": "layers.41.post_attention_layernorm"}}, {"index": 42, "intervention_points": {"attn_input": "layers.42.self_attn.", "attn_q_proj": "layers.42.self_attn.q_proj", "attn_k_proj": "layers.42.self_attn.k_proj", "attn_v_proj": "layers.42.self_attn.v_proj", "attn_output_proj": "layers.42.self_attn.o_proj", "mlp_input": "layers.42.mlp.", "mlp_gate_proj": "layers.42.mlp.gate_proj", "mlp_up_proj": "layers.42.mlp.up_proj", "mlp_down_proj": "layers.42.mlp.down_proj", "attn_pre_norm": "layers.42.input_layernorm", "attn_post_norm": "layers.42.post_attention_layernorm"}}, {"index": 43, "intervention_points": {"attn_input": "layers.43.self_attn.", "attn_q_proj": "layers.43.self_attn.q_proj", "attn_k_proj": "layers.43.self_attn.k_proj", "attn_v_proj": "layers.43.self_attn.v_proj", "attn_output_proj": "layers.43.self_attn.o_proj", "mlp_input": "layers.43.mlp.", "mlp_gate_proj": "layers.43.mlp.gate_proj", "mlp_up_proj": "layers.43.mlp.up_proj", "mlp_down_proj": "layers.43.mlp.down_proj", "attn_pre_norm": "layers.43.input_layernorm", "attn_post_norm": "layers.43.post_attention_layernorm"}}, {"index": 44, "intervention_points": {"attn_input": "layers.44.self_attn.", "attn_q_proj": "layers.44.self_attn.q_proj", "attn_k_proj": "layers.44.self_attn.k_proj", "attn_v_proj": "layers.44.self_attn.v_proj", "attn_output_proj": "layers.44.self_attn.o_proj", "mlp_input": "layers.44.mlp.", "mlp_gate_proj": "layers.44.mlp.gate_proj", "mlp_up_proj": "layers.44.mlp.up_proj", "mlp_down_proj": "layers.44.mlp.down_proj", "attn_pre_norm": "layers.44.input_layernorm", "attn_post_norm": "layers.44.post_attention_layernorm"}}, {"index": 45, "intervention_points": {"attn_input": "layers.45.self_attn.", "attn_q_proj": "layers.45.self_attn.q_proj", "attn_k_proj": "layers.45.self_attn.k_proj", "attn_v_proj": "layers.45.self_attn.v_proj", "attn_output_proj": "layers.45.self_attn.o_proj", "mlp_input": "layers.45.mlp.", "mlp_gate_proj": "layers.45.mlp.gate_proj", "mlp_up_proj": "layers.45.mlp.up_proj", "mlp_down_proj": "layers.45.mlp.down_proj", "attn_pre_norm": "layers.45.input_layernorm", "attn_post_norm": "layers.45.post_attention_layernorm"}}, {"index": 46, "intervention_points": {"attn_input": "layers.46.self_attn.", "attn_q_proj": "layers.46.self_attn.q_proj", "attn_k_proj": "layers.46.self_attn.k_proj", "attn_v_proj": "layers.46.self_attn.v_proj", "attn_output_proj": "layers.46.self_attn.o_proj", "mlp_input": "layers.46.mlp.", "mlp_gate_proj": "layers.46.mlp.gate_proj", "mlp_up_proj": "layers.46.mlp.up_proj", "mlp_down_proj": "layers.46.mlp.down_proj", "attn_pre_norm": "layers.46.input_layernorm", "attn_post_norm": "layers.46.post_attention_layernorm"}}, {"index": 47, "intervention_points": {"attn_input": "layers.47.self_attn.", "attn_q_proj": "layers.47.self_attn.q_proj", "attn_k_proj": "layers.47.self_attn.k_proj", "attn_v_proj": "layers.47.self_attn.v_proj", "attn_output_proj": "layers.47.self_attn.o_proj", "mlp_input": "layers.47.mlp.", "mlp_gate_proj": "layers.47.mlp.gate_proj", "mlp_up_proj": "layers.47.mlp.up_proj", "mlp_down_proj": "layers.47.mlp.down_proj", "attn_pre_norm": "layers.47.input_layernorm", "attn_post_norm": "layers.47.post_attention_layernorm"}}, {"index": 48, "intervention_points": {"attn_input": "layers.48.self_attn.", "attn_q_proj": "layers.48.self_attn.q_proj", "attn_k_proj": "layers.48.self_attn.k_proj", "attn_v_proj": "layers.48.self_attn.v_proj", "attn_output_proj": "layers.48.self_attn.o_proj", "mlp_input": "layers.48.mlp.", "mlp_gate_proj": "layers.48.mlp.gate_proj", "mlp_up_proj": "layers.48.mlp.up_proj", "mlp_down_proj": "layers.48.mlp.down_proj", "attn_pre_norm": "layers.48.input_layernorm", "attn_post_norm": "layers.48.post_attention_layernorm"}}, {"index": 49, "intervention_points": {"attn_input": "layers.49.self_attn.", "attn_q_proj": "layers.49.self_attn.q_proj", "attn_k_proj": "layers.49.self_attn.k_proj", "attn_v_proj": "layers.49.self_attn.v_proj", "attn_output_proj": "layers.49.self_attn.o_proj", "mlp_input": "layers.49.mlp.", "mlp_gate_proj": "layers.49.mlp.gate_proj", "mlp_up_proj": "layers.49.mlp.up_proj", "mlp_down_proj": "layers.49.mlp.down_proj", "attn_pre_norm": "layers.49.input_layernorm", "attn_post_norm": "layers.49.post_attention_layernorm"}}, {"index": 50, "intervention_points": {"attn_input": "layers.50.self_attn.", "attn_q_proj": "layers.50.self_attn.q_proj", "attn_k_proj": "layers.50.self_attn.k_proj", "attn_v_proj": "layers.50.self_attn.v_proj", "attn_output_proj": "layers.50.self_attn.o_proj", "mlp_input": "layers.50.mlp.", "mlp_gate_proj": "layers.50.mlp.gate_proj", "mlp_up_proj": "layers.50.mlp.up_proj", "mlp_down_proj": "layers.50.mlp.down_proj", "attn_pre_norm": "layers.50.input_layernorm", "attn_post_norm": "layers.50.post_attention_layernorm"}}, {"index": 51, "intervention_points": {"attn_input": "layers.51.self_attn.", "attn_q_proj": "layers.51.self_attn.q_proj", "attn_k_proj": "layers.51.self_attn.k_proj", "attn_v_proj": "layers.51.self_attn.v_proj", "attn_output_proj": "layers.51.self_attn.o_proj", "mlp_input": "layers.51.mlp.", "mlp_gate_proj": "layers.51.mlp.gate_proj", "mlp_up_proj": "layers.51.mlp.up_proj", "mlp_down_proj": "layers.51.mlp.down_proj", "attn_pre_norm": "layers.51.input_layernorm", "attn_post_norm": "layers.51.post_attention_layernorm"}}, {"index": 52, "intervention_points": {"attn_input": "layers.52.self_attn.", "attn_q_proj": "layers.52.self_attn.q_proj", "attn_k_proj": "layers.52.self_attn.k_proj", "attn_v_proj": "layers.52.self_attn.v_proj", "attn_output_proj": "layers.52.self_attn.o_proj", "mlp_input": "layers.52.mlp.", "mlp_gate_proj": "layers.52.mlp.gate_proj", "mlp_up_proj": "layers.52.mlp.up_proj", "mlp_down_proj": "layers.52.mlp.down_proj", "attn_pre_norm": "layers.52.input_layernorm", "attn_post_norm": "layers.52.post_attention_layernorm"}}, {"index": 53, "intervention_points": {"attn_input": "layers.53.self_attn.", "attn_q_proj": "layers.53.self_attn.q_proj", "attn_k_proj": "layers.53.self_attn.k_proj", "attn_v_proj": "layers.53.self_attn.v_proj", "attn_output_proj": "layers.53.self_attn.o_proj", "mlp_input": "layers.53.mlp.", "mlp_gate_proj": "layers.53.mlp.gate_proj", "mlp_up_proj": "layers.53.mlp.up_proj", "mlp_down_proj": "layers.53.mlp.down_proj", "attn_pre_norm": "layers.53.input_layernorm", "attn_post_norm": "layers.53.post_attention_layernorm"}}, {"index": 54, "intervention_points": {"attn_input": "layers.54.self_attn.", "attn_q_proj": "layers.54.self_attn.q_proj", "attn_k_proj": "layers.54.self_attn.k_proj", "attn_v_proj": "layers.54.self_attn.v_proj", "attn_output_proj": "layers.54.self_attn.o_proj", "mlp_input": "layers.54.mlp.", "mlp_gate_proj": "layers.54.mlp.gate_proj", "mlp_up_proj": "layers.54.mlp.up_proj", "mlp_down_proj": "layers.54.mlp.down_proj", "attn_pre_norm": "layers.54.input_layernorm", "attn_post_norm": "layers.54.post_attention_layernorm"}}, {"index": 55, "intervention_points": {"attn_input": "layers.55.self_attn.", "attn_q_proj": "layers.55.self_attn.q_proj", "attn_k_proj": "layers.55.self_attn.k_proj", "attn_v_proj": "layers.55.self_attn.v_proj", "attn_output_proj": "layers.55.self_attn.o_proj", "mlp_input": "layers.55.mlp.", "mlp_gate_proj": "layers.55.mlp.gate_proj", "mlp_up_proj": "layers.55.mlp.up_proj", "mlp_down_proj": "layers.55.mlp.down_proj", "attn_pre_norm": "layers.55.input_layernorm", "attn_post_norm": "layers.55.post_attention_layernorm"}}, {"index": 56, "intervention_points": {"attn_input": "layers.56.self_attn.", "attn_q_proj": "layers.56.self_attn.q_proj", "attn_k_proj": "layers.56.self_attn.k_proj", "attn_v_proj": "layers.56.self_attn.v_proj", "attn_output_proj": "layers.56.self_attn.o_proj", "mlp_input": "layers.56.mlp.", "mlp_gate_proj": "layers.56.mlp.gate_proj", "mlp_up_proj": "layers.56.mlp.up_proj", "mlp_down_proj": "layers.56.mlp.down_proj", "attn_pre_norm": "layers.56.input_layernorm", "attn_post_norm": "layers.56.post_attention_layernorm"}}, {"index": 57, "intervention_points": {"attn_input": "layers.57.self_attn.", "attn_q_proj": "layers.57.self_attn.q_proj", "attn_k_proj": "layers.57.self_attn.k_proj", "attn_v_proj": "layers.57.self_attn.v_proj", "attn_output_proj": "layers.57.self_attn.o_proj", "mlp_input": "layers.57.mlp.", "mlp_gate_proj": "layers.57.mlp.gate_proj", "mlp_up_proj": "layers.57.mlp.up_proj", "mlp_down_proj": "layers.57.mlp.down_proj", "attn_pre_norm": "layers.57.input_layernorm", "attn_post_norm": "layers.57.post_attention_layernorm"}}, {"index": 58, "intervention_points": {"attn_input": "layers.58.self_attn.", "attn_q_proj": "layers.58.self_attn.q_proj", "attn_k_proj": "layers.58.self_attn.k_proj", "attn_v_proj": "layers.58.self_attn.v_proj", "attn_output_proj": "layers.58.self_attn.o_proj", "mlp_input": "layers.58.mlp.", "mlp_gate_proj": "layers.58.mlp.gate_proj", "mlp_up_proj": "layers.58.mlp.up_proj", "mlp_down_proj": "layers.58.mlp.down_proj", "attn_pre_norm": "layers.58.input_layernorm", "attn_post_norm": "layers.58.post_attention_layernorm"}}, {"index": 59, "intervention_points": {"attn_input": "layers.59.self_attn.", "attn_q_proj": "layers.59.self_attn.q_proj", "attn_k_proj": "layers.59.self_attn.k_proj", "attn_v_proj": "layers.59.self_attn.v_proj", "attn_output_proj": "layers.59.self_attn.o_proj", "mlp_input": "layers.59.mlp.", "mlp_gate_proj": "layers.59.mlp.gate_proj", "mlp_up_proj": "layers.59.mlp.up_proj", "mlp_down_proj": "layers.59.mlp.down_proj", "attn_pre_norm": "layers.59.input_layernorm", "attn_post_norm": "layers.59.post_attention_layernorm"}}, {"index": 60, "intervention_points": {"attn_input": "layers.60.self_attn.", "attn_q_proj": "layers.60.self_attn.q_proj", "attn_k_proj": "layers.60.self_attn.k_proj", "attn_v_proj": "layers.60.self_attn.v_proj", "attn_output_proj": "layers.60.self_attn.o_proj", "mlp_input": "layers.60.mlp.", "mlp_gate_proj": "layers.60.mlp.gate_proj", "mlp_up_proj": "layers.60.mlp.up_proj", "mlp_down_proj": "layers.60.mlp.down_proj", "attn_pre_norm": "layers.60.input_layernorm", "attn_post_norm": "layers.60.post_attention_layernorm"}}, {"index": 61, "intervention_points": {"attn_input": "layers.61.self_attn.", "attn_q_proj": "layers.61.self_attn.q_proj", "attn_k_proj": "layers.61.self_attn.k_proj", "attn_v_proj": "layers.61.self_attn.v_proj", "attn_output_proj": "layers.61.self_attn.o_proj", "mlp_input": "layers.61.mlp.", "mlp_gate_proj": "layers.61.mlp.gate_proj", "mlp_up_proj": "layers.61.mlp.up_proj", "mlp_down_proj": "layers.61.mlp.down_proj", "attn_pre_norm": "layers.61.input_layernorm", "attn_post_norm": "layers.61.post_attention_layernorm"}}, {"index": 62, "intervention_points": {"attn_input": "layers.62.self_attn.", "attn_q_proj": "layers.62.self_attn.q_proj", "attn_k_proj": "layers.62.self_attn.k_proj", "attn_v_proj": "layers.62.self_attn.v_proj", "attn_output_proj": "layers.62.self_attn.o_proj", "mlp_input": "layers.62.mlp.", "mlp_gate_proj": "layers.62.mlp.gate_proj", "mlp_up_proj": "layers.62.mlp.up_proj", "mlp_down_proj": "layers.62.mlp.down_proj", "attn_pre_norm": "layers.62.input_layernorm", "attn_post_norm": "layers.62.post_attention_layernorm"}}, {"index": 63, "intervention_points": {"attn_input": "layers.63.self_attn.", "attn_q_proj": "layers.63.self_attn.q_proj", "attn_k_proj": "layers.63.self_attn.k_proj", "attn_v_proj": "layers.63.self_attn.v_proj", "attn_output_proj": "layers.63.self_attn.o_proj", "mlp_input": "layers.63.mlp.", "mlp_gate_proj": "layers.63.mlp.gate_proj", "mlp_up_proj": "layers.63.mlp.up_proj", "mlp_down_proj": "layers.63.mlp.down_proj", "attn_pre_norm": "layers.63.input_layernorm", "attn_post_norm": "layers.63.post_attention_layernorm"}}, {"index": 64, "intervention_points": {"attn_input": "layers.64.self_attn.", "attn_q_proj": "layers.64.self_attn.q_proj", "attn_k_proj": "layers.64.self_attn.k_proj", "attn_v_proj": "layers.64.self_attn.v_proj", "attn_output_proj": "layers.64.self_attn.o_proj", "mlp_input": "layers.64.mlp.", "mlp_gate_proj": "layers.64.mlp.gate_proj", "mlp_up_proj": "layers.64.mlp.up_proj", "mlp_down_proj": "layers.64.mlp.down_proj", "attn_pre_norm": "layers.64.input_layernorm", "attn_post_norm": "layers.64.post_attention_layernorm"}}, {"index": 65, "intervention_points": {"attn_input": "layers.65.self_attn.", "attn_q_proj": "layers.65.self_attn.q_proj", "attn_k_proj": "layers.65.self_attn.k_proj", "attn_v_proj": "layers.65.self_attn.v_proj", "attn_output_proj": "layers.65.self_attn.o_proj", "mlp_input": "layers.65.mlp.", "mlp_gate_proj": "layers.65.mlp.gate_proj", "mlp_up_proj": "layers.65.mlp.up_proj", "mlp_down_proj": "layers.65.mlp.down_proj", "attn_pre_norm": "layers.65.input_layernorm", "attn_post_norm": "layers.65.post_attention_layernorm"}}, {"index": 66, "intervention_points": {"attn_input": "layers.66.self_attn.", "attn_q_proj": "layers.66.self_attn.q_proj", "attn_k_proj": "layers.66.self_attn.k_proj", "attn_v_proj": "layers.66.self_attn.v_proj", "attn_output_proj": "layers.66.self_attn.o_proj", "mlp_input": "layers.66.mlp.", "mlp_gate_proj": "layers.66.mlp.gate_proj", "mlp_up_proj": "layers.66.mlp.up_proj", "mlp_down_proj": "layers.66.mlp.down_proj", "attn_pre_norm": "layers.66.input_layernorm", "attn_post_norm": "layers.66.post_attention_layernorm"}}, {"index": 67, "intervention_points": {"attn_input": "layers.67.self_attn.", "attn_q_proj": "layers.67.self_attn.q_proj", "attn_k_proj": "layers.67.self_attn.k_proj", "attn_v_proj": "layers.67.self_attn.v_proj", "attn_output_proj": "layers.67.self_attn.o_proj", "mlp_input": "layers.67.mlp.", "mlp_gate_proj": "layers.67.mlp.gate_proj", "mlp_up_proj": "layers.67.mlp.up_proj", "mlp_down_proj": "layers.67.mlp.down_proj", "attn_pre_norm": "layers.67.input_layernorm", "attn_post_norm": "layers.67.post_attention_layernorm"}}, {"index": 68, "intervention_points": {"attn_input": "layers.68.self_attn.", "attn_q_proj": "layers.68.self_attn.q_proj", "attn_k_proj": "layers.68.self_attn.k_proj", "attn_v_proj": "layers.68.self_attn.v_proj", "attn_output_proj": "layers.68.self_attn.o_proj", "mlp_input": "layers.68.mlp.", "mlp_gate_proj": "layers.68.mlp.gate_proj", "mlp_up_proj": "layers.68.mlp.up_proj", "mlp_down_proj": "layers.68.mlp.down_proj", "attn_pre_norm": "layers.68.input_layernorm", "attn_post_norm": "layers.68.post_attention_layernorm"}}, {"index": 69, "intervention_points": {"attn_input": "layers.69.self_attn.", "attn_q_proj": "layers.69.self_attn.q_proj", "attn_k_proj": "layers.69.self_attn.k_proj", "attn_v_proj": "layers.69.self_attn.v_proj", "attn_output_proj": "layers.69.self_attn.o_proj", "mlp_input": "layers.69.mlp.", "mlp_gate_proj": "layers.69.mlp.gate_proj", "mlp_up_proj": "layers.69.mlp.up_proj", "mlp_down_proj": "layers.69.mlp.down_proj", "attn_pre_norm": "layers.69.input_layernorm", "attn_post_norm": "layers.69.post_attention_layernorm"}}, {"index": 70, "intervention_points": {"attn_input": "layers.70.self_attn.", "attn_q_proj": "layers.70.self_attn.q_proj", "attn_k_proj": "layers.70.self_attn.k_proj", "attn_v_proj": "layers.70.self_attn.v_proj", "attn_output_proj": "layers.70.self_attn.o_proj", "mlp_input": "layers.70.mlp.", "mlp_gate_proj": "layers.70.mlp.gate_proj", "mlp_up_proj": "layers.70.mlp.up_proj", "mlp_down_proj": "layers.70.mlp.down_proj", "attn_pre_norm": "layers.70.input_layernorm", "attn_post_norm": "layers.70.post_attention_layernorm"}}, {"index": 71, "intervention_points": {"attn_input": "layers.71.self_attn.", "attn_q_proj": "layers.71.self_attn.q_proj", "attn_k_proj": "layers.71.self_attn.k_proj", "attn_v_proj": "layers.71.self_attn.v_proj", "attn_output_proj": "layers.71.self_attn.o_proj", "mlp_input": "layers.71.mlp.", "mlp_gate_proj": "layers.71.mlp.gate_proj", "mlp_up_proj": "layers.71.mlp.up_proj", "mlp_down_proj": "layers.71.mlp.down_proj", "attn_pre_norm": "layers.71.input_layernorm", "attn_post_norm": "layers.71.post_attention_layernorm"}}, {"index": 72, "intervention_points": {"attn_input": "layers.72.self_attn.", "attn_q_proj": "layers.72.self_attn.q_proj", "attn_k_proj": "layers.72.self_attn.k_proj", "attn_v_proj": "layers.72.self_attn.v_proj", "attn_output_proj": "layers.72.self_attn.o_proj", "mlp_input": "layers.72.mlp.", "mlp_gate_proj": "layers.72.mlp.gate_proj", "mlp_up_proj": "layers.72.mlp.up_proj", "mlp_down_proj": "layers.72.mlp.down_proj", "attn_pre_norm": "layers.72.input_layernorm", "attn_post_norm": "layers.72.post_attention_layernorm"}}, {"index": 73, "intervention_points": {"attn_input": "layers.73.self_attn.", "attn_q_proj": "layers.73.self_attn.q_proj", "attn_k_proj": "layers.73.self_attn.k_proj", "attn_v_proj": "layers.73.self_attn.v_proj", "attn_output_proj": "layers.73.self_attn.o_proj", "mlp_input": "layers.73.mlp.", "mlp_gate_proj": "layers.73.mlp.gate_proj", "mlp_up_proj": "layers.73.mlp.up_proj", "mlp_down_proj": "layers.73.mlp.down_proj", "attn_pre_norm": "layers.73.input_layernorm", "attn_post_norm": "layers.73.post_attention_layernorm"}}, {"index": 74, "intervention_points": {"attn_input": "layers.74.self_attn.", "attn_q_proj": "layers.74.self_attn.q_proj", "attn_k_proj": "layers.74.self_attn.k_proj", "attn_v_proj": "layers.74.self_attn.v_proj", "attn_output_proj": "layers.74.self_attn.o_proj", "mlp_input": "layers.74.mlp.", "mlp_gate_proj": "layers.74.mlp.gate_proj", "mlp_up_proj": "layers.74.mlp.up_proj", "mlp_down_proj": "layers.74.mlp.down_proj", "attn_pre_norm": "layers.74.input_layernorm", "attn_post_norm": "layers.74.post_attention_layernorm"}}, {"index": 75, "intervention_points": {"attn_input": "layers.75.self_attn.", "attn_q_proj": "layers.75.self_attn.q_proj", "attn_k_proj": "layers.75.self_attn.k_proj", "attn_v_proj": "layers.75.self_attn.v_proj", "attn_output_proj": "layers.75.self_attn.o_proj", "mlp_input": "layers.75.mlp.", "mlp_gate_proj": "layers.75.mlp.gate_proj", "mlp_up_proj": "layers.75.mlp.up_proj", "mlp_down_proj": "layers.75.mlp.down_proj", "attn_pre_norm": "layers.75.input_layernorm", "attn_post_norm": "layers.75.post_attention_layernorm"}}, {"index": 76, "intervention_points": {"attn_input": "layers.76.self_attn.", "attn_q_proj": "layers.76.self_attn.q_proj", "attn_k_proj": "layers.76.self_attn.k_proj", "attn_v_proj": "layers.76.self_attn.v_proj", "attn_output_proj": "layers.76.self_attn.o_proj", "mlp_input": "layers.76.mlp.", "mlp_gate_proj": "layers.76.mlp.gate_proj", "mlp_up_proj": "layers.76.mlp.up_proj", "mlp_down_proj": "layers.76.mlp.down_proj", "attn_pre_norm": "layers.76.input_layernorm", "attn_post_norm": "layers.76.post_attention_layernorm"}}, {"index": 77, "intervention_points": {"attn_input": "layers.77.self_attn.", "attn_q_proj": "layers.77.self_attn.q_proj", "attn_k_proj": "layers.77.self_attn.k_proj", "attn_v_proj": "layers.77.self_attn.v_proj", "attn_output_proj": "layers.77.self_attn.o_proj", "mlp_input": "layers.77.mlp.", "mlp_gate_proj": "layers.77.mlp.gate_proj", "mlp_up_proj": "layers.77.mlp.up_proj", "mlp_down_proj": "layers.77.mlp.down_proj", "attn_pre_norm": "layers.77.input_layernorm", "attn_post_norm": "layers.77.post_attention_layernorm"}}, {"index": 78, "intervention_points": {"attn_input": "layers.78.self_attn.", "attn_q_proj": "layers.78.self_attn.q_proj", "attn_k_proj": "layers.78.self_attn.k_proj", "attn_v_proj": "layers.78.self_attn.v_proj", "attn_output_proj": "layers.78.self_attn.o_proj", "mlp_input": "layers.78.mlp.", "mlp_gate_proj": "layers.78.mlp.gate_proj", "mlp_up_proj": "layers.78.mlp.up_proj", "mlp_down_proj": "layers.78.mlp.down_proj", "attn_pre_norm": "layers.78.input_layernorm", "attn_post_norm": "layers.78.post_attention_layernorm"}}, {"index": 79, "intervention_points": {"attn_input": "layers.79.self_attn.", "attn_q_proj": "layers.79.self_attn.q_proj", "attn_k_proj": "layers.79.self_attn.k_proj", "attn_v_proj": "layers.79.self_attn.v_proj", "attn_output_proj": "layers.79.self_attn.o_proj", "mlp_input": "layers.79.mlp.", "mlp_gate_proj": "layers.79.mlp.gate_proj", "mlp_up_proj": "layers.79.mlp.up_proj", "mlp_down_proj": "layers.79.mlp.down_proj", "attn_pre_norm": "layers.79.input_layernorm", "attn_post_norm": "layers.79.post_attention_layernorm"}}]} | {"contributed_at": "2026-04-01T19:41:14.510113", "contributor": "trohrbaugh", "attribution_level": "full", "schema_version": 1} | public | 0.95 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.