diff --git "a/config.json" "b/config.json" --- "a/config.json" +++ "b/config.json" @@ -1,5 +1,5 @@ { - "_name_or_path": "mistralai/Mistral-Large-Instruct-2407", + "_name_or_path": "/home/aiscuser/yangwang/convert_model/missed_models_unconvert/VPTQ-community--Mistral-Large-Instruct-2407-v8-k65536-65536-woft", "architectures": [ "LlamaForCausalLM" ], @@ -19,22 +19,12 @@ "num_hidden_layers": 88, "num_key_value_heads": 8, "pretraining_tp": 1, - "rms_norm_eps": 1e-05, - "rope_scaling": null, - "rope_theta": 1000000.0, - "sliding_window": null, - "tie_word_embeddings": false, - "torch_dtype": "bfloat16", - "transformers_version": "4.45.2", - "use_cache": true, - "vocab_size": 32768, "quantization_config": { - "quant_method": "vptq", "config_for_layers": { "model.layers.0.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -58,7 +48,7 @@ "model.layers.0.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -82,7 +72,7 @@ "model.layers.0.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -106,7 +96,7 @@ "model.layers.0.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -130,7 +120,7 @@ "model.layers.0.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -154,7 +144,7 @@ "model.layers.0.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -178,7 +168,7 @@ "model.layers.0.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -202,7 +192,7 @@ "model.layers.1.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -226,7 +216,7 @@ "model.layers.1.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -250,7 +240,7 @@ "model.layers.1.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -274,7 +264,7 @@ "model.layers.1.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -298,7 +288,7 @@ "model.layers.1.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -322,7 +312,7 @@ "model.layers.1.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -346,7 +336,7 @@ "model.layers.1.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -370,7 +360,7 @@ "model.layers.10.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -394,7 +384,7 @@ "model.layers.10.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -418,7 +408,7 @@ "model.layers.10.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -442,7 +432,7 @@ "model.layers.10.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -466,7 +456,7 @@ "model.layers.10.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -490,7 +480,7 @@ "model.layers.10.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -514,7 +504,7 @@ "model.layers.10.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -538,7 +528,7 @@ "model.layers.11.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -562,7 +552,7 @@ "model.layers.11.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -586,7 +576,7 @@ "model.layers.11.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -610,7 +600,7 @@ "model.layers.11.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -634,7 +624,7 @@ "model.layers.11.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -658,7 +648,7 @@ "model.layers.11.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -682,7 +672,7 @@ "model.layers.11.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -706,7 +696,7 @@ "model.layers.12.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -730,7 +720,7 @@ "model.layers.12.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -754,7 +744,7 @@ "model.layers.12.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -778,7 +768,7 @@ "model.layers.12.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -802,7 +792,7 @@ "model.layers.12.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -826,7 +816,7 @@ "model.layers.12.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -850,7 +840,7 @@ "model.layers.12.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -874,7 +864,7 @@ "model.layers.13.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -898,7 +888,7 @@ "model.layers.13.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -922,7 +912,7 @@ "model.layers.13.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -946,7 +936,7 @@ "model.layers.13.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -970,7 +960,7 @@ "model.layers.13.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -994,7 +984,7 @@ "model.layers.13.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1018,7 +1008,7 @@ "model.layers.13.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1042,7 +1032,7 @@ "model.layers.14.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -1066,7 +1056,7 @@ "model.layers.14.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1090,7 +1080,7 @@ "model.layers.14.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1114,7 +1104,7 @@ "model.layers.14.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1138,7 +1128,7 @@ "model.layers.14.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1162,7 +1152,7 @@ "model.layers.14.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1186,7 +1176,7 @@ "model.layers.14.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1210,7 +1200,7 @@ "model.layers.15.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -1234,7 +1224,7 @@ "model.layers.15.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1258,7 +1248,7 @@ "model.layers.15.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1282,7 +1272,7 @@ "model.layers.15.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1306,7 +1296,7 @@ "model.layers.15.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1330,7 +1320,7 @@ "model.layers.15.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1354,7 +1344,7 @@ "model.layers.15.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1378,7 +1368,7 @@ "model.layers.16.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -1402,7 +1392,7 @@ "model.layers.16.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1426,7 +1416,7 @@ "model.layers.16.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1450,7 +1440,7 @@ "model.layers.16.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1474,7 +1464,7 @@ "model.layers.16.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1498,7 +1488,7 @@ "model.layers.16.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1522,7 +1512,7 @@ "model.layers.16.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1546,7 +1536,7 @@ "model.layers.17.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -1570,7 +1560,7 @@ "model.layers.17.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1594,7 +1584,7 @@ "model.layers.17.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1618,7 +1608,7 @@ "model.layers.17.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1642,7 +1632,7 @@ "model.layers.17.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1666,7 +1656,7 @@ "model.layers.17.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1690,7 +1680,7 @@ "model.layers.17.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1714,7 +1704,7 @@ "model.layers.18.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -1738,7 +1728,7 @@ "model.layers.18.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1762,7 +1752,7 @@ "model.layers.18.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1786,7 +1776,7 @@ "model.layers.18.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1810,7 +1800,7 @@ "model.layers.18.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1834,7 +1824,7 @@ "model.layers.18.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1858,7 +1848,7 @@ "model.layers.18.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1882,7 +1872,7 @@ "model.layers.19.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -1906,7 +1896,7 @@ "model.layers.19.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1930,7 +1920,7 @@ "model.layers.19.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1954,7 +1944,7 @@ "model.layers.19.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -1978,7 +1968,7 @@ "model.layers.19.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2002,7 +1992,7 @@ "model.layers.19.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2026,7 +2016,7 @@ "model.layers.19.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2050,7 +2040,7 @@ "model.layers.2.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -2074,7 +2064,7 @@ "model.layers.2.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2098,7 +2088,7 @@ "model.layers.2.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2122,7 +2112,7 @@ "model.layers.2.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2146,7 +2136,7 @@ "model.layers.2.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2170,7 +2160,7 @@ "model.layers.2.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2194,7 +2184,7 @@ "model.layers.2.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2218,7 +2208,7 @@ "model.layers.20.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -2242,7 +2232,7 @@ "model.layers.20.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2266,7 +2256,7 @@ "model.layers.20.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2290,7 +2280,7 @@ "model.layers.20.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2314,7 +2304,7 @@ "model.layers.20.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2338,7 +2328,7 @@ "model.layers.20.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2362,7 +2352,7 @@ "model.layers.20.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2386,7 +2376,7 @@ "model.layers.21.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -2410,7 +2400,7 @@ "model.layers.21.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2434,7 +2424,7 @@ "model.layers.21.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2458,7 +2448,7 @@ "model.layers.21.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2482,7 +2472,7 @@ "model.layers.21.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2506,7 +2496,7 @@ "model.layers.21.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2530,7 +2520,7 @@ "model.layers.21.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2554,7 +2544,7 @@ "model.layers.22.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -2578,7 +2568,7 @@ "model.layers.22.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2602,7 +2592,7 @@ "model.layers.22.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2626,7 +2616,7 @@ "model.layers.22.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2650,7 +2640,7 @@ "model.layers.22.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2674,7 +2664,7 @@ "model.layers.22.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2698,7 +2688,7 @@ "model.layers.22.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2722,7 +2712,7 @@ "model.layers.23.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -2746,7 +2736,7 @@ "model.layers.23.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2770,7 +2760,7 @@ "model.layers.23.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2794,7 +2784,7 @@ "model.layers.23.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2818,7 +2808,7 @@ "model.layers.23.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2842,7 +2832,7 @@ "model.layers.23.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2866,7 +2856,7 @@ "model.layers.23.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2890,7 +2880,7 @@ "model.layers.24.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -2914,7 +2904,7 @@ "model.layers.24.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2938,7 +2928,7 @@ "model.layers.24.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2962,7 +2952,7 @@ "model.layers.24.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -2986,7 +2976,7 @@ "model.layers.24.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3010,7 +3000,7 @@ "model.layers.24.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3034,7 +3024,7 @@ "model.layers.24.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3058,7 +3048,7 @@ "model.layers.25.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -3082,7 +3072,7 @@ "model.layers.25.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3106,7 +3096,7 @@ "model.layers.25.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3130,7 +3120,7 @@ "model.layers.25.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3154,7 +3144,7 @@ "model.layers.25.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3178,7 +3168,7 @@ "model.layers.25.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3202,7 +3192,7 @@ "model.layers.25.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3226,7 +3216,7 @@ "model.layers.26.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -3250,7 +3240,7 @@ "model.layers.26.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3274,7 +3264,7 @@ "model.layers.26.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3298,7 +3288,7 @@ "model.layers.26.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3322,7 +3312,7 @@ "model.layers.26.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3346,7 +3336,7 @@ "model.layers.26.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3370,7 +3360,7 @@ "model.layers.26.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3394,7 +3384,7 @@ "model.layers.27.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -3418,7 +3408,7 @@ "model.layers.27.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3442,7 +3432,7 @@ "model.layers.27.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3466,7 +3456,7 @@ "model.layers.27.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3490,7 +3480,7 @@ "model.layers.27.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3514,7 +3504,7 @@ "model.layers.27.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3538,7 +3528,7 @@ "model.layers.27.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3562,7 +3552,7 @@ "model.layers.28.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -3586,7 +3576,7 @@ "model.layers.28.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3610,7 +3600,7 @@ "model.layers.28.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3634,7 +3624,7 @@ "model.layers.28.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3658,7 +3648,7 @@ "model.layers.28.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3682,7 +3672,7 @@ "model.layers.28.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3706,7 +3696,7 @@ "model.layers.28.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3730,7 +3720,7 @@ "model.layers.29.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -3754,7 +3744,7 @@ "model.layers.29.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3778,7 +3768,7 @@ "model.layers.29.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3802,7 +3792,7 @@ "model.layers.29.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3826,7 +3816,7 @@ "model.layers.29.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3850,7 +3840,7 @@ "model.layers.29.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3874,7 +3864,7 @@ "model.layers.29.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3898,7 +3888,7 @@ "model.layers.3.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -3922,7 +3912,7 @@ "model.layers.3.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3946,7 +3936,7 @@ "model.layers.3.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3970,7 +3960,7 @@ "model.layers.3.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -3994,7 +3984,7 @@ "model.layers.3.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4018,7 +4008,7 @@ "model.layers.3.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4042,7 +4032,7 @@ "model.layers.3.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4066,7 +4056,7 @@ "model.layers.30.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -4090,7 +4080,7 @@ "model.layers.30.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4114,7 +4104,7 @@ "model.layers.30.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4138,7 +4128,7 @@ "model.layers.30.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4162,7 +4152,7 @@ "model.layers.30.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4186,7 +4176,7 @@ "model.layers.30.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4210,7 +4200,7 @@ "model.layers.30.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4234,7 +4224,7 @@ "model.layers.31.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -4258,7 +4248,7 @@ "model.layers.31.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4282,7 +4272,7 @@ "model.layers.31.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4306,7 +4296,7 @@ "model.layers.31.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4330,7 +4320,7 @@ "model.layers.31.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4354,7 +4344,7 @@ "model.layers.31.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4378,7 +4368,7 @@ "model.layers.31.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4402,7 +4392,7 @@ "model.layers.32.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -4426,7 +4416,7 @@ "model.layers.32.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4450,7 +4440,7 @@ "model.layers.32.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4474,7 +4464,7 @@ "model.layers.32.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4498,7 +4488,7 @@ "model.layers.32.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4522,7 +4512,7 @@ "model.layers.32.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4546,7 +4536,7 @@ "model.layers.32.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4570,7 +4560,7 @@ "model.layers.33.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -4594,7 +4584,7 @@ "model.layers.33.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4618,7 +4608,7 @@ "model.layers.33.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4642,7 +4632,7 @@ "model.layers.33.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4666,7 +4656,7 @@ "model.layers.33.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4690,7 +4680,7 @@ "model.layers.33.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4714,7 +4704,7 @@ "model.layers.33.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4738,7 +4728,7 @@ "model.layers.34.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -4762,7 +4752,7 @@ "model.layers.34.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4786,7 +4776,7 @@ "model.layers.34.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4810,7 +4800,7 @@ "model.layers.34.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4834,7 +4824,7 @@ "model.layers.34.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4858,7 +4848,7 @@ "model.layers.34.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4882,7 +4872,7 @@ "model.layers.34.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4906,7 +4896,7 @@ "model.layers.35.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -4930,7 +4920,7 @@ "model.layers.35.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4954,7 +4944,7 @@ "model.layers.35.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -4978,7 +4968,7 @@ "model.layers.35.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5002,7 +4992,7 @@ "model.layers.35.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5026,7 +5016,7 @@ "model.layers.35.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5050,7 +5040,7 @@ "model.layers.35.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5074,7 +5064,7 @@ "model.layers.36.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -5098,7 +5088,7 @@ "model.layers.36.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5122,7 +5112,7 @@ "model.layers.36.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5146,7 +5136,7 @@ "model.layers.36.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5170,7 +5160,7 @@ "model.layers.36.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5194,7 +5184,7 @@ "model.layers.36.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5218,7 +5208,7 @@ "model.layers.36.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5242,7 +5232,7 @@ "model.layers.37.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -5266,7 +5256,7 @@ "model.layers.37.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5290,7 +5280,7 @@ "model.layers.37.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5314,7 +5304,7 @@ "model.layers.37.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5338,7 +5328,7 @@ "model.layers.37.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5362,7 +5352,7 @@ "model.layers.37.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5386,7 +5376,7 @@ "model.layers.37.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5410,7 +5400,7 @@ "model.layers.38.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -5434,7 +5424,7 @@ "model.layers.38.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5458,7 +5448,7 @@ "model.layers.38.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5482,7 +5472,7 @@ "model.layers.38.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5506,7 +5496,7 @@ "model.layers.38.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5530,7 +5520,7 @@ "model.layers.38.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5554,7 +5544,7 @@ "model.layers.38.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5578,7 +5568,7 @@ "model.layers.39.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -5602,7 +5592,7 @@ "model.layers.39.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5626,7 +5616,7 @@ "model.layers.39.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5650,7 +5640,7 @@ "model.layers.39.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5674,7 +5664,7 @@ "model.layers.39.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5698,7 +5688,7 @@ "model.layers.39.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5722,7 +5712,7 @@ "model.layers.39.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5746,7 +5736,7 @@ "model.layers.4.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -5770,7 +5760,7 @@ "model.layers.4.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5794,7 +5784,7 @@ "model.layers.4.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5818,7 +5808,7 @@ "model.layers.4.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5842,7 +5832,7 @@ "model.layers.4.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5866,7 +5856,7 @@ "model.layers.4.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5890,7 +5880,7 @@ "model.layers.4.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5914,7 +5904,7 @@ "model.layers.40.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -5938,7 +5928,7 @@ "model.layers.40.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5962,7 +5952,7 @@ "model.layers.40.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -5986,7 +5976,7 @@ "model.layers.40.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6010,7 +6000,7 @@ "model.layers.40.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6034,7 +6024,7 @@ "model.layers.40.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6058,7 +6048,7 @@ "model.layers.40.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6082,7 +6072,7 @@ "model.layers.41.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -6106,7 +6096,7 @@ "model.layers.41.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6130,7 +6120,7 @@ "model.layers.41.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6154,7 +6144,7 @@ "model.layers.41.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6178,7 +6168,7 @@ "model.layers.41.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6202,7 +6192,7 @@ "model.layers.41.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6226,7 +6216,7 @@ "model.layers.41.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6250,7 +6240,7 @@ "model.layers.42.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -6274,7 +6264,7 @@ "model.layers.42.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6298,7 +6288,7 @@ "model.layers.42.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6322,7 +6312,7 @@ "model.layers.42.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6346,7 +6336,7 @@ "model.layers.42.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6370,7 +6360,7 @@ "model.layers.42.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6394,7 +6384,7 @@ "model.layers.42.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6418,7 +6408,7 @@ "model.layers.43.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -6442,7 +6432,7 @@ "model.layers.43.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6466,7 +6456,7 @@ "model.layers.43.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6490,7 +6480,7 @@ "model.layers.43.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6514,7 +6504,7 @@ "model.layers.43.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6538,7 +6528,7 @@ "model.layers.43.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6562,7 +6552,7 @@ "model.layers.43.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6586,7 +6576,7 @@ "model.layers.44.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -6610,7 +6600,7 @@ "model.layers.44.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6634,7 +6624,7 @@ "model.layers.44.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6658,7 +6648,7 @@ "model.layers.44.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6682,7 +6672,7 @@ "model.layers.44.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6706,7 +6696,7 @@ "model.layers.44.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6730,7 +6720,7 @@ "model.layers.44.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6754,7 +6744,7 @@ "model.layers.45.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -6778,7 +6768,7 @@ "model.layers.45.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6802,7 +6792,7 @@ "model.layers.45.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6826,7 +6816,7 @@ "model.layers.45.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6850,7 +6840,7 @@ "model.layers.45.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6874,7 +6864,7 @@ "model.layers.45.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6898,7 +6888,7 @@ "model.layers.45.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6922,7 +6912,7 @@ "model.layers.46.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -6946,7 +6936,7 @@ "model.layers.46.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6970,7 +6960,7 @@ "model.layers.46.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -6994,7 +6984,7 @@ "model.layers.46.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7018,7 +7008,7 @@ "model.layers.46.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7042,7 +7032,7 @@ "model.layers.46.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7066,7 +7056,7 @@ "model.layers.46.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7090,7 +7080,7 @@ "model.layers.47.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -7114,7 +7104,7 @@ "model.layers.47.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7138,7 +7128,7 @@ "model.layers.47.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7162,7 +7152,7 @@ "model.layers.47.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7186,7 +7176,7 @@ "model.layers.47.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7210,7 +7200,7 @@ "model.layers.47.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7234,7 +7224,7 @@ "model.layers.47.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7258,7 +7248,7 @@ "model.layers.48.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -7282,7 +7272,7 @@ "model.layers.48.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7306,7 +7296,7 @@ "model.layers.48.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7330,7 +7320,7 @@ "model.layers.48.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7354,7 +7344,7 @@ "model.layers.48.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7378,7 +7368,7 @@ "model.layers.48.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7402,7 +7392,7 @@ "model.layers.48.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7426,7 +7416,7 @@ "model.layers.49.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -7450,7 +7440,7 @@ "model.layers.49.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7474,7 +7464,7 @@ "model.layers.49.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7498,7 +7488,7 @@ "model.layers.49.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7522,7 +7512,7 @@ "model.layers.49.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7546,7 +7536,7 @@ "model.layers.49.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7570,7 +7560,7 @@ "model.layers.49.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7594,7 +7584,7 @@ "model.layers.5.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -7618,7 +7608,7 @@ "model.layers.5.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7642,7 +7632,7 @@ "model.layers.5.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7666,7 +7656,7 @@ "model.layers.5.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7690,7 +7680,7 @@ "model.layers.5.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7714,7 +7704,7 @@ "model.layers.5.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7738,7 +7728,7 @@ "model.layers.5.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7762,7 +7752,7 @@ "model.layers.50.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -7786,7 +7776,7 @@ "model.layers.50.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7810,7 +7800,7 @@ "model.layers.50.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7834,7 +7824,7 @@ "model.layers.50.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7858,7 +7848,7 @@ "model.layers.50.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7882,7 +7872,7 @@ "model.layers.50.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7906,7 +7896,7 @@ "model.layers.50.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7930,7 +7920,7 @@ "model.layers.51.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -7954,7 +7944,7 @@ "model.layers.51.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -7978,7 +7968,7 @@ "model.layers.51.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8002,7 +7992,7 @@ "model.layers.51.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8026,7 +8016,7 @@ "model.layers.51.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8050,7 +8040,7 @@ "model.layers.51.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8074,7 +8064,7 @@ "model.layers.51.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8098,7 +8088,7 @@ "model.layers.52.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -8122,7 +8112,7 @@ "model.layers.52.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8146,7 +8136,7 @@ "model.layers.52.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8170,7 +8160,7 @@ "model.layers.52.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8194,7 +8184,7 @@ "model.layers.52.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8218,7 +8208,7 @@ "model.layers.52.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8242,7 +8232,7 @@ "model.layers.52.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8266,7 +8256,7 @@ "model.layers.53.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -8290,7 +8280,7 @@ "model.layers.53.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8314,7 +8304,7 @@ "model.layers.53.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8338,7 +8328,7 @@ "model.layers.53.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8362,7 +8352,7 @@ "model.layers.53.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8386,7 +8376,7 @@ "model.layers.53.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8410,7 +8400,7 @@ "model.layers.53.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8434,7 +8424,7 @@ "model.layers.54.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -8458,7 +8448,7 @@ "model.layers.54.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8482,7 +8472,7 @@ "model.layers.54.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8506,7 +8496,7 @@ "model.layers.54.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8530,7 +8520,7 @@ "model.layers.54.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8554,7 +8544,7 @@ "model.layers.54.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8578,7 +8568,7 @@ "model.layers.54.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8602,7 +8592,7 @@ "model.layers.55.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -8626,7 +8616,7 @@ "model.layers.55.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8650,7 +8640,7 @@ "model.layers.55.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8674,7 +8664,7 @@ "model.layers.55.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8698,7 +8688,7 @@ "model.layers.55.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8722,7 +8712,7 @@ "model.layers.55.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8746,7 +8736,7 @@ "model.layers.55.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8770,7 +8760,7 @@ "model.layers.56.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -8794,7 +8784,7 @@ "model.layers.56.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8818,7 +8808,7 @@ "model.layers.56.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8842,7 +8832,7 @@ "model.layers.56.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8866,7 +8856,7 @@ "model.layers.56.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8890,7 +8880,7 @@ "model.layers.56.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8914,7 +8904,7 @@ "model.layers.56.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8938,7 +8928,7 @@ "model.layers.57.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -8962,7 +8952,7 @@ "model.layers.57.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -8986,7 +8976,7 @@ "model.layers.57.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9010,7 +9000,7 @@ "model.layers.57.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9034,7 +9024,7 @@ "model.layers.57.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9058,7 +9048,7 @@ "model.layers.57.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9082,7 +9072,7 @@ "model.layers.57.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9106,7 +9096,7 @@ "model.layers.58.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -9130,7 +9120,7 @@ "model.layers.58.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9154,7 +9144,7 @@ "model.layers.58.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9178,7 +9168,7 @@ "model.layers.58.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9202,7 +9192,7 @@ "model.layers.58.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9226,7 +9216,7 @@ "model.layers.58.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9250,7 +9240,7 @@ "model.layers.58.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9274,7 +9264,7 @@ "model.layers.59.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -9298,7 +9288,7 @@ "model.layers.59.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9322,7 +9312,7 @@ "model.layers.59.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9346,7 +9336,7 @@ "model.layers.59.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9370,7 +9360,7 @@ "model.layers.59.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9394,7 +9384,7 @@ "model.layers.59.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9418,7 +9408,7 @@ "model.layers.59.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9442,7 +9432,7 @@ "model.layers.6.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -9466,7 +9456,7 @@ "model.layers.6.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9490,7 +9480,7 @@ "model.layers.6.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9514,7 +9504,7 @@ "model.layers.6.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9538,7 +9528,7 @@ "model.layers.6.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9562,7 +9552,7 @@ "model.layers.6.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9586,7 +9576,7 @@ "model.layers.6.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9610,7 +9600,7 @@ "model.layers.60.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -9634,7 +9624,7 @@ "model.layers.60.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9658,7 +9648,7 @@ "model.layers.60.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9682,7 +9672,7 @@ "model.layers.60.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9706,7 +9696,7 @@ "model.layers.60.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9730,7 +9720,7 @@ "model.layers.60.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9754,7 +9744,7 @@ "model.layers.60.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9778,7 +9768,7 @@ "model.layers.61.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -9802,7 +9792,7 @@ "model.layers.61.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9826,7 +9816,7 @@ "model.layers.61.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9850,7 +9840,7 @@ "model.layers.61.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9874,7 +9864,7 @@ "model.layers.61.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9898,7 +9888,7 @@ "model.layers.61.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9922,7 +9912,7 @@ "model.layers.61.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9946,7 +9936,7 @@ "model.layers.62.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -9970,7 +9960,7 @@ "model.layers.62.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -9994,7 +9984,7 @@ "model.layers.62.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10018,7 +10008,7 @@ "model.layers.62.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10042,7 +10032,7 @@ "model.layers.62.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10066,7 +10056,7 @@ "model.layers.62.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10090,7 +10080,7 @@ "model.layers.62.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10114,7 +10104,7 @@ "model.layers.63.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -10138,7 +10128,7 @@ "model.layers.63.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10162,7 +10152,7 @@ "model.layers.63.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10186,7 +10176,7 @@ "model.layers.63.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10210,7 +10200,7 @@ "model.layers.63.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10234,7 +10224,7 @@ "model.layers.63.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10258,7 +10248,7 @@ "model.layers.63.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10282,7 +10272,7 @@ "model.layers.64.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -10306,7 +10296,7 @@ "model.layers.64.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10330,7 +10320,7 @@ "model.layers.64.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10354,7 +10344,7 @@ "model.layers.64.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10378,7 +10368,7 @@ "model.layers.64.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10402,7 +10392,7 @@ "model.layers.64.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10426,7 +10416,7 @@ "model.layers.64.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10450,7 +10440,7 @@ "model.layers.65.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -10474,7 +10464,7 @@ "model.layers.65.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10498,7 +10488,7 @@ "model.layers.65.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10522,7 +10512,7 @@ "model.layers.65.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10546,7 +10536,7 @@ "model.layers.65.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10570,7 +10560,7 @@ "model.layers.65.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10594,7 +10584,7 @@ "model.layers.65.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10618,7 +10608,7 @@ "model.layers.66.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -10642,7 +10632,7 @@ "model.layers.66.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10666,7 +10656,7 @@ "model.layers.66.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10690,7 +10680,7 @@ "model.layers.66.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10714,7 +10704,7 @@ "model.layers.66.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10738,7 +10728,7 @@ "model.layers.66.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10762,7 +10752,7 @@ "model.layers.66.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10786,7 +10776,7 @@ "model.layers.67.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -10810,7 +10800,7 @@ "model.layers.67.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10834,7 +10824,7 @@ "model.layers.67.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10858,7 +10848,7 @@ "model.layers.67.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10882,7 +10872,7 @@ "model.layers.67.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10906,7 +10896,7 @@ "model.layers.67.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10930,7 +10920,7 @@ "model.layers.67.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -10954,7 +10944,7 @@ "model.layers.68.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -10978,7 +10968,7 @@ "model.layers.68.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11002,7 +10992,7 @@ "model.layers.68.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11026,7 +11016,7 @@ "model.layers.68.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11050,7 +11040,7 @@ "model.layers.68.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11074,7 +11064,7 @@ "model.layers.68.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11098,7 +11088,7 @@ "model.layers.68.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11122,7 +11112,7 @@ "model.layers.69.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -11146,7 +11136,7 @@ "model.layers.69.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11170,7 +11160,7 @@ "model.layers.69.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11194,7 +11184,7 @@ "model.layers.69.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11218,7 +11208,7 @@ "model.layers.69.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11242,7 +11232,7 @@ "model.layers.69.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11266,7 +11256,7 @@ "model.layers.69.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11290,7 +11280,7 @@ "model.layers.7.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -11314,7 +11304,7 @@ "model.layers.7.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11338,7 +11328,7 @@ "model.layers.7.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11362,7 +11352,7 @@ "model.layers.7.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11386,7 +11376,7 @@ "model.layers.7.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11410,7 +11400,7 @@ "model.layers.7.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11434,7 +11424,7 @@ "model.layers.7.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11458,7 +11448,7 @@ "model.layers.70.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -11482,7 +11472,7 @@ "model.layers.70.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11506,7 +11496,7 @@ "model.layers.70.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11530,7 +11520,7 @@ "model.layers.70.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11554,7 +11544,7 @@ "model.layers.70.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11578,7 +11568,7 @@ "model.layers.70.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11602,7 +11592,7 @@ "model.layers.70.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11626,7 +11616,7 @@ "model.layers.71.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -11650,7 +11640,7 @@ "model.layers.71.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11674,7 +11664,7 @@ "model.layers.71.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11698,7 +11688,7 @@ "model.layers.71.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11722,7 +11712,7 @@ "model.layers.71.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11746,7 +11736,7 @@ "model.layers.71.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11770,7 +11760,7 @@ "model.layers.71.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11794,7 +11784,7 @@ "model.layers.72.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -11818,7 +11808,7 @@ "model.layers.72.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11842,7 +11832,7 @@ "model.layers.72.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11866,7 +11856,7 @@ "model.layers.72.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11890,7 +11880,7 @@ "model.layers.72.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11914,7 +11904,7 @@ "model.layers.72.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11938,7 +11928,7 @@ "model.layers.72.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -11962,7 +11952,7 @@ "model.layers.73.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -11986,7 +11976,7 @@ "model.layers.73.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12010,7 +12000,7 @@ "model.layers.73.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12034,7 +12024,7 @@ "model.layers.73.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12058,7 +12048,7 @@ "model.layers.73.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12082,7 +12072,7 @@ "model.layers.73.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12106,7 +12096,7 @@ "model.layers.73.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12130,7 +12120,7 @@ "model.layers.74.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -12154,7 +12144,7 @@ "model.layers.74.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12178,7 +12168,7 @@ "model.layers.74.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12202,7 +12192,7 @@ "model.layers.74.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12226,7 +12216,7 @@ "model.layers.74.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12250,7 +12240,7 @@ "model.layers.74.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12274,7 +12264,7 @@ "model.layers.74.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12298,7 +12288,7 @@ "model.layers.75.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -12322,7 +12312,7 @@ "model.layers.75.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12346,7 +12336,7 @@ "model.layers.75.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12370,7 +12360,7 @@ "model.layers.75.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12394,7 +12384,7 @@ "model.layers.75.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12418,7 +12408,7 @@ "model.layers.75.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12442,7 +12432,7 @@ "model.layers.75.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12466,7 +12456,7 @@ "model.layers.76.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -12490,7 +12480,7 @@ "model.layers.76.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12514,7 +12504,7 @@ "model.layers.76.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12538,7 +12528,7 @@ "model.layers.76.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12562,7 +12552,7 @@ "model.layers.76.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12586,7 +12576,7 @@ "model.layers.76.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12610,7 +12600,7 @@ "model.layers.76.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12634,7 +12624,7 @@ "model.layers.77.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -12658,7 +12648,7 @@ "model.layers.77.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12682,7 +12672,7 @@ "model.layers.77.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12706,7 +12696,7 @@ "model.layers.77.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12730,7 +12720,7 @@ "model.layers.77.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12754,7 +12744,7 @@ "model.layers.77.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12778,7 +12768,7 @@ "model.layers.77.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12802,7 +12792,7 @@ "model.layers.78.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -12826,7 +12816,7 @@ "model.layers.78.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12850,7 +12840,7 @@ "model.layers.78.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12874,7 +12864,7 @@ "model.layers.78.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12898,7 +12888,7 @@ "model.layers.78.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12922,7 +12912,7 @@ "model.layers.78.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12946,7 +12936,7 @@ "model.layers.78.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -12970,7 +12960,7 @@ "model.layers.79.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -12994,7 +12984,7 @@ "model.layers.79.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13018,7 +13008,7 @@ "model.layers.79.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13042,7 +13032,7 @@ "model.layers.79.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13066,7 +13056,7 @@ "model.layers.79.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13090,7 +13080,7 @@ "model.layers.79.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13114,7 +13104,7 @@ "model.layers.79.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13138,7 +13128,7 @@ "model.layers.8.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -13162,7 +13152,7 @@ "model.layers.8.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13186,7 +13176,7 @@ "model.layers.8.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13210,7 +13200,7 @@ "model.layers.8.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13234,7 +13224,7 @@ "model.layers.8.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13258,7 +13248,7 @@ "model.layers.8.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13282,7 +13272,7 @@ "model.layers.8.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13306,7 +13296,7 @@ "model.layers.80.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -13330,7 +13320,7 @@ "model.layers.80.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13354,7 +13344,7 @@ "model.layers.80.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13378,7 +13368,7 @@ "model.layers.80.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13402,7 +13392,7 @@ "model.layers.80.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13426,7 +13416,7 @@ "model.layers.80.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13450,7 +13440,7 @@ "model.layers.80.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13474,7 +13464,7 @@ "model.layers.81.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -13498,7 +13488,7 @@ "model.layers.81.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13522,7 +13512,7 @@ "model.layers.81.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13546,7 +13536,7 @@ "model.layers.81.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13570,7 +13560,7 @@ "model.layers.81.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13594,7 +13584,7 @@ "model.layers.81.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13618,7 +13608,7 @@ "model.layers.81.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13642,7 +13632,7 @@ "model.layers.82.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -13666,7 +13656,7 @@ "model.layers.82.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13690,7 +13680,7 @@ "model.layers.82.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13714,7 +13704,7 @@ "model.layers.82.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13738,7 +13728,7 @@ "model.layers.82.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13762,7 +13752,7 @@ "model.layers.82.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13786,7 +13776,7 @@ "model.layers.82.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13810,7 +13800,7 @@ "model.layers.83.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -13834,7 +13824,7 @@ "model.layers.83.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13858,7 +13848,7 @@ "model.layers.83.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13882,7 +13872,7 @@ "model.layers.83.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13906,7 +13896,7 @@ "model.layers.83.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13930,7 +13920,7 @@ "model.layers.83.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13954,7 +13944,7 @@ "model.layers.83.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -13978,7 +13968,7 @@ "model.layers.84.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -14002,7 +13992,7 @@ "model.layers.84.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14026,7 +14016,7 @@ "model.layers.84.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14050,7 +14040,7 @@ "model.layers.84.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14074,7 +14064,7 @@ "model.layers.84.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14098,7 +14088,7 @@ "model.layers.84.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14122,7 +14112,7 @@ "model.layers.84.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14146,7 +14136,7 @@ "model.layers.85.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -14170,7 +14160,7 @@ "model.layers.85.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14194,7 +14184,7 @@ "model.layers.85.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14218,7 +14208,7 @@ "model.layers.85.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14242,7 +14232,7 @@ "model.layers.85.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14266,7 +14256,7 @@ "model.layers.85.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14290,7 +14280,7 @@ "model.layers.85.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14314,7 +14304,7 @@ "model.layers.86.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -14338,7 +14328,7 @@ "model.layers.86.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14362,7 +14352,7 @@ "model.layers.86.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14386,7 +14376,7 @@ "model.layers.86.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14410,7 +14400,7 @@ "model.layers.86.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14434,7 +14424,7 @@ "model.layers.86.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14458,7 +14448,7 @@ "model.layers.86.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14482,7 +14472,7 @@ "model.layers.87.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -14506,7 +14496,7 @@ "model.layers.87.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14530,7 +14520,7 @@ "model.layers.87.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14554,7 +14544,7 @@ "model.layers.87.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14578,7 +14568,7 @@ "model.layers.87.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14602,7 +14592,7 @@ "model.layers.87.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14626,7 +14616,7 @@ "model.layers.87.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14650,7 +14640,7 @@ "model.layers.9.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -14674,7 +14664,7 @@ "model.layers.9.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14698,7 +14688,7 @@ "model.layers.9.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14722,7 +14712,7 @@ "model.layers.9.self_attn.k_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14746,7 +14736,7 @@ "model.layers.9.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14770,7 +14760,7 @@ "model.layers.9.self_attn.q_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14794,7 +14784,7 @@ "model.layers.9.self_attn.v_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 12288, "in_features": 12288, @@ -14815,6 +14805,16 @@ 8 ] } - } - } -} \ No newline at end of file + }, + "quant_method": "vptq" + }, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 1000000.0, + "sliding_window": null, + "tie_word_embeddings": false, + "torch_dtype": "bfloat16", + "transformers_version": "4.49.0", + "use_cache": true, + "vocab_size": 32768 +}