wenhuach commited on
Commit
7d1625e
·
verified ·
1 Parent(s): 78b99c7

Regenerate the model to ensure successful execution on vLLM

Browse files
config.json CHANGED
@@ -25,13 +25,12 @@
25
  "num_key_value_heads": 4,
26
  "output_router_logits": false,
27
  "quantization_config": {
28
- "autoround_version": "0.6.0",
29
  "bits": 4,
30
  "data_type": "int",
31
  "extra_config": {
32
  "model.layers.0.mlp.gate": {
33
- "bits": 8,
34
- "group_size": 128
35
  },
36
  "model.layers.0.self_attn.k_proj": {
37
  "bits": 8,
@@ -50,8 +49,7 @@
50
  "group_size": 128
51
  },
52
  "model.layers.1.mlp.gate": {
53
- "bits": 8,
54
- "group_size": 128
55
  },
56
  "model.layers.1.self_attn.k_proj": {
57
  "bits": 8,
@@ -70,8 +68,7 @@
70
  "group_size": 128
71
  },
72
  "model.layers.10.mlp.gate": {
73
- "bits": 8,
74
- "group_size": 128
75
  },
76
  "model.layers.10.self_attn.k_proj": {
77
  "bits": 8,
@@ -90,8 +87,7 @@
90
  "group_size": 128
91
  },
92
  "model.layers.11.mlp.gate": {
93
- "bits": 8,
94
- "group_size": 128
95
  },
96
  "model.layers.11.self_attn.k_proj": {
97
  "bits": 8,
@@ -110,8 +106,7 @@
110
  "group_size": 128
111
  },
112
  "model.layers.12.mlp.gate": {
113
- "bits": 8,
114
- "group_size": 128
115
  },
116
  "model.layers.12.self_attn.k_proj": {
117
  "bits": 8,
@@ -130,8 +125,7 @@
130
  "group_size": 128
131
  },
132
  "model.layers.13.mlp.gate": {
133
- "bits": 8,
134
- "group_size": 128
135
  },
136
  "model.layers.13.self_attn.k_proj": {
137
  "bits": 8,
@@ -150,8 +144,7 @@
150
  "group_size": 128
151
  },
152
  "model.layers.14.mlp.gate": {
153
- "bits": 8,
154
- "group_size": 128
155
  },
156
  "model.layers.14.self_attn.k_proj": {
157
  "bits": 8,
@@ -170,8 +163,7 @@
170
  "group_size": 128
171
  },
172
  "model.layers.15.mlp.gate": {
173
- "bits": 8,
174
- "group_size": 128
175
  },
176
  "model.layers.15.self_attn.k_proj": {
177
  "bits": 8,
@@ -190,8 +182,7 @@
190
  "group_size": 128
191
  },
192
  "model.layers.16.mlp.gate": {
193
- "bits": 8,
194
- "group_size": 128
195
  },
196
  "model.layers.16.self_attn.k_proj": {
197
  "bits": 8,
@@ -210,8 +201,7 @@
210
  "group_size": 128
211
  },
212
  "model.layers.17.mlp.gate": {
213
- "bits": 8,
214
- "group_size": 128
215
  },
216
  "model.layers.17.self_attn.k_proj": {
217
  "bits": 8,
@@ -230,8 +220,7 @@
230
  "group_size": 128
231
  },
232
  "model.layers.18.mlp.gate": {
233
- "bits": 8,
234
- "group_size": 128
235
  },
236
  "model.layers.18.self_attn.k_proj": {
237
  "bits": 8,
@@ -250,8 +239,7 @@
250
  "group_size": 128
251
  },
252
  "model.layers.19.mlp.gate": {
253
- "bits": 8,
254
- "group_size": 128
255
  },
256
  "model.layers.19.self_attn.k_proj": {
257
  "bits": 8,
@@ -270,8 +258,7 @@
270
  "group_size": 128
271
  },
272
  "model.layers.2.mlp.gate": {
273
- "bits": 8,
274
- "group_size": 128
275
  },
276
  "model.layers.2.self_attn.k_proj": {
277
  "bits": 8,
@@ -290,8 +277,7 @@
290
  "group_size": 128
291
  },
292
  "model.layers.20.mlp.gate": {
293
- "bits": 8,
294
- "group_size": 128
295
  },
296
  "model.layers.20.self_attn.k_proj": {
297
  "bits": 8,
@@ -310,8 +296,7 @@
310
  "group_size": 128
311
  },
312
  "model.layers.21.mlp.gate": {
313
- "bits": 8,
314
- "group_size": 128
315
  },
316
  "model.layers.21.self_attn.k_proj": {
317
  "bits": 8,
@@ -330,8 +315,7 @@
330
  "group_size": 128
331
  },
332
  "model.layers.22.mlp.gate": {
333
- "bits": 8,
334
- "group_size": 128
335
  },
336
  "model.layers.22.self_attn.k_proj": {
337
  "bits": 8,
@@ -350,8 +334,7 @@
350
  "group_size": 128
351
  },
352
  "model.layers.23.mlp.gate": {
353
- "bits": 8,
354
- "group_size": 128
355
  },
356
  "model.layers.23.self_attn.k_proj": {
357
  "bits": 8,
@@ -370,8 +353,7 @@
370
  "group_size": 128
371
  },
372
  "model.layers.24.mlp.gate": {
373
- "bits": 8,
374
- "group_size": 128
375
  },
376
  "model.layers.24.self_attn.k_proj": {
377
  "bits": 8,
@@ -390,8 +372,7 @@
390
  "group_size": 128
391
  },
392
  "model.layers.25.mlp.gate": {
393
- "bits": 8,
394
- "group_size": 128
395
  },
396
  "model.layers.25.self_attn.k_proj": {
397
  "bits": 8,
@@ -410,8 +391,7 @@
410
  "group_size": 128
411
  },
412
  "model.layers.26.mlp.gate": {
413
- "bits": 8,
414
- "group_size": 128
415
  },
416
  "model.layers.26.self_attn.k_proj": {
417
  "bits": 8,
@@ -430,8 +410,7 @@
430
  "group_size": 128
431
  },
432
  "model.layers.27.mlp.gate": {
433
- "bits": 8,
434
- "group_size": 128
435
  },
436
  "model.layers.27.self_attn.k_proj": {
437
  "bits": 8,
@@ -450,8 +429,7 @@
450
  "group_size": 128
451
  },
452
  "model.layers.28.mlp.gate": {
453
- "bits": 8,
454
- "group_size": 128
455
  },
456
  "model.layers.28.self_attn.k_proj": {
457
  "bits": 8,
@@ -470,8 +448,7 @@
470
  "group_size": 128
471
  },
472
  "model.layers.29.mlp.gate": {
473
- "bits": 8,
474
- "group_size": 128
475
  },
476
  "model.layers.29.self_attn.k_proj": {
477
  "bits": 8,
@@ -490,8 +467,7 @@
490
  "group_size": 128
491
  },
492
  "model.layers.3.mlp.gate": {
493
- "bits": 8,
494
- "group_size": 128
495
  },
496
  "model.layers.3.self_attn.k_proj": {
497
  "bits": 8,
@@ -510,8 +486,7 @@
510
  "group_size": 128
511
  },
512
  "model.layers.30.mlp.gate": {
513
- "bits": 8,
514
- "group_size": 128
515
  },
516
  "model.layers.30.self_attn.k_proj": {
517
  "bits": 8,
@@ -530,8 +505,7 @@
530
  "group_size": 128
531
  },
532
  "model.layers.31.mlp.gate": {
533
- "bits": 8,
534
- "group_size": 128
535
  },
536
  "model.layers.31.self_attn.k_proj": {
537
  "bits": 8,
@@ -550,8 +524,7 @@
550
  "group_size": 128
551
  },
552
  "model.layers.32.mlp.gate": {
553
- "bits": 8,
554
- "group_size": 128
555
  },
556
  "model.layers.32.self_attn.k_proj": {
557
  "bits": 8,
@@ -570,8 +543,7 @@
570
  "group_size": 128
571
  },
572
  "model.layers.33.mlp.gate": {
573
- "bits": 8,
574
- "group_size": 128
575
  },
576
  "model.layers.33.self_attn.k_proj": {
577
  "bits": 8,
@@ -590,8 +562,7 @@
590
  "group_size": 128
591
  },
592
  "model.layers.34.mlp.gate": {
593
- "bits": 8,
594
- "group_size": 128
595
  },
596
  "model.layers.34.self_attn.k_proj": {
597
  "bits": 8,
@@ -610,8 +581,7 @@
610
  "group_size": 128
611
  },
612
  "model.layers.35.mlp.gate": {
613
- "bits": 8,
614
- "group_size": 128
615
  },
616
  "model.layers.35.self_attn.k_proj": {
617
  "bits": 8,
@@ -630,8 +600,7 @@
630
  "group_size": 128
631
  },
632
  "model.layers.36.mlp.gate": {
633
- "bits": 8,
634
- "group_size": 128
635
  },
636
  "model.layers.36.self_attn.k_proj": {
637
  "bits": 8,
@@ -650,8 +619,7 @@
650
  "group_size": 128
651
  },
652
  "model.layers.37.mlp.gate": {
653
- "bits": 8,
654
- "group_size": 128
655
  },
656
  "model.layers.37.self_attn.k_proj": {
657
  "bits": 8,
@@ -670,8 +638,7 @@
670
  "group_size": 128
671
  },
672
  "model.layers.38.mlp.gate": {
673
- "bits": 8,
674
- "group_size": 128
675
  },
676
  "model.layers.38.self_attn.k_proj": {
677
  "bits": 8,
@@ -690,8 +657,7 @@
690
  "group_size": 128
691
  },
692
  "model.layers.39.mlp.gate": {
693
- "bits": 8,
694
- "group_size": 128
695
  },
696
  "model.layers.39.self_attn.k_proj": {
697
  "bits": 8,
@@ -710,8 +676,7 @@
710
  "group_size": 128
711
  },
712
  "model.layers.4.mlp.gate": {
713
- "bits": 8,
714
- "group_size": 128
715
  },
716
  "model.layers.4.self_attn.k_proj": {
717
  "bits": 8,
@@ -730,8 +695,7 @@
730
  "group_size": 128
731
  },
732
  "model.layers.40.mlp.gate": {
733
- "bits": 8,
734
- "group_size": 128
735
  },
736
  "model.layers.40.self_attn.k_proj": {
737
  "bits": 8,
@@ -750,8 +714,7 @@
750
  "group_size": 128
751
  },
752
  "model.layers.41.mlp.gate": {
753
- "bits": 8,
754
- "group_size": 128
755
  },
756
  "model.layers.41.self_attn.k_proj": {
757
  "bits": 8,
@@ -770,8 +733,7 @@
770
  "group_size": 128
771
  },
772
  "model.layers.42.mlp.gate": {
773
- "bits": 8,
774
- "group_size": 128
775
  },
776
  "model.layers.42.self_attn.k_proj": {
777
  "bits": 8,
@@ -790,8 +752,7 @@
790
  "group_size": 128
791
  },
792
  "model.layers.43.mlp.gate": {
793
- "bits": 8,
794
- "group_size": 128
795
  },
796
  "model.layers.43.self_attn.k_proj": {
797
  "bits": 8,
@@ -810,8 +771,7 @@
810
  "group_size": 128
811
  },
812
  "model.layers.44.mlp.gate": {
813
- "bits": 8,
814
- "group_size": 128
815
  },
816
  "model.layers.44.self_attn.k_proj": {
817
  "bits": 8,
@@ -830,8 +790,7 @@
830
  "group_size": 128
831
  },
832
  "model.layers.45.mlp.gate": {
833
- "bits": 8,
834
- "group_size": 128
835
  },
836
  "model.layers.45.self_attn.k_proj": {
837
  "bits": 8,
@@ -850,8 +809,7 @@
850
  "group_size": 128
851
  },
852
  "model.layers.46.mlp.gate": {
853
- "bits": 8,
854
- "group_size": 128
855
  },
856
  "model.layers.46.self_attn.k_proj": {
857
  "bits": 8,
@@ -870,8 +828,7 @@
870
  "group_size": 128
871
  },
872
  "model.layers.47.mlp.gate": {
873
- "bits": 8,
874
- "group_size": 128
875
  },
876
  "model.layers.47.self_attn.k_proj": {
877
  "bits": 8,
@@ -890,8 +847,7 @@
890
  "group_size": 128
891
  },
892
  "model.layers.48.mlp.gate": {
893
- "bits": 8,
894
- "group_size": 128
895
  },
896
  "model.layers.48.self_attn.k_proj": {
897
  "bits": 8,
@@ -910,8 +866,7 @@
910
  "group_size": 128
911
  },
912
  "model.layers.49.mlp.gate": {
913
- "bits": 8,
914
- "group_size": 128
915
  },
916
  "model.layers.49.self_attn.k_proj": {
917
  "bits": 8,
@@ -930,8 +885,7 @@
930
  "group_size": 128
931
  },
932
  "model.layers.5.mlp.gate": {
933
- "bits": 8,
934
- "group_size": 128
935
  },
936
  "model.layers.5.self_attn.k_proj": {
937
  "bits": 8,
@@ -950,8 +904,7 @@
950
  "group_size": 128
951
  },
952
  "model.layers.50.mlp.gate": {
953
- "bits": 8,
954
- "group_size": 128
955
  },
956
  "model.layers.50.self_attn.k_proj": {
957
  "bits": 8,
@@ -970,8 +923,7 @@
970
  "group_size": 128
971
  },
972
  "model.layers.51.mlp.gate": {
973
- "bits": 8,
974
- "group_size": 128
975
  },
976
  "model.layers.51.self_attn.k_proj": {
977
  "bits": 8,
@@ -990,8 +942,7 @@
990
  "group_size": 128
991
  },
992
  "model.layers.52.mlp.gate": {
993
- "bits": 8,
994
- "group_size": 128
995
  },
996
  "model.layers.52.self_attn.k_proj": {
997
  "bits": 8,
@@ -1010,8 +961,7 @@
1010
  "group_size": 128
1011
  },
1012
  "model.layers.53.mlp.gate": {
1013
- "bits": 8,
1014
- "group_size": 128
1015
  },
1016
  "model.layers.53.self_attn.k_proj": {
1017
  "bits": 8,
@@ -1030,8 +980,7 @@
1030
  "group_size": 128
1031
  },
1032
  "model.layers.54.mlp.gate": {
1033
- "bits": 8,
1034
- "group_size": 128
1035
  },
1036
  "model.layers.54.self_attn.k_proj": {
1037
  "bits": 8,
@@ -1050,8 +999,7 @@
1050
  "group_size": 128
1051
  },
1052
  "model.layers.55.mlp.gate": {
1053
- "bits": 8,
1054
- "group_size": 128
1055
  },
1056
  "model.layers.55.self_attn.k_proj": {
1057
  "bits": 8,
@@ -1070,8 +1018,7 @@
1070
  "group_size": 128
1071
  },
1072
  "model.layers.56.mlp.gate": {
1073
- "bits": 8,
1074
- "group_size": 128
1075
  },
1076
  "model.layers.56.self_attn.k_proj": {
1077
  "bits": 8,
@@ -1090,8 +1037,7 @@
1090
  "group_size": 128
1091
  },
1092
  "model.layers.57.mlp.gate": {
1093
- "bits": 8,
1094
- "group_size": 128
1095
  },
1096
  "model.layers.57.self_attn.k_proj": {
1097
  "bits": 8,
@@ -1110,8 +1056,7 @@
1110
  "group_size": 128
1111
  },
1112
  "model.layers.58.mlp.gate": {
1113
- "bits": 8,
1114
- "group_size": 128
1115
  },
1116
  "model.layers.58.self_attn.k_proj": {
1117
  "bits": 8,
@@ -1130,8 +1075,7 @@
1130
  "group_size": 128
1131
  },
1132
  "model.layers.59.mlp.gate": {
1133
- "bits": 8,
1134
- "group_size": 128
1135
  },
1136
  "model.layers.59.self_attn.k_proj": {
1137
  "bits": 8,
@@ -1150,8 +1094,7 @@
1150
  "group_size": 128
1151
  },
1152
  "model.layers.6.mlp.gate": {
1153
- "bits": 8,
1154
- "group_size": 128
1155
  },
1156
  "model.layers.6.self_attn.k_proj": {
1157
  "bits": 8,
@@ -1170,8 +1113,7 @@
1170
  "group_size": 128
1171
  },
1172
  "model.layers.60.mlp.gate": {
1173
- "bits": 8,
1174
- "group_size": 128
1175
  },
1176
  "model.layers.60.self_attn.k_proj": {
1177
  "bits": 8,
@@ -1190,8 +1132,7 @@
1190
  "group_size": 128
1191
  },
1192
  "model.layers.61.mlp.gate": {
1193
- "bits": 8,
1194
- "group_size": 128
1195
  },
1196
  "model.layers.61.self_attn.k_proj": {
1197
  "bits": 8,
@@ -1210,8 +1151,7 @@
1210
  "group_size": 128
1211
  },
1212
  "model.layers.62.mlp.gate": {
1213
- "bits": 8,
1214
- "group_size": 128
1215
  },
1216
  "model.layers.62.self_attn.k_proj": {
1217
  "bits": 8,
@@ -1230,8 +1170,7 @@
1230
  "group_size": 128
1231
  },
1232
  "model.layers.63.mlp.gate": {
1233
- "bits": 8,
1234
- "group_size": 128
1235
  },
1236
  "model.layers.63.self_attn.k_proj": {
1237
  "bits": 8,
@@ -1250,8 +1189,7 @@
1250
  "group_size": 128
1251
  },
1252
  "model.layers.64.mlp.gate": {
1253
- "bits": 8,
1254
- "group_size": 128
1255
  },
1256
  "model.layers.64.self_attn.k_proj": {
1257
  "bits": 8,
@@ -1270,8 +1208,7 @@
1270
  "group_size": 128
1271
  },
1272
  "model.layers.65.mlp.gate": {
1273
- "bits": 8,
1274
- "group_size": 128
1275
  },
1276
  "model.layers.65.self_attn.k_proj": {
1277
  "bits": 8,
@@ -1290,8 +1227,7 @@
1290
  "group_size": 128
1291
  },
1292
  "model.layers.66.mlp.gate": {
1293
- "bits": 8,
1294
- "group_size": 128
1295
  },
1296
  "model.layers.66.self_attn.k_proj": {
1297
  "bits": 8,
@@ -1310,8 +1246,7 @@
1310
  "group_size": 128
1311
  },
1312
  "model.layers.67.mlp.gate": {
1313
- "bits": 8,
1314
- "group_size": 128
1315
  },
1316
  "model.layers.67.self_attn.k_proj": {
1317
  "bits": 8,
@@ -1330,8 +1265,7 @@
1330
  "group_size": 128
1331
  },
1332
  "model.layers.68.mlp.gate": {
1333
- "bits": 8,
1334
- "group_size": 128
1335
  },
1336
  "model.layers.68.self_attn.k_proj": {
1337
  "bits": 8,
@@ -1350,8 +1284,7 @@
1350
  "group_size": 128
1351
  },
1352
  "model.layers.69.mlp.gate": {
1353
- "bits": 8,
1354
- "group_size": 128
1355
  },
1356
  "model.layers.69.self_attn.k_proj": {
1357
  "bits": 8,
@@ -1370,8 +1303,7 @@
1370
  "group_size": 128
1371
  },
1372
  "model.layers.7.mlp.gate": {
1373
- "bits": 8,
1374
- "group_size": 128
1375
  },
1376
  "model.layers.7.self_attn.k_proj": {
1377
  "bits": 8,
@@ -1390,8 +1322,7 @@
1390
  "group_size": 128
1391
  },
1392
  "model.layers.70.mlp.gate": {
1393
- "bits": 8,
1394
- "group_size": 128
1395
  },
1396
  "model.layers.70.self_attn.k_proj": {
1397
  "bits": 8,
@@ -1410,8 +1341,7 @@
1410
  "group_size": 128
1411
  },
1412
  "model.layers.71.mlp.gate": {
1413
- "bits": 8,
1414
- "group_size": 128
1415
  },
1416
  "model.layers.71.self_attn.k_proj": {
1417
  "bits": 8,
@@ -1430,8 +1360,7 @@
1430
  "group_size": 128
1431
  },
1432
  "model.layers.72.mlp.gate": {
1433
- "bits": 8,
1434
- "group_size": 128
1435
  },
1436
  "model.layers.72.self_attn.k_proj": {
1437
  "bits": 8,
@@ -1450,8 +1379,7 @@
1450
  "group_size": 128
1451
  },
1452
  "model.layers.73.mlp.gate": {
1453
- "bits": 8,
1454
- "group_size": 128
1455
  },
1456
  "model.layers.73.self_attn.k_proj": {
1457
  "bits": 8,
@@ -1470,8 +1398,7 @@
1470
  "group_size": 128
1471
  },
1472
  "model.layers.74.mlp.gate": {
1473
- "bits": 8,
1474
- "group_size": 128
1475
  },
1476
  "model.layers.74.self_attn.k_proj": {
1477
  "bits": 8,
@@ -1490,8 +1417,7 @@
1490
  "group_size": 128
1491
  },
1492
  "model.layers.75.mlp.gate": {
1493
- "bits": 8,
1494
- "group_size": 128
1495
  },
1496
  "model.layers.75.self_attn.k_proj": {
1497
  "bits": 8,
@@ -1510,8 +1436,7 @@
1510
  "group_size": 128
1511
  },
1512
  "model.layers.76.mlp.gate": {
1513
- "bits": 8,
1514
- "group_size": 128
1515
  },
1516
  "model.layers.76.self_attn.k_proj": {
1517
  "bits": 8,
@@ -1530,8 +1455,7 @@
1530
  "group_size": 128
1531
  },
1532
  "model.layers.77.mlp.gate": {
1533
- "bits": 8,
1534
- "group_size": 128
1535
  },
1536
  "model.layers.77.self_attn.k_proj": {
1537
  "bits": 8,
@@ -1550,8 +1474,7 @@
1550
  "group_size": 128
1551
  },
1552
  "model.layers.78.mlp.gate": {
1553
- "bits": 8,
1554
- "group_size": 128
1555
  },
1556
  "model.layers.78.self_attn.k_proj": {
1557
  "bits": 8,
@@ -1570,8 +1493,7 @@
1570
  "group_size": 128
1571
  },
1572
  "model.layers.79.mlp.gate": {
1573
- "bits": 8,
1574
- "group_size": 128
1575
  },
1576
  "model.layers.79.self_attn.k_proj": {
1577
  "bits": 8,
@@ -1590,8 +1512,7 @@
1590
  "group_size": 128
1591
  },
1592
  "model.layers.8.mlp.gate": {
1593
- "bits": 8,
1594
- "group_size": 128
1595
  },
1596
  "model.layers.8.self_attn.k_proj": {
1597
  "bits": 8,
@@ -1610,8 +1531,7 @@
1610
  "group_size": 128
1611
  },
1612
  "model.layers.80.mlp.gate": {
1613
- "bits": 8,
1614
- "group_size": 128
1615
  },
1616
  "model.layers.80.self_attn.k_proj": {
1617
  "bits": 8,
@@ -1630,8 +1550,7 @@
1630
  "group_size": 128
1631
  },
1632
  "model.layers.81.mlp.gate": {
1633
- "bits": 8,
1634
- "group_size": 128
1635
  },
1636
  "model.layers.81.self_attn.k_proj": {
1637
  "bits": 8,
@@ -1650,8 +1569,7 @@
1650
  "group_size": 128
1651
  },
1652
  "model.layers.82.mlp.gate": {
1653
- "bits": 8,
1654
- "group_size": 128
1655
  },
1656
  "model.layers.82.self_attn.k_proj": {
1657
  "bits": 8,
@@ -1670,8 +1588,7 @@
1670
  "group_size": 128
1671
  },
1672
  "model.layers.83.mlp.gate": {
1673
- "bits": 8,
1674
- "group_size": 128
1675
  },
1676
  "model.layers.83.self_attn.k_proj": {
1677
  "bits": 8,
@@ -1690,8 +1607,7 @@
1690
  "group_size": 128
1691
  },
1692
  "model.layers.84.mlp.gate": {
1693
- "bits": 8,
1694
- "group_size": 128
1695
  },
1696
  "model.layers.84.self_attn.k_proj": {
1697
  "bits": 8,
@@ -1710,8 +1626,7 @@
1710
  "group_size": 128
1711
  },
1712
  "model.layers.85.mlp.gate": {
1713
- "bits": 8,
1714
- "group_size": 128
1715
  },
1716
  "model.layers.85.self_attn.k_proj": {
1717
  "bits": 8,
@@ -1730,8 +1645,7 @@
1730
  "group_size": 128
1731
  },
1732
  "model.layers.86.mlp.gate": {
1733
- "bits": 8,
1734
- "group_size": 128
1735
  },
1736
  "model.layers.86.self_attn.k_proj": {
1737
  "bits": 8,
@@ -1750,8 +1664,7 @@
1750
  "group_size": 128
1751
  },
1752
  "model.layers.87.mlp.gate": {
1753
- "bits": 8,
1754
- "group_size": 128
1755
  },
1756
  "model.layers.87.self_attn.k_proj": {
1757
  "bits": 8,
@@ -1770,8 +1683,7 @@
1770
  "group_size": 128
1771
  },
1772
  "model.layers.88.mlp.gate": {
1773
- "bits": 8,
1774
- "group_size": 128
1775
  },
1776
  "model.layers.88.self_attn.k_proj": {
1777
  "bits": 8,
@@ -1790,8 +1702,7 @@
1790
  "group_size": 128
1791
  },
1792
  "model.layers.89.mlp.gate": {
1793
- "bits": 8,
1794
- "group_size": 128
1795
  },
1796
  "model.layers.89.self_attn.k_proj": {
1797
  "bits": 8,
@@ -1810,8 +1721,7 @@
1810
  "group_size": 128
1811
  },
1812
  "model.layers.9.mlp.gate": {
1813
- "bits": 8,
1814
- "group_size": 128
1815
  },
1816
  "model.layers.9.self_attn.k_proj": {
1817
  "bits": 8,
@@ -1830,8 +1740,7 @@
1830
  "group_size": 128
1831
  },
1832
  "model.layers.90.mlp.gate": {
1833
- "bits": 8,
1834
- "group_size": 128
1835
  },
1836
  "model.layers.90.self_attn.k_proj": {
1837
  "bits": 8,
@@ -1850,8 +1759,7 @@
1850
  "group_size": 128
1851
  },
1852
  "model.layers.91.mlp.gate": {
1853
- "bits": 8,
1854
- "group_size": 128
1855
  },
1856
  "model.layers.91.self_attn.k_proj": {
1857
  "bits": 8,
@@ -1870,8 +1778,7 @@
1870
  "group_size": 128
1871
  },
1872
  "model.layers.92.mlp.gate": {
1873
- "bits": 8,
1874
- "group_size": 128
1875
  },
1876
  "model.layers.92.self_attn.k_proj": {
1877
  "bits": 8,
@@ -1890,8 +1797,7 @@
1890
  "group_size": 128
1891
  },
1892
  "model.layers.93.mlp.gate": {
1893
- "bits": 8,
1894
- "group_size": 128
1895
  },
1896
  "model.layers.93.self_attn.k_proj": {
1897
  "bits": 8,
@@ -1908,6 +1814,382 @@
1908
  "model.layers.93.self_attn.v_proj": {
1909
  "bits": 8,
1910
  "group_size": 128
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1911
  }
1912
  },
1913
  "group_size": 64,
@@ -1927,4 +2209,4 @@
1927
  "use_cache": true,
1928
  "use_sliding_window": false,
1929
  "vocab_size": 151936
1930
- }
 
25
  "num_key_value_heads": 4,
26
  "output_router_logits": false,
27
  "quantization_config": {
28
+ "autoround_version": "0.6.1.dev",
29
  "bits": 4,
30
  "data_type": "int",
31
  "extra_config": {
32
  "model.layers.0.mlp.gate": {
33
+ "bits": 16
 
34
  },
35
  "model.layers.0.self_attn.k_proj": {
36
  "bits": 8,
 
49
  "group_size": 128
50
  },
51
  "model.layers.1.mlp.gate": {
52
+ "bits": 16
 
53
  },
54
  "model.layers.1.self_attn.k_proj": {
55
  "bits": 8,
 
68
  "group_size": 128
69
  },
70
  "model.layers.10.mlp.gate": {
71
+ "bits": 16
 
72
  },
73
  "model.layers.10.self_attn.k_proj": {
74
  "bits": 8,
 
87
  "group_size": 128
88
  },
89
  "model.layers.11.mlp.gate": {
90
+ "bits": 16
 
91
  },
92
  "model.layers.11.self_attn.k_proj": {
93
  "bits": 8,
 
106
  "group_size": 128
107
  },
108
  "model.layers.12.mlp.gate": {
109
+ "bits": 16
 
110
  },
111
  "model.layers.12.self_attn.k_proj": {
112
  "bits": 8,
 
125
  "group_size": 128
126
  },
127
  "model.layers.13.mlp.gate": {
128
+ "bits": 16
 
129
  },
130
  "model.layers.13.self_attn.k_proj": {
131
  "bits": 8,
 
144
  "group_size": 128
145
  },
146
  "model.layers.14.mlp.gate": {
147
+ "bits": 16
 
148
  },
149
  "model.layers.14.self_attn.k_proj": {
150
  "bits": 8,
 
163
  "group_size": 128
164
  },
165
  "model.layers.15.mlp.gate": {
166
+ "bits": 16
 
167
  },
168
  "model.layers.15.self_attn.k_proj": {
169
  "bits": 8,
 
182
  "group_size": 128
183
  },
184
  "model.layers.16.mlp.gate": {
185
+ "bits": 16
 
186
  },
187
  "model.layers.16.self_attn.k_proj": {
188
  "bits": 8,
 
201
  "group_size": 128
202
  },
203
  "model.layers.17.mlp.gate": {
204
+ "bits": 16
 
205
  },
206
  "model.layers.17.self_attn.k_proj": {
207
  "bits": 8,
 
220
  "group_size": 128
221
  },
222
  "model.layers.18.mlp.gate": {
223
+ "bits": 16
 
224
  },
225
  "model.layers.18.self_attn.k_proj": {
226
  "bits": 8,
 
239
  "group_size": 128
240
  },
241
  "model.layers.19.mlp.gate": {
242
+ "bits": 16
 
243
  },
244
  "model.layers.19.self_attn.k_proj": {
245
  "bits": 8,
 
258
  "group_size": 128
259
  },
260
  "model.layers.2.mlp.gate": {
261
+ "bits": 16
 
262
  },
263
  "model.layers.2.self_attn.k_proj": {
264
  "bits": 8,
 
277
  "group_size": 128
278
  },
279
  "model.layers.20.mlp.gate": {
280
+ "bits": 16
 
281
  },
282
  "model.layers.20.self_attn.k_proj": {
283
  "bits": 8,
 
296
  "group_size": 128
297
  },
298
  "model.layers.21.mlp.gate": {
299
+ "bits": 16
 
300
  },
301
  "model.layers.21.self_attn.k_proj": {
302
  "bits": 8,
 
315
  "group_size": 128
316
  },
317
  "model.layers.22.mlp.gate": {
318
+ "bits": 16
 
319
  },
320
  "model.layers.22.self_attn.k_proj": {
321
  "bits": 8,
 
334
  "group_size": 128
335
  },
336
  "model.layers.23.mlp.gate": {
337
+ "bits": 16
 
338
  },
339
  "model.layers.23.self_attn.k_proj": {
340
  "bits": 8,
 
353
  "group_size": 128
354
  },
355
  "model.layers.24.mlp.gate": {
356
+ "bits": 16
 
357
  },
358
  "model.layers.24.self_attn.k_proj": {
359
  "bits": 8,
 
372
  "group_size": 128
373
  },
374
  "model.layers.25.mlp.gate": {
375
+ "bits": 16
 
376
  },
377
  "model.layers.25.self_attn.k_proj": {
378
  "bits": 8,
 
391
  "group_size": 128
392
  },
393
  "model.layers.26.mlp.gate": {
394
+ "bits": 16
 
395
  },
396
  "model.layers.26.self_attn.k_proj": {
397
  "bits": 8,
 
410
  "group_size": 128
411
  },
412
  "model.layers.27.mlp.gate": {
413
+ "bits": 16
 
414
  },
415
  "model.layers.27.self_attn.k_proj": {
416
  "bits": 8,
 
429
  "group_size": 128
430
  },
431
  "model.layers.28.mlp.gate": {
432
+ "bits": 16
 
433
  },
434
  "model.layers.28.self_attn.k_proj": {
435
  "bits": 8,
 
448
  "group_size": 128
449
  },
450
  "model.layers.29.mlp.gate": {
451
+ "bits": 16
 
452
  },
453
  "model.layers.29.self_attn.k_proj": {
454
  "bits": 8,
 
467
  "group_size": 128
468
  },
469
  "model.layers.3.mlp.gate": {
470
+ "bits": 16
 
471
  },
472
  "model.layers.3.self_attn.k_proj": {
473
  "bits": 8,
 
486
  "group_size": 128
487
  },
488
  "model.layers.30.mlp.gate": {
489
+ "bits": 16
 
490
  },
491
  "model.layers.30.self_attn.k_proj": {
492
  "bits": 8,
 
505
  "group_size": 128
506
  },
507
  "model.layers.31.mlp.gate": {
508
+ "bits": 16
 
509
  },
510
  "model.layers.31.self_attn.k_proj": {
511
  "bits": 8,
 
524
  "group_size": 128
525
  },
526
  "model.layers.32.mlp.gate": {
527
+ "bits": 16
 
528
  },
529
  "model.layers.32.self_attn.k_proj": {
530
  "bits": 8,
 
543
  "group_size": 128
544
  },
545
  "model.layers.33.mlp.gate": {
546
+ "bits": 16
 
547
  },
548
  "model.layers.33.self_attn.k_proj": {
549
  "bits": 8,
 
562
  "group_size": 128
563
  },
564
  "model.layers.34.mlp.gate": {
565
+ "bits": 16
 
566
  },
567
  "model.layers.34.self_attn.k_proj": {
568
  "bits": 8,
 
581
  "group_size": 128
582
  },
583
  "model.layers.35.mlp.gate": {
584
+ "bits": 16
 
585
  },
586
  "model.layers.35.self_attn.k_proj": {
587
  "bits": 8,
 
600
  "group_size": 128
601
  },
602
  "model.layers.36.mlp.gate": {
603
+ "bits": 16
 
604
  },
605
  "model.layers.36.self_attn.k_proj": {
606
  "bits": 8,
 
619
  "group_size": 128
620
  },
621
  "model.layers.37.mlp.gate": {
622
+ "bits": 16
 
623
  },
624
  "model.layers.37.self_attn.k_proj": {
625
  "bits": 8,
 
638
  "group_size": 128
639
  },
640
  "model.layers.38.mlp.gate": {
641
+ "bits": 16
 
642
  },
643
  "model.layers.38.self_attn.k_proj": {
644
  "bits": 8,
 
657
  "group_size": 128
658
  },
659
  "model.layers.39.mlp.gate": {
660
+ "bits": 16
 
661
  },
662
  "model.layers.39.self_attn.k_proj": {
663
  "bits": 8,
 
676
  "group_size": 128
677
  },
678
  "model.layers.4.mlp.gate": {
679
+ "bits": 16
 
680
  },
681
  "model.layers.4.self_attn.k_proj": {
682
  "bits": 8,
 
695
  "group_size": 128
696
  },
697
  "model.layers.40.mlp.gate": {
698
+ "bits": 16
 
699
  },
700
  "model.layers.40.self_attn.k_proj": {
701
  "bits": 8,
 
714
  "group_size": 128
715
  },
716
  "model.layers.41.mlp.gate": {
717
+ "bits": 16
 
718
  },
719
  "model.layers.41.self_attn.k_proj": {
720
  "bits": 8,
 
733
  "group_size": 128
734
  },
735
  "model.layers.42.mlp.gate": {
736
+ "bits": 16
 
737
  },
738
  "model.layers.42.self_attn.k_proj": {
739
  "bits": 8,
 
752
  "group_size": 128
753
  },
754
  "model.layers.43.mlp.gate": {
755
+ "bits": 16
 
756
  },
757
  "model.layers.43.self_attn.k_proj": {
758
  "bits": 8,
 
771
  "group_size": 128
772
  },
773
  "model.layers.44.mlp.gate": {
774
+ "bits": 16
 
775
  },
776
  "model.layers.44.self_attn.k_proj": {
777
  "bits": 8,
 
790
  "group_size": 128
791
  },
792
  "model.layers.45.mlp.gate": {
793
+ "bits": 16
 
794
  },
795
  "model.layers.45.self_attn.k_proj": {
796
  "bits": 8,
 
809
  "group_size": 128
810
  },
811
  "model.layers.46.mlp.gate": {
812
+ "bits": 16
 
813
  },
814
  "model.layers.46.self_attn.k_proj": {
815
  "bits": 8,
 
828
  "group_size": 128
829
  },
830
  "model.layers.47.mlp.gate": {
831
+ "bits": 16
 
832
  },
833
  "model.layers.47.self_attn.k_proj": {
834
  "bits": 8,
 
847
  "group_size": 128
848
  },
849
  "model.layers.48.mlp.gate": {
850
+ "bits": 16
 
851
  },
852
  "model.layers.48.self_attn.k_proj": {
853
  "bits": 8,
 
866
  "group_size": 128
867
  },
868
  "model.layers.49.mlp.gate": {
869
+ "bits": 16
 
870
  },
871
  "model.layers.49.self_attn.k_proj": {
872
  "bits": 8,
 
885
  "group_size": 128
886
  },
887
  "model.layers.5.mlp.gate": {
888
+ "bits": 16
 
889
  },
890
  "model.layers.5.self_attn.k_proj": {
891
  "bits": 8,
 
904
  "group_size": 128
905
  },
906
  "model.layers.50.mlp.gate": {
907
+ "bits": 16
 
908
  },
909
  "model.layers.50.self_attn.k_proj": {
910
  "bits": 8,
 
923
  "group_size": 128
924
  },
925
  "model.layers.51.mlp.gate": {
926
+ "bits": 16
 
927
  },
928
  "model.layers.51.self_attn.k_proj": {
929
  "bits": 8,
 
942
  "group_size": 128
943
  },
944
  "model.layers.52.mlp.gate": {
945
+ "bits": 16
 
946
  },
947
  "model.layers.52.self_attn.k_proj": {
948
  "bits": 8,
 
961
  "group_size": 128
962
  },
963
  "model.layers.53.mlp.gate": {
964
+ "bits": 16
 
965
  },
966
  "model.layers.53.self_attn.k_proj": {
967
  "bits": 8,
 
980
  "group_size": 128
981
  },
982
  "model.layers.54.mlp.gate": {
983
+ "bits": 16
 
984
  },
985
  "model.layers.54.self_attn.k_proj": {
986
  "bits": 8,
 
999
  "group_size": 128
1000
  },
1001
  "model.layers.55.mlp.gate": {
1002
+ "bits": 16
 
1003
  },
1004
  "model.layers.55.self_attn.k_proj": {
1005
  "bits": 8,
 
1018
  "group_size": 128
1019
  },
1020
  "model.layers.56.mlp.gate": {
1021
+ "bits": 16
 
1022
  },
1023
  "model.layers.56.self_attn.k_proj": {
1024
  "bits": 8,
 
1037
  "group_size": 128
1038
  },
1039
  "model.layers.57.mlp.gate": {
1040
+ "bits": 16
 
1041
  },
1042
  "model.layers.57.self_attn.k_proj": {
1043
  "bits": 8,
 
1056
  "group_size": 128
1057
  },
1058
  "model.layers.58.mlp.gate": {
1059
+ "bits": 16
 
1060
  },
1061
  "model.layers.58.self_attn.k_proj": {
1062
  "bits": 8,
 
1075
  "group_size": 128
1076
  },
1077
  "model.layers.59.mlp.gate": {
1078
+ "bits": 16
 
1079
  },
1080
  "model.layers.59.self_attn.k_proj": {
1081
  "bits": 8,
 
1094
  "group_size": 128
1095
  },
1096
  "model.layers.6.mlp.gate": {
1097
+ "bits": 16
 
1098
  },
1099
  "model.layers.6.self_attn.k_proj": {
1100
  "bits": 8,
 
1113
  "group_size": 128
1114
  },
1115
  "model.layers.60.mlp.gate": {
1116
+ "bits": 16
 
1117
  },
1118
  "model.layers.60.self_attn.k_proj": {
1119
  "bits": 8,
 
1132
  "group_size": 128
1133
  },
1134
  "model.layers.61.mlp.gate": {
1135
+ "bits": 16
 
1136
  },
1137
  "model.layers.61.self_attn.k_proj": {
1138
  "bits": 8,
 
1151
  "group_size": 128
1152
  },
1153
  "model.layers.62.mlp.gate": {
1154
+ "bits": 16
 
1155
  },
1156
  "model.layers.62.self_attn.k_proj": {
1157
  "bits": 8,
 
1170
  "group_size": 128
1171
  },
1172
  "model.layers.63.mlp.gate": {
1173
+ "bits": 16
 
1174
  },
1175
  "model.layers.63.self_attn.k_proj": {
1176
  "bits": 8,
 
1189
  "group_size": 128
1190
  },
1191
  "model.layers.64.mlp.gate": {
1192
+ "bits": 16
 
1193
  },
1194
  "model.layers.64.self_attn.k_proj": {
1195
  "bits": 8,
 
1208
  "group_size": 128
1209
  },
1210
  "model.layers.65.mlp.gate": {
1211
+ "bits": 16
 
1212
  },
1213
  "model.layers.65.self_attn.k_proj": {
1214
  "bits": 8,
 
1227
  "group_size": 128
1228
  },
1229
  "model.layers.66.mlp.gate": {
1230
+ "bits": 16
 
1231
  },
1232
  "model.layers.66.self_attn.k_proj": {
1233
  "bits": 8,
 
1246
  "group_size": 128
1247
  },
1248
  "model.layers.67.mlp.gate": {
1249
+ "bits": 16
 
1250
  },
1251
  "model.layers.67.self_attn.k_proj": {
1252
  "bits": 8,
 
1265
  "group_size": 128
1266
  },
1267
  "model.layers.68.mlp.gate": {
1268
+ "bits": 16
 
1269
  },
1270
  "model.layers.68.self_attn.k_proj": {
1271
  "bits": 8,
 
1284
  "group_size": 128
1285
  },
1286
  "model.layers.69.mlp.gate": {
1287
+ "bits": 16
 
1288
  },
1289
  "model.layers.69.self_attn.k_proj": {
1290
  "bits": 8,
 
1303
  "group_size": 128
1304
  },
1305
  "model.layers.7.mlp.gate": {
1306
+ "bits": 16
 
1307
  },
1308
  "model.layers.7.self_attn.k_proj": {
1309
  "bits": 8,
 
1322
  "group_size": 128
1323
  },
1324
  "model.layers.70.mlp.gate": {
1325
+ "bits": 16
 
1326
  },
1327
  "model.layers.70.self_attn.k_proj": {
1328
  "bits": 8,
 
1341
  "group_size": 128
1342
  },
1343
  "model.layers.71.mlp.gate": {
1344
+ "bits": 16
 
1345
  },
1346
  "model.layers.71.self_attn.k_proj": {
1347
  "bits": 8,
 
1360
  "group_size": 128
1361
  },
1362
  "model.layers.72.mlp.gate": {
1363
+ "bits": 16
 
1364
  },
1365
  "model.layers.72.self_attn.k_proj": {
1366
  "bits": 8,
 
1379
  "group_size": 128
1380
  },
1381
  "model.layers.73.mlp.gate": {
1382
+ "bits": 16
 
1383
  },
1384
  "model.layers.73.self_attn.k_proj": {
1385
  "bits": 8,
 
1398
  "group_size": 128
1399
  },
1400
  "model.layers.74.mlp.gate": {
1401
+ "bits": 16
 
1402
  },
1403
  "model.layers.74.self_attn.k_proj": {
1404
  "bits": 8,
 
1417
  "group_size": 128
1418
  },
1419
  "model.layers.75.mlp.gate": {
1420
+ "bits": 16
 
1421
  },
1422
  "model.layers.75.self_attn.k_proj": {
1423
  "bits": 8,
 
1436
  "group_size": 128
1437
  },
1438
  "model.layers.76.mlp.gate": {
1439
+ "bits": 16
 
1440
  },
1441
  "model.layers.76.self_attn.k_proj": {
1442
  "bits": 8,
 
1455
  "group_size": 128
1456
  },
1457
  "model.layers.77.mlp.gate": {
1458
+ "bits": 16
 
1459
  },
1460
  "model.layers.77.self_attn.k_proj": {
1461
  "bits": 8,
 
1474
  "group_size": 128
1475
  },
1476
  "model.layers.78.mlp.gate": {
1477
+ "bits": 16
 
1478
  },
1479
  "model.layers.78.self_attn.k_proj": {
1480
  "bits": 8,
 
1493
  "group_size": 128
1494
  },
1495
  "model.layers.79.mlp.gate": {
1496
+ "bits": 16
 
1497
  },
1498
  "model.layers.79.self_attn.k_proj": {
1499
  "bits": 8,
 
1512
  "group_size": 128
1513
  },
1514
  "model.layers.8.mlp.gate": {
1515
+ "bits": 16
 
1516
  },
1517
  "model.layers.8.self_attn.k_proj": {
1518
  "bits": 8,
 
1531
  "group_size": 128
1532
  },
1533
  "model.layers.80.mlp.gate": {
1534
+ "bits": 16
 
1535
  },
1536
  "model.layers.80.self_attn.k_proj": {
1537
  "bits": 8,
 
1550
  "group_size": 128
1551
  },
1552
  "model.layers.81.mlp.gate": {
1553
+ "bits": 16
 
1554
  },
1555
  "model.layers.81.self_attn.k_proj": {
1556
  "bits": 8,
 
1569
  "group_size": 128
1570
  },
1571
  "model.layers.82.mlp.gate": {
1572
+ "bits": 16
 
1573
  },
1574
  "model.layers.82.self_attn.k_proj": {
1575
  "bits": 8,
 
1588
  "group_size": 128
1589
  },
1590
  "model.layers.83.mlp.gate": {
1591
+ "bits": 16
 
1592
  },
1593
  "model.layers.83.self_attn.k_proj": {
1594
  "bits": 8,
 
1607
  "group_size": 128
1608
  },
1609
  "model.layers.84.mlp.gate": {
1610
+ "bits": 16
 
1611
  },
1612
  "model.layers.84.self_attn.k_proj": {
1613
  "bits": 8,
 
1626
  "group_size": 128
1627
  },
1628
  "model.layers.85.mlp.gate": {
1629
+ "bits": 16
 
1630
  },
1631
  "model.layers.85.self_attn.k_proj": {
1632
  "bits": 8,
 
1645
  "group_size": 128
1646
  },
1647
  "model.layers.86.mlp.gate": {
1648
+ "bits": 16
 
1649
  },
1650
  "model.layers.86.self_attn.k_proj": {
1651
  "bits": 8,
 
1664
  "group_size": 128
1665
  },
1666
  "model.layers.87.mlp.gate": {
1667
+ "bits": 16
 
1668
  },
1669
  "model.layers.87.self_attn.k_proj": {
1670
  "bits": 8,
 
1683
  "group_size": 128
1684
  },
1685
  "model.layers.88.mlp.gate": {
1686
+ "bits": 16
 
1687
  },
1688
  "model.layers.88.self_attn.k_proj": {
1689
  "bits": 8,
 
1702
  "group_size": 128
1703
  },
1704
  "model.layers.89.mlp.gate": {
1705
+ "bits": 16
 
1706
  },
1707
  "model.layers.89.self_attn.k_proj": {
1708
  "bits": 8,
 
1721
  "group_size": 128
1722
  },
1723
  "model.layers.9.mlp.gate": {
1724
+ "bits": 16
 
1725
  },
1726
  "model.layers.9.self_attn.k_proj": {
1727
  "bits": 8,
 
1740
  "group_size": 128
1741
  },
1742
  "model.layers.90.mlp.gate": {
1743
+ "bits": 16
 
1744
  },
1745
  "model.layers.90.self_attn.k_proj": {
1746
  "bits": 8,
 
1759
  "group_size": 128
1760
  },
1761
  "model.layers.91.mlp.gate": {
1762
+ "bits": 16
 
1763
  },
1764
  "model.layers.91.self_attn.k_proj": {
1765
  "bits": 8,
 
1778
  "group_size": 128
1779
  },
1780
  "model.layers.92.mlp.gate": {
1781
+ "bits": 16
 
1782
  },
1783
  "model.layers.92.self_attn.k_proj": {
1784
  "bits": 8,
 
1797
  "group_size": 128
1798
  },
1799
  "model.layers.93.mlp.gate": {
1800
+ "bits": 16
 
1801
  },
1802
  "model.layers.93.self_attn.k_proj": {
1803
  "bits": 8,
 
1814
  "model.layers.93.self_attn.v_proj": {
1815
  "bits": 8,
1816
  "group_size": 128
1817
+ },
1818
+ "model.layers.0.self_attn.qkv_proj": {
1819
+ "bits": 8,
1820
+ "group_size": 128
1821
+ },
1822
+ "model.layers.1.self_attn.qkv_proj": {
1823
+ "bits": 8,
1824
+ "group_size": 128
1825
+ },
1826
+ "model.layers.2.self_attn.qkv_proj": {
1827
+ "bits": 8,
1828
+ "group_size": 128
1829
+ },
1830
+ "model.layers.3.self_attn.qkv_proj": {
1831
+ "bits": 8,
1832
+ "group_size": 128
1833
+ },
1834
+ "model.layers.4.self_attn.qkv_proj": {
1835
+ "bits": 8,
1836
+ "group_size": 128
1837
+ },
1838
+ "model.layers.5.self_attn.qkv_proj": {
1839
+ "bits": 8,
1840
+ "group_size": 128
1841
+ },
1842
+ "model.layers.6.self_attn.qkv_proj": {
1843
+ "bits": 8,
1844
+ "group_size": 128
1845
+ },
1846
+ "model.layers.7.self_attn.qkv_proj": {
1847
+ "bits": 8,
1848
+ "group_size": 128
1849
+ },
1850
+ "model.layers.8.self_attn.qkv_proj": {
1851
+ "bits": 8,
1852
+ "group_size": 128
1853
+ },
1854
+ "model.layers.9.self_attn.qkv_proj": {
1855
+ "bits": 8,
1856
+ "group_size": 128
1857
+ },
1858
+ "model.layers.10.self_attn.qkv_proj": {
1859
+ "bits": 8,
1860
+ "group_size": 128
1861
+ },
1862
+ "model.layers.11.self_attn.qkv_proj": {
1863
+ "bits": 8,
1864
+ "group_size": 128
1865
+ },
1866
+ "model.layers.12.self_attn.qkv_proj": {
1867
+ "bits": 8,
1868
+ "group_size": 128
1869
+ },
1870
+ "model.layers.13.self_attn.qkv_proj": {
1871
+ "bits": 8,
1872
+ "group_size": 128
1873
+ },
1874
+ "model.layers.14.self_attn.qkv_proj": {
1875
+ "bits": 8,
1876
+ "group_size": 128
1877
+ },
1878
+ "model.layers.15.self_attn.qkv_proj": {
1879
+ "bits": 8,
1880
+ "group_size": 128
1881
+ },
1882
+ "model.layers.16.self_attn.qkv_proj": {
1883
+ "bits": 8,
1884
+ "group_size": 128
1885
+ },
1886
+ "model.layers.17.self_attn.qkv_proj": {
1887
+ "bits": 8,
1888
+ "group_size": 128
1889
+ },
1890
+ "model.layers.18.self_attn.qkv_proj": {
1891
+ "bits": 8,
1892
+ "group_size": 128
1893
+ },
1894
+ "model.layers.19.self_attn.qkv_proj": {
1895
+ "bits": 8,
1896
+ "group_size": 128
1897
+ },
1898
+ "model.layers.20.self_attn.qkv_proj": {
1899
+ "bits": 8,
1900
+ "group_size": 128
1901
+ },
1902
+ "model.layers.21.self_attn.qkv_proj": {
1903
+ "bits": 8,
1904
+ "group_size": 128
1905
+ },
1906
+ "model.layers.22.self_attn.qkv_proj": {
1907
+ "bits": 8,
1908
+ "group_size": 128
1909
+ },
1910
+ "model.layers.23.self_attn.qkv_proj": {
1911
+ "bits": 8,
1912
+ "group_size": 128
1913
+ },
1914
+ "model.layers.24.self_attn.qkv_proj": {
1915
+ "bits": 8,
1916
+ "group_size": 128
1917
+ },
1918
+ "model.layers.25.self_attn.qkv_proj": {
1919
+ "bits": 8,
1920
+ "group_size": 128
1921
+ },
1922
+ "model.layers.26.self_attn.qkv_proj": {
1923
+ "bits": 8,
1924
+ "group_size": 128
1925
+ },
1926
+ "model.layers.27.self_attn.qkv_proj": {
1927
+ "bits": 8,
1928
+ "group_size": 128
1929
+ },
1930
+ "model.layers.28.self_attn.qkv_proj": {
1931
+ "bits": 8,
1932
+ "group_size": 128
1933
+ },
1934
+ "model.layers.29.self_attn.qkv_proj": {
1935
+ "bits": 8,
1936
+ "group_size": 128
1937
+ },
1938
+ "model.layers.30.self_attn.qkv_proj": {
1939
+ "bits": 8,
1940
+ "group_size": 128
1941
+ },
1942
+ "model.layers.31.self_attn.qkv_proj": {
1943
+ "bits": 8,
1944
+ "group_size": 128
1945
+ },
1946
+ "model.layers.32.self_attn.qkv_proj": {
1947
+ "bits": 8,
1948
+ "group_size": 128
1949
+ },
1950
+ "model.layers.33.self_attn.qkv_proj": {
1951
+ "bits": 8,
1952
+ "group_size": 128
1953
+ },
1954
+ "model.layers.34.self_attn.qkv_proj": {
1955
+ "bits": 8,
1956
+ "group_size": 128
1957
+ },
1958
+ "model.layers.35.self_attn.qkv_proj": {
1959
+ "bits": 8,
1960
+ "group_size": 128
1961
+ },
1962
+ "model.layers.36.self_attn.qkv_proj": {
1963
+ "bits": 8,
1964
+ "group_size": 128
1965
+ },
1966
+ "model.layers.37.self_attn.qkv_proj": {
1967
+ "bits": 8,
1968
+ "group_size": 128
1969
+ },
1970
+ "model.layers.38.self_attn.qkv_proj": {
1971
+ "bits": 8,
1972
+ "group_size": 128
1973
+ },
1974
+ "model.layers.39.self_attn.qkv_proj": {
1975
+ "bits": 8,
1976
+ "group_size": 128
1977
+ },
1978
+ "model.layers.40.self_attn.qkv_proj": {
1979
+ "bits": 8,
1980
+ "group_size": 128
1981
+ },
1982
+ "model.layers.41.self_attn.qkv_proj": {
1983
+ "bits": 8,
1984
+ "group_size": 128
1985
+ },
1986
+ "model.layers.42.self_attn.qkv_proj": {
1987
+ "bits": 8,
1988
+ "group_size": 128
1989
+ },
1990
+ "model.layers.43.self_attn.qkv_proj": {
1991
+ "bits": 8,
1992
+ "group_size": 128
1993
+ },
1994
+ "model.layers.44.self_attn.qkv_proj": {
1995
+ "bits": 8,
1996
+ "group_size": 128
1997
+ },
1998
+ "model.layers.45.self_attn.qkv_proj": {
1999
+ "bits": 8,
2000
+ "group_size": 128
2001
+ },
2002
+ "model.layers.46.self_attn.qkv_proj": {
2003
+ "bits": 8,
2004
+ "group_size": 128
2005
+ },
2006
+ "model.layers.47.self_attn.qkv_proj": {
2007
+ "bits": 8,
2008
+ "group_size": 128
2009
+ },
2010
+ "model.layers.48.self_attn.qkv_proj": {
2011
+ "bits": 8,
2012
+ "group_size": 128
2013
+ },
2014
+ "model.layers.49.self_attn.qkv_proj": {
2015
+ "bits": 8,
2016
+ "group_size": 128
2017
+ },
2018
+ "model.layers.50.self_attn.qkv_proj": {
2019
+ "bits": 8,
2020
+ "group_size": 128
2021
+ },
2022
+ "model.layers.51.self_attn.qkv_proj": {
2023
+ "bits": 8,
2024
+ "group_size": 128
2025
+ },
2026
+ "model.layers.52.self_attn.qkv_proj": {
2027
+ "bits": 8,
2028
+ "group_size": 128
2029
+ },
2030
+ "model.layers.53.self_attn.qkv_proj": {
2031
+ "bits": 8,
2032
+ "group_size": 128
2033
+ },
2034
+ "model.layers.54.self_attn.qkv_proj": {
2035
+ "bits": 8,
2036
+ "group_size": 128
2037
+ },
2038
+ "model.layers.55.self_attn.qkv_proj": {
2039
+ "bits": 8,
2040
+ "group_size": 128
2041
+ },
2042
+ "model.layers.56.self_attn.qkv_proj": {
2043
+ "bits": 8,
2044
+ "group_size": 128
2045
+ },
2046
+ "model.layers.57.self_attn.qkv_proj": {
2047
+ "bits": 8,
2048
+ "group_size": 128
2049
+ },
2050
+ "model.layers.58.self_attn.qkv_proj": {
2051
+ "bits": 8,
2052
+ "group_size": 128
2053
+ },
2054
+ "model.layers.59.self_attn.qkv_proj": {
2055
+ "bits": 8,
2056
+ "group_size": 128
2057
+ },
2058
+ "model.layers.60.self_attn.qkv_proj": {
2059
+ "bits": 8,
2060
+ "group_size": 128
2061
+ },
2062
+ "model.layers.61.self_attn.qkv_proj": {
2063
+ "bits": 8,
2064
+ "group_size": 128
2065
+ },
2066
+ "model.layers.62.self_attn.qkv_proj": {
2067
+ "bits": 8,
2068
+ "group_size": 128
2069
+ },
2070
+ "model.layers.63.self_attn.qkv_proj": {
2071
+ "bits": 8,
2072
+ "group_size": 128
2073
+ },
2074
+ "model.layers.64.self_attn.qkv_proj": {
2075
+ "bits": 8,
2076
+ "group_size": 128
2077
+ },
2078
+ "model.layers.65.self_attn.qkv_proj": {
2079
+ "bits": 8,
2080
+ "group_size": 128
2081
+ },
2082
+ "model.layers.66.self_attn.qkv_proj": {
2083
+ "bits": 8,
2084
+ "group_size": 128
2085
+ },
2086
+ "model.layers.67.self_attn.qkv_proj": {
2087
+ "bits": 8,
2088
+ "group_size": 128
2089
+ },
2090
+ "model.layers.68.self_attn.qkv_proj": {
2091
+ "bits": 8,
2092
+ "group_size": 128
2093
+ },
2094
+ "model.layers.69.self_attn.qkv_proj": {
2095
+ "bits": 8,
2096
+ "group_size": 128
2097
+ },
2098
+ "model.layers.70.self_attn.qkv_proj": {
2099
+ "bits": 8,
2100
+ "group_size": 128
2101
+ },
2102
+ "model.layers.71.self_attn.qkv_proj": {
2103
+ "bits": 8,
2104
+ "group_size": 128
2105
+ },
2106
+ "model.layers.72.self_attn.qkv_proj": {
2107
+ "bits": 8,
2108
+ "group_size": 128
2109
+ },
2110
+ "model.layers.73.self_attn.qkv_proj": {
2111
+ "bits": 8,
2112
+ "group_size": 128
2113
+ },
2114
+ "model.layers.74.self_attn.qkv_proj": {
2115
+ "bits": 8,
2116
+ "group_size": 128
2117
+ },
2118
+ "model.layers.75.self_attn.qkv_proj": {
2119
+ "bits": 8,
2120
+ "group_size": 128
2121
+ },
2122
+ "model.layers.76.self_attn.qkv_proj": {
2123
+ "bits": 8,
2124
+ "group_size": 128
2125
+ },
2126
+ "model.layers.77.self_attn.qkv_proj": {
2127
+ "bits": 8,
2128
+ "group_size": 128
2129
+ },
2130
+ "model.layers.78.self_attn.qkv_proj": {
2131
+ "bits": 8,
2132
+ "group_size": 128
2133
+ },
2134
+ "model.layers.79.self_attn.qkv_proj": {
2135
+ "bits": 8,
2136
+ "group_size": 128
2137
+ },
2138
+ "model.layers.80.self_attn.qkv_proj": {
2139
+ "bits": 8,
2140
+ "group_size": 128
2141
+ },
2142
+ "model.layers.81.self_attn.qkv_proj": {
2143
+ "bits": 8,
2144
+ "group_size": 128
2145
+ },
2146
+ "model.layers.82.self_attn.qkv_proj": {
2147
+ "bits": 8,
2148
+ "group_size": 128
2149
+ },
2150
+ "model.layers.83.self_attn.qkv_proj": {
2151
+ "bits": 8,
2152
+ "group_size": 128
2153
+ },
2154
+ "model.layers.84.self_attn.qkv_proj": {
2155
+ "bits": 8,
2156
+ "group_size": 128
2157
+ },
2158
+ "model.layers.85.self_attn.qkv_proj": {
2159
+ "bits": 8,
2160
+ "group_size": 128
2161
+ },
2162
+ "model.layers.86.self_attn.qkv_proj": {
2163
+ "bits": 8,
2164
+ "group_size": 128
2165
+ },
2166
+ "model.layers.87.self_attn.qkv_proj": {
2167
+ "bits": 8,
2168
+ "group_size": 128
2169
+ },
2170
+ "model.layers.88.self_attn.qkv_proj": {
2171
+ "bits": 8,
2172
+ "group_size": 128
2173
+ },
2174
+ "model.layers.89.self_attn.qkv_proj": {
2175
+ "bits": 8,
2176
+ "group_size": 128
2177
+ },
2178
+ "model.layers.90.self_attn.qkv_proj": {
2179
+ "bits": 8,
2180
+ "group_size": 128
2181
+ },
2182
+ "model.layers.91.self_attn.qkv_proj": {
2183
+ "bits": 8,
2184
+ "group_size": 128
2185
+ },
2186
+ "model.layers.92.self_attn.qkv_proj": {
2187
+ "bits": 8,
2188
+ "group_size": 128
2189
+ },
2190
+ "model.layers.93.self_attn.qkv_proj": {
2191
+ "bits": 8,
2192
+ "group_size": 128
2193
  }
2194
  },
2195
  "group_size": 64,
 
2209
  "use_cache": true,
2210
  "use_sliding_window": false,
2211
  "vocab_size": 151936
2212
+ }
model-00001-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2db4454021cbb415af060909c07d7e867237e798074255d5a00f28935e44f4a6
3
- size 4999542768
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:709da264a348a57aed3d46e32b8b82f59936729748ea12d67f8640cd82bcc858
3
+ size 4997686288
model-00002-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8d9abe67c97534e243edf48f7afe97ae866fb2a7fdc43d47ea1abd3a642ef421
3
- size 4998619208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f6cc4f5228b2e72cd77956322d78f2e4682d52e999a380cf77c78674725f032
3
+ size 5000469648
model-00003-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:de6f6dcb5b823e0d03a31132f7b7f6aa4f72d1435c6ecb59a39a35f532846875
3
- size 4999710112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b800b474eabd6517790d29c25a770e04ed0ef19951314cfee4012acb949e0e3
3
+ size 4998050352
model-00004-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:69db0731313b75802f8aa8ed4c3c3463ceb86fe321125db8083e748bd3af4894
3
- size 4998623192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7625687947be1c92f45f5759aea6e6c8daa2f2c003b8b599c848a44bf6ae3ff0
3
+ size 5000473624
model-00005-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bda66646cd162bb0c1eaa20a15b97f0f7be3015089da3c21145f626bc179438
3
- size 4998623456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f67b73e401c2306ae6c69fb3a3567e0f0d466d13233faf3cc5b02c28a711e76f
3
+ size 4997475480
model-00006-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6606e710e95b0e2bbeeb4754b1b78df51331189c84483b0ac7a21889b9b8e207
3
- size 4999714272
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31b6e169a2927f79284433c0d09c3fae2dc70c3c82bc46f5dc85ea0bc85b3aba
3
+ size 4997857776
model-00007-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7d545a9d33959ac17bb553bb890df43985e41932206680ca1770a7b176bcc19c
3
- size 4998623328
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1854cf497f8a84e2ae95d1412a22f2004b351c71b753b2cac88dd2273b451530
3
+ size 5000473760
model-00008-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:625f92aeb6b6126cfe7e6ef5b6219886cf25f79b1319d2e739f2ffaa792202e3
3
- size 4998623456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:670c80d0fb33f18555477b711f8d2ce5f33c77b74109636c7f3b1f5823516285
3
+ size 4997475480
model-00009-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a72a2bf31b6c03c3e5e2a2328aae8590b83c2681512c0ffeae3ce798babe8b3
3
- size 4999714144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33fe8f8ac4cfe9a986fd6952543c5c40429a7f626d88202515bf88359153349f
3
+ size 4997857664
model-00010-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64e95499bf82f134f87e2c05ae7380c18150c71f4697baf942dab8c1203d0184
3
- size 4998623456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0676ba2db25493eb09b04cb26e29b7b64de60b5ea1ecfc2eac7758602ad28b09
3
+ size 5000473880
model-00011-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:46bc13560510498aa0fc9bd83f9815256f4622370531cc76d9a87cf3d40327a0
3
- size 4998623528
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df7f3210476cf18ff837c6a0d488f134925f41714f79c36468efb175421734c8
3
+ size 4979044216
model-00012-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f5ba0da091b5866788b8a350f0a90eb84c3e24ea24459412cef59ab4294392a
3
- size 4999714072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e851692e845c7ba89b897c4f80f29ffe438600ce32e9363b2290aa0e17167ae9
3
+ size 4999329640
model-00013-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ccf5369e611cc507eb538a183727a62643a3cbf829ce0548d5b818a7ce6b32f9
3
- size 4998623448
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:964971f1c221f88a72a4486b959e460ccc56a90dcd98aafc6e6e98336aec703f
3
+ size 5000473880
model-00014-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f7509d5f435abaa2d31c4a6e796135c4aa30bbdd430688cb1110f95e15fe51d
3
- size 4999714352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d0e15cfaf2acf5408c15228b28572bf7ef609f54d07578b2defc4dcd0fe0696
3
+ size 4998054544
model-00015-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9efe25b316fb6b3d699245e3d5a5c5027c8c136a123c3138e4d499e91cd7b376
3
- size 4998623248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6b69923387f0fbf303b742683fc93f907c9c36467120c5dbac84308ffe2c54a
3
+ size 5000473720
model-00016-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9345bded90fbe487712c7aa4500f7da635c850d392c02e0eb5d4146ba3c02964
3
- size 4998623456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:402be03a708628f4858dbcf72aca8129b758a257a066e3c23d12086dc633767a
3
+ size 4997475480
model-00017-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:459b3ebefe1c381215f757889bdc1b23cb5fa59bd51c690188fe3df3243a437a
3
- size 4999714224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a36e86af947290ff8731709b57b77cc7243d1de3950d0a6b57a06333567bbe35
3
+ size 4997857696
model-00018-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84341a7c112b0016989ba68a21b993027f0f222ce06b3a5b59a33af2fd7bbb55
3
- size 4998623376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff72b29a06af55b80df3fec195527f5768be66062a5159a014e1d608ef05c566
3
+ size 5000473840
model-00019-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6aa129e9a82157d02d9ef3a0884e17d81a3c9b120a477d545d112b0b64201dae
3
- size 4998623456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b58b3cfdcd64748de5305eff0088a4361fc179c1bba9351d2e5a22417b4db060
3
+ size 4997475536
model-00020-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06f3481304c2b888f77354970248badb5af6998b21669801992117e26660de16
3
- size 4999714144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd1252267a4b0cb460e249218b7a7f2fb92f6dc23117cc0832e8d64074c53c00
3
+ size 4997857608
model-00021-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cef8fe568b1a8dff29723b48649fa8d9f85e9bd459c34c011d47b7c3ba2481cf
3
- size 4998623456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af9ca368691115d938cb090812bd1e4bf99b2be9288d9e56a7ee89b21e2d94de
3
+ size 5000473880
model-00022-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5bacc4e28d075a2b26da4045afc3dfc3ce7c5bf4d43c5a8795147dc4d22d243f
3
- size 4969204312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6692a6a622d3ff5cdf2ab6306d6b902b05738693d36ecc6395c4eb00746fba7
3
+ size 4998054608
model-00023-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:382ee5ce9ef03ece231055f63464a43c13aa8495d3901f52d7b75703af582d8c
3
- size 4998606584
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdc587274e59ed106b411be6d8a32c2bda0ed0f688a9995b4e30592182e48175
3
+ size 5000473656
model-00024-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac04d2d937796544cade0381c2ef8869ff6335e50bec0d12aaa18489c7072b44
3
- size 4998623448
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40204216497cfc2ad0b106281e942a5394cd8479a95b3d875469beaf4402b873
3
+ size 4997475480
model-00025-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:09f16a68a303a1aadc7efc9450c9f13e11bb5d2f792aae1cac73e68288f61099
3
- size 4999714280
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1f38345ad63ad35ec89aaa0892ca70f419fb33691466929632c80fae45bc3a6
3
+ size 4997857752
model-00026-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ff89e38ff71fd4db6c6e5b081ad1dd0eaa98b3f76913f4961879de43f68ddcc
3
- size 4998623320
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:228366122aa9fa751c301f9d75af2e8efb0f6c59ae5a07baac7dc18ff47da9cc
3
+ size 5000473784
model-00027-of-00027.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6437f564ed0b86a445e61cc7f07fb3d6be42d3d155944e3abc4f2d87fad2f86
3
- size 1889136736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:722e40d3b7a0f7755937dd90e5325818c9c5e4019c6605f79435b10ab8bdbbc9
3
+ size 1930035696
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff
 
quantization_config.json CHANGED
@@ -4,7 +4,7 @@
4
  "sym": true,
5
  "data_type": "int",
6
  "iters": 0,
7
- "autoround_version": "0.6.0",
8
  "quant_method": "auto-round",
9
  "packing_format": "auto_round:auto_gptq",
10
  "extra_config": {
@@ -25,8 +25,7 @@
25
  "group_size": 128
26
  },
27
  "model.layers.0.mlp.gate": {
28
- "bits": 8,
29
- "group_size": 128
30
  },
31
  "model.layers.1.self_attn.q_proj": {
32
  "bits": 8,
@@ -45,8 +44,7 @@
45
  "group_size": 128
46
  },
47
  "model.layers.1.mlp.gate": {
48
- "bits": 8,
49
- "group_size": 128
50
  },
51
  "model.layers.2.self_attn.q_proj": {
52
  "bits": 8,
@@ -65,8 +63,7 @@
65
  "group_size": 128
66
  },
67
  "model.layers.2.mlp.gate": {
68
- "bits": 8,
69
- "group_size": 128
70
  },
71
  "model.layers.3.self_attn.q_proj": {
72
  "bits": 8,
@@ -85,8 +82,7 @@
85
  "group_size": 128
86
  },
87
  "model.layers.3.mlp.gate": {
88
- "bits": 8,
89
- "group_size": 128
90
  },
91
  "model.layers.4.self_attn.q_proj": {
92
  "bits": 8,
@@ -105,8 +101,7 @@
105
  "group_size": 128
106
  },
107
  "model.layers.4.mlp.gate": {
108
- "bits": 8,
109
- "group_size": 128
110
  },
111
  "model.layers.5.self_attn.q_proj": {
112
  "bits": 8,
@@ -125,8 +120,7 @@
125
  "group_size": 128
126
  },
127
  "model.layers.5.mlp.gate": {
128
- "bits": 8,
129
- "group_size": 128
130
  },
131
  "model.layers.6.self_attn.q_proj": {
132
  "bits": 8,
@@ -145,8 +139,7 @@
145
  "group_size": 128
146
  },
147
  "model.layers.6.mlp.gate": {
148
- "bits": 8,
149
- "group_size": 128
150
  },
151
  "model.layers.7.self_attn.q_proj": {
152
  "bits": 8,
@@ -165,8 +158,7 @@
165
  "group_size": 128
166
  },
167
  "model.layers.7.mlp.gate": {
168
- "bits": 8,
169
- "group_size": 128
170
  },
171
  "model.layers.8.self_attn.q_proj": {
172
  "bits": 8,
@@ -185,8 +177,7 @@
185
  "group_size": 128
186
  },
187
  "model.layers.8.mlp.gate": {
188
- "bits": 8,
189
- "group_size": 128
190
  },
191
  "model.layers.9.self_attn.q_proj": {
192
  "bits": 8,
@@ -205,8 +196,7 @@
205
  "group_size": 128
206
  },
207
  "model.layers.9.mlp.gate": {
208
- "bits": 8,
209
- "group_size": 128
210
  },
211
  "model.layers.10.self_attn.q_proj": {
212
  "bits": 8,
@@ -225,8 +215,7 @@
225
  "group_size": 128
226
  },
227
  "model.layers.10.mlp.gate": {
228
- "bits": 8,
229
- "group_size": 128
230
  },
231
  "model.layers.11.self_attn.q_proj": {
232
  "bits": 8,
@@ -245,8 +234,7 @@
245
  "group_size": 128
246
  },
247
  "model.layers.11.mlp.gate": {
248
- "bits": 8,
249
- "group_size": 128
250
  },
251
  "model.layers.12.self_attn.q_proj": {
252
  "bits": 8,
@@ -265,8 +253,7 @@
265
  "group_size": 128
266
  },
267
  "model.layers.12.mlp.gate": {
268
- "bits": 8,
269
- "group_size": 128
270
  },
271
  "model.layers.13.self_attn.q_proj": {
272
  "bits": 8,
@@ -285,8 +272,7 @@
285
  "group_size": 128
286
  },
287
  "model.layers.13.mlp.gate": {
288
- "bits": 8,
289
- "group_size": 128
290
  },
291
  "model.layers.14.self_attn.q_proj": {
292
  "bits": 8,
@@ -305,8 +291,7 @@
305
  "group_size": 128
306
  },
307
  "model.layers.14.mlp.gate": {
308
- "bits": 8,
309
- "group_size": 128
310
  },
311
  "model.layers.15.self_attn.q_proj": {
312
  "bits": 8,
@@ -325,8 +310,7 @@
325
  "group_size": 128
326
  },
327
  "model.layers.15.mlp.gate": {
328
- "bits": 8,
329
- "group_size": 128
330
  },
331
  "model.layers.16.self_attn.q_proj": {
332
  "bits": 8,
@@ -345,8 +329,7 @@
345
  "group_size": 128
346
  },
347
  "model.layers.16.mlp.gate": {
348
- "bits": 8,
349
- "group_size": 128
350
  },
351
  "model.layers.17.self_attn.q_proj": {
352
  "bits": 8,
@@ -365,8 +348,7 @@
365
  "group_size": 128
366
  },
367
  "model.layers.17.mlp.gate": {
368
- "bits": 8,
369
- "group_size": 128
370
  },
371
  "model.layers.18.self_attn.q_proj": {
372
  "bits": 8,
@@ -385,8 +367,7 @@
385
  "group_size": 128
386
  },
387
  "model.layers.18.mlp.gate": {
388
- "bits": 8,
389
- "group_size": 128
390
  },
391
  "model.layers.19.self_attn.q_proj": {
392
  "bits": 8,
@@ -405,8 +386,7 @@
405
  "group_size": 128
406
  },
407
  "model.layers.19.mlp.gate": {
408
- "bits": 8,
409
- "group_size": 128
410
  },
411
  "model.layers.20.self_attn.q_proj": {
412
  "bits": 8,
@@ -425,8 +405,7 @@
425
  "group_size": 128
426
  },
427
  "model.layers.20.mlp.gate": {
428
- "bits": 8,
429
- "group_size": 128
430
  },
431
  "model.layers.21.self_attn.q_proj": {
432
  "bits": 8,
@@ -445,8 +424,7 @@
445
  "group_size": 128
446
  },
447
  "model.layers.21.mlp.gate": {
448
- "bits": 8,
449
- "group_size": 128
450
  },
451
  "model.layers.22.self_attn.q_proj": {
452
  "bits": 8,
@@ -465,8 +443,7 @@
465
  "group_size": 128
466
  },
467
  "model.layers.22.mlp.gate": {
468
- "bits": 8,
469
- "group_size": 128
470
  },
471
  "model.layers.23.self_attn.q_proj": {
472
  "bits": 8,
@@ -485,8 +462,7 @@
485
  "group_size": 128
486
  },
487
  "model.layers.23.mlp.gate": {
488
- "bits": 8,
489
- "group_size": 128
490
  },
491
  "model.layers.24.self_attn.q_proj": {
492
  "bits": 8,
@@ -505,8 +481,7 @@
505
  "group_size": 128
506
  },
507
  "model.layers.24.mlp.gate": {
508
- "bits": 8,
509
- "group_size": 128
510
  },
511
  "model.layers.25.self_attn.q_proj": {
512
  "bits": 8,
@@ -525,8 +500,7 @@
525
  "group_size": 128
526
  },
527
  "model.layers.25.mlp.gate": {
528
- "bits": 8,
529
- "group_size": 128
530
  },
531
  "model.layers.26.self_attn.q_proj": {
532
  "bits": 8,
@@ -545,8 +519,7 @@
545
  "group_size": 128
546
  },
547
  "model.layers.26.mlp.gate": {
548
- "bits": 8,
549
- "group_size": 128
550
  },
551
  "model.layers.27.self_attn.q_proj": {
552
  "bits": 8,
@@ -565,8 +538,7 @@
565
  "group_size": 128
566
  },
567
  "model.layers.27.mlp.gate": {
568
- "bits": 8,
569
- "group_size": 128
570
  },
571
  "model.layers.28.self_attn.q_proj": {
572
  "bits": 8,
@@ -585,8 +557,7 @@
585
  "group_size": 128
586
  },
587
  "model.layers.28.mlp.gate": {
588
- "bits": 8,
589
- "group_size": 128
590
  },
591
  "model.layers.29.self_attn.q_proj": {
592
  "bits": 8,
@@ -605,8 +576,7 @@
605
  "group_size": 128
606
  },
607
  "model.layers.29.mlp.gate": {
608
- "bits": 8,
609
- "group_size": 128
610
  },
611
  "model.layers.30.self_attn.q_proj": {
612
  "bits": 8,
@@ -625,8 +595,7 @@
625
  "group_size": 128
626
  },
627
  "model.layers.30.mlp.gate": {
628
- "bits": 8,
629
- "group_size": 128
630
  },
631
  "model.layers.31.self_attn.q_proj": {
632
  "bits": 8,
@@ -645,8 +614,7 @@
645
  "group_size": 128
646
  },
647
  "model.layers.31.mlp.gate": {
648
- "bits": 8,
649
- "group_size": 128
650
  },
651
  "model.layers.32.self_attn.q_proj": {
652
  "bits": 8,
@@ -665,8 +633,7 @@
665
  "group_size": 128
666
  },
667
  "model.layers.32.mlp.gate": {
668
- "bits": 8,
669
- "group_size": 128
670
  },
671
  "model.layers.33.self_attn.q_proj": {
672
  "bits": 8,
@@ -685,8 +652,7 @@
685
  "group_size": 128
686
  },
687
  "model.layers.33.mlp.gate": {
688
- "bits": 8,
689
- "group_size": 128
690
  },
691
  "model.layers.34.self_attn.q_proj": {
692
  "bits": 8,
@@ -705,8 +671,7 @@
705
  "group_size": 128
706
  },
707
  "model.layers.34.mlp.gate": {
708
- "bits": 8,
709
- "group_size": 128
710
  },
711
  "model.layers.35.self_attn.q_proj": {
712
  "bits": 8,
@@ -725,8 +690,7 @@
725
  "group_size": 128
726
  },
727
  "model.layers.35.mlp.gate": {
728
- "bits": 8,
729
- "group_size": 128
730
  },
731
  "model.layers.36.self_attn.q_proj": {
732
  "bits": 8,
@@ -745,8 +709,7 @@
745
  "group_size": 128
746
  },
747
  "model.layers.36.mlp.gate": {
748
- "bits": 8,
749
- "group_size": 128
750
  },
751
  "model.layers.37.self_attn.q_proj": {
752
  "bits": 8,
@@ -765,8 +728,7 @@
765
  "group_size": 128
766
  },
767
  "model.layers.37.mlp.gate": {
768
- "bits": 8,
769
- "group_size": 128
770
  },
771
  "model.layers.38.self_attn.q_proj": {
772
  "bits": 8,
@@ -785,8 +747,7 @@
785
  "group_size": 128
786
  },
787
  "model.layers.38.mlp.gate": {
788
- "bits": 8,
789
- "group_size": 128
790
  },
791
  "model.layers.39.self_attn.q_proj": {
792
  "bits": 8,
@@ -805,8 +766,7 @@
805
  "group_size": 128
806
  },
807
  "model.layers.39.mlp.gate": {
808
- "bits": 8,
809
- "group_size": 128
810
  },
811
  "model.layers.40.self_attn.q_proj": {
812
  "bits": 8,
@@ -825,8 +785,7 @@
825
  "group_size": 128
826
  },
827
  "model.layers.40.mlp.gate": {
828
- "bits": 8,
829
- "group_size": 128
830
  },
831
  "model.layers.41.self_attn.q_proj": {
832
  "bits": 8,
@@ -845,8 +804,7 @@
845
  "group_size": 128
846
  },
847
  "model.layers.41.mlp.gate": {
848
- "bits": 8,
849
- "group_size": 128
850
  },
851
  "model.layers.42.self_attn.q_proj": {
852
  "bits": 8,
@@ -865,8 +823,7 @@
865
  "group_size": 128
866
  },
867
  "model.layers.42.mlp.gate": {
868
- "bits": 8,
869
- "group_size": 128
870
  },
871
  "model.layers.43.self_attn.q_proj": {
872
  "bits": 8,
@@ -885,8 +842,7 @@
885
  "group_size": 128
886
  },
887
  "model.layers.43.mlp.gate": {
888
- "bits": 8,
889
- "group_size": 128
890
  },
891
  "model.layers.44.self_attn.q_proj": {
892
  "bits": 8,
@@ -905,8 +861,7 @@
905
  "group_size": 128
906
  },
907
  "model.layers.44.mlp.gate": {
908
- "bits": 8,
909
- "group_size": 128
910
  },
911
  "model.layers.45.self_attn.q_proj": {
912
  "bits": 8,
@@ -925,8 +880,7 @@
925
  "group_size": 128
926
  },
927
  "model.layers.45.mlp.gate": {
928
- "bits": 8,
929
- "group_size": 128
930
  },
931
  "model.layers.46.self_attn.q_proj": {
932
  "bits": 8,
@@ -945,8 +899,7 @@
945
  "group_size": 128
946
  },
947
  "model.layers.46.mlp.gate": {
948
- "bits": 8,
949
- "group_size": 128
950
  },
951
  "model.layers.47.self_attn.q_proj": {
952
  "bits": 8,
@@ -965,8 +918,7 @@
965
  "group_size": 128
966
  },
967
  "model.layers.47.mlp.gate": {
968
- "bits": 8,
969
- "group_size": 128
970
  },
971
  "model.layers.48.self_attn.q_proj": {
972
  "bits": 8,
@@ -985,8 +937,7 @@
985
  "group_size": 128
986
  },
987
  "model.layers.48.mlp.gate": {
988
- "bits": 8,
989
- "group_size": 128
990
  },
991
  "model.layers.49.self_attn.q_proj": {
992
  "bits": 8,
@@ -1005,8 +956,7 @@
1005
  "group_size": 128
1006
  },
1007
  "model.layers.49.mlp.gate": {
1008
- "bits": 8,
1009
- "group_size": 128
1010
  },
1011
  "model.layers.50.self_attn.q_proj": {
1012
  "bits": 8,
@@ -1025,8 +975,7 @@
1025
  "group_size": 128
1026
  },
1027
  "model.layers.50.mlp.gate": {
1028
- "bits": 8,
1029
- "group_size": 128
1030
  },
1031
  "model.layers.51.self_attn.q_proj": {
1032
  "bits": 8,
@@ -1045,8 +994,7 @@
1045
  "group_size": 128
1046
  },
1047
  "model.layers.51.mlp.gate": {
1048
- "bits": 8,
1049
- "group_size": 128
1050
  },
1051
  "model.layers.52.self_attn.q_proj": {
1052
  "bits": 8,
@@ -1065,8 +1013,7 @@
1065
  "group_size": 128
1066
  },
1067
  "model.layers.52.mlp.gate": {
1068
- "bits": 8,
1069
- "group_size": 128
1070
  },
1071
  "model.layers.53.self_attn.q_proj": {
1072
  "bits": 8,
@@ -1085,8 +1032,7 @@
1085
  "group_size": 128
1086
  },
1087
  "model.layers.53.mlp.gate": {
1088
- "bits": 8,
1089
- "group_size": 128
1090
  },
1091
  "model.layers.54.self_attn.q_proj": {
1092
  "bits": 8,
@@ -1105,8 +1051,7 @@
1105
  "group_size": 128
1106
  },
1107
  "model.layers.54.mlp.gate": {
1108
- "bits": 8,
1109
- "group_size": 128
1110
  },
1111
  "model.layers.55.self_attn.q_proj": {
1112
  "bits": 8,
@@ -1125,8 +1070,7 @@
1125
  "group_size": 128
1126
  },
1127
  "model.layers.55.mlp.gate": {
1128
- "bits": 8,
1129
- "group_size": 128
1130
  },
1131
  "model.layers.56.self_attn.q_proj": {
1132
  "bits": 8,
@@ -1145,8 +1089,7 @@
1145
  "group_size": 128
1146
  },
1147
  "model.layers.56.mlp.gate": {
1148
- "bits": 8,
1149
- "group_size": 128
1150
  },
1151
  "model.layers.57.self_attn.q_proj": {
1152
  "bits": 8,
@@ -1165,8 +1108,7 @@
1165
  "group_size": 128
1166
  },
1167
  "model.layers.57.mlp.gate": {
1168
- "bits": 8,
1169
- "group_size": 128
1170
  },
1171
  "model.layers.58.self_attn.q_proj": {
1172
  "bits": 8,
@@ -1185,8 +1127,7 @@
1185
  "group_size": 128
1186
  },
1187
  "model.layers.58.mlp.gate": {
1188
- "bits": 8,
1189
- "group_size": 128
1190
  },
1191
  "model.layers.59.self_attn.q_proj": {
1192
  "bits": 8,
@@ -1205,8 +1146,7 @@
1205
  "group_size": 128
1206
  },
1207
  "model.layers.59.mlp.gate": {
1208
- "bits": 8,
1209
- "group_size": 128
1210
  },
1211
  "model.layers.60.self_attn.q_proj": {
1212
  "bits": 8,
@@ -1225,8 +1165,7 @@
1225
  "group_size": 128
1226
  },
1227
  "model.layers.60.mlp.gate": {
1228
- "bits": 8,
1229
- "group_size": 128
1230
  },
1231
  "model.layers.61.self_attn.q_proj": {
1232
  "bits": 8,
@@ -1245,8 +1184,7 @@
1245
  "group_size": 128
1246
  },
1247
  "model.layers.61.mlp.gate": {
1248
- "bits": 8,
1249
- "group_size": 128
1250
  },
1251
  "model.layers.62.self_attn.q_proj": {
1252
  "bits": 8,
@@ -1265,8 +1203,7 @@
1265
  "group_size": 128
1266
  },
1267
  "model.layers.62.mlp.gate": {
1268
- "bits": 8,
1269
- "group_size": 128
1270
  },
1271
  "model.layers.63.self_attn.q_proj": {
1272
  "bits": 8,
@@ -1285,8 +1222,7 @@
1285
  "group_size": 128
1286
  },
1287
  "model.layers.63.mlp.gate": {
1288
- "bits": 8,
1289
- "group_size": 128
1290
  },
1291
  "model.layers.64.self_attn.q_proj": {
1292
  "bits": 8,
@@ -1305,8 +1241,7 @@
1305
  "group_size": 128
1306
  },
1307
  "model.layers.64.mlp.gate": {
1308
- "bits": 8,
1309
- "group_size": 128
1310
  },
1311
  "model.layers.65.self_attn.q_proj": {
1312
  "bits": 8,
@@ -1325,8 +1260,7 @@
1325
  "group_size": 128
1326
  },
1327
  "model.layers.65.mlp.gate": {
1328
- "bits": 8,
1329
- "group_size": 128
1330
  },
1331
  "model.layers.66.self_attn.q_proj": {
1332
  "bits": 8,
@@ -1345,8 +1279,7 @@
1345
  "group_size": 128
1346
  },
1347
  "model.layers.66.mlp.gate": {
1348
- "bits": 8,
1349
- "group_size": 128
1350
  },
1351
  "model.layers.67.self_attn.q_proj": {
1352
  "bits": 8,
@@ -1365,8 +1298,7 @@
1365
  "group_size": 128
1366
  },
1367
  "model.layers.67.mlp.gate": {
1368
- "bits": 8,
1369
- "group_size": 128
1370
  },
1371
  "model.layers.68.self_attn.q_proj": {
1372
  "bits": 8,
@@ -1385,8 +1317,7 @@
1385
  "group_size": 128
1386
  },
1387
  "model.layers.68.mlp.gate": {
1388
- "bits": 8,
1389
- "group_size": 128
1390
  },
1391
  "model.layers.69.self_attn.q_proj": {
1392
  "bits": 8,
@@ -1405,8 +1336,7 @@
1405
  "group_size": 128
1406
  },
1407
  "model.layers.69.mlp.gate": {
1408
- "bits": 8,
1409
- "group_size": 128
1410
  },
1411
  "model.layers.70.self_attn.q_proj": {
1412
  "bits": 8,
@@ -1425,8 +1355,7 @@
1425
  "group_size": 128
1426
  },
1427
  "model.layers.70.mlp.gate": {
1428
- "bits": 8,
1429
- "group_size": 128
1430
  },
1431
  "model.layers.71.self_attn.q_proj": {
1432
  "bits": 8,
@@ -1445,8 +1374,7 @@
1445
  "group_size": 128
1446
  },
1447
  "model.layers.71.mlp.gate": {
1448
- "bits": 8,
1449
- "group_size": 128
1450
  },
1451
  "model.layers.72.self_attn.q_proj": {
1452
  "bits": 8,
@@ -1465,8 +1393,7 @@
1465
  "group_size": 128
1466
  },
1467
  "model.layers.72.mlp.gate": {
1468
- "bits": 8,
1469
- "group_size": 128
1470
  },
1471
  "model.layers.73.self_attn.q_proj": {
1472
  "bits": 8,
@@ -1485,8 +1412,7 @@
1485
  "group_size": 128
1486
  },
1487
  "model.layers.73.mlp.gate": {
1488
- "bits": 8,
1489
- "group_size": 128
1490
  },
1491
  "model.layers.74.self_attn.q_proj": {
1492
  "bits": 8,
@@ -1505,8 +1431,7 @@
1505
  "group_size": 128
1506
  },
1507
  "model.layers.74.mlp.gate": {
1508
- "bits": 8,
1509
- "group_size": 128
1510
  },
1511
  "model.layers.75.self_attn.q_proj": {
1512
  "bits": 8,
@@ -1525,8 +1450,7 @@
1525
  "group_size": 128
1526
  },
1527
  "model.layers.75.mlp.gate": {
1528
- "bits": 8,
1529
- "group_size": 128
1530
  },
1531
  "model.layers.76.self_attn.q_proj": {
1532
  "bits": 8,
@@ -1545,8 +1469,7 @@
1545
  "group_size": 128
1546
  },
1547
  "model.layers.76.mlp.gate": {
1548
- "bits": 8,
1549
- "group_size": 128
1550
  },
1551
  "model.layers.77.self_attn.q_proj": {
1552
  "bits": 8,
@@ -1565,8 +1488,7 @@
1565
  "group_size": 128
1566
  },
1567
  "model.layers.77.mlp.gate": {
1568
- "bits": 8,
1569
- "group_size": 128
1570
  },
1571
  "model.layers.78.self_attn.q_proj": {
1572
  "bits": 8,
@@ -1585,8 +1507,7 @@
1585
  "group_size": 128
1586
  },
1587
  "model.layers.78.mlp.gate": {
1588
- "bits": 8,
1589
- "group_size": 128
1590
  },
1591
  "model.layers.79.self_attn.q_proj": {
1592
  "bits": 8,
@@ -1605,8 +1526,7 @@
1605
  "group_size": 128
1606
  },
1607
  "model.layers.79.mlp.gate": {
1608
- "bits": 8,
1609
- "group_size": 128
1610
  },
1611
  "model.layers.80.self_attn.q_proj": {
1612
  "bits": 8,
@@ -1625,8 +1545,7 @@
1625
  "group_size": 128
1626
  },
1627
  "model.layers.80.mlp.gate": {
1628
- "bits": 8,
1629
- "group_size": 128
1630
  },
1631
  "model.layers.81.self_attn.q_proj": {
1632
  "bits": 8,
@@ -1645,8 +1564,7 @@
1645
  "group_size": 128
1646
  },
1647
  "model.layers.81.mlp.gate": {
1648
- "bits": 8,
1649
- "group_size": 128
1650
  },
1651
  "model.layers.82.self_attn.q_proj": {
1652
  "bits": 8,
@@ -1665,8 +1583,7 @@
1665
  "group_size": 128
1666
  },
1667
  "model.layers.82.mlp.gate": {
1668
- "bits": 8,
1669
- "group_size": 128
1670
  },
1671
  "model.layers.83.self_attn.q_proj": {
1672
  "bits": 8,
@@ -1685,8 +1602,7 @@
1685
  "group_size": 128
1686
  },
1687
  "model.layers.83.mlp.gate": {
1688
- "bits": 8,
1689
- "group_size": 128
1690
  },
1691
  "model.layers.84.self_attn.q_proj": {
1692
  "bits": 8,
@@ -1705,8 +1621,7 @@
1705
  "group_size": 128
1706
  },
1707
  "model.layers.84.mlp.gate": {
1708
- "bits": 8,
1709
- "group_size": 128
1710
  },
1711
  "model.layers.85.self_attn.q_proj": {
1712
  "bits": 8,
@@ -1725,8 +1640,7 @@
1725
  "group_size": 128
1726
  },
1727
  "model.layers.85.mlp.gate": {
1728
- "bits": 8,
1729
- "group_size": 128
1730
  },
1731
  "model.layers.86.self_attn.q_proj": {
1732
  "bits": 8,
@@ -1745,8 +1659,7 @@
1745
  "group_size": 128
1746
  },
1747
  "model.layers.86.mlp.gate": {
1748
- "bits": 8,
1749
- "group_size": 128
1750
  },
1751
  "model.layers.87.self_attn.q_proj": {
1752
  "bits": 8,
@@ -1765,8 +1678,7 @@
1765
  "group_size": 128
1766
  },
1767
  "model.layers.87.mlp.gate": {
1768
- "bits": 8,
1769
- "group_size": 128
1770
  },
1771
  "model.layers.88.self_attn.q_proj": {
1772
  "bits": 8,
@@ -1785,8 +1697,7 @@
1785
  "group_size": 128
1786
  },
1787
  "model.layers.88.mlp.gate": {
1788
- "bits": 8,
1789
- "group_size": 128
1790
  },
1791
  "model.layers.89.self_attn.q_proj": {
1792
  "bits": 8,
@@ -1805,8 +1716,7 @@
1805
  "group_size": 128
1806
  },
1807
  "model.layers.89.mlp.gate": {
1808
- "bits": 8,
1809
- "group_size": 128
1810
  },
1811
  "model.layers.90.self_attn.q_proj": {
1812
  "bits": 8,
@@ -1825,8 +1735,7 @@
1825
  "group_size": 128
1826
  },
1827
  "model.layers.90.mlp.gate": {
1828
- "bits": 8,
1829
- "group_size": 128
1830
  },
1831
  "model.layers.91.self_attn.q_proj": {
1832
  "bits": 8,
@@ -1845,8 +1754,7 @@
1845
  "group_size": 128
1846
  },
1847
  "model.layers.91.mlp.gate": {
1848
- "bits": 8,
1849
- "group_size": 128
1850
  },
1851
  "model.layers.92.self_attn.q_proj": {
1852
  "bits": 8,
@@ -1865,8 +1773,7 @@
1865
  "group_size": 128
1866
  },
1867
  "model.layers.92.mlp.gate": {
1868
- "bits": 8,
1869
- "group_size": 128
1870
  },
1871
  "model.layers.93.self_attn.q_proj": {
1872
  "bits": 8,
@@ -1885,8 +1792,7 @@
1885
  "group_size": 128
1886
  },
1887
  "model.layers.93.mlp.gate": {
1888
- "bits": 8,
1889
- "group_size": 128
1890
  }
1891
  }
1892
  }
 
4
  "sym": true,
5
  "data_type": "int",
6
  "iters": 0,
7
+ "autoround_version": "0.6.1.dev",
8
  "quant_method": "auto-round",
9
  "packing_format": "auto_round:auto_gptq",
10
  "extra_config": {
 
25
  "group_size": 128
26
  },
27
  "model.layers.0.mlp.gate": {
28
+ "bits": 16
 
29
  },
30
  "model.layers.1.self_attn.q_proj": {
31
  "bits": 8,
 
44
  "group_size": 128
45
  },
46
  "model.layers.1.mlp.gate": {
47
+ "bits": 16
 
48
  },
49
  "model.layers.2.self_attn.q_proj": {
50
  "bits": 8,
 
63
  "group_size": 128
64
  },
65
  "model.layers.2.mlp.gate": {
66
+ "bits": 16
 
67
  },
68
  "model.layers.3.self_attn.q_proj": {
69
  "bits": 8,
 
82
  "group_size": 128
83
  },
84
  "model.layers.3.mlp.gate": {
85
+ "bits": 16
 
86
  },
87
  "model.layers.4.self_attn.q_proj": {
88
  "bits": 8,
 
101
  "group_size": 128
102
  },
103
  "model.layers.4.mlp.gate": {
104
+ "bits": 16
 
105
  },
106
  "model.layers.5.self_attn.q_proj": {
107
  "bits": 8,
 
120
  "group_size": 128
121
  },
122
  "model.layers.5.mlp.gate": {
123
+ "bits": 16
 
124
  },
125
  "model.layers.6.self_attn.q_proj": {
126
  "bits": 8,
 
139
  "group_size": 128
140
  },
141
  "model.layers.6.mlp.gate": {
142
+ "bits": 16
 
143
  },
144
  "model.layers.7.self_attn.q_proj": {
145
  "bits": 8,
 
158
  "group_size": 128
159
  },
160
  "model.layers.7.mlp.gate": {
161
+ "bits": 16
 
162
  },
163
  "model.layers.8.self_attn.q_proj": {
164
  "bits": 8,
 
177
  "group_size": 128
178
  },
179
  "model.layers.8.mlp.gate": {
180
+ "bits": 16
 
181
  },
182
  "model.layers.9.self_attn.q_proj": {
183
  "bits": 8,
 
196
  "group_size": 128
197
  },
198
  "model.layers.9.mlp.gate": {
199
+ "bits": 16
 
200
  },
201
  "model.layers.10.self_attn.q_proj": {
202
  "bits": 8,
 
215
  "group_size": 128
216
  },
217
  "model.layers.10.mlp.gate": {
218
+ "bits": 16
 
219
  },
220
  "model.layers.11.self_attn.q_proj": {
221
  "bits": 8,
 
234
  "group_size": 128
235
  },
236
  "model.layers.11.mlp.gate": {
237
+ "bits": 16
 
238
  },
239
  "model.layers.12.self_attn.q_proj": {
240
  "bits": 8,
 
253
  "group_size": 128
254
  },
255
  "model.layers.12.mlp.gate": {
256
+ "bits": 16
 
257
  },
258
  "model.layers.13.self_attn.q_proj": {
259
  "bits": 8,
 
272
  "group_size": 128
273
  },
274
  "model.layers.13.mlp.gate": {
275
+ "bits": 16
 
276
  },
277
  "model.layers.14.self_attn.q_proj": {
278
  "bits": 8,
 
291
  "group_size": 128
292
  },
293
  "model.layers.14.mlp.gate": {
294
+ "bits": 16
 
295
  },
296
  "model.layers.15.self_attn.q_proj": {
297
  "bits": 8,
 
310
  "group_size": 128
311
  },
312
  "model.layers.15.mlp.gate": {
313
+ "bits": 16
 
314
  },
315
  "model.layers.16.self_attn.q_proj": {
316
  "bits": 8,
 
329
  "group_size": 128
330
  },
331
  "model.layers.16.mlp.gate": {
332
+ "bits": 16
 
333
  },
334
  "model.layers.17.self_attn.q_proj": {
335
  "bits": 8,
 
348
  "group_size": 128
349
  },
350
  "model.layers.17.mlp.gate": {
351
+ "bits": 16
 
352
  },
353
  "model.layers.18.self_attn.q_proj": {
354
  "bits": 8,
 
367
  "group_size": 128
368
  },
369
  "model.layers.18.mlp.gate": {
370
+ "bits": 16
 
371
  },
372
  "model.layers.19.self_attn.q_proj": {
373
  "bits": 8,
 
386
  "group_size": 128
387
  },
388
  "model.layers.19.mlp.gate": {
389
+ "bits": 16
 
390
  },
391
  "model.layers.20.self_attn.q_proj": {
392
  "bits": 8,
 
405
  "group_size": 128
406
  },
407
  "model.layers.20.mlp.gate": {
408
+ "bits": 16
 
409
  },
410
  "model.layers.21.self_attn.q_proj": {
411
  "bits": 8,
 
424
  "group_size": 128
425
  },
426
  "model.layers.21.mlp.gate": {
427
+ "bits": 16
 
428
  },
429
  "model.layers.22.self_attn.q_proj": {
430
  "bits": 8,
 
443
  "group_size": 128
444
  },
445
  "model.layers.22.mlp.gate": {
446
+ "bits": 16
 
447
  },
448
  "model.layers.23.self_attn.q_proj": {
449
  "bits": 8,
 
462
  "group_size": 128
463
  },
464
  "model.layers.23.mlp.gate": {
465
+ "bits": 16
 
466
  },
467
  "model.layers.24.self_attn.q_proj": {
468
  "bits": 8,
 
481
  "group_size": 128
482
  },
483
  "model.layers.24.mlp.gate": {
484
+ "bits": 16
 
485
  },
486
  "model.layers.25.self_attn.q_proj": {
487
  "bits": 8,
 
500
  "group_size": 128
501
  },
502
  "model.layers.25.mlp.gate": {
503
+ "bits": 16
 
504
  },
505
  "model.layers.26.self_attn.q_proj": {
506
  "bits": 8,
 
519
  "group_size": 128
520
  },
521
  "model.layers.26.mlp.gate": {
522
+ "bits": 16
 
523
  },
524
  "model.layers.27.self_attn.q_proj": {
525
  "bits": 8,
 
538
  "group_size": 128
539
  },
540
  "model.layers.27.mlp.gate": {
541
+ "bits": 16
 
542
  },
543
  "model.layers.28.self_attn.q_proj": {
544
  "bits": 8,
 
557
  "group_size": 128
558
  },
559
  "model.layers.28.mlp.gate": {
560
+ "bits": 16
 
561
  },
562
  "model.layers.29.self_attn.q_proj": {
563
  "bits": 8,
 
576
  "group_size": 128
577
  },
578
  "model.layers.29.mlp.gate": {
579
+ "bits": 16
 
580
  },
581
  "model.layers.30.self_attn.q_proj": {
582
  "bits": 8,
 
595
  "group_size": 128
596
  },
597
  "model.layers.30.mlp.gate": {
598
+ "bits": 16
 
599
  },
600
  "model.layers.31.self_attn.q_proj": {
601
  "bits": 8,
 
614
  "group_size": 128
615
  },
616
  "model.layers.31.mlp.gate": {
617
+ "bits": 16
 
618
  },
619
  "model.layers.32.self_attn.q_proj": {
620
  "bits": 8,
 
633
  "group_size": 128
634
  },
635
  "model.layers.32.mlp.gate": {
636
+ "bits": 16
 
637
  },
638
  "model.layers.33.self_attn.q_proj": {
639
  "bits": 8,
 
652
  "group_size": 128
653
  },
654
  "model.layers.33.mlp.gate": {
655
+ "bits": 16
 
656
  },
657
  "model.layers.34.self_attn.q_proj": {
658
  "bits": 8,
 
671
  "group_size": 128
672
  },
673
  "model.layers.34.mlp.gate": {
674
+ "bits": 16
 
675
  },
676
  "model.layers.35.self_attn.q_proj": {
677
  "bits": 8,
 
690
  "group_size": 128
691
  },
692
  "model.layers.35.mlp.gate": {
693
+ "bits": 16
 
694
  },
695
  "model.layers.36.self_attn.q_proj": {
696
  "bits": 8,
 
709
  "group_size": 128
710
  },
711
  "model.layers.36.mlp.gate": {
712
+ "bits": 16
 
713
  },
714
  "model.layers.37.self_attn.q_proj": {
715
  "bits": 8,
 
728
  "group_size": 128
729
  },
730
  "model.layers.37.mlp.gate": {
731
+ "bits": 16
 
732
  },
733
  "model.layers.38.self_attn.q_proj": {
734
  "bits": 8,
 
747
  "group_size": 128
748
  },
749
  "model.layers.38.mlp.gate": {
750
+ "bits": 16
 
751
  },
752
  "model.layers.39.self_attn.q_proj": {
753
  "bits": 8,
 
766
  "group_size": 128
767
  },
768
  "model.layers.39.mlp.gate": {
769
+ "bits": 16
 
770
  },
771
  "model.layers.40.self_attn.q_proj": {
772
  "bits": 8,
 
785
  "group_size": 128
786
  },
787
  "model.layers.40.mlp.gate": {
788
+ "bits": 16
 
789
  },
790
  "model.layers.41.self_attn.q_proj": {
791
  "bits": 8,
 
804
  "group_size": 128
805
  },
806
  "model.layers.41.mlp.gate": {
807
+ "bits": 16
 
808
  },
809
  "model.layers.42.self_attn.q_proj": {
810
  "bits": 8,
 
823
  "group_size": 128
824
  },
825
  "model.layers.42.mlp.gate": {
826
+ "bits": 16
 
827
  },
828
  "model.layers.43.self_attn.q_proj": {
829
  "bits": 8,
 
842
  "group_size": 128
843
  },
844
  "model.layers.43.mlp.gate": {
845
+ "bits": 16
 
846
  },
847
  "model.layers.44.self_attn.q_proj": {
848
  "bits": 8,
 
861
  "group_size": 128
862
  },
863
  "model.layers.44.mlp.gate": {
864
+ "bits": 16
 
865
  },
866
  "model.layers.45.self_attn.q_proj": {
867
  "bits": 8,
 
880
  "group_size": 128
881
  },
882
  "model.layers.45.mlp.gate": {
883
+ "bits": 16
 
884
  },
885
  "model.layers.46.self_attn.q_proj": {
886
  "bits": 8,
 
899
  "group_size": 128
900
  },
901
  "model.layers.46.mlp.gate": {
902
+ "bits": 16
 
903
  },
904
  "model.layers.47.self_attn.q_proj": {
905
  "bits": 8,
 
918
  "group_size": 128
919
  },
920
  "model.layers.47.mlp.gate": {
921
+ "bits": 16
 
922
  },
923
  "model.layers.48.self_attn.q_proj": {
924
  "bits": 8,
 
937
  "group_size": 128
938
  },
939
  "model.layers.48.mlp.gate": {
940
+ "bits": 16
 
941
  },
942
  "model.layers.49.self_attn.q_proj": {
943
  "bits": 8,
 
956
  "group_size": 128
957
  },
958
  "model.layers.49.mlp.gate": {
959
+ "bits": 16
 
960
  },
961
  "model.layers.50.self_attn.q_proj": {
962
  "bits": 8,
 
975
  "group_size": 128
976
  },
977
  "model.layers.50.mlp.gate": {
978
+ "bits": 16
 
979
  },
980
  "model.layers.51.self_attn.q_proj": {
981
  "bits": 8,
 
994
  "group_size": 128
995
  },
996
  "model.layers.51.mlp.gate": {
997
+ "bits": 16
 
998
  },
999
  "model.layers.52.self_attn.q_proj": {
1000
  "bits": 8,
 
1013
  "group_size": 128
1014
  },
1015
  "model.layers.52.mlp.gate": {
1016
+ "bits": 16
 
1017
  },
1018
  "model.layers.53.self_attn.q_proj": {
1019
  "bits": 8,
 
1032
  "group_size": 128
1033
  },
1034
  "model.layers.53.mlp.gate": {
1035
+ "bits": 16
 
1036
  },
1037
  "model.layers.54.self_attn.q_proj": {
1038
  "bits": 8,
 
1051
  "group_size": 128
1052
  },
1053
  "model.layers.54.mlp.gate": {
1054
+ "bits": 16
 
1055
  },
1056
  "model.layers.55.self_attn.q_proj": {
1057
  "bits": 8,
 
1070
  "group_size": 128
1071
  },
1072
  "model.layers.55.mlp.gate": {
1073
+ "bits": 16
 
1074
  },
1075
  "model.layers.56.self_attn.q_proj": {
1076
  "bits": 8,
 
1089
  "group_size": 128
1090
  },
1091
  "model.layers.56.mlp.gate": {
1092
+ "bits": 16
 
1093
  },
1094
  "model.layers.57.self_attn.q_proj": {
1095
  "bits": 8,
 
1108
  "group_size": 128
1109
  },
1110
  "model.layers.57.mlp.gate": {
1111
+ "bits": 16
 
1112
  },
1113
  "model.layers.58.self_attn.q_proj": {
1114
  "bits": 8,
 
1127
  "group_size": 128
1128
  },
1129
  "model.layers.58.mlp.gate": {
1130
+ "bits": 16
 
1131
  },
1132
  "model.layers.59.self_attn.q_proj": {
1133
  "bits": 8,
 
1146
  "group_size": 128
1147
  },
1148
  "model.layers.59.mlp.gate": {
1149
+ "bits": 16
 
1150
  },
1151
  "model.layers.60.self_attn.q_proj": {
1152
  "bits": 8,
 
1165
  "group_size": 128
1166
  },
1167
  "model.layers.60.mlp.gate": {
1168
+ "bits": 16
 
1169
  },
1170
  "model.layers.61.self_attn.q_proj": {
1171
  "bits": 8,
 
1184
  "group_size": 128
1185
  },
1186
  "model.layers.61.mlp.gate": {
1187
+ "bits": 16
 
1188
  },
1189
  "model.layers.62.self_attn.q_proj": {
1190
  "bits": 8,
 
1203
  "group_size": 128
1204
  },
1205
  "model.layers.62.mlp.gate": {
1206
+ "bits": 16
 
1207
  },
1208
  "model.layers.63.self_attn.q_proj": {
1209
  "bits": 8,
 
1222
  "group_size": 128
1223
  },
1224
  "model.layers.63.mlp.gate": {
1225
+ "bits": 16
 
1226
  },
1227
  "model.layers.64.self_attn.q_proj": {
1228
  "bits": 8,
 
1241
  "group_size": 128
1242
  },
1243
  "model.layers.64.mlp.gate": {
1244
+ "bits": 16
 
1245
  },
1246
  "model.layers.65.self_attn.q_proj": {
1247
  "bits": 8,
 
1260
  "group_size": 128
1261
  },
1262
  "model.layers.65.mlp.gate": {
1263
+ "bits": 16
 
1264
  },
1265
  "model.layers.66.self_attn.q_proj": {
1266
  "bits": 8,
 
1279
  "group_size": 128
1280
  },
1281
  "model.layers.66.mlp.gate": {
1282
+ "bits": 16
 
1283
  },
1284
  "model.layers.67.self_attn.q_proj": {
1285
  "bits": 8,
 
1298
  "group_size": 128
1299
  },
1300
  "model.layers.67.mlp.gate": {
1301
+ "bits": 16
 
1302
  },
1303
  "model.layers.68.self_attn.q_proj": {
1304
  "bits": 8,
 
1317
  "group_size": 128
1318
  },
1319
  "model.layers.68.mlp.gate": {
1320
+ "bits": 16
 
1321
  },
1322
  "model.layers.69.self_attn.q_proj": {
1323
  "bits": 8,
 
1336
  "group_size": 128
1337
  },
1338
  "model.layers.69.mlp.gate": {
1339
+ "bits": 16
 
1340
  },
1341
  "model.layers.70.self_attn.q_proj": {
1342
  "bits": 8,
 
1355
  "group_size": 128
1356
  },
1357
  "model.layers.70.mlp.gate": {
1358
+ "bits": 16
 
1359
  },
1360
  "model.layers.71.self_attn.q_proj": {
1361
  "bits": 8,
 
1374
  "group_size": 128
1375
  },
1376
  "model.layers.71.mlp.gate": {
1377
+ "bits": 16
 
1378
  },
1379
  "model.layers.72.self_attn.q_proj": {
1380
  "bits": 8,
 
1393
  "group_size": 128
1394
  },
1395
  "model.layers.72.mlp.gate": {
1396
+ "bits": 16
 
1397
  },
1398
  "model.layers.73.self_attn.q_proj": {
1399
  "bits": 8,
 
1412
  "group_size": 128
1413
  },
1414
  "model.layers.73.mlp.gate": {
1415
+ "bits": 16
 
1416
  },
1417
  "model.layers.74.self_attn.q_proj": {
1418
  "bits": 8,
 
1431
  "group_size": 128
1432
  },
1433
  "model.layers.74.mlp.gate": {
1434
+ "bits": 16
 
1435
  },
1436
  "model.layers.75.self_attn.q_proj": {
1437
  "bits": 8,
 
1450
  "group_size": 128
1451
  },
1452
  "model.layers.75.mlp.gate": {
1453
+ "bits": 16
 
1454
  },
1455
  "model.layers.76.self_attn.q_proj": {
1456
  "bits": 8,
 
1469
  "group_size": 128
1470
  },
1471
  "model.layers.76.mlp.gate": {
1472
+ "bits": 16
 
1473
  },
1474
  "model.layers.77.self_attn.q_proj": {
1475
  "bits": 8,
 
1488
  "group_size": 128
1489
  },
1490
  "model.layers.77.mlp.gate": {
1491
+ "bits": 16
 
1492
  },
1493
  "model.layers.78.self_attn.q_proj": {
1494
  "bits": 8,
 
1507
  "group_size": 128
1508
  },
1509
  "model.layers.78.mlp.gate": {
1510
+ "bits": 16
 
1511
  },
1512
  "model.layers.79.self_attn.q_proj": {
1513
  "bits": 8,
 
1526
  "group_size": 128
1527
  },
1528
  "model.layers.79.mlp.gate": {
1529
+ "bits": 16
 
1530
  },
1531
  "model.layers.80.self_attn.q_proj": {
1532
  "bits": 8,
 
1545
  "group_size": 128
1546
  },
1547
  "model.layers.80.mlp.gate": {
1548
+ "bits": 16
 
1549
  },
1550
  "model.layers.81.self_attn.q_proj": {
1551
  "bits": 8,
 
1564
  "group_size": 128
1565
  },
1566
  "model.layers.81.mlp.gate": {
1567
+ "bits": 16
 
1568
  },
1569
  "model.layers.82.self_attn.q_proj": {
1570
  "bits": 8,
 
1583
  "group_size": 128
1584
  },
1585
  "model.layers.82.mlp.gate": {
1586
+ "bits": 16
 
1587
  },
1588
  "model.layers.83.self_attn.q_proj": {
1589
  "bits": 8,
 
1602
  "group_size": 128
1603
  },
1604
  "model.layers.83.mlp.gate": {
1605
+ "bits": 16
 
1606
  },
1607
  "model.layers.84.self_attn.q_proj": {
1608
  "bits": 8,
 
1621
  "group_size": 128
1622
  },
1623
  "model.layers.84.mlp.gate": {
1624
+ "bits": 16
 
1625
  },
1626
  "model.layers.85.self_attn.q_proj": {
1627
  "bits": 8,
 
1640
  "group_size": 128
1641
  },
1642
  "model.layers.85.mlp.gate": {
1643
+ "bits": 16
 
1644
  },
1645
  "model.layers.86.self_attn.q_proj": {
1646
  "bits": 8,
 
1659
  "group_size": 128
1660
  },
1661
  "model.layers.86.mlp.gate": {
1662
+ "bits": 16
 
1663
  },
1664
  "model.layers.87.self_attn.q_proj": {
1665
  "bits": 8,
 
1678
  "group_size": 128
1679
  },
1680
  "model.layers.87.mlp.gate": {
1681
+ "bits": 16
 
1682
  },
1683
  "model.layers.88.self_attn.q_proj": {
1684
  "bits": 8,
 
1697
  "group_size": 128
1698
  },
1699
  "model.layers.88.mlp.gate": {
1700
+ "bits": 16
 
1701
  },
1702
  "model.layers.89.self_attn.q_proj": {
1703
  "bits": 8,
 
1716
  "group_size": 128
1717
  },
1718
  "model.layers.89.mlp.gate": {
1719
+ "bits": 16
 
1720
  },
1721
  "model.layers.90.self_attn.q_proj": {
1722
  "bits": 8,
 
1735
  "group_size": 128
1736
  },
1737
  "model.layers.90.mlp.gate": {
1738
+ "bits": 16
 
1739
  },
1740
  "model.layers.91.self_attn.q_proj": {
1741
  "bits": 8,
 
1754
  "group_size": 128
1755
  },
1756
  "model.layers.91.mlp.gate": {
1757
+ "bits": 16
 
1758
  },
1759
  "model.layers.92.self_attn.q_proj": {
1760
  "bits": 8,
 
1773
  "group_size": 128
1774
  },
1775
  "model.layers.92.mlp.gate": {
1776
+ "bits": 16
 
1777
  },
1778
  "model.layers.93.self_attn.q_proj": {
1779
  "bits": 8,
 
1792
  "group_size": 128
1793
  },
1794
  "model.layers.93.mlp.gate": {
1795
+ "bits": 16
 
1796
  }
1797
  }
1798
  }