lejelly commited on
Commit
a975df7
·
verified ·
1 Parent(s): 5c2e131

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - merge
4
+ - parameter_wise
5
+ - llm-adamerge
6
+ base_model: mistralai/Mistral-7B-v0.1
7
+ ---
8
+
9
+ # Merged Model using LLM-AdaMerge (parameter_wise)
10
+
11
+ This model was created by merging multiple fine-tuned models using the LLM-AdaMerge approach with parameter_wise merging.
12
+
13
+ ## Merge Details
14
+
15
+ - **Merge Type**: parameter_wise
16
+ - **Base Model**: mistralai/Mistral-7B-v0.1
17
+ - **Number of Models Merged**: 3
18
+ - **Models Merged**: instruct, math, code
19
+ - **Final Training Loss**: N/A
20
+ - **Training Epochs**: 0
21
+
22
+ ## Lambda Coefficients
23
+
24
+ The following lambda coefficients were learned during training:
25
+
26
+
27
+ ### Parameter-wise Lambdas
28
+ This model uses parameter-wise lambda coefficients. Total parameters with individual lambdas: 291
29
+
30
+
31
+ See the uploaded `learned_lambdas.json` file for detailed parameter-wise coefficients.
32
+
33
+ ## Usage
34
+
35
+ ```python
36
+ from transformers import AutoModelForCausalLM, AutoTokenizer
37
+
38
+ model = AutoModelForCausalLM.from_pretrained("your-username/model-name")
39
+ tokenizer = AutoTokenizer.from_pretrained("your-username/model-name")
40
+
41
+ # Use the model
42
+ inputs = tokenizer("Hello, how are you?", return_tensors="pt")
43
+ outputs = model.generate(**inputs)
44
+ print(tokenizer.decode(outputs[0]))
45
+ ```
46
+
47
+ ## Training Configuration
48
+
49
+ See the uploaded `training_config.json` file for detailed training configuration.
50
+
51
+ ## Citation
52
+
53
+ If you use this model, please cite the LLM-AdaMerge paper:
54
+
55
+ ```bibtex
56
+ @article{llmadamerge2024,
57
+ title={LLM-AdaMerge: Adaptive Model Merging for Large Language Models},
58
+ author={...},
59
+ year={2024}
60
+ }
61
+ ```
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MistralForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "head_dim": null,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 4096,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 14336,
13
+ "max_position_embeddings": 32768,
14
+ "model_type": "mistral",
15
+ "num_attention_heads": 32,
16
+ "num_hidden_layers": 32,
17
+ "num_key_value_heads": 8,
18
+ "rms_norm_eps": 1e-05,
19
+ "rope_theta": 10000.0,
20
+ "sliding_window": 4096,
21
+ "tie_word_embeddings": false,
22
+ "torch_dtype": "float16",
23
+ "transformers_version": "4.52.4",
24
+ "use_cache": true,
25
+ "vocab_size": 32000
26
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.52.4"
6
+ }
learned_lambdas.json ADDED
@@ -0,0 +1,1759 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lambdas": [
3
+ [
4
+ 0.4511440098285675,
5
+ 0.34568217396736145,
6
+ 0.4310983717441559
7
+ ],
8
+ [
9
+ 0.4077865779399872,
10
+ 0.44282540678977966,
11
+ 0.3368643522262573
12
+ ],
13
+ [
14
+ 0.28722721338272095,
15
+ 0.4189178943634033,
16
+ 0.36558520793914795
17
+ ],
18
+ [
19
+ 0.4073631763458252,
20
+ 0.3240983784198761,
21
+ 0.38805025815963745
22
+ ],
23
+ [
24
+ 0.4540562629699707,
25
+ 0.2138141542673111,
26
+ 0.2885076403617859
27
+ ],
28
+ [
29
+ 0.45886796712875366,
30
+ 0.4091651439666748,
31
+ 0.4662487804889679
32
+ ],
33
+ [
34
+ 0.46575337648391724,
35
+ 0.3446827530860901,
36
+ 0.4880160689353943
37
+ ],
38
+ [
39
+ 0.3579305112361908,
40
+ 0.32643458247184753,
41
+ 0.43178242444992065
42
+ ],
43
+ [
44
+ 0.23109537363052368,
45
+ 0.33650973439216614,
46
+ 0.40308934450149536
47
+ ],
48
+ [
49
+ 0.3454071283340454,
50
+ 0.3584277331829071,
51
+ 0.27936890721321106
52
+ ],
53
+ [
54
+ 0.35491490364074707,
55
+ 0.3612421751022339,
56
+ 0.20739391446113586
57
+ ],
58
+ [
59
+ 0.25850096344947815,
60
+ 0.22921274602413177,
61
+ 0.2169143110513687
62
+ ],
63
+ [
64
+ 0.2233383059501648,
65
+ 0.4430111348628998,
66
+ 0.4659053683280945
67
+ ],
68
+ [
69
+ 0.20520174503326416,
70
+ 0.24486230313777924,
71
+ 0.23429881036281586
72
+ ],
73
+ [
74
+ 0.2931152880191803,
75
+ 0.4284198582172394,
76
+ 0.42642292380332947
77
+ ],
78
+ [
79
+ 0.4972710609436035,
80
+ 0.44421976804733276,
81
+ 0.38601377606391907
82
+ ],
83
+ [
84
+ 0.2682636082172394,
85
+ 0.17335446178913116,
86
+ 0.41383957862854004
87
+ ],
88
+ [
89
+ 0.2809653878211975,
90
+ 0.3548043966293335,
91
+ 0.25039395689964294
92
+ ],
93
+ [
94
+ 0.3761029541492462,
95
+ 0.26408830285072327,
96
+ 0.22779043018817902
97
+ ],
98
+ [
99
+ 0.27649909257888794,
100
+ 0.4074185788631439,
101
+ 0.25979748368263245
102
+ ],
103
+ [
104
+ 0.3501487076282501,
105
+ 0.3458843231201172,
106
+ 0.2438654899597168
107
+ ],
108
+ [
109
+ 0.12980426847934723,
110
+ 0.27279528975486755,
111
+ 0.12390352785587311
112
+ ],
113
+ [
114
+ 0.4389776885509491,
115
+ 0.4472886323928833,
116
+ 0.15493842959403992
117
+ ],
118
+ [
119
+ 0.4947451055049896,
120
+ 0.1215558797121048,
121
+ 0.46804678440093994
122
+ ],
123
+ [
124
+ 0.4076070189476013,
125
+ 0.3182707726955414,
126
+ 0.36709439754486084
127
+ ],
128
+ [
129
+ 0.18712618947029114,
130
+ 0.1271672546863556,
131
+ 0.19190722703933716
132
+ ],
133
+ [
134
+ 0.3089291751384735,
135
+ 0.22327086329460144,
136
+ 0.14891940355300903
137
+ ],
138
+ [
139
+ 0.378889799118042,
140
+ 0.29507654905319214,
141
+ 0.4630947411060333
142
+ ],
143
+ [
144
+ 0.14441412687301636,
145
+ 0.18343178927898407,
146
+ 0.2112361043691635
147
+ ],
148
+ [
149
+ 0.2136223167181015,
150
+ 0.2010475993156433,
151
+ 0.23488080501556396
152
+ ],
153
+ [
154
+ 0.1485104262828827,
155
+ 0.23044322431087494,
156
+ 0.1440969556570053
157
+ ],
158
+ [
159
+ 0.40699833631515503,
160
+ 0.19921451807022095,
161
+ 0.21528899669647217
162
+ ],
163
+ [
164
+ 0.16625864803791046,
165
+ 0.16209496557712555,
166
+ 0.41099050641059875
167
+ ],
168
+ [
169
+ 0.4706342816352844,
170
+ 0.4607185423374176,
171
+ 0.38538724184036255
172
+ ],
173
+ [
174
+ 0.15955328941345215,
175
+ 0.2526731789112091,
176
+ 0.1258438378572464
177
+ ],
178
+ [
179
+ 0.27645596861839294,
180
+ 0.2471887171268463,
181
+ 0.3655036389827728
182
+ ],
183
+ [
184
+ 0.48072558641433716,
185
+ 0.18587033450603485,
186
+ 0.11874216049909592
187
+ ],
188
+ [
189
+ 0.32947680354118347,
190
+ 0.3421177566051483,
191
+ 0.2826484739780426
192
+ ],
193
+ [
194
+ 0.22274364531040192,
195
+ 0.4459400177001953,
196
+ 0.4375954568386078
197
+ ],
198
+ [
199
+ 0.16980119049549103,
200
+ 0.17654569447040558,
201
+ 0.14438512921333313
202
+ ],
203
+ [
204
+ 0.3976966142654419,
205
+ 0.3390005826950073,
206
+ 0.16800610721111298
207
+ ],
208
+ [
209
+ 0.3277134597301483,
210
+ 0.402604877948761,
211
+ 0.23107308149337769
212
+ ],
213
+ [
214
+ 0.3207840323448181,
215
+ 0.3076019585132599,
216
+ 0.24418573081493378
217
+ ],
218
+ [
219
+ 0.31352096796035767,
220
+ 0.2760259211063385,
221
+ 0.18287041783332825
222
+ ],
223
+ [
224
+ 0.19400371611118317,
225
+ 0.2872091233730316,
226
+ 0.4200877249240875
227
+ ],
228
+ [
229
+ 0.4157150685787201,
230
+ 0.16084441542625427,
231
+ 0.47111842036247253
232
+ ],
233
+ [
234
+ 0.21216757595539093,
235
+ 0.21133485436439514,
236
+ 0.25343239307403564
237
+ ],
238
+ [
239
+ 0.28364571928977966,
240
+ 0.30081626772880554,
241
+ 0.35991379618644714
242
+ ],
243
+ [
244
+ 0.15380434691905975,
245
+ 0.29351890087127686,
246
+ 0.15854692459106445
247
+ ],
248
+ [
249
+ 0.11507381498813629,
250
+ 0.1427544206380844,
251
+ 0.3868664503097534
252
+ ],
253
+ [
254
+ 0.1273266226053238,
255
+ 0.10876727104187012,
256
+ 0.11684992909431458
257
+ ],
258
+ [
259
+ 0.432993620634079,
260
+ 0.16948646306991577,
261
+ 0.3123820126056671
262
+ ],
263
+ [
264
+ 0.23365753889083862,
265
+ 0.14285247027873993,
266
+ 0.117621511220932
267
+ ],
268
+ [
269
+ 0.3091544508934021,
270
+ 0.18454627692699432,
271
+ 0.22001750767230988
272
+ ],
273
+ [
274
+ 0.20372647047042847,
275
+ 0.19373637437820435,
276
+ 0.35549601912498474
277
+ ],
278
+ [
279
+ 0.4604543447494507,
280
+ 0.4596346318721771,
281
+ 0.44899746775627136
282
+ ],
283
+ [
284
+ 0.283939003944397,
285
+ 0.21922054886817932,
286
+ 0.19192583858966827
287
+ ],
288
+ [
289
+ 0.12823466956615448,
290
+ 0.2267867624759674,
291
+ 0.154268279671669
292
+ ],
293
+ [
294
+ 0.11155832558870316,
295
+ 0.1461743265390396,
296
+ 0.21099990606307983
297
+ ],
298
+ [
299
+ 0.1693519800901413,
300
+ 0.39470532536506653,
301
+ 0.11266002804040909
302
+ ],
303
+ [
304
+ 0.13164618611335754,
305
+ 0.1770446002483368,
306
+ 0.15961745381355286
307
+ ],
308
+ [
309
+ 0.4330594539642334,
310
+ 0.08785655349493027,
311
+ 0.18969190120697021
312
+ ],
313
+ [
314
+ 0.1738586127758026,
315
+ 0.4532787501811981,
316
+ 0.3413299322128296
317
+ ],
318
+ [
319
+ 0.14526906609535217,
320
+ 0.1388753354549408,
321
+ 0.2560168504714966
322
+ ],
323
+ [
324
+ 0.1856858879327774,
325
+ 0.24550595879554749,
326
+ 0.22082236409187317
327
+ ],
328
+ [
329
+ 0.4369717240333557,
330
+ 0.425886332988739,
331
+ 0.37043431401252747
332
+ ],
333
+ [
334
+ 0.35394248366355896,
335
+ 0.2707023620605469,
336
+ 0.16234971582889557
337
+ ],
338
+ [
339
+ 0.17669916152954102,
340
+ 0.1573595404624939,
341
+ 0.4139661490917206
342
+ ],
343
+ [
344
+ 0.2815914750099182,
345
+ 0.16102179884910583,
346
+ 0.26710090041160583
347
+ ],
348
+ [
349
+ 0.29185590147972107,
350
+ 0.17331469058990479,
351
+ 0.3098703920841217
352
+ ],
353
+ [
354
+ 0.4079897701740265,
355
+ 0.272733211517334,
356
+ 0.33182939887046814
357
+ ],
358
+ [
359
+ 0.34515219926834106,
360
+ 0.29923370480537415,
361
+ 0.37515610456466675
362
+ ],
363
+ [
364
+ 0.22567912936210632,
365
+ 0.42377275228500366,
366
+ 0.2018435001373291
367
+ ],
368
+ [
369
+ 0.2051728218793869,
370
+ 0.11492697149515152,
371
+ 0.14809797704219818
372
+ ],
373
+ [
374
+ 0.35379916429519653,
375
+ 0.2150591015815735,
376
+ 0.1706886887550354
377
+ ],
378
+ [
379
+ 0.41625112295150757,
380
+ 0.4670504331588745,
381
+ 0.32151931524276733
382
+ ],
383
+ [
384
+ 0.14866942167282104,
385
+ 0.24495528638362885,
386
+ 0.46335816383361816
387
+ ],
388
+ [
389
+ 0.12006267160177231,
390
+ 0.34981709718704224,
391
+ 0.10970097035169601
392
+ ],
393
+ [
394
+ 0.279428631067276,
395
+ 0.37996864318847656,
396
+ 0.12361448258161545
397
+ ],
398
+ [
399
+ 0.24313272535800934,
400
+ 0.25561684370040894,
401
+ 0.13757745921611786
402
+ ],
403
+ [
404
+ 0.2363184094429016,
405
+ 0.29923370480537415,
406
+ 0.13216632604599
407
+ ],
408
+ [
409
+ 0.2571988105773926,
410
+ 0.1526828557252884,
411
+ 0.3124895989894867
412
+ ],
413
+ [
414
+ 0.41544994711875916,
415
+ 0.2990501821041107,
416
+ 0.4286174178123474
417
+ ],
418
+ [
419
+ 0.2243904322385788,
420
+ 0.1854301393032074,
421
+ 0.28971952199935913
422
+ ],
423
+ [
424
+ 0.22121213376522064,
425
+ 0.3917357325553894,
426
+ 0.16811366379261017
427
+ ],
428
+ [
429
+ 0.20991240441799164,
430
+ 0.3683031499385834,
431
+ 0.17156505584716797
432
+ ],
433
+ [
434
+ 0.4028129279613495,
435
+ 0.3321937918663025,
436
+ 0.16572663187980652
437
+ ],
438
+ [
439
+ 0.1638612598180771,
440
+ 0.48852384090423584,
441
+ 0.1588045209646225
442
+ ],
443
+ [
444
+ 0.24374844133853912,
445
+ 0.22875361144542694,
446
+ 0.35431236028671265
447
+ ],
448
+ [
449
+ 0.2844041883945465,
450
+ 0.4644933342933655,
451
+ 0.39079946279525757
452
+ ],
453
+ [
454
+ 0.1330053210258484,
455
+ 0.36112070083618164,
456
+ 0.22321827709674835
457
+ ],
458
+ [
459
+ 0.39910584688186646,
460
+ 0.20124942064285278,
461
+ 0.18411631882190704
462
+ ],
463
+ [
464
+ 0.4278510510921478,
465
+ 0.46304693818092346,
466
+ 0.4296000003814697
467
+ ],
468
+ [
469
+ 0.22069533169269562,
470
+ 0.1936628520488739,
471
+ 0.2681259214878082
472
+ ],
473
+ [
474
+ 0.44691622257232666,
475
+ 0.43557244539260864,
476
+ 0.42169421911239624
477
+ ],
478
+ [
479
+ 0.31918004155158997,
480
+ 0.3289930820465088,
481
+ 0.1294279396533966
482
+ ],
483
+ [
484
+ 0.145906463265419,
485
+ 0.13503797352313995,
486
+ 0.1452464610338211
487
+ ],
488
+ [
489
+ 0.474565327167511,
490
+ 0.17498759925365448,
491
+ 0.08103786408901215
492
+ ],
493
+ [
494
+ 0.23951378464698792,
495
+ 0.29923370480537415,
496
+ 0.12034349143505096
497
+ ],
498
+ [
499
+ 0.21072536706924438,
500
+ 0.22345665097236633,
501
+ 0.41911208629608154
502
+ ],
503
+ [
504
+ 0.450216144323349,
505
+ 0.21926067769527435,
506
+ 0.4052693545818329
507
+ ],
508
+ [
509
+ 0.43164557218551636,
510
+ 0.42432743310928345,
511
+ 0.3966224789619446
512
+ ],
513
+ [
514
+ 0.44685351848602295,
515
+ 0.4619434177875519,
516
+ 0.36645224690437317
517
+ ],
518
+ [
519
+ 0.13675516843795776,
520
+ 0.14978928864002228,
521
+ 0.22030705213546753
522
+ ],
523
+ [
524
+ 0.12139047682285309,
525
+ 0.38780781626701355,
526
+ 0.47421255707740784
527
+ ],
528
+ [
529
+ 0.46543893218040466,
530
+ 0.1662176251411438,
531
+ 0.105337955057621
532
+ ],
533
+ [
534
+ 0.31978920102119446,
535
+ 0.353864461183548,
536
+ 0.4400956928730011
537
+ ],
538
+ [
539
+ 0.30786028504371643,
540
+ 0.42015624046325684,
541
+ 0.3595878481864929
542
+ ],
543
+ [
544
+ 0.4008002281188965,
545
+ 0.3998754024505615,
546
+ 0.16971978545188904
547
+ ],
548
+ [
549
+ 0.20247533917427063,
550
+ 0.2598496973514557,
551
+ 0.31049075722694397
552
+ ],
553
+ [
554
+ 0.15641634166240692,
555
+ 0.18051019310951233,
556
+ 0.2142690122127533
557
+ ],
558
+ [
559
+ 0.2825934588909149,
560
+ 0.13605156540870667,
561
+ 0.18161936104297638
562
+ ],
563
+ [
564
+ 0.25422942638397217,
565
+ 0.3929879367351532,
566
+ 0.33059224486351013
567
+ ],
568
+ [
569
+ 0.4474930465221405,
570
+ 0.4542589783668518,
571
+ 0.4316270351409912
572
+ ],
573
+ [
574
+ 0.45026838779449463,
575
+ 0.42323270440101624,
576
+ 0.3742210865020752
577
+ ],
578
+ [
579
+ 0.46585074067115784,
580
+ 0.37513217329978943,
581
+ 0.24602507054805756
582
+ ],
583
+ [
584
+ 0.4334811568260193,
585
+ 0.19446375966072083,
586
+ 0.16675102710723877
587
+ ],
588
+ [
589
+ 0.35880813002586365,
590
+ 0.43830639123916626,
591
+ 0.3471340835094452
592
+ ],
593
+ [
594
+ 0.40446630120277405,
595
+ 0.45337149500846863,
596
+ 0.2590368986129761
597
+ ],
598
+ [
599
+ 0.28164413571357727,
600
+ 0.33701270818710327,
601
+ 0.38862496614456177
602
+ ],
603
+ [
604
+ 0.21139982342720032,
605
+ 0.385542631149292,
606
+ 0.164750337600708
607
+ ],
608
+ [
609
+ 0.2833697199821472,
610
+ 0.1660548895597458,
611
+ 0.16905063390731812
612
+ ],
613
+ [
614
+ 0.48628145456314087,
615
+ 0.39683374762535095,
616
+ 0.4549018144607544
617
+ ],
618
+ [
619
+ 0.42255058884620667,
620
+ 0.12292030453681946,
621
+ 0.2693982422351837
622
+ ],
623
+ [
624
+ 0.4466954469680786,
625
+ 0.4508359134197235,
626
+ 0.42110398411750793
627
+ ],
628
+ [
629
+ 0.3344976007938385,
630
+ 0.29923370480537415,
631
+ 0.3550164997577667
632
+ ],
633
+ [
634
+ 0.22152170538902283,
635
+ 0.25506770610809326,
636
+ 0.19674324989318848
637
+ ],
638
+ [
639
+ 0.2944912910461426,
640
+ 0.40148842334747314,
641
+ 0.4079500734806061
642
+ ],
643
+ [
644
+ 0.31500399112701416,
645
+ 0.264303594827652,
646
+ 0.24273818731307983
647
+ ],
648
+ [
649
+ 0.24160929024219513,
650
+ 0.43865495920181274,
651
+ 0.17754635214805603
652
+ ],
653
+ [
654
+ 0.35719066858291626,
655
+ 0.4569880962371826,
656
+ 0.10947935283184052
657
+ ],
658
+ [
659
+ 0.17870254814624786,
660
+ 0.15666398406028748,
661
+ 0.1511623114347458
662
+ ],
663
+ [
664
+ 0.1309283971786499,
665
+ 0.39916273951530457,
666
+ 0.35891109704971313
667
+ ],
668
+ [
669
+ 0.41414639353752136,
670
+ 0.40458348393440247,
671
+ 0.4373874068260193
672
+ ],
673
+ [
674
+ 0.223582461476326,
675
+ 0.20872588455677032,
676
+ 0.20849093794822693
677
+ ],
678
+ [
679
+ 0.15978248417377472,
680
+ 0.3871738910675049,
681
+ 0.20173275470733643
682
+ ],
683
+ [
684
+ 0.40426936745643616,
685
+ 0.17517821490764618,
686
+ 0.15639956295490265
687
+ ],
688
+ [
689
+ 0.2445727288722992,
690
+ 0.24165970087051392,
691
+ 0.26430395245552063
692
+ ],
693
+ [
694
+ 0.27058425545692444,
695
+ 0.45792025327682495,
696
+ 0.22470559179782867
697
+ ],
698
+ [
699
+ 0.1343778669834137,
700
+ 0.23025086522102356,
701
+ 0.4482075273990631
702
+ ],
703
+ [
704
+ 0.301599383354187,
705
+ 0.4401021897792816,
706
+ 0.30802470445632935
707
+ ],
708
+ [
709
+ 0.3678591847419739,
710
+ 0.41126516461372375,
711
+ 0.12729883193969727
712
+ ],
713
+ [
714
+ 0.4265497028827667,
715
+ 0.3064115643501282,
716
+ 0.12861239910125732
717
+ ],
718
+ [
719
+ 0.25590309500694275,
720
+ 0.29923370480537415,
721
+ 0.40055426955223083
722
+ ],
723
+ [
724
+ 0.21432910859584808,
725
+ 0.35219594836235046,
726
+ 0.3894593119621277
727
+ ],
728
+ [
729
+ 0.16293656826019287,
730
+ 0.20112387835979462,
731
+ 0.1641329973936081
732
+ ],
733
+ [
734
+ 0.23716431856155396,
735
+ 0.23458261787891388,
736
+ 0.22277328372001648
737
+ ],
738
+ [
739
+ 0.16283032298088074,
740
+ 0.24371109902858734,
741
+ 0.18812552094459534
742
+ ],
743
+ [
744
+ 0.17482346296310425,
745
+ 0.27810290455818176,
746
+ 0.2653261721134186
747
+ ],
748
+ [
749
+ 0.44124141335487366,
750
+ 0.4653926193714142,
751
+ 0.18024171888828278
752
+ ],
753
+ [
754
+ 0.4573570489883423,
755
+ 0.23765107989311218,
756
+ 0.223820760846138
757
+ ],
758
+ [
759
+ 0.48433545231819153,
760
+ 0.45144760608673096,
761
+ 0.2777790427207947
762
+ ],
763
+ [
764
+ 0.1559525728225708,
765
+ 0.29923370480537415,
766
+ 0.4903254210948944
767
+ ],
768
+ [
769
+ 0.3679139316082001,
770
+ 0.40922513604164124,
771
+ 0.14046727120876312
772
+ ],
773
+ [
774
+ 0.3210708498954773,
775
+ 0.16060097515583038,
776
+ 0.435181587934494
777
+ ],
778
+ [
779
+ 0.2536516785621643,
780
+ 0.23436033725738525,
781
+ 0.27494579553604126
782
+ ],
783
+ [
784
+ 0.14041517674922943,
785
+ 0.45182326436042786,
786
+ 0.16119162738323212
787
+ ],
788
+ [
789
+ 0.14711785316467285,
790
+ 0.1674029380083084,
791
+ 0.15063847601413727
792
+ ],
793
+ [
794
+ 0.36876657605171204,
795
+ 0.48956817388534546,
796
+ 0.3209514319896698
797
+ ],
798
+ [
799
+ 0.3705935776233673,
800
+ 0.21650540828704834,
801
+ 0.4331192374229431
802
+ ],
803
+ [
804
+ 0.4736544191837311,
805
+ 0.4322001338005066,
806
+ 0.12574197351932526
807
+ ],
808
+ [
809
+ 0.245101198554039,
810
+ 0.29923370480537415,
811
+ 0.40264415740966797
812
+ ],
813
+ [
814
+ 0.44746536016464233,
815
+ 0.33772048354148865,
816
+ 0.1483197808265686
817
+ ],
818
+ [
819
+ 0.3145650029182434,
820
+ 0.44416722655296326,
821
+ 0.35693588852882385
822
+ ],
823
+ [
824
+ 0.37297913432121277,
825
+ 0.18011485040187836,
826
+ 0.16238778829574585
827
+ ],
828
+ [
829
+ 0.26034584641456604,
830
+ 0.4297695755958557,
831
+ 0.32993659377098083
832
+ ],
833
+ [
834
+ 0.37756991386413574,
835
+ 0.44577649235725403,
836
+ 0.44806650280952454
837
+ ],
838
+ [
839
+ 0.0845801904797554,
840
+ 0.23957164585590363,
841
+ 0.12926174700260162
842
+ ],
843
+ [
844
+ 0.13752779364585876,
845
+ 0.18119001388549805,
846
+ 0.19948138296604156
847
+ ],
848
+ [
849
+ 0.10815407335758209,
850
+ 0.46910011768341064,
851
+ 0.4763980805873871
852
+ ],
853
+ [
854
+ 0.19809484481811523,
855
+ 0.29923370480537415,
856
+ 0.366342157125473
857
+ ],
858
+ [
859
+ 0.18643337488174438,
860
+ 0.45779895782470703,
861
+ 0.48765724897384644
862
+ ],
863
+ [
864
+ 0.13408148288726807,
865
+ 0.41982096433639526,
866
+ 0.19417880475521088
867
+ ],
868
+ [
869
+ 0.2252083420753479,
870
+ 0.16844086349010468,
871
+ 0.26387283205986023
872
+ ],
873
+ [
874
+ 0.2037154585123062,
875
+ 0.20130690932273865,
876
+ 0.19506138563156128
877
+ ],
878
+ [
879
+ 0.18133416771888733,
880
+ 0.3889558017253876,
881
+ 0.17258818447589874
882
+ ],
883
+ [
884
+ 0.4873719811439514,
885
+ 0.43230339884757996,
886
+ 0.21695755422115326
887
+ ],
888
+ [
889
+ 0.4797218143939972,
890
+ 0.10397535562515259,
891
+ 0.12909311056137085
892
+ ],
893
+ [
894
+ 0.23951224982738495,
895
+ 0.10424715280532837,
896
+ 0.45742061734199524
897
+ ],
898
+ [
899
+ 0.2841903865337372,
900
+ 0.29923370480537415,
901
+ 0.30746060609817505
902
+ ],
903
+ [
904
+ 0.2450842559337616,
905
+ 0.23246590793132782,
906
+ 0.48158928751945496
907
+ ],
908
+ [
909
+ 0.3935782015323639,
910
+ 0.3982652723789215,
911
+ 0.44564884901046753
912
+ ],
913
+ [
914
+ 0.42746949195861816,
915
+ 0.43627870082855225,
916
+ 0.41870421171188354
917
+ ],
918
+ [
919
+ 0.1621244102716446,
920
+ 0.18475289642810822,
921
+ 0.1475932002067566
922
+ ],
923
+ [
924
+ 0.21315304934978485,
925
+ 0.17851777374744415,
926
+ 0.4540901780128479
927
+ ],
928
+ [
929
+ 0.459518700838089,
930
+ 0.46814388036727905,
931
+ 0.12805859744548798
932
+ ],
933
+ [
934
+ 0.4473622441291809,
935
+ 0.1991201937198639,
936
+ 0.45949587225914
937
+ ],
938
+ [
939
+ 0.3688053488731384,
940
+ 0.2689209282398224,
941
+ 0.490568071603775
942
+ ],
943
+ [
944
+ 0.4463302791118622,
945
+ 0.29923370480537415,
946
+ 0.363638699054718
947
+ ],
948
+ [
949
+ 0.07843068242073059,
950
+ 0.1257428079843521,
951
+ 0.47206512093544006
952
+ ],
953
+ [
954
+ 0.2896142303943634,
955
+ 0.48443368077278137,
956
+ 0.1995886117219925
957
+ ],
958
+ [
959
+ 0.23569075763225555,
960
+ 0.17438194155693054,
961
+ 0.16555899381637573
962
+ ],
963
+ [
964
+ 0.3339311480522156,
965
+ 0.138455331325531,
966
+ 0.4986150860786438
967
+ ],
968
+ [
969
+ 0.4561089277267456,
970
+ 0.13907161355018616,
971
+ 0.37022289633750916
972
+ ],
973
+ [
974
+ 0.11571026593446732,
975
+ 0.1735377013683319,
976
+ 0.3909721076488495
977
+ ],
978
+ [
979
+ 0.41012808680534363,
980
+ 0.12148356437683105,
981
+ 0.13725464046001434
982
+ ],
983
+ [
984
+ 0.4633272588253021,
985
+ 0.1216013953089714,
986
+ 0.36428725719451904
987
+ ],
988
+ [
989
+ 0.13796786963939667,
990
+ 0.29923370480537415,
991
+ 0.46274399757385254
992
+ ],
993
+ [
994
+ 0.45114168524742126,
995
+ 0.14195981621742249,
996
+ 0.46917733550071716
997
+ ],
998
+ [
999
+ 0.4262934625148773,
1000
+ 0.11466571688652039,
1001
+ 0.19186291098594666
1002
+ ],
1003
+ [
1004
+ 0.44846612215042114,
1005
+ 0.14936070144176483,
1006
+ 0.29932039976119995
1007
+ ],
1008
+ [
1009
+ 0.1297784298658371,
1010
+ 0.36136552691459656,
1011
+ 0.1830536276102066
1012
+ ],
1013
+ [
1014
+ 0.25918319821357727,
1015
+ 0.3930901885032654,
1016
+ 0.42164939641952515
1017
+ ],
1018
+ [
1019
+ 0.454575777053833,
1020
+ 0.48423320055007935,
1021
+ 0.18333756923675537
1022
+ ],
1023
+ [
1024
+ 0.4566199481487274,
1025
+ 0.2628677785396576,
1026
+ 0.07379919290542603
1027
+ ],
1028
+ [
1029
+ 0.46558135747909546,
1030
+ 0.4114809036254883,
1031
+ 0.467506468296051
1032
+ ],
1033
+ [
1034
+ 0.3389800786972046,
1035
+ 0.28094711899757385,
1036
+ 0.23497305810451508
1037
+ ],
1038
+ [
1039
+ 0.46289947628974915,
1040
+ 0.17742177844047546,
1041
+ 0.14021877944469452
1042
+ ],
1043
+ [
1044
+ 0.09659633785486221,
1045
+ 0.13616475462913513,
1046
+ 0.14357511699199677
1047
+ ],
1048
+ [
1049
+ 0.350506991147995,
1050
+ 0.10470135509967804,
1051
+ 0.12466616928577423
1052
+ ],
1053
+ [
1054
+ 0.1182580217719078,
1055
+ 0.42097800970077515,
1056
+ 0.13212710618972778
1057
+ ],
1058
+ [
1059
+ 0.5112250447273254,
1060
+ 0.15854305028915405,
1061
+ 0.1426139622926712
1062
+ ],
1063
+ [
1064
+ 0.20451374351978302,
1065
+ 0.4863293766975403,
1066
+ 0.294251412153244
1067
+ ],
1068
+ [
1069
+ 0.48365139961242676,
1070
+ 0.41005825996398926,
1071
+ 0.11694485694169998
1072
+ ],
1073
+ [
1074
+ 0.24377882480621338,
1075
+ 0.1884140521287918,
1076
+ 0.48632267117500305
1077
+ ],
1078
+ [
1079
+ 0.31411013007164,
1080
+ 0.29923370480537415,
1081
+ 0.18183420598506927
1082
+ ],
1083
+ [
1084
+ 0.19102278351783752,
1085
+ 0.20111915469169617,
1086
+ 0.1339617818593979
1087
+ ],
1088
+ [
1089
+ 0.26393866539001465,
1090
+ 0.22674188017845154,
1091
+ 0.4447513222694397
1092
+ ],
1093
+ [
1094
+ 0.1592191904783249,
1095
+ 0.2599426805973053,
1096
+ 0.21397040784358978
1097
+ ],
1098
+ [
1099
+ 0.10093661397695541,
1100
+ 0.31607648730278015,
1101
+ 0.2935095429420471
1102
+ ],
1103
+ [
1104
+ 0.4517417848110199,
1105
+ 0.11916311085224152,
1106
+ 0.4487104117870331
1107
+ ],
1108
+ [
1109
+ 0.414072185754776,
1110
+ 0.4277026653289795,
1111
+ 0.4792312979698181
1112
+ ],
1113
+ [
1114
+ 0.4864226281642914,
1115
+ 0.37999510765075684,
1116
+ 0.4680911898612976
1117
+ ],
1118
+ [
1119
+ 0.4162227213382721,
1120
+ 0.496494323015213,
1121
+ 0.4781576097011566
1122
+ ],
1123
+ [
1124
+ 0.1922696977853775,
1125
+ 0.41444700956344604,
1126
+ 0.457570344209671
1127
+ ],
1128
+ [
1129
+ 0.47464364767074585,
1130
+ 0.40973153710365295,
1131
+ 0.4741663932800293
1132
+ ],
1133
+ [
1134
+ 0.39086464047431946,
1135
+ 0.4758678078651428,
1136
+ 0.37202033400535583
1137
+ ],
1138
+ [
1139
+ 0.45488241314888,
1140
+ 0.38954630494117737,
1141
+ 0.35717692971229553
1142
+ ],
1143
+ [
1144
+ 0.4128759801387787,
1145
+ 0.18781141936779022,
1146
+ 0.18855643272399902
1147
+ ],
1148
+ [
1149
+ 0.4449164867401123,
1150
+ 0.24285383522510529,
1151
+ 0.48864275217056274
1152
+ ],
1153
+ [
1154
+ 0.08543943613767624,
1155
+ 0.4856216013431549,
1156
+ 0.4056006669998169
1157
+ ],
1158
+ [
1159
+ 0.4745456576347351,
1160
+ 0.08682437986135483,
1161
+ 0.4459170699119568
1162
+ ],
1163
+ [
1164
+ 0.10274919122457504,
1165
+ 0.48529237508773804,
1166
+ 0.47913098335266113
1167
+ ],
1168
+ [
1169
+ 0.46872222423553467,
1170
+ 0.18523962795734406,
1171
+ 0.15047748386859894
1172
+ ],
1173
+ [
1174
+ 0.39252278208732605,
1175
+ 0.39467936754226685,
1176
+ 0.5171827077865601
1177
+ ],
1178
+ [
1179
+ 0.46476298570632935,
1180
+ 0.47057217359542847,
1181
+ 0.13309550285339355
1182
+ ],
1183
+ [
1184
+ 0.44297391176223755,
1185
+ 0.23084476590156555,
1186
+ 0.1847122311592102
1187
+ ],
1188
+ [
1189
+ 0.4511570930480957,
1190
+ 0.37978699803352356,
1191
+ 0.4634334444999695
1192
+ ],
1193
+ [
1194
+ 0.4489362835884094,
1195
+ 0.4209038019180298,
1196
+ 0.1906983107328415
1197
+ ],
1198
+ [
1199
+ 0.292703241109848,
1200
+ 0.48317569494247437,
1201
+ 0.4351253807544708
1202
+ ],
1203
+ [
1204
+ 0.4513009488582611,
1205
+ 0.11310037225484848,
1206
+ 0.41090840101242065
1207
+ ],
1208
+ [
1209
+ 0.15975885093212128,
1210
+ 0.13482822477817535,
1211
+ 0.4434134364128113
1212
+ ],
1213
+ [
1214
+ 0.46823757886886597,
1215
+ 0.288373738527298,
1216
+ 0.36955058574676514
1217
+ ],
1218
+ [
1219
+ 0.32527464628219604,
1220
+ 0.3661269247531891,
1221
+ 0.1951110064983368
1222
+ ],
1223
+ [
1224
+ 0.4789314270019531,
1225
+ 0.07736703008413315,
1226
+ 0.2624352276325226
1227
+ ],
1228
+ [
1229
+ 0.3614385724067688,
1230
+ 0.13425564765930176,
1231
+ 0.2201119214296341
1232
+ ],
1233
+ [
1234
+ 0.13200026750564575,
1235
+ 0.322633296251297,
1236
+ 0.3131023943424225
1237
+ ],
1238
+ [
1239
+ 0.4229698181152344,
1240
+ 0.175356924533844,
1241
+ 0.18913114070892334
1242
+ ],
1243
+ [
1244
+ 0.44432491064071655,
1245
+ 0.49774083495140076,
1246
+ 0.22631461918354034
1247
+ ],
1248
+ [
1249
+ 0.4541386663913727,
1250
+ 0.10495408624410629,
1251
+ 0.43360456824302673
1252
+ ],
1253
+ [
1254
+ 0.15666554868221283,
1255
+ 0.15558116137981415,
1256
+ 0.2319335788488388
1257
+ ],
1258
+ [
1259
+ 0.23508504033088684,
1260
+ 0.29923370480537415,
1261
+ 0.1499142199754715
1262
+ ],
1263
+ [
1264
+ 0.39173614978790283,
1265
+ 0.15001919865608215,
1266
+ 0.1313292533159256
1267
+ ],
1268
+ [
1269
+ 0.41128212213516235,
1270
+ 0.4478446841239929,
1271
+ 0.3802175223827362
1272
+ ],
1273
+ [
1274
+ 0.4341309368610382,
1275
+ 0.15578855574131012,
1276
+ 0.20559746026992798
1277
+ ],
1278
+ [
1279
+ 0.14219129085540771,
1280
+ 0.14060185849666595,
1281
+ 0.20822055637836456
1282
+ ],
1283
+ [
1284
+ 0.4654639959335327,
1285
+ 0.3616563379764557,
1286
+ 0.3889743983745575
1287
+ ],
1288
+ [
1289
+ 0.24674555659294128,
1290
+ 0.42591360211372375,
1291
+ 0.09854333102703094
1292
+ ],
1293
+ [
1294
+ 0.4760911762714386,
1295
+ 0.08101151138544083,
1296
+ 0.44511404633522034
1297
+ ],
1298
+ [
1299
+ 0.15141984820365906,
1300
+ 0.4241148829460144,
1301
+ 0.4395207166671753
1302
+ ],
1303
+ [
1304
+ 0.40674036741256714,
1305
+ 0.3935491144657135,
1306
+ 0.1797652244567871
1307
+ ],
1308
+ [
1309
+ 0.4640699625015259,
1310
+ 0.35606488585472107,
1311
+ 0.48520034551620483
1312
+ ],
1313
+ [
1314
+ 0.45517098903656006,
1315
+ 0.10837940871715546,
1316
+ 0.3634328544139862
1317
+ ],
1318
+ [
1319
+ 0.42892348766326904,
1320
+ 0.3894287049770355,
1321
+ 0.4179135262966156
1322
+ ],
1323
+ [
1324
+ 0.1619490683078766,
1325
+ 0.1275283247232437,
1326
+ 0.4848206639289856
1327
+ ],
1328
+ [
1329
+ 0.4587709605693817,
1330
+ 0.3761414885520935,
1331
+ 0.4999237358570099
1332
+ ],
1333
+ [
1334
+ 0.45937579870224,
1335
+ 0.10048946738243103,
1336
+ 0.3620624542236328
1337
+ ],
1338
+ [
1339
+ 0.1622796505689621,
1340
+ 0.06675637513399124,
1341
+ 0.4629978835582733
1342
+ ],
1343
+ [
1344
+ 0.14350159466266632,
1345
+ 0.3432171046733856,
1346
+ 0.08686857670545578
1347
+ ],
1348
+ [
1349
+ 0.2418466955423355,
1350
+ 0.29923370480537415,
1351
+ 0.15336613357067108
1352
+ ],
1353
+ [
1354
+ 0.10652916133403778,
1355
+ 0.3400610387325287,
1356
+ 0.5012139081954956
1357
+ ],
1358
+ [
1359
+ 0.470577597618103,
1360
+ 0.4068159759044647,
1361
+ 0.18197748064994812
1362
+ ],
1363
+ [
1364
+ 0.16513541340827942,
1365
+ 0.3945123255252838,
1366
+ 0.3649037182331085
1367
+ ],
1368
+ [
1369
+ 0.169801726937294,
1370
+ 0.5046741366386414,
1371
+ 0.4489770233631134
1372
+ ],
1373
+ [
1374
+ 0.381733775138855,
1375
+ 0.42710191011428833,
1376
+ 0.46337172389030457
1377
+ ],
1378
+ [
1379
+ 0.2477995753288269,
1380
+ 0.1106489896774292,
1381
+ 0.13583549857139587
1382
+ ],
1383
+ [
1384
+ 0.09561775624752045,
1385
+ 0.48951175808906555,
1386
+ 0.19908744096755981
1387
+ ],
1388
+ [
1389
+ 0.13161425292491913,
1390
+ 0.1405383050441742,
1391
+ 0.3670024275779724
1392
+ ],
1393
+ [
1394
+ 0.27499905228614807,
1395
+ 0.28686726093292236,
1396
+ 0.2851425111293793
1397
+ ],
1398
+ [
1399
+ 0.10630062967538834,
1400
+ 0.4811836779117584,
1401
+ 0.47763416171073914
1402
+ ],
1403
+ [
1404
+ 0.11440788954496384,
1405
+ 0.31618937849998474,
1406
+ 0.3314228653907776
1407
+ ],
1408
+ [
1409
+ 0.19934608042240143,
1410
+ 0.1537989377975464,
1411
+ 0.1396467685699463
1412
+ ],
1413
+ [
1414
+ 0.12085483968257904,
1415
+ 0.49329978227615356,
1416
+ 0.4896600842475891
1417
+ ],
1418
+ [
1419
+ 0.1220579594373703,
1420
+ 0.4949476718902588,
1421
+ 0.5103369951248169
1422
+ ],
1423
+ [
1424
+ 0.4522629380226135,
1425
+ 0.16088415682315826,
1426
+ 0.23696646094322205
1427
+ ],
1428
+ [
1429
+ 0.4699283838272095,
1430
+ 0.17811758816242218,
1431
+ 0.12464112788438797
1432
+ ],
1433
+ [
1434
+ 0.11091146618127823,
1435
+ 0.48623692989349365,
1436
+ 0.4849981367588043
1437
+ ],
1438
+ [
1439
+ 0.4842935800552368,
1440
+ 0.29923370480537415,
1441
+ 0.06224353238940239
1442
+ ],
1443
+ [
1444
+ 0.06589362770318985,
1445
+ 0.29923370480537415,
1446
+ 0.44713032245635986
1447
+ ],
1448
+ [
1449
+ 0.5308268070220947,
1450
+ 0.1069527193903923,
1451
+ 0.07115095853805542
1452
+ ],
1453
+ [
1454
+ 0.510724663734436,
1455
+ 0.4425855875015259,
1456
+ 0.49699831008911133
1457
+ ]
1458
+ ],
1459
+ "model_names": [
1460
+ "instruct",
1461
+ "math",
1462
+ "code"
1463
+ ],
1464
+ "num_models": 3,
1465
+ "num_params": 291,
1466
+ "param_names": [
1467
+ "model.embed_tokens.weight",
1468
+ "model.layers.0.self_attn.q_proj.weight",
1469
+ "model.layers.0.self_attn.k_proj.weight",
1470
+ "model.layers.0.self_attn.v_proj.weight",
1471
+ "model.layers.0.self_attn.o_proj.weight",
1472
+ "model.layers.0.mlp.gate_proj.weight",
1473
+ "model.layers.0.mlp.up_proj.weight",
1474
+ "model.layers.0.mlp.down_proj.weight",
1475
+ "model.layers.0.input_layernorm.weight",
1476
+ "model.layers.0.post_attention_layernorm.weight",
1477
+ "model.layers.1.self_attn.q_proj.weight",
1478
+ "model.layers.1.self_attn.k_proj.weight",
1479
+ "model.layers.1.self_attn.v_proj.weight",
1480
+ "model.layers.1.self_attn.o_proj.weight",
1481
+ "model.layers.1.mlp.gate_proj.weight",
1482
+ "model.layers.1.mlp.up_proj.weight",
1483
+ "model.layers.1.mlp.down_proj.weight",
1484
+ "model.layers.1.input_layernorm.weight",
1485
+ "model.layers.1.post_attention_layernorm.weight",
1486
+ "model.layers.2.self_attn.q_proj.weight",
1487
+ "model.layers.2.self_attn.k_proj.weight",
1488
+ "model.layers.2.self_attn.v_proj.weight",
1489
+ "model.layers.2.self_attn.o_proj.weight",
1490
+ "model.layers.2.mlp.gate_proj.weight",
1491
+ "model.layers.2.mlp.up_proj.weight",
1492
+ "model.layers.2.mlp.down_proj.weight",
1493
+ "model.layers.2.input_layernorm.weight",
1494
+ "model.layers.2.post_attention_layernorm.weight",
1495
+ "model.layers.3.self_attn.q_proj.weight",
1496
+ "model.layers.3.self_attn.k_proj.weight",
1497
+ "model.layers.3.self_attn.v_proj.weight",
1498
+ "model.layers.3.self_attn.o_proj.weight",
1499
+ "model.layers.3.mlp.gate_proj.weight",
1500
+ "model.layers.3.mlp.up_proj.weight",
1501
+ "model.layers.3.mlp.down_proj.weight",
1502
+ "model.layers.3.input_layernorm.weight",
1503
+ "model.layers.3.post_attention_layernorm.weight",
1504
+ "model.layers.4.self_attn.q_proj.weight",
1505
+ "model.layers.4.self_attn.k_proj.weight",
1506
+ "model.layers.4.self_attn.v_proj.weight",
1507
+ "model.layers.4.self_attn.o_proj.weight",
1508
+ "model.layers.4.mlp.gate_proj.weight",
1509
+ "model.layers.4.mlp.up_proj.weight",
1510
+ "model.layers.4.mlp.down_proj.weight",
1511
+ "model.layers.4.input_layernorm.weight",
1512
+ "model.layers.4.post_attention_layernorm.weight",
1513
+ "model.layers.5.self_attn.q_proj.weight",
1514
+ "model.layers.5.self_attn.k_proj.weight",
1515
+ "model.layers.5.self_attn.v_proj.weight",
1516
+ "model.layers.5.self_attn.o_proj.weight",
1517
+ "model.layers.5.mlp.gate_proj.weight",
1518
+ "model.layers.5.mlp.up_proj.weight",
1519
+ "model.layers.5.mlp.down_proj.weight",
1520
+ "model.layers.5.input_layernorm.weight",
1521
+ "model.layers.5.post_attention_layernorm.weight",
1522
+ "model.layers.6.self_attn.q_proj.weight",
1523
+ "model.layers.6.self_attn.k_proj.weight",
1524
+ "model.layers.6.self_attn.v_proj.weight",
1525
+ "model.layers.6.self_attn.o_proj.weight",
1526
+ "model.layers.6.mlp.gate_proj.weight",
1527
+ "model.layers.6.mlp.up_proj.weight",
1528
+ "model.layers.6.mlp.down_proj.weight",
1529
+ "model.layers.6.input_layernorm.weight",
1530
+ "model.layers.6.post_attention_layernorm.weight",
1531
+ "model.layers.7.self_attn.q_proj.weight",
1532
+ "model.layers.7.self_attn.k_proj.weight",
1533
+ "model.layers.7.self_attn.v_proj.weight",
1534
+ "model.layers.7.self_attn.o_proj.weight",
1535
+ "model.layers.7.mlp.gate_proj.weight",
1536
+ "model.layers.7.mlp.up_proj.weight",
1537
+ "model.layers.7.mlp.down_proj.weight",
1538
+ "model.layers.7.input_layernorm.weight",
1539
+ "model.layers.7.post_attention_layernorm.weight",
1540
+ "model.layers.8.self_attn.q_proj.weight",
1541
+ "model.layers.8.self_attn.k_proj.weight",
1542
+ "model.layers.8.self_attn.v_proj.weight",
1543
+ "model.layers.8.self_attn.o_proj.weight",
1544
+ "model.layers.8.mlp.gate_proj.weight",
1545
+ "model.layers.8.mlp.up_proj.weight",
1546
+ "model.layers.8.mlp.down_proj.weight",
1547
+ "model.layers.8.input_layernorm.weight",
1548
+ "model.layers.8.post_attention_layernorm.weight",
1549
+ "model.layers.9.self_attn.q_proj.weight",
1550
+ "model.layers.9.self_attn.k_proj.weight",
1551
+ "model.layers.9.self_attn.v_proj.weight",
1552
+ "model.layers.9.self_attn.o_proj.weight",
1553
+ "model.layers.9.mlp.gate_proj.weight",
1554
+ "model.layers.9.mlp.up_proj.weight",
1555
+ "model.layers.9.mlp.down_proj.weight",
1556
+ "model.layers.9.input_layernorm.weight",
1557
+ "model.layers.9.post_attention_layernorm.weight",
1558
+ "model.layers.10.self_attn.q_proj.weight",
1559
+ "model.layers.10.self_attn.k_proj.weight",
1560
+ "model.layers.10.self_attn.v_proj.weight",
1561
+ "model.layers.10.self_attn.o_proj.weight",
1562
+ "model.layers.10.mlp.gate_proj.weight",
1563
+ "model.layers.10.mlp.up_proj.weight",
1564
+ "model.layers.10.mlp.down_proj.weight",
1565
+ "model.layers.10.input_layernorm.weight",
1566
+ "model.layers.10.post_attention_layernorm.weight",
1567
+ "model.layers.11.self_attn.q_proj.weight",
1568
+ "model.layers.11.self_attn.k_proj.weight",
1569
+ "model.layers.11.self_attn.v_proj.weight",
1570
+ "model.layers.11.self_attn.o_proj.weight",
1571
+ "model.layers.11.mlp.gate_proj.weight",
1572
+ "model.layers.11.mlp.up_proj.weight",
1573
+ "model.layers.11.mlp.down_proj.weight",
1574
+ "model.layers.11.input_layernorm.weight",
1575
+ "model.layers.11.post_attention_layernorm.weight",
1576
+ "model.layers.12.self_attn.q_proj.weight",
1577
+ "model.layers.12.self_attn.k_proj.weight",
1578
+ "model.layers.12.self_attn.v_proj.weight",
1579
+ "model.layers.12.self_attn.o_proj.weight",
1580
+ "model.layers.12.mlp.gate_proj.weight",
1581
+ "model.layers.12.mlp.up_proj.weight",
1582
+ "model.layers.12.mlp.down_proj.weight",
1583
+ "model.layers.12.input_layernorm.weight",
1584
+ "model.layers.12.post_attention_layernorm.weight",
1585
+ "model.layers.13.self_attn.q_proj.weight",
1586
+ "model.layers.13.self_attn.k_proj.weight",
1587
+ "model.layers.13.self_attn.v_proj.weight",
1588
+ "model.layers.13.self_attn.o_proj.weight",
1589
+ "model.layers.13.mlp.gate_proj.weight",
1590
+ "model.layers.13.mlp.up_proj.weight",
1591
+ "model.layers.13.mlp.down_proj.weight",
1592
+ "model.layers.13.input_layernorm.weight",
1593
+ "model.layers.13.post_attention_layernorm.weight",
1594
+ "model.layers.14.self_attn.q_proj.weight",
1595
+ "model.layers.14.self_attn.k_proj.weight",
1596
+ "model.layers.14.self_attn.v_proj.weight",
1597
+ "model.layers.14.self_attn.o_proj.weight",
1598
+ "model.layers.14.mlp.gate_proj.weight",
1599
+ "model.layers.14.mlp.up_proj.weight",
1600
+ "model.layers.14.mlp.down_proj.weight",
1601
+ "model.layers.14.input_layernorm.weight",
1602
+ "model.layers.14.post_attention_layernorm.weight",
1603
+ "model.layers.15.self_attn.q_proj.weight",
1604
+ "model.layers.15.self_attn.k_proj.weight",
1605
+ "model.layers.15.self_attn.v_proj.weight",
1606
+ "model.layers.15.self_attn.o_proj.weight",
1607
+ "model.layers.15.mlp.gate_proj.weight",
1608
+ "model.layers.15.mlp.up_proj.weight",
1609
+ "model.layers.15.mlp.down_proj.weight",
1610
+ "model.layers.15.input_layernorm.weight",
1611
+ "model.layers.15.post_attention_layernorm.weight",
1612
+ "model.layers.16.self_attn.q_proj.weight",
1613
+ "model.layers.16.self_attn.k_proj.weight",
1614
+ "model.layers.16.self_attn.v_proj.weight",
1615
+ "model.layers.16.self_attn.o_proj.weight",
1616
+ "model.layers.16.mlp.gate_proj.weight",
1617
+ "model.layers.16.mlp.up_proj.weight",
1618
+ "model.layers.16.mlp.down_proj.weight",
1619
+ "model.layers.16.input_layernorm.weight",
1620
+ "model.layers.16.post_attention_layernorm.weight",
1621
+ "model.layers.17.self_attn.q_proj.weight",
1622
+ "model.layers.17.self_attn.k_proj.weight",
1623
+ "model.layers.17.self_attn.v_proj.weight",
1624
+ "model.layers.17.self_attn.o_proj.weight",
1625
+ "model.layers.17.mlp.gate_proj.weight",
1626
+ "model.layers.17.mlp.up_proj.weight",
1627
+ "model.layers.17.mlp.down_proj.weight",
1628
+ "model.layers.17.input_layernorm.weight",
1629
+ "model.layers.17.post_attention_layernorm.weight",
1630
+ "model.layers.18.self_attn.q_proj.weight",
1631
+ "model.layers.18.self_attn.k_proj.weight",
1632
+ "model.layers.18.self_attn.v_proj.weight",
1633
+ "model.layers.18.self_attn.o_proj.weight",
1634
+ "model.layers.18.mlp.gate_proj.weight",
1635
+ "model.layers.18.mlp.up_proj.weight",
1636
+ "model.layers.18.mlp.down_proj.weight",
1637
+ "model.layers.18.input_layernorm.weight",
1638
+ "model.layers.18.post_attention_layernorm.weight",
1639
+ "model.layers.19.self_attn.q_proj.weight",
1640
+ "model.layers.19.self_attn.k_proj.weight",
1641
+ "model.layers.19.self_attn.v_proj.weight",
1642
+ "model.layers.19.self_attn.o_proj.weight",
1643
+ "model.layers.19.mlp.gate_proj.weight",
1644
+ "model.layers.19.mlp.up_proj.weight",
1645
+ "model.layers.19.mlp.down_proj.weight",
1646
+ "model.layers.19.input_layernorm.weight",
1647
+ "model.layers.19.post_attention_layernorm.weight",
1648
+ "model.layers.20.self_attn.q_proj.weight",
1649
+ "model.layers.20.self_attn.k_proj.weight",
1650
+ "model.layers.20.self_attn.v_proj.weight",
1651
+ "model.layers.20.self_attn.o_proj.weight",
1652
+ "model.layers.20.mlp.gate_proj.weight",
1653
+ "model.layers.20.mlp.up_proj.weight",
1654
+ "model.layers.20.mlp.down_proj.weight",
1655
+ "model.layers.20.input_layernorm.weight",
1656
+ "model.layers.20.post_attention_layernorm.weight",
1657
+ "model.layers.21.self_attn.q_proj.weight",
1658
+ "model.layers.21.self_attn.k_proj.weight",
1659
+ "model.layers.21.self_attn.v_proj.weight",
1660
+ "model.layers.21.self_attn.o_proj.weight",
1661
+ "model.layers.21.mlp.gate_proj.weight",
1662
+ "model.layers.21.mlp.up_proj.weight",
1663
+ "model.layers.21.mlp.down_proj.weight",
1664
+ "model.layers.21.input_layernorm.weight",
1665
+ "model.layers.21.post_attention_layernorm.weight",
1666
+ "model.layers.22.self_attn.q_proj.weight",
1667
+ "model.layers.22.self_attn.k_proj.weight",
1668
+ "model.layers.22.self_attn.v_proj.weight",
1669
+ "model.layers.22.self_attn.o_proj.weight",
1670
+ "model.layers.22.mlp.gate_proj.weight",
1671
+ "model.layers.22.mlp.up_proj.weight",
1672
+ "model.layers.22.mlp.down_proj.weight",
1673
+ "model.layers.22.input_layernorm.weight",
1674
+ "model.layers.22.post_attention_layernorm.weight",
1675
+ "model.layers.23.self_attn.q_proj.weight",
1676
+ "model.layers.23.self_attn.k_proj.weight",
1677
+ "model.layers.23.self_attn.v_proj.weight",
1678
+ "model.layers.23.self_attn.o_proj.weight",
1679
+ "model.layers.23.mlp.gate_proj.weight",
1680
+ "model.layers.23.mlp.up_proj.weight",
1681
+ "model.layers.23.mlp.down_proj.weight",
1682
+ "model.layers.23.input_layernorm.weight",
1683
+ "model.layers.23.post_attention_layernorm.weight",
1684
+ "model.layers.24.self_attn.q_proj.weight",
1685
+ "model.layers.24.self_attn.k_proj.weight",
1686
+ "model.layers.24.self_attn.v_proj.weight",
1687
+ "model.layers.24.self_attn.o_proj.weight",
1688
+ "model.layers.24.mlp.gate_proj.weight",
1689
+ "model.layers.24.mlp.up_proj.weight",
1690
+ "model.layers.24.mlp.down_proj.weight",
1691
+ "model.layers.24.input_layernorm.weight",
1692
+ "model.layers.24.post_attention_layernorm.weight",
1693
+ "model.layers.25.self_attn.q_proj.weight",
1694
+ "model.layers.25.self_attn.k_proj.weight",
1695
+ "model.layers.25.self_attn.v_proj.weight",
1696
+ "model.layers.25.self_attn.o_proj.weight",
1697
+ "model.layers.25.mlp.gate_proj.weight",
1698
+ "model.layers.25.mlp.up_proj.weight",
1699
+ "model.layers.25.mlp.down_proj.weight",
1700
+ "model.layers.25.input_layernorm.weight",
1701
+ "model.layers.25.post_attention_layernorm.weight",
1702
+ "model.layers.26.self_attn.q_proj.weight",
1703
+ "model.layers.26.self_attn.k_proj.weight",
1704
+ "model.layers.26.self_attn.v_proj.weight",
1705
+ "model.layers.26.self_attn.o_proj.weight",
1706
+ "model.layers.26.mlp.gate_proj.weight",
1707
+ "model.layers.26.mlp.up_proj.weight",
1708
+ "model.layers.26.mlp.down_proj.weight",
1709
+ "model.layers.26.input_layernorm.weight",
1710
+ "model.layers.26.post_attention_layernorm.weight",
1711
+ "model.layers.27.self_attn.q_proj.weight",
1712
+ "model.layers.27.self_attn.k_proj.weight",
1713
+ "model.layers.27.self_attn.v_proj.weight",
1714
+ "model.layers.27.self_attn.o_proj.weight",
1715
+ "model.layers.27.mlp.gate_proj.weight",
1716
+ "model.layers.27.mlp.up_proj.weight",
1717
+ "model.layers.27.mlp.down_proj.weight",
1718
+ "model.layers.27.input_layernorm.weight",
1719
+ "model.layers.27.post_attention_layernorm.weight",
1720
+ "model.layers.28.self_attn.q_proj.weight",
1721
+ "model.layers.28.self_attn.k_proj.weight",
1722
+ "model.layers.28.self_attn.v_proj.weight",
1723
+ "model.layers.28.self_attn.o_proj.weight",
1724
+ "model.layers.28.mlp.gate_proj.weight",
1725
+ "model.layers.28.mlp.up_proj.weight",
1726
+ "model.layers.28.mlp.down_proj.weight",
1727
+ "model.layers.28.input_layernorm.weight",
1728
+ "model.layers.28.post_attention_layernorm.weight",
1729
+ "model.layers.29.self_attn.q_proj.weight",
1730
+ "model.layers.29.self_attn.k_proj.weight",
1731
+ "model.layers.29.self_attn.v_proj.weight",
1732
+ "model.layers.29.self_attn.o_proj.weight",
1733
+ "model.layers.29.mlp.gate_proj.weight",
1734
+ "model.layers.29.mlp.up_proj.weight",
1735
+ "model.layers.29.mlp.down_proj.weight",
1736
+ "model.layers.29.input_layernorm.weight",
1737
+ "model.layers.29.post_attention_layernorm.weight",
1738
+ "model.layers.30.self_attn.q_proj.weight",
1739
+ "model.layers.30.self_attn.k_proj.weight",
1740
+ "model.layers.30.self_attn.v_proj.weight",
1741
+ "model.layers.30.self_attn.o_proj.weight",
1742
+ "model.layers.30.mlp.gate_proj.weight",
1743
+ "model.layers.30.mlp.up_proj.weight",
1744
+ "model.layers.30.mlp.down_proj.weight",
1745
+ "model.layers.30.input_layernorm.weight",
1746
+ "model.layers.30.post_attention_layernorm.weight",
1747
+ "model.layers.31.self_attn.q_proj.weight",
1748
+ "model.layers.31.self_attn.k_proj.weight",
1749
+ "model.layers.31.self_attn.v_proj.weight",
1750
+ "model.layers.31.self_attn.o_proj.weight",
1751
+ "model.layers.31.mlp.gate_proj.weight",
1752
+ "model.layers.31.mlp.up_proj.weight",
1753
+ "model.layers.31.mlp.down_proj.weight",
1754
+ "model.layers.31.input_layernorm.weight",
1755
+ "model.layers.31.post_attention_layernorm.weight",
1756
+ "model.norm.weight",
1757
+ "lm_head.weight"
1758
+ ]
1759
+ }
logs/save_merged_model_20250616_121007.log ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-06-16 12:10:07 - experiment_save_merged_model - INFO - Starting merged model save process
2
+ 2025-06-16 12:10:07 - experiment_save_merged_model - INFO - Arguments: {'lambdas_path': '/work/gj26/b20042/LLM-AdaMerge/outputs/mistral-7b/parameter-wise/cross-entropy-loss/llm_adamerge_parameterwise_lambdas.json', 'model_config': '/work/gj26/b20042/LLM-AdaMerge/src/configs/model_config.yaml', 'output_dir': '/work/gj26/b20042/LLM-AdaMerge/mergekit/outputs/mistral-7b/llmadamerge/parameter-wise-crossentropy', 'model_name': 'merged-model', 'push_to_hub': False, 'hub_repo_id': 'lejelly/parameter-wise-llm-adamerge-crossentropy-mistral-7b-instrcut-math-code', 'private': False, 'device': 'cuda', 'debug': False}
3
+ 2025-06-16 12:10:07 - experiment_save_merged_model - INFO - Loading lambdas from /work/gj26/b20042/LLM-AdaMerge/outputs/mistral-7b/parameter-wise/cross-entropy-loss/llm_adamerge_parameterwise_lambdas.json
4
+ 2025-06-16 12:10:07 - experiment_save_merged_model - INFO - Auto-detected parameter-wise merge from JSON structure
5
+ 2025-06-16 12:10:07 - experiment_save_merged_model - INFO - Merge type: parameter_wise
6
+ 2025-06-16 12:10:07 - experiment_save_merged_model - INFO - [Initial] Memory Usage:
7
+ 2025-06-16 12:10:07 - experiment_save_merged_model - INFO - Process: 0.38 GB (0.2%)
8
+ 2025-06-16 12:10:07 - experiment_save_merged_model - INFO - System: 9.52 GB / 212.52 GB (9.1%)
9
+ 2025-06-16 12:10:07 - experiment_save_merged_model - INFO - Available: 193.16 GB
10
+ 2025-06-16 12:10:07 - experiment_save_merged_model - INFO - GPU 0: Allocated: 0.00 GB, Reserved: 0.00 GB, Total: 94.50 GB
11
+ 2025-06-16 12:10:07 - experiment_save_merged_model - INFO - Loading models
12
+ 2025-06-16 12:10:25 - experiment_save_merged_model - INFO - [After loading models] Memory Usage:
13
+ 2025-06-16 12:10:25 - experiment_save_merged_model - INFO - Process: 40.60 GB (19.1%)
14
+ 2025-06-16 12:10:25 - experiment_save_merged_model - INFO - System: 48.75 GB / 212.52 GB (30.9%)
15
+ 2025-06-16 12:10:25 - experiment_save_merged_model - INFO - Available: 146.93 GB
16
+ 2025-06-16 12:10:25 - experiment_save_merged_model - INFO - GPU 0: Allocated: 0.00 GB, Reserved: 0.00 GB, Total: 94.50 GB
17
+ 2025-06-16 12:10:25 - experiment_save_merged_model - INFO - Initializing parameter_wise AdaMerge
18
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Loading learned lambdas
19
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Deleting original models to free memory (task vectors already computed)
20
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - [Before deleting models] Memory Usage:
21
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Process: 94.70 GB (44.6%)
22
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - System: 89.79 GB / 212.52 GB (50.2%)
23
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Available: 105.82 GB
24
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - GPU 0: Allocated: 0.00 GB, Reserved: 0.00 GB, Total: 94.50 GB
25
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Clearing model_loader references
26
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Deleting model variables
27
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Running garbage collection
28
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - [After deleting models and GC] Memory Usage:
29
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Process: 55.38 GB (26.1%)
30
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - System: 64.31 GB / 212.52 GB (38.2%)
31
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Available: 131.30 GB
32
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - GPU 0: Allocated: 0.00 GB, Reserved: 0.00 GB, Total: 94.50 GB
33
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - [After loading lambdas] Memory Usage:
34
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Process: 55.38 GB (26.1%)
35
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - System: 64.31 GB / 212.52 GB (38.2%)
36
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Available: 131.30 GB
37
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - GPU 0: Allocated: 0.00 GB, Reserved: 0.00 GB, Total: 94.50 GB
38
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Creating merged model with learned lambdas
39
+ 2025-06-16 12:11:40 - experiment_save_merged_model - INFO - Using merge_models_for_save() for parameter-wise merge
40
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - [After merging models] Memory Usage:
41
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - Process: 57.71 GB (27.2%)
42
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - System: 93.30 GB / 212.52 GB (48.7%)
43
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - Available: 109.02 GB
44
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - GPU 0: Allocated: 13.49 GB, Reserved: 27.23 GB, Total: 94.50 GB
45
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - Freeing memory from AdaMerge object (task vectors and base params no longer needed)
46
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - Deleting task vectors
47
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - Deleting base params
48
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - Deleting functional model
49
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - [After freeing AdaMerge memory] Memory Usage:
50
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - Process: 5.72 GB (2.7%)
51
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - System: 27.36 GB / 212.52 GB (17.7%)
52
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - Available: 174.96 GB
53
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - GPU 0: Allocated: 13.49 GB, Reserved: 13.62 GB, Total: 94.50 GB
54
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - Saving merged model to /work/gj26/b20042/LLM-AdaMerge/mergekit/outputs/mistral-7b/llmadamerge/parameter-wise-crossentropy
55
+ 2025-06-16 12:13:34 - experiment_save_merged_model - INFO - Moving parameter-wise merged model to CPU for saving
56
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - Successfully saved 3 safetensors files:
57
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - - model-00001-of-00003.safetensors (4714.17 MB)
58
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - - model-00003-of-00003.safetensors (4330.17 MB)
59
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - - model-00002-of-00003.safetensors (4768.20 MB)
60
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - [After saving model] Memory Usage:
61
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - Process: 15.98 GB (7.5%)
62
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - System: 24.02 GB / 212.52 GB (19.3%)
63
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - Available: 171.60 GB
64
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - GPU 0: Allocated: 0.00 GB, Reserved: 0.00 GB, Total: 94.50 GB
65
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - Saving tokenizer
66
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - Copied lambdas file to /work/gj26/b20042/LLM-AdaMerge/mergekit/outputs/mistral-7b/llmadamerge/parameter-wise-crossentropy/learned_lambdas.json
67
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - Creating model card
68
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - Cleaning up models
69
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - [After cleanup] Memory Usage:
70
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - Process: 4.67 GB (2.2%)
71
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - System: 12.73 GB / 212.52 GB (13.9%)
72
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - Available: 182.89 GB
73
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - GPU 0: Allocated: 0.00 GB, Reserved: 0.00 GB, Total: 94.50 GB
74
+ 2025-06-16 12:14:12 - experiment_save_merged_model - INFO - Model saved successfully to /work/gj26/b20042/LLM-AdaMerge/mergekit/outputs/mistral-7b/llmadamerge/parameter-wise-crossentropy
model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:172f1687869a7dfe0b1a1631d1454cd5f6060085e7d4bfc5426a7ae9553e7185
3
+ size 4943162240
model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4a7bdf4a970fded8b782219c2367a4cab17bcc215b18c98d06298d2795528b5
3
+ size 4999819232
model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7adbbcb01b8e6f1f4edcbd71602fbad4b330f216dc740b5f6057a8c3c820d335
3
+ size 4540516256
model.safetensors.index.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 14483464192
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00003-of-00003.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00003.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
15
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
16
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
17
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00003.safetensors",
18
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
19
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
20
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
21
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
22
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
23
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
24
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
25
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
26
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00003.safetensors",
27
+ "model.layers.10.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
28
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
29
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
30
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
31
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
32
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
33
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
34
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
35
+ "model.layers.11.input_layernorm.weight": "model-00002-of-00003.safetensors",
36
+ "model.layers.11.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
37
+ "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
38
+ "model.layers.11.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
39
+ "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
40
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
41
+ "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
42
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
43
+ "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
44
+ "model.layers.12.input_layernorm.weight": "model-00002-of-00003.safetensors",
45
+ "model.layers.12.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
46
+ "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
47
+ "model.layers.12.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
48
+ "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
49
+ "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
50
+ "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
51
+ "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
52
+ "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
53
+ "model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
54
+ "model.layers.13.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
55
+ "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
56
+ "model.layers.13.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
57
+ "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
58
+ "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
59
+ "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
60
+ "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
61
+ "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
62
+ "model.layers.14.input_layernorm.weight": "model-00002-of-00003.safetensors",
63
+ "model.layers.14.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
64
+ "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
65
+ "model.layers.14.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
66
+ "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
67
+ "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
68
+ "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
69
+ "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
70
+ "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
71
+ "model.layers.15.input_layernorm.weight": "model-00002-of-00003.safetensors",
72
+ "model.layers.15.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
73
+ "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
74
+ "model.layers.15.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
75
+ "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
76
+ "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
77
+ "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
78
+ "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
79
+ "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
80
+ "model.layers.16.input_layernorm.weight": "model-00002-of-00003.safetensors",
81
+ "model.layers.16.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
82
+ "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
83
+ "model.layers.16.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
84
+ "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
85
+ "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
86
+ "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
87
+ "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
88
+ "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
89
+ "model.layers.17.input_layernorm.weight": "model-00002-of-00003.safetensors",
90
+ "model.layers.17.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
91
+ "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
92
+ "model.layers.17.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
93
+ "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
94
+ "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
95
+ "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
96
+ "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
97
+ "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
98
+ "model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
99
+ "model.layers.18.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
100
+ "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
101
+ "model.layers.18.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
102
+ "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
103
+ "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
104
+ "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
105
+ "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
106
+ "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
107
+ "model.layers.19.input_layernorm.weight": "model-00002-of-00003.safetensors",
108
+ "model.layers.19.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
109
+ "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
110
+ "model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
111
+ "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
112
+ "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
113
+ "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
114
+ "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
115
+ "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
116
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00003.safetensors",
117
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
118
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
119
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
120
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
121
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
122
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
123
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
124
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
125
+ "model.layers.20.input_layernorm.weight": "model-00002-of-00003.safetensors",
126
+ "model.layers.20.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
127
+ "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
128
+ "model.layers.20.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
129
+ "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
130
+ "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
131
+ "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
132
+ "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
133
+ "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
134
+ "model.layers.21.input_layernorm.weight": "model-00002-of-00003.safetensors",
135
+ "model.layers.21.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
136
+ "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
137
+ "model.layers.21.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
138
+ "model.layers.21.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
139
+ "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
140
+ "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
141
+ "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
142
+ "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
143
+ "model.layers.22.input_layernorm.weight": "model-00003-of-00003.safetensors",
144
+ "model.layers.22.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
145
+ "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
146
+ "model.layers.22.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
147
+ "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
148
+ "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
149
+ "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
150
+ "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
151
+ "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
152
+ "model.layers.23.input_layernorm.weight": "model-00003-of-00003.safetensors",
153
+ "model.layers.23.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
154
+ "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
155
+ "model.layers.23.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
156
+ "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
157
+ "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
158
+ "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
159
+ "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
160
+ "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
161
+ "model.layers.24.input_layernorm.weight": "model-00003-of-00003.safetensors",
162
+ "model.layers.24.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
163
+ "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
164
+ "model.layers.24.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
165
+ "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
166
+ "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
167
+ "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
168
+ "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
169
+ "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
170
+ "model.layers.25.input_layernorm.weight": "model-00003-of-00003.safetensors",
171
+ "model.layers.25.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
172
+ "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
173
+ "model.layers.25.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
174
+ "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
175
+ "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
176
+ "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
177
+ "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
178
+ "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
179
+ "model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
180
+ "model.layers.26.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
181
+ "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
182
+ "model.layers.26.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
183
+ "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
184
+ "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
185
+ "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
186
+ "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
187
+ "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
188
+ "model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
189
+ "model.layers.27.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
190
+ "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
191
+ "model.layers.27.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
192
+ "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
193
+ "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
194
+ "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
195
+ "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
196
+ "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
197
+ "model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
198
+ "model.layers.28.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
199
+ "model.layers.28.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
200
+ "model.layers.28.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
201
+ "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
202
+ "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
203
+ "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
204
+ "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
205
+ "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
206
+ "model.layers.29.input_layernorm.weight": "model-00003-of-00003.safetensors",
207
+ "model.layers.29.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
208
+ "model.layers.29.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
209
+ "model.layers.29.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
210
+ "model.layers.29.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
211
+ "model.layers.29.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
212
+ "model.layers.29.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
213
+ "model.layers.29.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
214
+ "model.layers.29.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
215
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
216
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
217
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
218
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
219
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
220
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
221
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
222
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
223
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
224
+ "model.layers.30.input_layernorm.weight": "model-00003-of-00003.safetensors",
225
+ "model.layers.30.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
226
+ "model.layers.30.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
227
+ "model.layers.30.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
228
+ "model.layers.30.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
229
+ "model.layers.30.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
230
+ "model.layers.30.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
231
+ "model.layers.30.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
232
+ "model.layers.30.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
233
+ "model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
234
+ "model.layers.31.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
235
+ "model.layers.31.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
236
+ "model.layers.31.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
237
+ "model.layers.31.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
238
+ "model.layers.31.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
239
+ "model.layers.31.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
240
+ "model.layers.31.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
241
+ "model.layers.31.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
242
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00003.safetensors",
243
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
244
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
245
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
246
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
247
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
248
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
249
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
250
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
251
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00003.safetensors",
252
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
253
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
254
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
255
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
256
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
257
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
258
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
259
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
260
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00003.safetensors",
261
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
262
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
263
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
264
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
265
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
266
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
267
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
268
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
269
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00003.safetensors",
270
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
271
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
272
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
273
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
274
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
275
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
276
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
277
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
278
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00003.safetensors",
279
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
280
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
281
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
282
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
283
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
284
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
285
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
286
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
287
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00003.safetensors",
288
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
289
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
290
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
291
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
292
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
293
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
294
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
295
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
296
+ "model.norm.weight": "model-00003-of-00003.safetensors"
297
+ }
298
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "additional_special_tokens": [],
32
+ "bos_token": "<s>",
33
+ "clean_up_tokenization_spaces": false,
34
+ "eos_token": "</s>",
35
+ "extra_special_tokens": {},
36
+ "legacy": false,
37
+ "model_max_length": 1000000000000000019884624838656,
38
+ "pad_token": "</s>",
39
+ "sp_model_kwargs": {},
40
+ "spaces_between_special_tokens": false,
41
+ "tokenizer_class": "LlamaTokenizerFast",
42
+ "unk_token": "<unk>",
43
+ "use_default_system_prompt": false
44
+ }