tomaarsen HF Staff commited on
Commit
830199e
·
verified ·
1 Parent(s): 9094f3b

Add new SentenceTransformer model

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 1024,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": false,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": true,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,941 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
+ tags:
6
+ - sentence-transformers
7
+ - sentence-similarity
8
+ - feature-extraction
9
+ - dense
10
+ - generated_from_trainer
11
+ - dataset_size:99231
12
+ - loss:CachedMultipleNegativesRankingLoss
13
+ widget:
14
+ - source_sentence: who ordered the charge of the light brigade
15
+ sentences:
16
+ - Charge of the Light Brigade The Charge of the Light Brigade was a charge of British
17
+ light cavalry led by Lord Cardigan against Russian forces during the Battle of
18
+ Balaclava on 25 October 1854 in the Crimean War. Lord Raglan, overall commander
19
+ of the British forces, had intended to send the Light Brigade to prevent the Russians
20
+ from removing captured guns from overrun Turkish positions, a task well-suited
21
+ to light cavalry.
22
+ - UNICEF The United Nations International Children's Emergency Fund was created
23
+ by the United Nations General Assembly on 11 December 1946, to provide emergency
24
+ food and healthcare to children in countries that had been devastated by World
25
+ War II. The Polish physician Ludwik Rajchman is widely regarded as the founder
26
+ of UNICEF and served as its first chairman from 1946. On Rajchman's suggestion,
27
+ the American Maurice Pate was appointed its first executive director, serving
28
+ from 1947 until his death in 1965.[5][6] In 1950, UNICEF's mandate was extended
29
+ to address the long-term needs of children and women in developing countries everywhere.
30
+ In 1953 it became a permanent part of the United Nations System, and the words
31
+ "international" and "emergency" were dropped from the organization's name, making
32
+ it simply the United Nations Children's Fund, retaining the original acronym,
33
+ "UNICEF".[3]
34
+ - Marcus Jordan Marcus James Jordan (born December 24, 1990) is an American former
35
+ college basketball player who played for the UCF Knights men's basketball team
36
+ of Conference USA.[1] He is the son of retired Hall of Fame basketball player
37
+ Michael Jordan.
38
+ - source_sentence: what part of the cow is the rib roast
39
+ sentences:
40
+ - Standing rib roast A standing rib roast, also known as prime rib, is a cut of
41
+ beef from the primal rib, one of the nine primal cuts of beef. While the entire
42
+ rib section comprises ribs six through 12, a standing rib roast may contain anywhere
43
+ from two to seven ribs.
44
+ - Blaine Anderson Kurt begins to mend their relationship in "Thanksgiving", just
45
+ before New Directions loses at Sectionals to the Warblers, and they spend Christmas
46
+ together in New York City.[29][30] Though he and Kurt continue to be on good terms,
47
+ Blaine finds himself developing a crush on his best friend, Sam, which he knows
48
+ will come to nothing as he knows Sam is not gay; the two of them team up to find
49
+ evidence that the Warblers cheated at Sectionals, which means New Directions will
50
+ be competing at Regionals. He ends up going to the Sadie Hawkins dance with Tina
51
+ Cohen-Chang (Jenna Ushkowitz), who has developed a crush on him, but as friends
52
+ only.[31] When Kurt comes to Lima for the wedding of glee club director Will (Matthew
53
+ Morrison) and Emma (Jayma Mays)—which Emma flees—he and Blaine make out beforehand,
54
+ and sleep together afterward, though they do not resume a permanent relationship.[32]
55
+ - 'Soviet Union The Soviet Union (Russian: Сове́тский Сою́з, tr. Sovétsky Soyúz,
56
+ IPA: [sɐˈvʲɛt͡skʲɪj sɐˈjus] ( listen)), officially the Union of Soviet Socialist
57
+ Republics (Russian: Сою́з Сове́тских Социалисти́ческих Респу́блик, tr. Soyúz Sovétskikh
58
+ Sotsialistícheskikh Respúblik, IPA: [sɐˈjus sɐˈvʲɛtskʲɪx sətsɨəlʲɪsˈtʲitɕɪskʲɪx
59
+ rʲɪˈspublʲɪk] ( listen)), abbreviated as the USSR (Russian: СССР, tr. SSSR), was
60
+ a socialist state in Eurasia that existed from 1922 to 1991. Nominally a union
61
+ of multiple national Soviet republics,[a] its government and economy were highly
62
+ centralized. The country was a one-party state, governed by the Communist Party
63
+ with Moscow as its capital in its largest republic, the Russian Soviet Federative
64
+ Socialist Republic. The Russian nation had constitutionally equal status among
65
+ the many nations of the union but exerted de facto dominance in various respects.[7]
66
+ Other major urban centres were Leningrad, Kiev, Minsk, Alma-Ata and Novosibirsk.
67
+ The Soviet Union was one of the five recognized nuclear weapons states and possessed
68
+ the largest stockpile of weapons of mass destruction.[8] It was a founding permanent
69
+ member of the United Nations Security Council, as well as a member of the Organization
70
+ for Security and Co-operation in Europe (OSCE) and the leading member of the Council
71
+ for Mutual Economic Assistance (CMEA) and the Warsaw Pact.'
72
+ - source_sentence: what is the current big bang theory season
73
+ sentences:
74
+ - Byzantine army From the seventh to the 12th centuries, the Byzantine army was
75
+ among the most powerful and effective military forces in the world – neither
76
+ Middle Ages Europe nor (following its early successes) the fracturing Caliphate
77
+ could match the strategies and the efficiency of the Byzantine army. Restricted
78
+ to a largely defensive role in the 7th to mid-9th centuries, the Byzantines developed
79
+ the theme-system to counter the more powerful Caliphate. From the mid-9th century,
80
+ however, they gradually went on the offensive, culminating in the great conquests
81
+ of the 10th century under a series of soldier-emperors such as Nikephoros II Phokas,
82
+ John Tzimiskes and Basil II. The army they led was less reliant on the militia
83
+ of the themes; it was by now a largely professional force, with a strong and well-drilled
84
+ infantry at its core and augmented by a revived heavy cavalry arm. With one of
85
+ the most powerful economies in the world at the time, the Empire had the resources
86
+ to put to the field a powerful host when needed, in order to reclaim its long-lost
87
+ territories.
88
+ - The Big Bang Theory The Big Bang Theory is an American television sitcom created
89
+ by Chuck Lorre and Bill Prady, both of whom serve as executive producers on the
90
+ series, along with Steven Molaro. All three also serve as head writers. The show
91
+ premiered on CBS on September 24, 2007.[3] The series' tenth season premiered
92
+ on September 19, 2016.[4] In March 2017, the series was renewed for two additional
93
+ seasons, bringing its total to twelve, and running through the 2018–19 television
94
+ season. The eleventh season is set to premiere on September 25, 2017.[5]
95
+ - 2016 NCAA Division I Softball Tournament The 2016 NCAA Division I Softball Tournament
96
+ was held from May 20 through June 8, 2016 as the final part of the 2016 NCAA Division
97
+ I softball season. The 64 NCAA Division I college softball teams were to be selected
98
+ out of an eligible 293 teams on May 15, 2016. Thirty-two teams were awarded an
99
+ automatic bid as champions of their conference, and thirty-two teams were selected
100
+ at-large by the NCAA Division I softball selection committee. The tournament culminated
101
+ with eight teams playing in the 2016 Women's College World Series at ASA Hall
102
+ of Fame Stadium in Oklahoma City in which the Oklahoma Sooners were crowned the
103
+ champions.
104
+ - source_sentence: what happened to tates mom on days of our lives
105
+ sentences:
106
+ - 'Paige O''Hara Donna Paige Helmintoller, better known as Paige O''Hara (born May
107
+ 10, 1956),[1] is an American actress, voice actress, singer and painter. O''Hara
108
+ began her career as a Broadway actress in 1983 when she portrayed Ellie May Chipley
109
+ in the musical Showboat. In 1991, she made her motion picture debut in Disney''s
110
+ Beauty and the Beast, in which she voiced the film''s heroine, Belle. Following
111
+ the critical and commercial success of Beauty and the Beast, O''Hara reprised
112
+ her role as Belle in the film''s two direct-to-video follow-ups, Beauty and the
113
+ Beast: The Enchanted Christmas and Belle''s Magical World.'
114
+ - M. Shadows Matthew Charles Sanders (born July 31, 1981), better known as M. Shadows,
115
+ is an American singer, songwriter, and musician. He is best known as the lead
116
+ vocalist, songwriter, and a founding member of the American heavy metal band Avenged
117
+ Sevenfold. In 2017, he was voted 3rd in the list of Top 25 Greatest Modern Frontmen
118
+ by Ultimate Guitar.[1]
119
+ - Theresa Donovan In July 2013, Jeannie returns to Salem, this time going by her
120
+ middle name, Theresa. Initially, she strikes up a connection with resident bad
121
+ boy JJ Deveraux (Casey Moss) while trying to secure some pot.[28] During a confrontation
122
+ with JJ and his mother Jennifer Horton (Melissa Reeves) in her office, her aunt
123
+ Kayla confirms that Theresa is in fact Jeannie and that Jen promised to hire her
124
+ as her assistant, a promise she reluctantly agrees to. Kayla reminds Theresa it
125
+ is her last chance at a fresh start.[29] Theresa also strikes up a bad first impression
126
+ with Jennifer's daughter Abigail Deveraux (Kate Mansi) when Abigail smells pot
127
+ on Theresa in her mother's office.[30] To continue to battle against Jennifer,
128
+ she teams up with Anne Milbauer (Meredith Scott Lynn) in hopes of exacting her
129
+ perfect revenge. In a ploy, Theresa reveals her intentions to hopefully woo Dr.
130
+ Daniel Jonas (Shawn Christian). After sleeping with JJ, Theresa overdoses on marijuana
131
+ and GHB. Upon hearing of their daughter's overdose and continuing problems, Shane
132
+ and Kimberly return to town in the hopes of handling their daughter's problem,
133
+ together. After believing that Theresa has a handle on her addictions, Shane and
134
+ Kimberly leave town together. Theresa then teams up with hospital co-worker Anne
135
+ Milbauer (Meredith Scott Lynn) to conspire against Jennifer, using Daniel as a
136
+ way to hurt their relationship. In early 2014, following a Narcotics Anonymous
137
+ (NA) meeting, she begins a sexual and drugged-fused relationship with Brady Black
138
+ (Eric Martsolf). In 2015, after it is found that Kristen DiMera (Eileen Davidson)
139
+ stole Theresa's embryo and carried it to term, Brady and Melanie Jonas return
140
+ her son, Christopher, to her and Brady, and the pair rename him Tate. When Theresa
141
+ moves into the Kiriakis mansion, tensions arise between her and Victor. She eventually
142
+ expresses her interest in purchasing Basic Black and running it as her own fashion
143
+ company, with financial backing from Maggie Horton (Suzanne Rogers). In the hopes
144
+ of finding the right partner, she teams up with Kate Roberts (Lauren Koslow) and
145
+ Nicole Walker (Arianne Zucker) to achieve the goal of purchasing Basic Black,
146
+ with Kate and Nicole's business background and her own interest in fashion design.
147
+ As she and Brady share several instances of rekindling their romance, she is kicked
148
+ out of the mansion by Victor; as a result, Brady quits Titan and moves in with
149
+ Theresa and Tate, in their own penthouse.
150
+ - source_sentence: where does the last name francisco come from
151
+ sentences:
152
+ - Francisco Francisco is the Spanish and Portuguese form of the masculine given
153
+ name Franciscus (corresponding to English Francis).
154
+ - 'Book of Esther The Book of Esther, also known in Hebrew as "the Scroll" (Megillah),
155
+ is a book in the third section (Ketuvim, "Writings") of the Jewish Tanakh (the
156
+ Hebrew Bible) and in the Christian Old Testament. It is one of the five Scrolls
157
+ (Megillot) in the Hebrew Bible. It relates the story of a Hebrew woman in Persia,
158
+ born as Hadassah but known as Esther, who becomes queen of Persia and thwarts
159
+ a genocide of her people. The story forms the core of the Jewish festival of Purim,
160
+ during which it is read aloud twice: once in the evening and again the following
161
+ morning. The books of Esther and Song of Songs are the only books in the Hebrew
162
+ Bible that do not explicitly mention God.[2]'
163
+ - Times Square Times Square is a major commercial intersection, tourist destination,
164
+ entertainment center and neighborhood in the Midtown Manhattan section of New
165
+ York City at the junction of Broadway and Seventh Avenue. It stretches from West
166
+ 42nd to West 47th Streets.[1] Brightly adorned with billboards and advertisements,
167
+ Times Square is sometimes referred to as "The Crossroads of the World",[2] "The
168
+ Center of the Universe",[3] "the heart of The Great White Way",[4][5][6] and the
169
+ "heart of the world".[7] One of the world's busiest pedestrian areas,[8] it is
170
+ also the hub of the Broadway Theater District[9] and a major center of the world's
171
+ entertainment industry.[10] Times Square is one of the world's most visited tourist
172
+ attractions, drawing an estimated 50 million visitors annually.[11] Approximately
173
+ 330,000 people pass through Times Square daily,[12] many of them tourists,[13]
174
+ while over 460,000 pedestrians walk through Times Square on its busiest days.[7]
175
+ datasets:
176
+ - sentence-transformers/natural-questions
177
+ pipeline_tag: sentence-similarity
178
+ library_name: sentence-transformers
179
+ metrics:
180
+ - cosine_accuracy@1
181
+ - cosine_accuracy@3
182
+ - cosine_accuracy@5
183
+ - cosine_accuracy@10
184
+ - cosine_precision@1
185
+ - cosine_precision@3
186
+ - cosine_precision@5
187
+ - cosine_precision@10
188
+ - cosine_recall@1
189
+ - cosine_recall@3
190
+ - cosine_recall@5
191
+ - cosine_recall@10
192
+ - cosine_ndcg@10
193
+ - cosine_mrr@10
194
+ - cosine_map@100
195
+ co2_eq_emissions:
196
+ emissions: 405.46378000747745
197
+ energy_consumed: 1.043122443433472
198
+ source: codecarbon
199
+ training_type: fine-tuning
200
+ on_cloud: false
201
+ cpu_model: 13th Gen Intel(R) Core(TM) i7-13700K
202
+ ram_total_size: 31.777088165283203
203
+ hours_used: 3.425
204
+ hardware_used: 1 x NVIDIA GeForce RTX 3090
205
+ model-index:
206
+ - name: LiquidAI/LFM2-350M trained on Natural Questions pairs
207
+ results:
208
+ - task:
209
+ type: information-retrieval
210
+ name: Information Retrieval
211
+ dataset:
212
+ name: NanoMSMARCO
213
+ type: NanoMSMARCO
214
+ metrics:
215
+ - type: cosine_accuracy@1
216
+ value: 0.28
217
+ name: Cosine Accuracy@1
218
+ - type: cosine_accuracy@3
219
+ value: 0.46
220
+ name: Cosine Accuracy@3
221
+ - type: cosine_accuracy@5
222
+ value: 0.64
223
+ name: Cosine Accuracy@5
224
+ - type: cosine_accuracy@10
225
+ value: 0.74
226
+ name: Cosine Accuracy@10
227
+ - type: cosine_precision@1
228
+ value: 0.28
229
+ name: Cosine Precision@1
230
+ - type: cosine_precision@3
231
+ value: 0.15333333333333332
232
+ name: Cosine Precision@3
233
+ - type: cosine_precision@5
234
+ value: 0.128
235
+ name: Cosine Precision@5
236
+ - type: cosine_precision@10
237
+ value: 0.07400000000000001
238
+ name: Cosine Precision@10
239
+ - type: cosine_recall@1
240
+ value: 0.28
241
+ name: Cosine Recall@1
242
+ - type: cosine_recall@3
243
+ value: 0.46
244
+ name: Cosine Recall@3
245
+ - type: cosine_recall@5
246
+ value: 0.64
247
+ name: Cosine Recall@5
248
+ - type: cosine_recall@10
249
+ value: 0.74
250
+ name: Cosine Recall@10
251
+ - type: cosine_ndcg@10
252
+ value: 0.4909415698599729
253
+ name: Cosine Ndcg@10
254
+ - type: cosine_mrr@10
255
+ value: 0.4130714285714285
256
+ name: Cosine Mrr@10
257
+ - type: cosine_map@100
258
+ value: 0.42354966538209315
259
+ name: Cosine Map@100
260
+ - task:
261
+ type: information-retrieval
262
+ name: Information Retrieval
263
+ dataset:
264
+ name: NanoNFCorpus
265
+ type: NanoNFCorpus
266
+ metrics:
267
+ - type: cosine_accuracy@1
268
+ value: 0.4
269
+ name: Cosine Accuracy@1
270
+ - type: cosine_accuracy@3
271
+ value: 0.5
272
+ name: Cosine Accuracy@3
273
+ - type: cosine_accuracy@5
274
+ value: 0.58
275
+ name: Cosine Accuracy@5
276
+ - type: cosine_accuracy@10
277
+ value: 0.68
278
+ name: Cosine Accuracy@10
279
+ - type: cosine_precision@1
280
+ value: 0.4
281
+ name: Cosine Precision@1
282
+ - type: cosine_precision@3
283
+ value: 0.35999999999999993
284
+ name: Cosine Precision@3
285
+ - type: cosine_precision@5
286
+ value: 0.324
287
+ name: Cosine Precision@5
288
+ - type: cosine_precision@10
289
+ value: 0.26599999999999996
290
+ name: Cosine Precision@10
291
+ - type: cosine_recall@1
292
+ value: 0.02298357366763854
293
+ name: Cosine Recall@1
294
+ - type: cosine_recall@3
295
+ value: 0.061632366484571384
296
+ name: Cosine Recall@3
297
+ - type: cosine_recall@5
298
+ value: 0.09750915762412557
299
+ name: Cosine Recall@5
300
+ - type: cosine_recall@10
301
+ value: 0.13301219077618073
302
+ name: Cosine Recall@10
303
+ - type: cosine_ndcg@10
304
+ value: 0.32361002047039217
305
+ name: Cosine Ndcg@10
306
+ - type: cosine_mrr@10
307
+ value: 0.47583333333333333
308
+ name: Cosine Mrr@10
309
+ - type: cosine_map@100
310
+ value: 0.12539829347446158
311
+ name: Cosine Map@100
312
+ - task:
313
+ type: information-retrieval
314
+ name: Information Retrieval
315
+ dataset:
316
+ name: NanoNQ
317
+ type: NanoNQ
318
+ metrics:
319
+ - type: cosine_accuracy@1
320
+ value: 0.48
321
+ name: Cosine Accuracy@1
322
+ - type: cosine_accuracy@3
323
+ value: 0.68
324
+ name: Cosine Accuracy@3
325
+ - type: cosine_accuracy@5
326
+ value: 0.78
327
+ name: Cosine Accuracy@5
328
+ - type: cosine_accuracy@10
329
+ value: 0.82
330
+ name: Cosine Accuracy@10
331
+ - type: cosine_precision@1
332
+ value: 0.48
333
+ name: Cosine Precision@1
334
+ - type: cosine_precision@3
335
+ value: 0.22666666666666666
336
+ name: Cosine Precision@3
337
+ - type: cosine_precision@5
338
+ value: 0.15600000000000003
339
+ name: Cosine Precision@5
340
+ - type: cosine_precision@10
341
+ value: 0.086
342
+ name: Cosine Precision@10
343
+ - type: cosine_recall@1
344
+ value: 0.47
345
+ name: Cosine Recall@1
346
+ - type: cosine_recall@3
347
+ value: 0.64
348
+ name: Cosine Recall@3
349
+ - type: cosine_recall@5
350
+ value: 0.72
351
+ name: Cosine Recall@5
352
+ - type: cosine_recall@10
353
+ value: 0.78
354
+ name: Cosine Recall@10
355
+ - type: cosine_ndcg@10
356
+ value: 0.632163202477609
357
+ name: Cosine Ndcg@10
358
+ - type: cosine_mrr@10
359
+ value: 0.5983571428571428
360
+ name: Cosine Mrr@10
361
+ - type: cosine_map@100
362
+ value: 0.5837963147038205
363
+ name: Cosine Map@100
364
+ - task:
365
+ type: nano-beir
366
+ name: Nano BEIR
367
+ dataset:
368
+ name: NanoBEIR mean
369
+ type: NanoBEIR_mean
370
+ metrics:
371
+ - type: cosine_accuracy@1
372
+ value: 0.3866666666666667
373
+ name: Cosine Accuracy@1
374
+ - type: cosine_accuracy@3
375
+ value: 0.5466666666666667
376
+ name: Cosine Accuracy@3
377
+ - type: cosine_accuracy@5
378
+ value: 0.6666666666666666
379
+ name: Cosine Accuracy@5
380
+ - type: cosine_accuracy@10
381
+ value: 0.7466666666666666
382
+ name: Cosine Accuracy@10
383
+ - type: cosine_precision@1
384
+ value: 0.3866666666666667
385
+ name: Cosine Precision@1
386
+ - type: cosine_precision@3
387
+ value: 0.24666666666666662
388
+ name: Cosine Precision@3
389
+ - type: cosine_precision@5
390
+ value: 0.2026666666666667
391
+ name: Cosine Precision@5
392
+ - type: cosine_precision@10
393
+ value: 0.142
394
+ name: Cosine Precision@10
395
+ - type: cosine_recall@1
396
+ value: 0.25766119122254616
397
+ name: Cosine Recall@1
398
+ - type: cosine_recall@3
399
+ value: 0.38721078882819054
400
+ name: Cosine Recall@3
401
+ - type: cosine_recall@5
402
+ value: 0.48583638587470857
403
+ name: Cosine Recall@5
404
+ - type: cosine_recall@10
405
+ value: 0.5510040635920602
406
+ name: Cosine Recall@10
407
+ - type: cosine_ndcg@10
408
+ value: 0.48223826426932465
409
+ name: Cosine Ndcg@10
410
+ - type: cosine_mrr@10
411
+ value: 0.4957539682539682
412
+ name: Cosine Mrr@10
413
+ - type: cosine_map@100
414
+ value: 0.37758142452012505
415
+ name: Cosine Map@100
416
+ ---
417
+
418
+ # LiquidAI/LFM2-350M trained on Natural Questions pairs
419
+
420
+ This is a [sentence-transformers](https://www.SBERT.net) model trained on the [natural-questions](https://huggingface.co/datasets/sentence-transformers/natural-questions) dataset. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
421
+
422
+ ## Model Details
423
+
424
+ ### Model Description
425
+ - **Model Type:** Sentence Transformer
426
+ <!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
427
+ - **Maximum Sequence Length:** 128000 tokens
428
+ - **Output Dimensionality:** 1024 dimensions
429
+ - **Similarity Function:** Cosine Similarity
430
+ - **Training Dataset:**
431
+ - [natural-questions](https://huggingface.co/datasets/sentence-transformers/natural-questions)
432
+ - **Language:** en
433
+ - **License:** apache-2.0
434
+
435
+ ### Model Sources
436
+
437
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
438
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
439
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
440
+
441
+ ### Full Model Architecture
442
+
443
+ ```
444
+ SentenceTransformer(
445
+ (0): Transformer({'max_seq_length': 128000, 'do_lower_case': False, 'architecture': 'LFM2Model'})
446
+ (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': True, 'include_prompt': True})
447
+ )
448
+ ```
449
+
450
+ ## Usage
451
+
452
+ ### Direct Usage (Sentence Transformers)
453
+
454
+ First install the Sentence Transformers library:
455
+
456
+ ```bash
457
+ pip install -U sentence-transformers
458
+ ```
459
+
460
+ Then you can load this model and run inference.
461
+ ```python
462
+ from sentence_transformers import SentenceTransformer
463
+
464
+ # Download from the 🤗 Hub
465
+ model = SentenceTransformer("tomaarsen/LFM2-350M-nq-prompts")
466
+ # Run inference
467
+ queries = [
468
+ "where does the last name francisco come from",
469
+ ]
470
+ documents = [
471
+ 'Francisco Francisco is the Spanish and Portuguese form of the masculine given name Franciscus (corresponding to English Francis).',
472
+ 'Book of Esther The Book of Esther, also known in Hebrew as "the Scroll" (Megillah), is a book in the third section (Ketuvim, "Writings") of the Jewish Tanakh (the Hebrew Bible) and in the Christian Old Testament. It is one of the five Scrolls (Megillot) in the Hebrew Bible. It relates the story of a Hebrew woman in Persia, born as Hadassah but known as Esther, who becomes queen of Persia and thwarts a genocide of her people. The story forms the core of the Jewish festival of Purim, during which it is read aloud twice: once in the evening and again the following morning. The books of Esther and Song of Songs are the only books in the Hebrew Bible that do not explicitly mention God.[2]',
473
+ 'Times Square Times Square is a major commercial intersection, tourist destination, entertainment center and neighborhood in the Midtown Manhattan section of New York City at the junction of Broadway and Seventh Avenue. It stretches from West 42nd to West 47th Streets.[1] Brightly adorned with billboards and advertisements, Times Square is sometimes referred to as "The Crossroads of the World",[2] "The Center of the Universe",[3] "the heart of The Great White Way",[4][5][6] and the "heart of the world".[7] One of the world\'s busiest pedestrian areas,[8] it is also the hub of the Broadway Theater District[9] and a major center of the world\'s entertainment industry.[10] Times Square is one of the world\'s most visited tourist attractions, drawing an estimated 50 million visitors annually.[11] Approximately 330,000 people pass through Times Square daily,[12] many of them tourists,[13] while over 460,000 pedestrians walk through Times Square on its busiest days.[7]',
474
+ ]
475
+ query_embeddings = model.encode_query(queries)
476
+ document_embeddings = model.encode_document(documents)
477
+ print(query_embeddings.shape, document_embeddings.shape)
478
+ # [1, 1024] [3, 1024]
479
+
480
+ # Get the similarity scores for the embeddings
481
+ similarities = model.similarity(query_embeddings, document_embeddings)
482
+ print(similarities)
483
+ # tensor([[ 0.7825, -0.0811, -0.0414]])
484
+ ```
485
+
486
+ <!--
487
+ ### Direct Usage (Transformers)
488
+
489
+ <details><summary>Click to see the direct usage in Transformers</summary>
490
+
491
+ </details>
492
+ -->
493
+
494
+ <!--
495
+ ### Downstream Usage (Sentence Transformers)
496
+
497
+ You can finetune this model on your own dataset.
498
+
499
+ <details><summary>Click to expand</summary>
500
+
501
+ </details>
502
+ -->
503
+
504
+ <!--
505
+ ### Out-of-Scope Use
506
+
507
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
508
+ -->
509
+
510
+ ## Evaluation
511
+
512
+ ### Metrics
513
+
514
+ #### Information Retrieval
515
+
516
+ * Datasets: `NanoMSMARCO`, `NanoNFCorpus` and `NanoNQ`
517
+ * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator) with these parameters:
518
+ ```json
519
+ {
520
+ "query_prompt": "query: ",
521
+ "corpus_prompt": "document: "
522
+ }
523
+ ```
524
+
525
+ | Metric | NanoMSMARCO | NanoNFCorpus | NanoNQ |
526
+ |:--------------------|:------------|:-------------|:-----------|
527
+ | cosine_accuracy@1 | 0.28 | 0.4 | 0.48 |
528
+ | cosine_accuracy@3 | 0.46 | 0.5 | 0.68 |
529
+ | cosine_accuracy@5 | 0.64 | 0.58 | 0.78 |
530
+ | cosine_accuracy@10 | 0.74 | 0.68 | 0.82 |
531
+ | cosine_precision@1 | 0.28 | 0.4 | 0.48 |
532
+ | cosine_precision@3 | 0.1533 | 0.36 | 0.2267 |
533
+ | cosine_precision@5 | 0.128 | 0.324 | 0.156 |
534
+ | cosine_precision@10 | 0.074 | 0.266 | 0.086 |
535
+ | cosine_recall@1 | 0.28 | 0.023 | 0.47 |
536
+ | cosine_recall@3 | 0.46 | 0.0616 | 0.64 |
537
+ | cosine_recall@5 | 0.64 | 0.0975 | 0.72 |
538
+ | cosine_recall@10 | 0.74 | 0.133 | 0.78 |
539
+ | **cosine_ndcg@10** | **0.4909** | **0.3236** | **0.6322** |
540
+ | cosine_mrr@10 | 0.4131 | 0.4758 | 0.5984 |
541
+ | cosine_map@100 | 0.4235 | 0.1254 | 0.5838 |
542
+
543
+ #### Nano BEIR
544
+
545
+ * Dataset: `NanoBEIR_mean`
546
+ * Evaluated with [<code>NanoBEIREvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.NanoBEIREvaluator) with these parameters:
547
+ ```json
548
+ {
549
+ "dataset_names": [
550
+ "msmarco",
551
+ "nfcorpus",
552
+ "nq"
553
+ ],
554
+ "query_prompts": {
555
+ "msmarco": "query: ",
556
+ "nfcorpus": "query: ",
557
+ "nq": "query: "
558
+ },
559
+ "corpus_prompts": {
560
+ "msmarco": "document: ",
561
+ "nfcorpus": "document: ",
562
+ "nq": "document: "
563
+ }
564
+ }
565
+ ```
566
+
567
+ | Metric | Value |
568
+ |:--------------------|:-----------|
569
+ | cosine_accuracy@1 | 0.3867 |
570
+ | cosine_accuracy@3 | 0.5467 |
571
+ | cosine_accuracy@5 | 0.6667 |
572
+ | cosine_accuracy@10 | 0.7467 |
573
+ | cosine_precision@1 | 0.3867 |
574
+ | cosine_precision@3 | 0.2467 |
575
+ | cosine_precision@5 | 0.2027 |
576
+ | cosine_precision@10 | 0.142 |
577
+ | cosine_recall@1 | 0.2577 |
578
+ | cosine_recall@3 | 0.3872 |
579
+ | cosine_recall@5 | 0.4858 |
580
+ | cosine_recall@10 | 0.551 |
581
+ | **cosine_ndcg@10** | **0.4822** |
582
+ | cosine_mrr@10 | 0.4958 |
583
+ | cosine_map@100 | 0.3776 |
584
+
585
+ <!--
586
+ ## Bias, Risks and Limitations
587
+
588
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
589
+ -->
590
+
591
+ <!--
592
+ ### Recommendations
593
+
594
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
595
+ -->
596
+
597
+ ## Training Details
598
+
599
+ ### Training Dataset
600
+
601
+ #### natural-questions
602
+
603
+ * Dataset: [natural-questions](https://huggingface.co/datasets/sentence-transformers/natural-questions) at [f9e894e](https://huggingface.co/datasets/sentence-transformers/natural-questions/tree/f9e894e1081e206e577b4eaa9ee6de2b06ae6f17)
604
+ * Size: 99,231 training samples
605
+ * Columns: <code>query</code> and <code>answer</code>
606
+ * Approximate statistics based on the first 1000 samples:
607
+ | | query | answer |
608
+ |:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
609
+ | type | string | string |
610
+ | details | <ul><li>min: 9 tokens</li><li>mean: 11.59 tokens</li><li>max: 26 tokens</li></ul> | <ul><li>min: 16 tokens</li><li>mean: 142.35 tokens</li><li>max: 559 tokens</li></ul> |
611
+ * Samples:
612
+ | query | answer |
613
+ |:------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
614
+ | <code>who is required to report according to the hmda</code> | <code>Home Mortgage Disclosure Act US financial institutions must report HMDA data to their regulator if they meet certain criteria, such as having assets above a specific threshold. The criteria is different for depository and non-depository institutions and are available on the FFIEC website.[4] In 2012, there were 7,400 institutions that reported a total of 18.7 million HMDA records.[5]</code> |
615
+ | <code>what is the definition of endoplasmic reticulum in biology</code> | <code>Endoplasmic reticulum The endoplasmic reticulum (ER) is a type of organelle in eukaryotic cells that forms an interconnected network of flattened, membrane-enclosed sacs or tube-like structures known as cisternae. The membranes of the ER are continuous with the outer nuclear membrane. The endoplasmic reticulum occurs in most types of eukaryotic cells, but is absent from red blood cells and spermatozoa. There are two types of endoplasmic reticulum: rough and smooth. The outer (cytosolic) face of the rough endoplasmic reticulum is studded with ribosomes that are the sites of protein synthesis. The rough endoplasmic reticulum is especially prominent in cells such as hepatocytes. The smooth endoplasmic reticulum lacks ribosomes and functions in lipid manufacture and metabolism, the production of steroid hormones, and detoxification.[1] The smooth ER is especially abundant in mammalian liver and gonad cells. The lacy membranes of the endoplasmic reticulum were first seen in 1945 using elect...</code> |
616
+ | <code>what does the ski mean in polish names</code> | <code>Polish name Since the High Middle Ages, Polish-sounding surnames ending with the masculine -ski suffix, including -cki and -dzki, and the corresponding feminine suffix -ska/-cka/-dzka were associated with the nobility (Polish szlachta), which alone, in the early years, had such suffix distinctions.[1] They are widely popular today.</code> |
617
+ * Loss: [<code>CachedMultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedmultiplenegativesrankingloss) with these parameters:
618
+ ```json
619
+ {
620
+ "scale": 20.0,
621
+ "similarity_fct": "cos_sim",
622
+ "mini_batch_size": 4
623
+ }
624
+ ```
625
+
626
+ ### Evaluation Dataset
627
+
628
+ #### natural-questions
629
+
630
+ * Dataset: [natural-questions](https://huggingface.co/datasets/sentence-transformers/natural-questions) at [f9e894e](https://huggingface.co/datasets/sentence-transformers/natural-questions/tree/f9e894e1081e206e577b4eaa9ee6de2b06ae6f17)
631
+ * Size: 1,000 evaluation samples
632
+ * Columns: <code>query</code> and <code>answer</code>
633
+ * Approximate statistics based on the first 1000 samples:
634
+ | | query | answer |
635
+ |:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
636
+ | type | string | string |
637
+ | details | <ul><li>min: 9 tokens</li><li>mean: 11.62 tokens</li><li>max: 23 tokens</li></ul> | <ul><li>min: 9 tokens</li><li>mean: 141.66 tokens</li><li>max: 664 tokens</li></ul> |
638
+ * Samples:
639
+ | query | answer |
640
+ |:------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
641
+ | <code>difference between russian blue and british blue cat</code> | <code>Russian Blue The coat is known as a "double coat", with the undercoat being soft, downy and equal in length to the guard hairs, which are an even blue with silver tips. However, the tail may have a few very dull, almost unnoticeable stripes. The coat is described as thick, plush and soft to the touch. The feeling is softer than the softest silk. The silver tips give the coat a shimmering appearance. Its eyes are almost always a dark and vivid green. Any white patches of fur or yellow eyes in adulthood are seen as flaws in show cats.[3] Russian Blues should not be confused with British Blues (which are not a distinct breed, but rather a British Shorthair with a blue coat as the British Shorthair breed itself comes in a wide variety of colors and patterns), nor the Chartreux or Korat which are two other naturally occurring breeds of blue cats, although they have similar traits.</code> |
642
+ | <code>who played the little girl on mrs doubtfire</code> | <code>Mara Wilson Mara Elizabeth Wilson[2] (born July 24, 1987) is an American writer and former child actress. She is known for playing Natalie Hillard in Mrs. Doubtfire (1993), Susan Walker in Miracle on 34th Street (1994), Matilda Wormwood in Matilda (1996) and Lily Stone in Thomas and the Magic Railroad (2000). Since retiring from film acting, Wilson has focused on writing.</code> |
643
+ | <code>what year did the movie the sound of music come out</code> | <code>The Sound of Music (film) The film was released on March 2, 1965 in the United States, initially as a limited roadshow theatrical release. Although critical response to the film was widely mixed, the film was a major commercial success, becoming the number one box office movie after four weeks, and the highest-grossing film of 1965. By November 1966, The Sound of Music had become the highest-grossing film of all-time—surpassing Gone with the Wind—and held that distinction for five years. The film was just as popular throughout the world, breaking previous box-office records in twenty-nine countries. Following an initial theatrical release that lasted four and a half years, and two successful re-releases, the film sold 283 million admissions worldwide and earned a total worldwide gross of $286,000,000.</code> |
644
+ * Loss: [<code>CachedMultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedmultiplenegativesrankingloss) with these parameters:
645
+ ```json
646
+ {
647
+ "scale": 20.0,
648
+ "similarity_fct": "cos_sim",
649
+ "mini_batch_size": 4
650
+ }
651
+ ```
652
+
653
+ ### Training Hyperparameters
654
+ #### Non-Default Hyperparameters
655
+
656
+ - `eval_strategy`: steps
657
+ - `per_device_train_batch_size`: 256
658
+ - `per_device_eval_batch_size`: 256
659
+ - `learning_rate`: 2e-05
660
+ - `num_train_epochs`: 1
661
+ - `warmup_ratio`: 0.1
662
+ - `seed`: 12
663
+ - `bf16`: True
664
+ - `prompts`: {'query': 'query: ', 'answer': 'document: '}
665
+ - `batch_sampler`: no_duplicates
666
+
667
+ #### All Hyperparameters
668
+ <details><summary>Click to expand</summary>
669
+
670
+ - `overwrite_output_dir`: False
671
+ - `do_predict`: False
672
+ - `eval_strategy`: steps
673
+ - `prediction_loss_only`: True
674
+ - `per_device_train_batch_size`: 256
675
+ - `per_device_eval_batch_size`: 256
676
+ - `per_gpu_train_batch_size`: None
677
+ - `per_gpu_eval_batch_size`: None
678
+ - `gradient_accumulation_steps`: 1
679
+ - `eval_accumulation_steps`: None
680
+ - `torch_empty_cache_steps`: None
681
+ - `learning_rate`: 2e-05
682
+ - `weight_decay`: 0.0
683
+ - `adam_beta1`: 0.9
684
+ - `adam_beta2`: 0.999
685
+ - `adam_epsilon`: 1e-08
686
+ - `max_grad_norm`: 1.0
687
+ - `num_train_epochs`: 1
688
+ - `max_steps`: -1
689
+ - `lr_scheduler_type`: linear
690
+ - `lr_scheduler_kwargs`: {}
691
+ - `warmup_ratio`: 0.1
692
+ - `warmup_steps`: 0
693
+ - `log_level`: passive
694
+ - `log_level_replica`: warning
695
+ - `log_on_each_node`: True
696
+ - `logging_nan_inf_filter`: True
697
+ - `save_safetensors`: True
698
+ - `save_on_each_node`: False
699
+ - `save_only_model`: False
700
+ - `restore_callback_states_from_checkpoint`: False
701
+ - `no_cuda`: False
702
+ - `use_cpu`: False
703
+ - `use_mps_device`: False
704
+ - `seed`: 12
705
+ - `data_seed`: None
706
+ - `jit_mode_eval`: False
707
+ - `use_ipex`: False
708
+ - `bf16`: True
709
+ - `fp16`: False
710
+ - `fp16_opt_level`: O1
711
+ - `half_precision_backend`: auto
712
+ - `bf16_full_eval`: False
713
+ - `fp16_full_eval`: False
714
+ - `tf32`: None
715
+ - `local_rank`: 0
716
+ - `ddp_backend`: None
717
+ - `tpu_num_cores`: None
718
+ - `tpu_metrics_debug`: False
719
+ - `debug`: []
720
+ - `dataloader_drop_last`: False
721
+ - `dataloader_num_workers`: 0
722
+ - `dataloader_prefetch_factor`: None
723
+ - `past_index`: -1
724
+ - `disable_tqdm`: False
725
+ - `remove_unused_columns`: True
726
+ - `label_names`: None
727
+ - `load_best_model_at_end`: False
728
+ - `ignore_data_skip`: False
729
+ - `fsdp`: []
730
+ - `fsdp_min_num_params`: 0
731
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
732
+ - `fsdp_transformer_layer_cls_to_wrap`: None
733
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
734
+ - `deepspeed`: None
735
+ - `label_smoothing_factor`: 0.0
736
+ - `optim`: adamw_torch
737
+ - `optim_args`: None
738
+ - `adafactor`: False
739
+ - `group_by_length`: False
740
+ - `length_column_name`: length
741
+ - `ddp_find_unused_parameters`: None
742
+ - `ddp_bucket_cap_mb`: None
743
+ - `ddp_broadcast_buffers`: False
744
+ - `dataloader_pin_memory`: True
745
+ - `dataloader_persistent_workers`: False
746
+ - `skip_memory_metrics`: True
747
+ - `use_legacy_prediction_loop`: False
748
+ - `push_to_hub`: False
749
+ - `resume_from_checkpoint`: None
750
+ - `hub_model_id`: None
751
+ - `hub_strategy`: every_save
752
+ - `hub_private_repo`: None
753
+ - `hub_always_push`: False
754
+ - `hub_revision`: None
755
+ - `gradient_checkpointing`: False
756
+ - `gradient_checkpointing_kwargs`: None
757
+ - `include_inputs_for_metrics`: False
758
+ - `include_for_metrics`: []
759
+ - `eval_do_concat_batches`: True
760
+ - `fp16_backend`: auto
761
+ - `push_to_hub_model_id`: None
762
+ - `push_to_hub_organization`: None
763
+ - `mp_parameters`:
764
+ - `auto_find_batch_size`: False
765
+ - `full_determinism`: False
766
+ - `torchdynamo`: None
767
+ - `ray_scope`: last
768
+ - `ddp_timeout`: 1800
769
+ - `torch_compile`: False
770
+ - `torch_compile_backend`: None
771
+ - `torch_compile_mode`: None
772
+ - `include_tokens_per_second`: False
773
+ - `include_num_input_tokens_seen`: False
774
+ - `neftune_noise_alpha`: None
775
+ - `optim_target_modules`: None
776
+ - `batch_eval_metrics`: False
777
+ - `eval_on_start`: False
778
+ - `use_liger_kernel`: False
779
+ - `liger_kernel_config`: None
780
+ - `eval_use_gather_object`: False
781
+ - `average_tokens_across_devices`: False
782
+ - `prompts`: {'query': 'query: ', 'answer': 'document: '}
783
+ - `batch_sampler`: no_duplicates
784
+ - `multi_dataset_batch_sampler`: proportional
785
+ - `router_mapping`: {}
786
+ - `learning_rate_mapping`: {}
787
+
788
+ </details>
789
+
790
+ ### Training Logs
791
+ | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNFCorpus_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
792
+ |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------------:|:---------------------:|:----------------------------:|
793
+ | -1 | -1 | - | - | 0.0086 | 0.0233 | 0.0063 | 0.0128 |
794
+ | 0.0026 | 1 | 4.6189 | - | - | - | - | - |
795
+ | 0.0129 | 5 | 4.1284 | - | - | - | - | - |
796
+ | 0.0258 | 10 | 3.6638 | - | - | - | - | - |
797
+ | 0.0387 | 15 | 2.3118 | - | - | - | - | - |
798
+ | 0.0515 | 20 | 1.0986 | - | - | - | - | - |
799
+ | 0.0644 | 25 | 0.5063 | - | - | - | - | - |
800
+ | 0.0773 | 30 | 0.2891 | - | - | - | - | - |
801
+ | 0.0902 | 35 | 0.2138 | - | - | - | - | - |
802
+ | 0.1031 | 40 | 0.1967 | - | - | - | - | - |
803
+ | 0.1160 | 45 | 0.1745 | - | - | - | - | - |
804
+ | 0.1289 | 50 | 0.1479 | 0.1425 | 0.4927 | 0.3162 | 0.5375 | 0.4488 |
805
+ | 0.1418 | 55 | 0.1257 | - | - | - | - | - |
806
+ | 0.1546 | 60 | 0.1215 | - | - | - | - | - |
807
+ | 0.1675 | 65 | 0.1475 | - | - | - | - | - |
808
+ | 0.1804 | 70 | 0.1066 | - | - | - | - | - |
809
+ | 0.1933 | 75 | 0.1056 | - | - | - | - | - |
810
+ | 0.2062 | 80 | 0.1181 | - | - | - | - | - |
811
+ | 0.2191 | 85 | 0.118 | - | - | - | - | - |
812
+ | 0.2320 | 90 | 0.1031 | - | - | - | - | - |
813
+ | 0.2448 | 95 | 0.0775 | - | - | - | - | - |
814
+ | 0.2577 | 100 | 0.0906 | 0.1009 | 0.4791 | 0.3151 | 0.6007 | 0.4650 |
815
+ | 0.2706 | 105 | 0.0921 | - | - | - | - | - |
816
+ | 0.2835 | 110 | 0.1105 | - | - | - | - | - |
817
+ | 0.2964 | 115 | 0.0906 | - | - | - | - | - |
818
+ | 0.3093 | 120 | 0.1002 | - | - | - | - | - |
819
+ | 0.3222 | 125 | 0.0952 | - | - | - | - | - |
820
+ | 0.3351 | 130 | 0.0652 | - | - | - | - | - |
821
+ | 0.3479 | 135 | 0.079 | - | - | - | - | - |
822
+ | 0.3608 | 140 | 0.0951 | - | - | - | - | - |
823
+ | 0.3737 | 145 | 0.0918 | - | - | - | - | - |
824
+ | 0.3866 | 150 | 0.065 | 0.0772 | 0.5115 | 0.3070 | 0.6105 | 0.4763 |
825
+ | 0.3995 | 155 | 0.1065 | - | - | - | - | - |
826
+ | 0.4124 | 160 | 0.0871 | - | - | - | - | - |
827
+ | 0.4253 | 165 | 0.0623 | - | - | - | - | - |
828
+ | 0.4381 | 170 | 0.0771 | - | - | - | - | - |
829
+ | 0.4510 | 175 | 0.0795 | - | - | - | - | - |
830
+ | 0.4639 | 180 | 0.0814 | - | - | - | - | - |
831
+ | 0.4768 | 185 | 0.0794 | - | - | - | - | - |
832
+ | 0.4897 | 190 | 0.0744 | - | - | - | - | - |
833
+ | 0.5026 | 195 | 0.0612 | - | - | - | - | - |
834
+ | 0.5155 | 200 | 0.0684 | 0.0692 | 0.4818 | 0.3173 | 0.6161 | 0.4717 |
835
+ | 0.5284 | 205 | 0.0635 | - | - | - | - | - |
836
+ | 0.5412 | 210 | 0.0768 | - | - | - | - | - |
837
+ | 0.5541 | 215 | 0.0544 | - | - | - | - | - |
838
+ | 0.5670 | 220 | 0.0654 | - | - | - | - | - |
839
+ | 0.5799 | 225 | 0.0729 | - | - | - | - | - |
840
+ | 0.5928 | 230 | 0.0923 | - | - | - | - | - |
841
+ | 0.6057 | 235 | 0.0763 | - | - | - | - | - |
842
+ | 0.6186 | 240 | 0.0687 | - | - | - | - | - |
843
+ | 0.6314 | 245 | 0.0657 | - | - | - | - | - |
844
+ | 0.6443 | 250 | 0.0708 | 0.0643 | 0.4843 | 0.3152 | 0.6023 | 0.4673 |
845
+ | 0.6572 | 255 | 0.0555 | - | - | - | - | - |
846
+ | 0.6701 | 260 | 0.0792 | - | - | - | - | - |
847
+ | 0.6830 | 265 | 0.0681 | - | - | - | - | - |
848
+ | 0.6959 | 270 | 0.0855 | - | - | - | - | - |
849
+ | 0.7088 | 275 | 0.0788 | - | - | - | - | - |
850
+ | 0.7216 | 280 | 0.0631 | - | - | - | - | - |
851
+ | 0.7345 | 285 | 0.0676 | - | - | - | - | - |
852
+ | 0.7474 | 290 | 0.0536 | - | - | - | - | - |
853
+ | 0.7603 | 295 | 0.0814 | - | - | - | - | - |
854
+ | 0.7732 | 300 | 0.062 | 0.0606 | 0.4630 | 0.3235 | 0.6256 | 0.4707 |
855
+ | 0.7861 | 305 | 0.0777 | - | - | - | - | - |
856
+ | 0.7990 | 310 | 0.0801 | - | - | - | - | - |
857
+ | 0.8119 | 315 | 0.0566 | - | - | - | - | - |
858
+ | 0.8247 | 320 | 0.0711 | - | - | - | - | - |
859
+ | 0.8376 | 325 | 0.0643 | - | - | - | - | - |
860
+ | 0.8505 | 330 | 0.0422 | - | - | - | - | - |
861
+ | 0.8634 | 335 | 0.0614 | - | - | - | - | - |
862
+ | 0.8763 | 340 | 0.06 | - | - | - | - | - |
863
+ | 0.8892 | 345 | 0.0584 | - | - | - | - | - |
864
+ | 0.9021 | 350 | 0.0457 | 0.0583 | 0.4952 | 0.3214 | 0.6268 | 0.4811 |
865
+ | 0.9149 | 355 | 0.0838 | - | - | - | - | - |
866
+ | 0.9278 | 360 | 0.0657 | - | - | - | - | - |
867
+ | 0.9407 | 365 | 0.0658 | - | - | - | - | - |
868
+ | 0.9536 | 370 | 0.0757 | - | - | - | - | - |
869
+ | 0.9665 | 375 | 0.0603 | - | - | - | - | - |
870
+ | 0.9794 | 380 | 0.0647 | - | - | - | - | - |
871
+ | 0.9923 | 385 | 0.0575 | - | - | - | - | - |
872
+ | -1 | -1 | - | - | 0.4909 | 0.3236 | 0.6322 | 0.4822 |
873
+
874
+
875
+ ### Environmental Impact
876
+ Carbon emissions were measured using [CodeCarbon](https://github.com/mlco2/codecarbon).
877
+ - **Energy Consumed**: 1.043 kWh
878
+ - **Carbon Emitted**: 0.405 kg of CO2
879
+ - **Hours Used**: 3.425 hours
880
+
881
+ ### Training Hardware
882
+ - **On Cloud**: No
883
+ - **GPU Model**: 1 x NVIDIA GeForce RTX 3090
884
+ - **CPU Model**: 13th Gen Intel(R) Core(TM) i7-13700K
885
+ - **RAM Size**: 31.78 GB
886
+
887
+ ### Framework Versions
888
+ - Python: 3.11.6
889
+ - Sentence Transformers: 5.1.0.dev0
890
+ - Transformers: 4.53.0
891
+ - PyTorch: 2.7.1+cu126
892
+ - Accelerate: 1.5.1
893
+ - Datasets: 2.21.0
894
+ - Tokenizers: 0.21.1
895
+
896
+ ## Citation
897
+
898
+ ### BibTeX
899
+
900
+ #### Sentence Transformers
901
+ ```bibtex
902
+ @inproceedings{reimers-2019-sentence-bert,
903
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
904
+ author = "Reimers, Nils and Gurevych, Iryna",
905
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
906
+ month = "11",
907
+ year = "2019",
908
+ publisher = "Association for Computational Linguistics",
909
+ url = "https://arxiv.org/abs/1908.10084",
910
+ }
911
+ ```
912
+
913
+ #### CachedMultipleNegativesRankingLoss
914
+ ```bibtex
915
+ @misc{gao2021scaling,
916
+ title={Scaling Deep Contrastive Learning Batch Size under Memory Limited Setup},
917
+ author={Luyu Gao and Yunyi Zhang and Jiawei Han and Jamie Callan},
918
+ year={2021},
919
+ eprint={2101.06983},
920
+ archivePrefix={arXiv},
921
+ primaryClass={cs.LG}
922
+ }
923
+ ```
924
+
925
+ <!--
926
+ ## Glossary
927
+
928
+ *Clearly define terms in order to be accessible across audiences.*
929
+ -->
930
+
931
+ <!--
932
+ ## Model Card Authors
933
+
934
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
935
+ -->
936
+
937
+ <!--
938
+ ## Model Card Contact
939
+
940
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
941
+ -->
chat_template.jinja ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '
2
+ ' + message['content'] + '<|im_end|>' + '
3
+ '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
4
+ ' }}{% endif %}
config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LFM2Model"
4
+ ],
5
+ "auto_map": {
6
+ "AutoConfig": "modeling_lfm2.LFM2Config",
7
+ "AutoModel": "modeling_lfm2.LFM2Model",
8
+ "AutoModelForCausalLM": "modeling_lfm2.LFM2ForCausalLM"
9
+ },
10
+ "block_auto_adjust_ff_dim": true,
11
+ "block_dim": 1024,
12
+ "block_ff_dim": 6656,
13
+ "block_ffn_dim_multiplier": 1.0,
14
+ "block_mlp_init_scale": 1.0,
15
+ "block_multiple_of": 256,
16
+ "block_norm_eps": 1e-05,
17
+ "block_out_init_scale": 1.0,
18
+ "block_use_swiglu": true,
19
+ "block_use_xavier_init": true,
20
+ "bos_token_id": 1,
21
+ "conv_L_cache": 3,
22
+ "conv_bias": false,
23
+ "conv_dim": 1024,
24
+ "conv_dim_out": 1024,
25
+ "conv_use_xavier_init": true,
26
+ "eos_token_id": 7,
27
+ "full_attn_idxs": [
28
+ 2,
29
+ 5,
30
+ 8,
31
+ 10,
32
+ 12,
33
+ 14
34
+ ],
35
+ "hidden_size": 1024,
36
+ "initializer_range": 0.02,
37
+ "max_position_embeddings": 128000,
38
+ "model_type": "lfm2",
39
+ "norm_eps": 1e-05,
40
+ "num_attention_heads": 16,
41
+ "num_heads": 16,
42
+ "num_hidden_layers": 16,
43
+ "num_key_value_heads": 8,
44
+ "pad_token_id": 0,
45
+ "rope_theta": 1000000.0,
46
+ "torch_dtype": "float32",
47
+ "transformers_version": "4.53.0",
48
+ "use_cache": true,
49
+ "use_pos_enc": true,
50
+ "vocab_size": 65536
51
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "SentenceTransformer",
3
+ "__version__": {
4
+ "sentence_transformers": "5.1.0.dev0",
5
+ "transformers": "4.53.0",
6
+ "pytorch": "2.7.1+cu126"
7
+ },
8
+ "prompts": {
9
+ "query": "",
10
+ "document": ""
11
+ },
12
+ "default_prompt_name": null,
13
+ "similarity_fn_name": "cosine"
14
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5fd26d7b8c8324fed628003486cb73bda756fb45b7dac7c53a325a33d7879ca
3
+ size 1417951464
modeling_lfm2.py ADDED
@@ -0,0 +1,919 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from typing import Any, Callable, ClassVar, Optional, Union
2
+
3
+ import torch
4
+ import torch.nn as nn
5
+ import torch.nn.functional as F
6
+ from transformers.cache_utils import DynamicCache
7
+ from transformers.configuration_utils import PretrainedConfig
8
+ from transformers.generation import GenerationMixin
9
+ from transformers.masking_utils import create_causal_mask
10
+ from transformers.modeling_flash_attention_utils import FlashAttentionKwargs
11
+ from transformers.modeling_layers import GradientCheckpointingLayer
12
+ from transformers.modeling_outputs import BaseModelOutputWithPast, CausalLMOutputWithPast
13
+ from transformers.modeling_rope_utils import ROPE_INIT_FUNCTIONS, dynamic_rope_update
14
+ from transformers.modeling_utils import ALL_ATTENTION_FUNCTIONS, PreTrainedModel
15
+ from transformers.processing_utils import Unpack
16
+ from transformers.utils import LossKwargs, auto_docstring, can_return_tuple, logging
17
+
18
+
19
+ causal_conv1d_fn, causal_conv1d_update = None, None
20
+ kernel_modules = (causal_conv1d_fn, causal_conv1d_update)
21
+ is_fast_path_available = all(kernel_modules)
22
+
23
+ logger = logging.get_logger(__name__)
24
+
25
+
26
+ # ========================================================
27
+ # Config Class (to be removed) once integrated into
28
+ # `transformers`. For now, allows for dynamic importing.
29
+ # ========================================================s
30
+ # from .configuration_lfm2 import LFM2Config
31
+
32
+
33
+ class LFM2Config(PretrainedConfig):
34
+ model_type = "lfm2"
35
+ keys_to_ignore_at_inference: ClassVar = ["past_key_values"]
36
+
37
+ def __init__(
38
+ self,
39
+ vocab_size: int = 65536,
40
+ hidden_size: int = 2560,
41
+ num_hidden_layers: int = 32,
42
+ pad_token_id: int = 0,
43
+ bos_token_id: int = 1,
44
+ eos_token_id: int = 2,
45
+ tie_embedding: bool = True,
46
+ theta: float = 1000000.0,
47
+ max_position_embeddings: int = 128_000,
48
+ use_cache: bool = True,
49
+ norm_eps: float = 0.00001,
50
+ initializer_range: float = 0.02,
51
+ num_attention_heads: int = 32,
52
+ num_key_value_heads: int = 8,
53
+ conv_bias: bool = False,
54
+ conv_dim: int = 2560,
55
+ conv_L_cache: int = 3,
56
+ block_dim: int = 2560,
57
+ block_ff_dim: int = 12288,
58
+ block_multiple_of: int = 256,
59
+ block_ffn_dim_multiplier: float = 1.0,
60
+ block_auto_adjust_ff_dim: bool = True,
61
+ full_attn_idxs: Optional[list[int]] = None,
62
+ **kwargs,
63
+ ):
64
+ self.vocab_size = vocab_size
65
+ self.hidden_size = hidden_size
66
+ self.num_hidden_layers = num_hidden_layers
67
+ self.rope_theta = theta
68
+ self.max_position_embeddings = max_position_embeddings
69
+ self.use_cache = use_cache
70
+ self.norm_eps = norm_eps
71
+ self.initializer_range = initializer_range
72
+
73
+ # attn operator config
74
+ self.num_attention_heads = num_attention_heads
75
+ self.num_key_value_heads = num_key_value_heads
76
+ self.full_attn_idxs = full_attn_idxs
77
+
78
+ # custom operator config
79
+ self.conv_bias = conv_bias
80
+ self.conv_dim = conv_dim
81
+ self.conv_L_cache = conv_L_cache
82
+
83
+ # block config
84
+ self.block_dim = block_dim
85
+ self.block_ff_dim = block_ff_dim
86
+ self.block_multiple_of = block_multiple_of
87
+ self.block_ffn_dim_multiplier = block_ffn_dim_multiplier
88
+ self.block_auto_adjust_ff_dim = block_auto_adjust_ff_dim
89
+
90
+ super().__init__(
91
+ pad_token_id=pad_token_id,
92
+ bos_token_id=bos_token_id,
93
+ eos_token_id=eos_token_id,
94
+ tie_word_embeddings=tie_embedding,
95
+ **kwargs,
96
+ )
97
+
98
+ @property
99
+ def layers_block_type(self):
100
+ return ["attention" if i in self.full_attn_idxs else "conv" for i in range(self.num_hidden_layers)]
101
+
102
+
103
+ class LFM2RMSNorm(torch.nn.Module):
104
+ def __init__(self, dim: int, eps: float = 1e-6):
105
+ super().__init__()
106
+ self.eps = eps
107
+ self.weight = nn.Parameter(torch.ones(dim))
108
+
109
+ def _norm(self, x):
110
+ return x * torch.rsqrt(x.pow(2).mean(-1, keepdim=True) + self.eps)
111
+
112
+ def forward(self, x):
113
+ output = self._norm(x.float())
114
+ return output.type_as(x) * self.weight
115
+
116
+
117
+ def rotate_half(x):
118
+ """Rotates half the hidden dims of the input."""
119
+ x1 = x[..., : x.shape[-1] // 2]
120
+ x2 = x[..., x.shape[-1] // 2 :]
121
+ return torch.cat((-x2, x1), dim=-1)
122
+
123
+
124
+ def apply_rotary_pos_emb(q, k, cos, sin, position_ids=None, unsqueeze_dim=1):
125
+ """Applies Rotary Position Embedding to the query and key tensors."""
126
+ cos = cos.unsqueeze(unsqueeze_dim)
127
+ sin = sin.unsqueeze(unsqueeze_dim)
128
+ q_embed = (q * cos) + (rotate_half(q) * sin)
129
+ k_embed = (k * cos) + (rotate_half(k) * sin)
130
+ return q_embed, k_embed
131
+
132
+
133
+ class LFM2RotaryEmbedding(nn.Module):
134
+ def __init__(self, config: LFM2Config, device=None):
135
+ super().__init__()
136
+ # BC: "rope_type" was originally "type"
137
+ if hasattr(config, "rope_scaling") and config.rope_scaling is not None:
138
+ self.rope_type = config.rope_scaling.get("rope_type", config.rope_scaling.get("type"))
139
+ else:
140
+ self.rope_type = "default"
141
+ self.max_seq_len_cached = config.max_position_embeddings
142
+ self.original_max_seq_len = config.max_position_embeddings
143
+
144
+ self.config = config
145
+ self.rope_init_fn = ROPE_INIT_FUNCTIONS[self.rope_type]
146
+
147
+ inv_freq, self.attention_scaling = self.rope_init_fn(self.config, device)
148
+ self.register_buffer("inv_freq", inv_freq, persistent=False)
149
+ self.original_inv_freq = self.inv_freq
150
+
151
+ @torch.no_grad()
152
+ @dynamic_rope_update # power user: used with advanced RoPE types (e.g. dynamic rope)
153
+ def forward(self, x, position_ids):
154
+ inv_freq_expanded = self.inv_freq[None, :, None].float().expand(position_ids.shape[0], -1, 1).to(x.device)
155
+ position_ids_expanded = position_ids[:, None, :].float()
156
+
157
+ device_type = x.device.type if isinstance(x.device.type, str) and x.device.type != "mps" else "cpu"
158
+ with torch.autocast(device_type=device_type, enabled=False): # Force float32
159
+ freqs = (inv_freq_expanded.float() @ position_ids_expanded.float()).transpose(1, 2)
160
+ emb = torch.cat((freqs, freqs), dim=-1)
161
+ cos = emb.cos() * self.attention_scaling
162
+ sin = emb.sin() * self.attention_scaling
163
+
164
+ return cos.to(dtype=x.dtype), sin.to(dtype=x.dtype)
165
+
166
+
167
+ def repeat_kv(hidden_states: torch.Tensor, n_rep: int) -> torch.Tensor:
168
+ """
169
+ This is the equivalent of torch.repeat_interleave(x, dim=1, repeats=n_rep). The hidden states go from (batch,
170
+ num_key_value_heads, seqlen, head_dim) to (batch, num_attention_heads, seqlen, head_dim)
171
+ """
172
+ batch, num_key_value_heads, slen, head_dim = hidden_states.shape
173
+ if n_rep == 1:
174
+ return hidden_states
175
+ hidden_states = hidden_states[:, :, None, :, :].expand(batch, num_key_value_heads, n_rep, slen, head_dim)
176
+ return hidden_states.reshape(batch, num_key_value_heads * n_rep, slen, head_dim)
177
+
178
+
179
+ def eager_attention_forward(
180
+ module: nn.Module,
181
+ query: torch.Tensor,
182
+ key: torch.Tensor,
183
+ value: torch.Tensor,
184
+ attention_mask: Optional[torch.Tensor],
185
+ scaling: float,
186
+ dropout: float = 0.0,
187
+ **kwargs,
188
+ ):
189
+ num_key_value_groups = query.shape[1] // key.shape[1]
190
+ key_states = repeat_kv(key, num_key_value_groups)
191
+ value_states = repeat_kv(value, num_key_value_groups)
192
+
193
+ attn_weights = torch.matmul(query, key_states.transpose(2, 3)) * scaling
194
+ if attention_mask is not None:
195
+ causal_mask = attention_mask[:, :, :, : key_states.shape[-2]]
196
+ attn_weights = attn_weights + causal_mask
197
+ else:
198
+ seq_len = key_states.shape[-2]
199
+ causal_mask = torch.triu(
200
+ torch.full((seq_len, seq_len), float("-inf"), device=attn_weights.device),
201
+ diagonal=1,
202
+ )
203
+ attn_weights = attn_weights + causal_mask
204
+
205
+ attn_weights = nn.functional.softmax(attn_weights, dim=-1, dtype=torch.float32).to(query.dtype)
206
+ attn_weights = nn.functional.dropout(attn_weights, p=dropout, training=module.training)
207
+ attn_output = torch.matmul(attn_weights, value_states)
208
+ attn_output = attn_output.transpose(1, 2).contiguous()
209
+
210
+ return attn_output, attn_weights
211
+
212
+
213
+ class LFM2MLP(nn.Module):
214
+ def __init__(
215
+ self,
216
+ dim: int,
217
+ ff_dim: int,
218
+ multiple_of: int,
219
+ auto_adjust_ff_dim: bool,
220
+ ffn_dim_multiplier: Optional[float],
221
+ ):
222
+ super().__init__()
223
+ if auto_adjust_ff_dim:
224
+ ff_dim = int(2 * ff_dim / 3)
225
+ # custom dim factor multiplier
226
+ if ffn_dim_multiplier is not None:
227
+ ff_dim = int(ffn_dim_multiplier * ff_dim)
228
+ ff_dim = multiple_of * ((ff_dim + multiple_of - 1) // multiple_of)
229
+
230
+ self.w1 = nn.Linear(dim, ff_dim, bias=False)
231
+ self.w3 = nn.Linear(dim, ff_dim, bias=False)
232
+ self.w2 = nn.Linear(ff_dim, dim, bias=False)
233
+
234
+ def forward(self, x):
235
+ return self.w2(F.silu(self.w1(x)) * self.w3(x))
236
+
237
+
238
+ class LFM2Cache(DynamicCache):
239
+ """
240
+ Attention and conv cache for LFM2.
241
+
242
+ It stores the Key and Value states as a list of tensors, one for each layer.
243
+ Attention layer cache shape: `[batch_size, num_heads, seq_len, head_dim]`.
244
+ Conv layer cache shape: `[batch_size, conv_dim, L_cache-1]`.
245
+ """
246
+
247
+ def __init__(
248
+ self,
249
+ config: LFM2Config,
250
+ max_batch_size: int,
251
+ dtype: torch.dtype = torch.float32,
252
+ device: Union[torch.device, str, None] = None,
253
+ ):
254
+ super().__init__() # initialize key and value cache
255
+ self.max_batch_size = max_batch_size
256
+ self.full_attn_idxs = config.full_attn_idxs
257
+ self.conv_L_cache = config.conv_L_cache
258
+ self._dtype = dtype
259
+
260
+ self.conv_cache: list[torch.Tensor] = []
261
+ device = torch.device(device) if device is not None else None
262
+
263
+ for _ in range(config.num_hidden_layers):
264
+ conv_state = torch.zeros(
265
+ self.max_batch_size,
266
+ config.conv_dim,
267
+ self.conv_L_cache,
268
+ dtype=self._dtype,
269
+ device=device,
270
+ )
271
+ torch._dynamo.mark_static_address(conv_state)
272
+ self.conv_cache.append(conv_state)
273
+
274
+ def update(
275
+ self,
276
+ key_states: torch.Tensor,
277
+ value_states: torch.Tensor,
278
+ layer_idx: int,
279
+ cache_kwargs: Optional[dict[str, Any]] = None,
280
+ ) -> tuple[torch.Tensor, torch.Tensor]:
281
+ """
282
+ Updates the cache with the new `key_states` and `value_states` for the layer `layer_idx`.
283
+
284
+ Parameters:
285
+ key_states (`torch.Tensor`):
286
+ The new key states to cache.
287
+ value_states (`torch.Tensor`):
288
+ The new value states to cache.
289
+ layer_idx (`int`):
290
+ The index of the layer to cache the states for.
291
+ cache_kwargs (`Dict[str, Any]`, `optional`):
292
+ Additional arguments for the cache subclass. No additional arguments are used in `DynamicCache`.
293
+
294
+ Return:
295
+ A tuple containing the updated key and value states.
296
+ """
297
+ # Update the number of seen tokens
298
+ # if layer_idx == 0:
299
+ if layer_idx == self.full_attn_idxs[0]:
300
+ self._seen_tokens += key_states.shape[-2]
301
+
302
+ # Update the cache
303
+ if key_states is not None:
304
+ if len(self.key_cache) <= layer_idx:
305
+ # There may be skipped layers, fill them with empty lists
306
+ for _ in range(len(self.key_cache), layer_idx):
307
+ self.key_cache.append(torch.tensor([]))
308
+ self.value_cache.append(torch.tensor([]))
309
+ self.key_cache.append(key_states)
310
+ self.value_cache.append(value_states)
311
+ elif (
312
+ not self.key_cache[layer_idx].numel() # prefers not t.numel() to len(t) == 0 to export the model
313
+ ): # fills previously skipped layers; checking for tensor causes errors
314
+ self.key_cache[layer_idx] = key_states
315
+ self.value_cache[layer_idx] = value_states
316
+ else:
317
+ self.key_cache[layer_idx] = torch.cat([self.key_cache[layer_idx], key_states], dim=-2)
318
+ self.value_cache[layer_idx] = torch.cat([self.value_cache[layer_idx], value_states], dim=-2)
319
+
320
+ return self.key_cache[layer_idx], self.value_cache[layer_idx]
321
+
322
+ def reorder_cache(self, beam_idx: torch.LongTensor):
323
+ """Reorders the cache for beam search, given the selected beam indices."""
324
+ for layer_idx in range(len(self.key_cache)):
325
+ device = self.key_cache[layer_idx].device
326
+ self.key_cache[layer_idx] = self.key_cache[layer_idx].index_select(0, beam_idx.to(device))
327
+ device = self.value_cache[layer_idx].device
328
+ self.value_cache[layer_idx] = self.value_cache[layer_idx].index_select(0, beam_idx.to(device))
329
+
330
+ device = self.conv_cache[layer_idx].device
331
+ self.conv_cache[layer_idx] = self.conv_cache[layer_idx].index_select(0, beam_idx.to(device))
332
+
333
+ def get_seq_length(self, layer_idx: Optional[int] = 0) -> int:
334
+ """Returns the sequence length of the cached states. A layer index can be optionally passed."""
335
+ # take any layer that contains cache and not empty tensor
336
+ layer_idx = self.full_attn_idxs[0] if layer_idx not in self.full_attn_idxs else layer_idx
337
+ if len(self.key_cache) <= layer_idx or self.key_cache[layer_idx].numel() == 0:
338
+ return 0
339
+ return self.key_cache[layer_idx].shape[-2]
340
+
341
+ def reset(self):
342
+ for layer_idx in range(len(self.conv_cache)):
343
+ # In-place ops prevent breaking the static address
344
+ self.conv_cache[layer_idx].zero_()
345
+
346
+
347
+ class LFM2Attention(nn.Module):
348
+ def __init__(self, config: LFM2Config, layer_idx: Optional[int] = None, **kwargs):
349
+ super().__init__()
350
+ self.config = config
351
+ self.layer_idx = layer_idx
352
+ if layer_idx is None:
353
+ logger.warning_once(
354
+ f"Instantiating {self.__class__.__name__} without passing a `layer_idx` is not recommended and "
355
+ "will lead to errors during the forward call if caching is used. Please make sure to provide a "
356
+ "`layer_idx` when creating this class."
357
+ )
358
+ self.head_dim = config.hidden_size // config.num_attention_heads
359
+ self.num_key_value_heads = config.num_key_value_heads
360
+ self.num_key_value_groups = config.num_attention_heads // config.num_key_value_heads
361
+ self.scaling = self.head_dim**-0.5
362
+ self.is_causal = True
363
+
364
+ self.q_layernorm = LFM2RMSNorm(self.head_dim, eps=config.norm_eps)
365
+ self.k_layernorm = LFM2RMSNorm(self.head_dim, eps=config.norm_eps)
366
+
367
+ self.q_proj = nn.Linear(config.hidden_size, config.num_attention_heads * self.head_dim, bias=False)
368
+ self.k_proj = nn.Linear(config.hidden_size, config.num_key_value_heads * self.head_dim, bias=False)
369
+ self.v_proj = nn.Linear(config.hidden_size, config.num_key_value_heads * self.head_dim, bias=False)
370
+ self.out_proj = nn.Linear(config.num_attention_heads * self.head_dim, config.hidden_size, bias=False)
371
+
372
+ def forward(
373
+ self,
374
+ hidden_states: torch.Tensor,
375
+ position_embeddings: tuple[torch.Tensor, torch.Tensor],
376
+ attention_mask: Optional[torch.Tensor],
377
+ past_key_value: Optional[LFM2Cache] = None,
378
+ cache_position: Optional[torch.LongTensor] = None,
379
+ **kwargs,
380
+ ) -> tuple[torch.Tensor, Optional[torch.Tensor], Optional[tuple[torch.Tensor]]]:
381
+ input_shape = hidden_states.shape[:-1]
382
+ hidden_shape = (*input_shape, -1, self.head_dim)
383
+
384
+ q = self.q_layernorm(self.q_proj(hidden_states).view(*hidden_shape)).transpose(1, 2)
385
+ k = self.k_layernorm(self.k_proj(hidden_states).view(*hidden_shape)).transpose(1, 2)
386
+ v = self.v_proj(hidden_states).view(*hidden_shape).transpose(1, 2)
387
+
388
+ cos, sin = position_embeddings
389
+ q, k = apply_rotary_pos_emb(q, k, cos, sin)
390
+
391
+ if past_key_value is not None:
392
+ cache_kwargs = {"sin": sin, "cos": cos, "cache_position": cache_position}
393
+ k, v = past_key_value.update(
394
+ key_states=k, value_states=v, layer_idx=self.layer_idx, cache_kwargs=cache_kwargs
395
+ )
396
+
397
+ attention_interface: Callable = eager_attention_forward
398
+ if self.config._attn_implementation != "eager":
399
+ attention_interface = ALL_ATTENTION_FUNCTIONS[self.config._attn_implementation]
400
+
401
+ attn_output, attn_weights = attention_interface(
402
+ self,
403
+ q,
404
+ k,
405
+ v,
406
+ attention_mask,
407
+ dropout=0.0,
408
+ scaling=self.scaling,
409
+ **kwargs,
410
+ )
411
+ attn_output = attn_output.reshape(*input_shape, -1).contiguous()
412
+ output = self.out_proj(attn_output)
413
+ return output, attn_weights
414
+
415
+
416
+ class LFM2ShortConv(nn.Module):
417
+ def __init__(
418
+ self,
419
+ config: LFM2Config,
420
+ dim: int,
421
+ layer_idx: int,
422
+ ):
423
+ super().__init__()
424
+ self.config = config
425
+ self.layer_idx = layer_idx
426
+ self.L_cache = config.conv_L_cache
427
+ self.bias = config.conv_bias
428
+
429
+ self.conv = nn.Conv1d(
430
+ in_channels=dim,
431
+ out_channels=dim,
432
+ kernel_size=self.L_cache,
433
+ groups=dim,
434
+ bias=self.bias,
435
+ padding=self.L_cache - 1,
436
+ )
437
+ self.in_proj = nn.Linear(dim, 3 * dim, bias=self.bias)
438
+ self.out_proj = nn.Linear(dim, dim, bias=self.bias)
439
+
440
+ def cuda_kernels_forward(
441
+ self,
442
+ x: torch.Tensor,
443
+ cache_params: Optional[LFM2Cache] = None,
444
+ cache_position: Optional[torch.LongTensor] = None,
445
+ attention_mask: Optional[torch.Tensor] = None,
446
+ ):
447
+ BCx = self.in_proj(x).transpose(-1, -2)
448
+ B, C, x = BCx.chunk(3, dim=-2)
449
+
450
+ Bx = B * x
451
+
452
+ conv_weights = self.conv.weight.view(self.conv.weight.size(0), self.conv.weight.size(2))
453
+ if cache_params is not None and cache_position[0] > 0:
454
+ conv_out = causal_conv1d_update(
455
+ Bx.squeeze(-1),
456
+ cache_params.conv_cache[self.layer_idx],
457
+ conv_weights,
458
+ self.conv.bias,
459
+ None,
460
+ )
461
+ conv_out = conv_out.unsqueeze(-1)
462
+ else:
463
+ if cache_params is not None:
464
+ conv_state = nn.functional.pad(Bx, (self.L_cache - Bx.shape[-1], 0))
465
+ cache_params.conv_cache[self.layer_idx].copy_(conv_state)
466
+
467
+ conv_out = causal_conv1d_fn(Bx, conv_weights, self.conv.bias, activation=None)
468
+
469
+ y = C * conv_out
470
+ y = self.out_proj(y.transpose(-1, -2).contiguous())
471
+ return y
472
+
473
+ def slow_forward(
474
+ self,
475
+ x: torch.Tensor,
476
+ cache_params: Optional[LFM2Cache] = None,
477
+ cache_position: Optional[torch.LongTensor] = None,
478
+ attention_mask: Optional[torch.Tensor] = None,
479
+ ):
480
+ seqlen = x.shape[1]
481
+ BCx = self.in_proj(x).transpose(-1, -2)
482
+ B, C, x = BCx.chunk(3, dim=-2)
483
+
484
+ Bx = B * x
485
+
486
+ if cache_params is not None and cache_position[0] > 0:
487
+ conv_state = cache_params.conv_cache[self.layer_idx]
488
+ cache_position = cache_position.clamp(0, self.L_cache - 1)
489
+ conv_state = conv_state.roll(shifts=-1, dims=-1)
490
+ conv_state[:, :, cache_position] = Bx.to(device=conv_state.device, dtype=conv_state.dtype)
491
+ cache_params.conv_cache[self.layer_idx].copy_(conv_state)
492
+ conv_out = torch.sum(conv_state.to(Bx.device) * self.conv.weight[:, 0, :], dim=-1)
493
+ if self.bias:
494
+ conv_out += self.conv.bias
495
+
496
+ conv_out = conv_out.unsqueeze(-1)
497
+ else:
498
+ if cache_params is not None:
499
+ conv_state = nn.functional.pad(Bx, (self.L_cache - Bx.shape[-1], 0))
500
+ cache_params.conv_cache[self.layer_idx].copy_(conv_state)
501
+
502
+ conv_out = self.conv(Bx)[..., :seqlen]
503
+
504
+ y = C * conv_out
505
+ y = y.transpose(-1, -2).contiguous()
506
+ y = self.out_proj(y)
507
+ return y
508
+
509
+ def forward(
510
+ self,
511
+ x: torch.Tensor,
512
+ cache_params: Optional[LFM2Cache] = None,
513
+ cache_position: Optional[torch.LongTensor] = None,
514
+ attention_mask: Optional[torch.Tensor] = None,
515
+ ):
516
+ if is_fast_path_available and "cuda" in x.device.type and not torch._dynamo.is_compiling():
517
+ return self.cuda_kernels_forward(x, cache_params, cache_position, attention_mask)
518
+ return self.slow_forward(x, cache_params, cache_position, attention_mask)
519
+
520
+
521
+ class LFM2AttentionDecoderLayer(GradientCheckpointingLayer):
522
+ def __init__(self, config: LFM2Config, layer_idx: int):
523
+ super().__init__()
524
+ self.self_attn = LFM2Attention(config, layer_idx)
525
+ self.feed_forward = LFM2MLP(
526
+ dim=config.block_dim,
527
+ ff_dim=config.block_ff_dim,
528
+ multiple_of=config.block_multiple_of,
529
+ auto_adjust_ff_dim=config.block_auto_adjust_ff_dim,
530
+ ffn_dim_multiplier=config.block_ffn_dim_multiplier,
531
+ )
532
+ self.operator_norm = LFM2RMSNorm(config.hidden_size, eps=config.norm_eps)
533
+ self.ffn_norm = LFM2RMSNorm(config.hidden_size, eps=config.norm_eps)
534
+
535
+ def forward(
536
+ self,
537
+ hidden_states: torch.Tensor,
538
+ position_embeddings: tuple[torch.Tensor, torch.Tensor],
539
+ attention_mask: Optional[torch.Tensor] = None,
540
+ position_ids: Optional[torch.LongTensor] = None,
541
+ past_key_value: Optional[tuple[torch.Tensor]] = None,
542
+ output_attentions: Optional[bool] = False,
543
+ cache_position: Optional[torch.LongTensor] = None,
544
+ **kwargs,
545
+ ) -> tuple[torch.FloatTensor, Optional[tuple[torch.FloatTensor, torch.FloatTensor]]]:
546
+ h, self_attn_weights = self.self_attn(
547
+ hidden_states=self.operator_norm(hidden_states),
548
+ position_embeddings=position_embeddings,
549
+ attention_mask=attention_mask,
550
+ position_ids=position_ids,
551
+ past_key_value=past_key_value,
552
+ cache_position=cache_position,
553
+ **kwargs,
554
+ )
555
+ h += hidden_states
556
+ out = h + self.feed_forward.forward(self.ffn_norm(h))
557
+
558
+ outputs = (out,)
559
+ if output_attentions:
560
+ outputs += (self_attn_weights,)
561
+
562
+ return outputs
563
+
564
+
565
+ class LFM2ShortConvDecoderLayer(GradientCheckpointingLayer):
566
+ def __init__(self, config: LFM2Config, layer_idx: int):
567
+ super().__init__()
568
+ self.conv = LFM2ShortConv(
569
+ config=config,
570
+ dim=config.conv_dim,
571
+ layer_idx=layer_idx,
572
+ )
573
+ self.feed_forward = LFM2MLP(
574
+ dim=config.block_dim,
575
+ ff_dim=config.block_ff_dim,
576
+ multiple_of=config.block_multiple_of,
577
+ auto_adjust_ff_dim=config.block_auto_adjust_ff_dim,
578
+ ffn_dim_multiplier=config.block_ffn_dim_multiplier,
579
+ )
580
+ self.operator_norm = LFM2RMSNorm(config.hidden_size, eps=config.norm_eps)
581
+ self.ffn_norm = LFM2RMSNorm(config.hidden_size, eps=config.norm_eps)
582
+
583
+ def forward(
584
+ self,
585
+ hidden_states: torch.Tensor,
586
+ past_key_value: Optional[LFM2Cache] = None,
587
+ cache_position: Optional[torch.LongTensor] = None,
588
+ attention_mask: Optional[torch.Tensor] = None,
589
+ output_attentions: Optional[bool] = False,
590
+ **kwargs,
591
+ ) -> tuple[torch.FloatTensor, Optional[tuple[torch.FloatTensor, torch.FloatTensor]]]:
592
+ h = self.conv(
593
+ self.operator_norm(hidden_states),
594
+ cache_params=past_key_value,
595
+ cache_position=cache_position,
596
+ attention_mask=attention_mask,
597
+ )
598
+ self_attn_weights = None
599
+
600
+ h += hidden_states
601
+ out = h + self.feed_forward.forward(self.ffn_norm(h))
602
+
603
+ outputs = (out,)
604
+ if output_attentions:
605
+ outputs += (self_attn_weights,)
606
+
607
+ return outputs
608
+
609
+
610
+ @auto_docstring
611
+ class LFM2PretrainedModel(PreTrainedModel):
612
+ config_class = LFM2Config
613
+ base_model_prefix = "model"
614
+ supports_gradient_checkpointing = True
615
+ _no_split_modules: ClassVar = ["LFM2AttentionDecoderLayer", "LFM2ShortConvDecoderLayer"]
616
+ _skip_keys_device_placement = "past_key_values"
617
+ _supports_flash_attn_2 = True
618
+ _supports_sdpa = True
619
+ _supports_flex_attn = True
620
+ _supports_cache_class = True
621
+ _supports_quantized_cache = True
622
+ _supports_static_cache = True
623
+ _supports_attention_backend = True
624
+
625
+ def _init_weights(self, module):
626
+ std = self.config.initializer_range
627
+ if isinstance(module, (nn.Linear, nn.Conv1d)):
628
+ module.weight.data.normal_(mean=0.0, std=std)
629
+ if module.bias is not None:
630
+ module.bias.data.zero_()
631
+ elif isinstance(module, nn.Embedding):
632
+ module.weight.data.normal_(mean=0.0, std=std)
633
+ if module.padding_idx is not None:
634
+ module.weight.data[module.padding_idx].zero_()
635
+ elif isinstance(module, LFM2RMSNorm):
636
+ module.weight.data.fill_(1.0)
637
+
638
+
639
+ class LFM2Model(LFM2PretrainedModel):
640
+ def __init__(self, config: LFM2Config):
641
+ super().__init__(config)
642
+ self.padding_idx = config.pad_token_id
643
+ self.vocab_size = config.vocab_size
644
+
645
+ self.embed_tokens = nn.Embedding(config.vocab_size, config.hidden_size, self.padding_idx)
646
+
647
+ self.pos_emb = LFM2RotaryEmbedding(config)
648
+
649
+ decoder_layers = []
650
+ for i in range(config.num_hidden_layers):
651
+ if i in config.full_attn_idxs:
652
+ decoder_layers.append(LFM2AttentionDecoderLayer(config, layer_idx=i))
653
+ else:
654
+ decoder_layers.append(LFM2ShortConvDecoderLayer(config, layer_idx=i))
655
+ self.layers = nn.ModuleList(decoder_layers)
656
+
657
+ self.embedding_norm = LFM2RMSNorm(config.hidden_size, eps=config.norm_eps)
658
+
659
+ self.gradient_checkpointing = False
660
+
661
+ # Initialize weights and apply final processing
662
+ self.post_init()
663
+
664
+ def get_input_embeddings(self):
665
+ return self.embed_tokens
666
+
667
+ def set_input_embeddings(self, value):
668
+ self.embed_tokens = value
669
+
670
+ @can_return_tuple
671
+ @auto_docstring
672
+ def forward(
673
+ self,
674
+ input_ids: torch.LongTensor = None,
675
+ attention_mask: Optional[torch.Tensor] = None,
676
+ position_ids: Optional[torch.LongTensor] = None,
677
+ past_key_values: Optional[LFM2Cache] = None,
678
+ inputs_embeds: Optional[torch.FloatTensor] = None,
679
+ use_cache: Optional[bool] = None,
680
+ output_attentions: Optional[bool] = None,
681
+ output_hidden_states: Optional[bool] = None,
682
+ return_dict: Optional[bool] = None,
683
+ cache_position: Optional[torch.LongTensor] = None,
684
+ **flash_attn_kwargs: Unpack[FlashAttentionKwargs],
685
+ ) -> BaseModelOutputWithPast:
686
+ output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions
687
+ output_hidden_states = (
688
+ output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states
689
+ )
690
+ use_cache = use_cache if use_cache is not None else self.config.use_cache
691
+
692
+ return_dict = return_dict if return_dict is not None else self.config.use_return_dict
693
+
694
+ if (input_ids is None) ^ (inputs_embeds is not None):
695
+ raise ValueError("You must specify exactly one of input_ids or inputs_embeds")
696
+
697
+ if self.gradient_checkpointing and self.training and use_cache:
698
+ logger.warning_once(
699
+ "`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
700
+ )
701
+ use_cache = False
702
+
703
+ if inputs_embeds is None:
704
+ inputs_embeds = self.embed_tokens(input_ids)
705
+
706
+ if use_cache and past_key_values is None:
707
+ batch_size = inputs_embeds.shape[0]
708
+ past_key_values = LFM2Cache(
709
+ config=self.config, max_batch_size=batch_size, dtype=self.dtype, device=self.device
710
+ )
711
+
712
+ if cache_position is None:
713
+ past_seen_tokens = past_key_values.get_seq_length() if past_key_values is not None else 0
714
+ cache_position = torch.arange(
715
+ past_seen_tokens, past_seen_tokens + inputs_embeds.shape[1], device=inputs_embeds.device
716
+ )
717
+
718
+ if position_ids is None:
719
+ position_ids = cache_position.unsqueeze(0)
720
+
721
+ causal_mask = create_causal_mask(
722
+ config=self.config,
723
+ input_embeds=inputs_embeds,
724
+ attention_mask=attention_mask,
725
+ cache_position=cache_position,
726
+ past_key_values=past_key_values,
727
+ )
728
+ hidden_states = inputs_embeds
729
+
730
+ position_embeddings = self.pos_emb(hidden_states, position_ids)
731
+
732
+ # decoder layers
733
+ all_hidden_states = () if output_hidden_states else None
734
+ all_self_attns = () if output_attentions else None
735
+ for decoder_layer in self.layers:
736
+ if output_hidden_states:
737
+ all_hidden_states += (hidden_states,)
738
+
739
+ layer_outputs = decoder_layer(
740
+ hidden_states,
741
+ attention_mask=causal_mask,
742
+ position_ids=position_ids,
743
+ past_key_value=past_key_values,
744
+ output_attentions=output_attentions,
745
+ use_cache=use_cache,
746
+ cache_position=cache_position,
747
+ position_embeddings=position_embeddings,
748
+ **flash_attn_kwargs,
749
+ )
750
+
751
+ hidden_states = layer_outputs[0]
752
+
753
+ if output_attentions:
754
+ all_self_attns += (layer_outputs[1],)
755
+
756
+ hidden_states = self.embedding_norm(hidden_states)
757
+
758
+ # add hidden states from the last decoder layer
759
+ if output_hidden_states:
760
+ all_hidden_states += (hidden_states,)
761
+
762
+ output = BaseModelOutputWithPast(
763
+ last_hidden_state=hidden_states,
764
+ past_key_values=past_key_values if use_cache else None,
765
+ hidden_states=all_hidden_states,
766
+ attentions=all_self_attns,
767
+ )
768
+ return output if return_dict else output.to_tuple()
769
+
770
+
771
+ class KwargsForCausalLM(FlashAttentionKwargs, LossKwargs): ...
772
+
773
+
774
+ @auto_docstring
775
+ class LFM2ForCausalLM(LFM2PretrainedModel, GenerationMixin):
776
+ _tied_weights_keys = ["lm_head.weight"]
777
+
778
+ def __init__(self, config: LFM2Config):
779
+ super().__init__(config)
780
+ self.model = LFM2Model(config)
781
+ self.vocab_size = config.vocab_size
782
+ self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False)
783
+ self.post_init()
784
+
785
+ def get_input_embeddings(self):
786
+ return self.model.embed_tokens
787
+
788
+ def set_input_embeddings(self, value):
789
+ self.model.embed_tokens = value
790
+
791
+ def get_output_embeddings(self):
792
+ return self.lm_head
793
+
794
+ def set_output_embeddings(self, new_embeddings):
795
+ self.lm_head = new_embeddings
796
+
797
+ def set_decoder(self, decoder):
798
+ self.model = decoder
799
+
800
+ def get_decoder(self):
801
+ return self.model
802
+
803
+ def forward(
804
+ self,
805
+ input_ids: torch.LongTensor = None,
806
+ attention_mask: Optional[torch.Tensor] = None,
807
+ position_ids: Optional[torch.LongTensor] = None,
808
+ past_key_values: Optional[LFM2Cache] = None,
809
+ inputs_embeds: Optional[torch.FloatTensor] = None,
810
+ labels: Optional[torch.LongTensor] = None,
811
+ use_cache: Optional[bool] = None,
812
+ output_attentions: Optional[bool] = None,
813
+ output_hidden_states: Optional[bool] = None,
814
+ return_dict: Optional[bool] = None,
815
+ cache_position: Optional[torch.LongTensor] = None,
816
+ logits_to_keep: Union[int, torch.Tensor] = 0,
817
+ **kwargs: Unpack[KwargsForCausalLM],
818
+ ) -> Union[tuple, CausalLMOutputWithPast]:
819
+ output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions
820
+ output_hidden_states = (
821
+ output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states
822
+ )
823
+ return_dict = return_dict if return_dict is not None else self.config.use_return_dict
824
+
825
+ outputs: BaseModelOutputWithPast = self.model(
826
+ input_ids=input_ids,
827
+ attention_mask=attention_mask,
828
+ position_ids=position_ids,
829
+ past_key_values=past_key_values,
830
+ inputs_embeds=inputs_embeds,
831
+ use_cache=use_cache,
832
+ output_attentions=output_attentions,
833
+ output_hidden_states=output_hidden_states,
834
+ cache_position=cache_position,
835
+ return_dict=return_dict,
836
+ **kwargs,
837
+ )
838
+
839
+ hidden_states = outputs.last_hidden_state
840
+ # Only compute necessary logits, and do not upcast them to float if we are not computing the loss
841
+ slice_indices = slice(-logits_to_keep, None) if isinstance(logits_to_keep, int) else logits_to_keep
842
+ logits = self.lm_head(hidden_states[:, slice_indices, :])
843
+
844
+ loss = None
845
+ if labels is not None:
846
+ loss = self.loss_function(logits=logits, labels=labels, vocab_size=self.config.vocab_size, **kwargs)
847
+
848
+ if not return_dict:
849
+ output = (logits,) + outputs[1:]
850
+ return (loss,) + output if loss is not None else output
851
+
852
+ return CausalLMOutputWithPast(
853
+ loss=loss,
854
+ logits=logits,
855
+ past_key_values=outputs.past_key_values,
856
+ hidden_states=outputs.hidden_states,
857
+ attentions=outputs.attentions,
858
+ )
859
+
860
+ def prepare_inputs_for_generation(
861
+ self,
862
+ input_ids,
863
+ past_key_values=None,
864
+ attention_mask=None,
865
+ inputs_embeds=None,
866
+ cache_position=None,
867
+ position_ids=None,
868
+ use_cache=True,
869
+ **kwargs,
870
+ ):
871
+ # Overwritten -- Support custom LFM2Cache.
872
+
873
+ empty_past_kv = past_key_values is None or (
874
+ isinstance(past_key_values, DynamicCache) and past_key_values._seen_tokens == 0
875
+ )
876
+
877
+ # Omit tokens covered by past_key_values.
878
+ if not empty_past_kv:
879
+ # If we have cache: let's slice `input_ids` through `cache_position`, to keep only the unprocessed tokens
880
+ # Exception 1: when passing input_embeds, input_ids may be missing entries
881
+ # Exception 2: some generation methods do special slicing of input_ids, so we don't need to do it here
882
+ # Exception 3: with synced GPUs cache_position may go out of bounds, but we only want dummy token in that case.
883
+ # (we can't check exception 3 while compiling)
884
+ if (
885
+ inputs_embeds is not None # Exception 1
886
+ or cache_position[-1] >= input_ids.shape[1] # Exception 3
887
+ ):
888
+ input_ids = input_ids[:, -cache_position.shape[0] :]
889
+ elif input_ids.shape[1] != cache_position.shape[0]: # Default case (the "else", a no op, is Exception 2)
890
+ input_ids = input_ids[:, cache_position]
891
+ else:
892
+ past_key_values = LFM2Cache(self.config, input_ids.shape[0], dtype=self.dtype, device=self.device)
893
+
894
+ # if attention_mask is not None and position_ids is None:
895
+ # # create position_ids on the fly for batch generation
896
+ # position_ids = attention_mask.long().cumsum(-1) - 1
897
+ # position_ids.masked_fill_(attention_mask == 0, 1)
898
+ # if not empty_past_kv:
899
+ # position_ids = position_ids[:, -input_ids.shape[1] :]
900
+
901
+ # if `inputs_embeds` are passed, we only want to use them in the 1st generation step
902
+ if inputs_embeds is not None and empty_past_kv:
903
+ model_inputs = {"inputs_embeds": inputs_embeds}
904
+ else:
905
+ model_inputs = {"input_ids": input_ids.contiguous()} # `contiguous()` needed for compilation use cases
906
+
907
+ model_inputs.update(
908
+ {
909
+ # "position_ids": position_ids,
910
+ "past_key_values": past_key_values,
911
+ "use_cache": use_cache,
912
+ "attention_mask": attention_mask,
913
+ "cache_position": cache_position,
914
+ }
915
+ )
916
+ return model_inputs
917
+
918
+
919
+ __all__ = ["LFM2ForCausalLM", "LFM2Model", "LFM2PretrainedModel"]
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 128000,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|im_end|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|pad|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,4074 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<|pad|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<|startoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "<|endoftext|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "3": {
30
+ "content": "<|fim_pre|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "4": {
38
+ "content": "<|fim_mid|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "5": {
46
+ "content": "<|fim_suf|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "6": {
54
+ "content": "<|im_start|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "7": {
62
+ "content": "<|im_end|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "8": {
70
+ "content": "<|tool_list_start|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "9": {
78
+ "content": "<|tool_list_end|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "10": {
86
+ "content": "<|tool_call_start|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "11": {
94
+ "content": "<|tool_call_end|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "12": {
102
+ "content": "<|tool_response_start|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "13": {
110
+ "content": "<|tool_response_end|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "16": {
118
+ "content": "<|reserved_6|>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": true
124
+ },
125
+ "17": {
126
+ "content": "<|reserved_7|>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": true
132
+ },
133
+ "18": {
134
+ "content": "<|reserved_8|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": true
140
+ },
141
+ "19": {
142
+ "content": "<|reserved_9|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": true
148
+ },
149
+ "20": {
150
+ "content": "<|reserved_10|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": true
156
+ },
157
+ "21": {
158
+ "content": "<|reserved_11|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": true
164
+ },
165
+ "22": {
166
+ "content": "<|reserved_12|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": true
172
+ },
173
+ "23": {
174
+ "content": "<|reserved_13|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": true
180
+ },
181
+ "24": {
182
+ "content": "<|reserved_14|>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": true
188
+ },
189
+ "25": {
190
+ "content": "<|reserved_15|>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": true
196
+ },
197
+ "26": {
198
+ "content": "<|reserved_16|>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": true
204
+ },
205
+ "27": {
206
+ "content": "<|reserved_17|>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": true
212
+ },
213
+ "28": {
214
+ "content": "<|reserved_18|>",
215
+ "lstrip": false,
216
+ "normalized": false,
217
+ "rstrip": false,
218
+ "single_word": false,
219
+ "special": true
220
+ },
221
+ "29": {
222
+ "content": "<|reserved_19|>",
223
+ "lstrip": false,
224
+ "normalized": false,
225
+ "rstrip": false,
226
+ "single_word": false,
227
+ "special": true
228
+ },
229
+ "30": {
230
+ "content": "<|reserved_20|>",
231
+ "lstrip": false,
232
+ "normalized": false,
233
+ "rstrip": false,
234
+ "single_word": false,
235
+ "special": true
236
+ },
237
+ "31": {
238
+ "content": "<|reserved_21|>",
239
+ "lstrip": false,
240
+ "normalized": false,
241
+ "rstrip": false,
242
+ "single_word": false,
243
+ "special": true
244
+ },
245
+ "32": {
246
+ "content": "<|reserved_22|>",
247
+ "lstrip": false,
248
+ "normalized": false,
249
+ "rstrip": false,
250
+ "single_word": false,
251
+ "special": true
252
+ },
253
+ "33": {
254
+ "content": "<|reserved_23|>",
255
+ "lstrip": false,
256
+ "normalized": false,
257
+ "rstrip": false,
258
+ "single_word": false,
259
+ "special": true
260
+ },
261
+ "34": {
262
+ "content": "<|reserved_24|>",
263
+ "lstrip": false,
264
+ "normalized": false,
265
+ "rstrip": false,
266
+ "single_word": false,
267
+ "special": true
268
+ },
269
+ "35": {
270
+ "content": "<|reserved_25|>",
271
+ "lstrip": false,
272
+ "normalized": false,
273
+ "rstrip": false,
274
+ "single_word": false,
275
+ "special": true
276
+ },
277
+ "36": {
278
+ "content": "<|reserved_26|>",
279
+ "lstrip": false,
280
+ "normalized": false,
281
+ "rstrip": false,
282
+ "single_word": false,
283
+ "special": true
284
+ },
285
+ "37": {
286
+ "content": "<|reserved_27|>",
287
+ "lstrip": false,
288
+ "normalized": false,
289
+ "rstrip": false,
290
+ "single_word": false,
291
+ "special": true
292
+ },
293
+ "38": {
294
+ "content": "<|reserved_28|>",
295
+ "lstrip": false,
296
+ "normalized": false,
297
+ "rstrip": false,
298
+ "single_word": false,
299
+ "special": true
300
+ },
301
+ "39": {
302
+ "content": "<|reserved_29|>",
303
+ "lstrip": false,
304
+ "normalized": false,
305
+ "rstrip": false,
306
+ "single_word": false,
307
+ "special": true
308
+ },
309
+ "40": {
310
+ "content": "<|reserved_30|>",
311
+ "lstrip": false,
312
+ "normalized": false,
313
+ "rstrip": false,
314
+ "single_word": false,
315
+ "special": true
316
+ },
317
+ "41": {
318
+ "content": "<|reserved_31|>",
319
+ "lstrip": false,
320
+ "normalized": false,
321
+ "rstrip": false,
322
+ "single_word": false,
323
+ "special": true
324
+ },
325
+ "42": {
326
+ "content": "<|reserved_32|>",
327
+ "lstrip": false,
328
+ "normalized": false,
329
+ "rstrip": false,
330
+ "single_word": false,
331
+ "special": true
332
+ },
333
+ "43": {
334
+ "content": "<|reserved_33|>",
335
+ "lstrip": false,
336
+ "normalized": false,
337
+ "rstrip": false,
338
+ "single_word": false,
339
+ "special": true
340
+ },
341
+ "44": {
342
+ "content": "<|reserved_34|>",
343
+ "lstrip": false,
344
+ "normalized": false,
345
+ "rstrip": false,
346
+ "single_word": false,
347
+ "special": true
348
+ },
349
+ "45": {
350
+ "content": "<|reserved_35|>",
351
+ "lstrip": false,
352
+ "normalized": false,
353
+ "rstrip": false,
354
+ "single_word": false,
355
+ "special": true
356
+ },
357
+ "46": {
358
+ "content": "<|reserved_36|>",
359
+ "lstrip": false,
360
+ "normalized": false,
361
+ "rstrip": false,
362
+ "single_word": false,
363
+ "special": true
364
+ },
365
+ "47": {
366
+ "content": "<|reserved_37|>",
367
+ "lstrip": false,
368
+ "normalized": false,
369
+ "rstrip": false,
370
+ "single_word": false,
371
+ "special": true
372
+ },
373
+ "48": {
374
+ "content": "<|reserved_38|>",
375
+ "lstrip": false,
376
+ "normalized": false,
377
+ "rstrip": false,
378
+ "single_word": false,
379
+ "special": true
380
+ },
381
+ "49": {
382
+ "content": "<|reserved_39|>",
383
+ "lstrip": false,
384
+ "normalized": false,
385
+ "rstrip": false,
386
+ "single_word": false,
387
+ "special": true
388
+ },
389
+ "50": {
390
+ "content": "<|reserved_40|>",
391
+ "lstrip": false,
392
+ "normalized": false,
393
+ "rstrip": false,
394
+ "single_word": false,
395
+ "special": true
396
+ },
397
+ "51": {
398
+ "content": "<|reserved_41|>",
399
+ "lstrip": false,
400
+ "normalized": false,
401
+ "rstrip": false,
402
+ "single_word": false,
403
+ "special": true
404
+ },
405
+ "52": {
406
+ "content": "<|reserved_42|>",
407
+ "lstrip": false,
408
+ "normalized": false,
409
+ "rstrip": false,
410
+ "single_word": false,
411
+ "special": true
412
+ },
413
+ "53": {
414
+ "content": "<|reserved_43|>",
415
+ "lstrip": false,
416
+ "normalized": false,
417
+ "rstrip": false,
418
+ "single_word": false,
419
+ "special": true
420
+ },
421
+ "54": {
422
+ "content": "<|reserved_44|>",
423
+ "lstrip": false,
424
+ "normalized": false,
425
+ "rstrip": false,
426
+ "single_word": false,
427
+ "special": true
428
+ },
429
+ "55": {
430
+ "content": "<|reserved_45|>",
431
+ "lstrip": false,
432
+ "normalized": false,
433
+ "rstrip": false,
434
+ "single_word": false,
435
+ "special": true
436
+ },
437
+ "56": {
438
+ "content": "<|reserved_46|>",
439
+ "lstrip": false,
440
+ "normalized": false,
441
+ "rstrip": false,
442
+ "single_word": false,
443
+ "special": true
444
+ },
445
+ "57": {
446
+ "content": "<|reserved_47|>",
447
+ "lstrip": false,
448
+ "normalized": false,
449
+ "rstrip": false,
450
+ "single_word": false,
451
+ "special": true
452
+ },
453
+ "58": {
454
+ "content": "<|reserved_48|>",
455
+ "lstrip": false,
456
+ "normalized": false,
457
+ "rstrip": false,
458
+ "single_word": false,
459
+ "special": true
460
+ },
461
+ "59": {
462
+ "content": "<|reserved_49|>",
463
+ "lstrip": false,
464
+ "normalized": false,
465
+ "rstrip": false,
466
+ "single_word": false,
467
+ "special": true
468
+ },
469
+ "60": {
470
+ "content": "<|reserved_50|>",
471
+ "lstrip": false,
472
+ "normalized": false,
473
+ "rstrip": false,
474
+ "single_word": false,
475
+ "special": true
476
+ },
477
+ "61": {
478
+ "content": "<|reserved_51|>",
479
+ "lstrip": false,
480
+ "normalized": false,
481
+ "rstrip": false,
482
+ "single_word": false,
483
+ "special": true
484
+ },
485
+ "62": {
486
+ "content": "<|reserved_52|>",
487
+ "lstrip": false,
488
+ "normalized": false,
489
+ "rstrip": false,
490
+ "single_word": false,
491
+ "special": true
492
+ },
493
+ "63": {
494
+ "content": "<|reserved_53|>",
495
+ "lstrip": false,
496
+ "normalized": false,
497
+ "rstrip": false,
498
+ "single_word": false,
499
+ "special": true
500
+ },
501
+ "64": {
502
+ "content": "<|reserved_54|>",
503
+ "lstrip": false,
504
+ "normalized": false,
505
+ "rstrip": false,
506
+ "single_word": false,
507
+ "special": true
508
+ },
509
+ "65": {
510
+ "content": "<|reserved_55|>",
511
+ "lstrip": false,
512
+ "normalized": false,
513
+ "rstrip": false,
514
+ "single_word": false,
515
+ "special": true
516
+ },
517
+ "66": {
518
+ "content": "<|reserved_56|>",
519
+ "lstrip": false,
520
+ "normalized": false,
521
+ "rstrip": false,
522
+ "single_word": false,
523
+ "special": true
524
+ },
525
+ "67": {
526
+ "content": "<|reserved_57|>",
527
+ "lstrip": false,
528
+ "normalized": false,
529
+ "rstrip": false,
530
+ "single_word": false,
531
+ "special": true
532
+ },
533
+ "68": {
534
+ "content": "<|reserved_58|>",
535
+ "lstrip": false,
536
+ "normalized": false,
537
+ "rstrip": false,
538
+ "single_word": false,
539
+ "special": true
540
+ },
541
+ "69": {
542
+ "content": "<|reserved_59|>",
543
+ "lstrip": false,
544
+ "normalized": false,
545
+ "rstrip": false,
546
+ "single_word": false,
547
+ "special": true
548
+ },
549
+ "70": {
550
+ "content": "<|reserved_60|>",
551
+ "lstrip": false,
552
+ "normalized": false,
553
+ "rstrip": false,
554
+ "single_word": false,
555
+ "special": true
556
+ },
557
+ "71": {
558
+ "content": "<|reserved_61|>",
559
+ "lstrip": false,
560
+ "normalized": false,
561
+ "rstrip": false,
562
+ "single_word": false,
563
+ "special": true
564
+ },
565
+ "72": {
566
+ "content": "<|reserved_62|>",
567
+ "lstrip": false,
568
+ "normalized": false,
569
+ "rstrip": false,
570
+ "single_word": false,
571
+ "special": true
572
+ },
573
+ "73": {
574
+ "content": "<|reserved_63|>",
575
+ "lstrip": false,
576
+ "normalized": false,
577
+ "rstrip": false,
578
+ "single_word": false,
579
+ "special": true
580
+ },
581
+ "74": {
582
+ "content": "<|reserved_64|>",
583
+ "lstrip": false,
584
+ "normalized": false,
585
+ "rstrip": false,
586
+ "single_word": false,
587
+ "special": true
588
+ },
589
+ "75": {
590
+ "content": "<|reserved_65|>",
591
+ "lstrip": false,
592
+ "normalized": false,
593
+ "rstrip": false,
594
+ "single_word": false,
595
+ "special": true
596
+ },
597
+ "76": {
598
+ "content": "<|reserved_66|>",
599
+ "lstrip": false,
600
+ "normalized": false,
601
+ "rstrip": false,
602
+ "single_word": false,
603
+ "special": true
604
+ },
605
+ "77": {
606
+ "content": "<|reserved_67|>",
607
+ "lstrip": false,
608
+ "normalized": false,
609
+ "rstrip": false,
610
+ "single_word": false,
611
+ "special": true
612
+ },
613
+ "78": {
614
+ "content": "<|reserved_68|>",
615
+ "lstrip": false,
616
+ "normalized": false,
617
+ "rstrip": false,
618
+ "single_word": false,
619
+ "special": true
620
+ },
621
+ "79": {
622
+ "content": "<|reserved_69|>",
623
+ "lstrip": false,
624
+ "normalized": false,
625
+ "rstrip": false,
626
+ "single_word": false,
627
+ "special": true
628
+ },
629
+ "80": {
630
+ "content": "<|reserved_70|>",
631
+ "lstrip": false,
632
+ "normalized": false,
633
+ "rstrip": false,
634
+ "single_word": false,
635
+ "special": true
636
+ },
637
+ "81": {
638
+ "content": "<|reserved_71|>",
639
+ "lstrip": false,
640
+ "normalized": false,
641
+ "rstrip": false,
642
+ "single_word": false,
643
+ "special": true
644
+ },
645
+ "82": {
646
+ "content": "<|reserved_72|>",
647
+ "lstrip": false,
648
+ "normalized": false,
649
+ "rstrip": false,
650
+ "single_word": false,
651
+ "special": true
652
+ },
653
+ "83": {
654
+ "content": "<|reserved_73|>",
655
+ "lstrip": false,
656
+ "normalized": false,
657
+ "rstrip": false,
658
+ "single_word": false,
659
+ "special": true
660
+ },
661
+ "84": {
662
+ "content": "<|reserved_74|>",
663
+ "lstrip": false,
664
+ "normalized": false,
665
+ "rstrip": false,
666
+ "single_word": false,
667
+ "special": true
668
+ },
669
+ "85": {
670
+ "content": "<|reserved_75|>",
671
+ "lstrip": false,
672
+ "normalized": false,
673
+ "rstrip": false,
674
+ "single_word": false,
675
+ "special": true
676
+ },
677
+ "86": {
678
+ "content": "<|reserved_76|>",
679
+ "lstrip": false,
680
+ "normalized": false,
681
+ "rstrip": false,
682
+ "single_word": false,
683
+ "special": true
684
+ },
685
+ "87": {
686
+ "content": "<|reserved_77|>",
687
+ "lstrip": false,
688
+ "normalized": false,
689
+ "rstrip": false,
690
+ "single_word": false,
691
+ "special": true
692
+ },
693
+ "88": {
694
+ "content": "<|reserved_78|>",
695
+ "lstrip": false,
696
+ "normalized": false,
697
+ "rstrip": false,
698
+ "single_word": false,
699
+ "special": true
700
+ },
701
+ "89": {
702
+ "content": "<|reserved_79|>",
703
+ "lstrip": false,
704
+ "normalized": false,
705
+ "rstrip": false,
706
+ "single_word": false,
707
+ "special": true
708
+ },
709
+ "90": {
710
+ "content": "<|reserved_80|>",
711
+ "lstrip": false,
712
+ "normalized": false,
713
+ "rstrip": false,
714
+ "single_word": false,
715
+ "special": true
716
+ },
717
+ "91": {
718
+ "content": "<|reserved_81|>",
719
+ "lstrip": false,
720
+ "normalized": false,
721
+ "rstrip": false,
722
+ "single_word": false,
723
+ "special": true
724
+ },
725
+ "92": {
726
+ "content": "<|reserved_82|>",
727
+ "lstrip": false,
728
+ "normalized": false,
729
+ "rstrip": false,
730
+ "single_word": false,
731
+ "special": true
732
+ },
733
+ "93": {
734
+ "content": "<|reserved_83|>",
735
+ "lstrip": false,
736
+ "normalized": false,
737
+ "rstrip": false,
738
+ "single_word": false,
739
+ "special": true
740
+ },
741
+ "94": {
742
+ "content": "<|reserved_84|>",
743
+ "lstrip": false,
744
+ "normalized": false,
745
+ "rstrip": false,
746
+ "single_word": false,
747
+ "special": true
748
+ },
749
+ "95": {
750
+ "content": "<|reserved_85|>",
751
+ "lstrip": false,
752
+ "normalized": false,
753
+ "rstrip": false,
754
+ "single_word": false,
755
+ "special": true
756
+ },
757
+ "96": {
758
+ "content": "<|reserved_86|>",
759
+ "lstrip": false,
760
+ "normalized": false,
761
+ "rstrip": false,
762
+ "single_word": false,
763
+ "special": true
764
+ },
765
+ "97": {
766
+ "content": "<|reserved_87|>",
767
+ "lstrip": false,
768
+ "normalized": false,
769
+ "rstrip": false,
770
+ "single_word": false,
771
+ "special": true
772
+ },
773
+ "98": {
774
+ "content": "<|reserved_88|>",
775
+ "lstrip": false,
776
+ "normalized": false,
777
+ "rstrip": false,
778
+ "single_word": false,
779
+ "special": true
780
+ },
781
+ "99": {
782
+ "content": "<|reserved_89|>",
783
+ "lstrip": false,
784
+ "normalized": false,
785
+ "rstrip": false,
786
+ "single_word": false,
787
+ "special": true
788
+ },
789
+ "100": {
790
+ "content": "<|reserved_90|>",
791
+ "lstrip": false,
792
+ "normalized": false,
793
+ "rstrip": false,
794
+ "single_word": false,
795
+ "special": true
796
+ },
797
+ "101": {
798
+ "content": "<|reserved_91|>",
799
+ "lstrip": false,
800
+ "normalized": false,
801
+ "rstrip": false,
802
+ "single_word": false,
803
+ "special": true
804
+ },
805
+ "102": {
806
+ "content": "<|reserved_92|>",
807
+ "lstrip": false,
808
+ "normalized": false,
809
+ "rstrip": false,
810
+ "single_word": false,
811
+ "special": true
812
+ },
813
+ "103": {
814
+ "content": "<|reserved_93|>",
815
+ "lstrip": false,
816
+ "normalized": false,
817
+ "rstrip": false,
818
+ "single_word": false,
819
+ "special": true
820
+ },
821
+ "104": {
822
+ "content": "<|reserved_94|>",
823
+ "lstrip": false,
824
+ "normalized": false,
825
+ "rstrip": false,
826
+ "single_word": false,
827
+ "special": true
828
+ },
829
+ "105": {
830
+ "content": "<|reserved_95|>",
831
+ "lstrip": false,
832
+ "normalized": false,
833
+ "rstrip": false,
834
+ "single_word": false,
835
+ "special": true
836
+ },
837
+ "106": {
838
+ "content": "<|reserved_96|>",
839
+ "lstrip": false,
840
+ "normalized": false,
841
+ "rstrip": false,
842
+ "single_word": false,
843
+ "special": true
844
+ },
845
+ "107": {
846
+ "content": "<|reserved_97|>",
847
+ "lstrip": false,
848
+ "normalized": false,
849
+ "rstrip": false,
850
+ "single_word": false,
851
+ "special": true
852
+ },
853
+ "108": {
854
+ "content": "<|reserved_98|>",
855
+ "lstrip": false,
856
+ "normalized": false,
857
+ "rstrip": false,
858
+ "single_word": false,
859
+ "special": true
860
+ },
861
+ "109": {
862
+ "content": "<|reserved_99|>",
863
+ "lstrip": false,
864
+ "normalized": false,
865
+ "rstrip": false,
866
+ "single_word": false,
867
+ "special": true
868
+ },
869
+ "110": {
870
+ "content": "<|reserved_100|>",
871
+ "lstrip": false,
872
+ "normalized": false,
873
+ "rstrip": false,
874
+ "single_word": false,
875
+ "special": true
876
+ },
877
+ "111": {
878
+ "content": "<|reserved_101|>",
879
+ "lstrip": false,
880
+ "normalized": false,
881
+ "rstrip": false,
882
+ "single_word": false,
883
+ "special": true
884
+ },
885
+ "112": {
886
+ "content": "<|reserved_102|>",
887
+ "lstrip": false,
888
+ "normalized": false,
889
+ "rstrip": false,
890
+ "single_word": false,
891
+ "special": true
892
+ },
893
+ "113": {
894
+ "content": "<|reserved_103|>",
895
+ "lstrip": false,
896
+ "normalized": false,
897
+ "rstrip": false,
898
+ "single_word": false,
899
+ "special": true
900
+ },
901
+ "114": {
902
+ "content": "<|reserved_104|>",
903
+ "lstrip": false,
904
+ "normalized": false,
905
+ "rstrip": false,
906
+ "single_word": false,
907
+ "special": true
908
+ },
909
+ "115": {
910
+ "content": "<|reserved_105|>",
911
+ "lstrip": false,
912
+ "normalized": false,
913
+ "rstrip": false,
914
+ "single_word": false,
915
+ "special": true
916
+ },
917
+ "116": {
918
+ "content": "<|reserved_106|>",
919
+ "lstrip": false,
920
+ "normalized": false,
921
+ "rstrip": false,
922
+ "single_word": false,
923
+ "special": true
924
+ },
925
+ "117": {
926
+ "content": "<|reserved_107|>",
927
+ "lstrip": false,
928
+ "normalized": false,
929
+ "rstrip": false,
930
+ "single_word": false,
931
+ "special": true
932
+ },
933
+ "118": {
934
+ "content": "<|reserved_108|>",
935
+ "lstrip": false,
936
+ "normalized": false,
937
+ "rstrip": false,
938
+ "single_word": false,
939
+ "special": true
940
+ },
941
+ "119": {
942
+ "content": "<|reserved_109|>",
943
+ "lstrip": false,
944
+ "normalized": false,
945
+ "rstrip": false,
946
+ "single_word": false,
947
+ "special": true
948
+ },
949
+ "120": {
950
+ "content": "<|reserved_110|>",
951
+ "lstrip": false,
952
+ "normalized": false,
953
+ "rstrip": false,
954
+ "single_word": false,
955
+ "special": true
956
+ },
957
+ "121": {
958
+ "content": "<|reserved_111|>",
959
+ "lstrip": false,
960
+ "normalized": false,
961
+ "rstrip": false,
962
+ "single_word": false,
963
+ "special": true
964
+ },
965
+ "122": {
966
+ "content": "<|reserved_112|>",
967
+ "lstrip": false,
968
+ "normalized": false,
969
+ "rstrip": false,
970
+ "single_word": false,
971
+ "special": true
972
+ },
973
+ "123": {
974
+ "content": "<|reserved_113|>",
975
+ "lstrip": false,
976
+ "normalized": false,
977
+ "rstrip": false,
978
+ "single_word": false,
979
+ "special": true
980
+ },
981
+ "124": {
982
+ "content": "<|reserved_114|>",
983
+ "lstrip": false,
984
+ "normalized": false,
985
+ "rstrip": false,
986
+ "single_word": false,
987
+ "special": true
988
+ },
989
+ "125": {
990
+ "content": "<|reserved_115|>",
991
+ "lstrip": false,
992
+ "normalized": false,
993
+ "rstrip": false,
994
+ "single_word": false,
995
+ "special": true
996
+ },
997
+ "126": {
998
+ "content": "<|reserved_116|>",
999
+ "lstrip": false,
1000
+ "normalized": false,
1001
+ "rstrip": false,
1002
+ "single_word": false,
1003
+ "special": true
1004
+ },
1005
+ "127": {
1006
+ "content": "<|reserved_117|>",
1007
+ "lstrip": false,
1008
+ "normalized": false,
1009
+ "rstrip": false,
1010
+ "single_word": false,
1011
+ "special": true
1012
+ },
1013
+ "128": {
1014
+ "content": "<|reserved_118|>",
1015
+ "lstrip": false,
1016
+ "normalized": false,
1017
+ "rstrip": false,
1018
+ "single_word": false,
1019
+ "special": true
1020
+ },
1021
+ "129": {
1022
+ "content": "<|reserved_119|>",
1023
+ "lstrip": false,
1024
+ "normalized": false,
1025
+ "rstrip": false,
1026
+ "single_word": false,
1027
+ "special": true
1028
+ },
1029
+ "130": {
1030
+ "content": "<|reserved_120|>",
1031
+ "lstrip": false,
1032
+ "normalized": false,
1033
+ "rstrip": false,
1034
+ "single_word": false,
1035
+ "special": true
1036
+ },
1037
+ "131": {
1038
+ "content": "<|reserved_121|>",
1039
+ "lstrip": false,
1040
+ "normalized": false,
1041
+ "rstrip": false,
1042
+ "single_word": false,
1043
+ "special": true
1044
+ },
1045
+ "132": {
1046
+ "content": "<|reserved_122|>",
1047
+ "lstrip": false,
1048
+ "normalized": false,
1049
+ "rstrip": false,
1050
+ "single_word": false,
1051
+ "special": true
1052
+ },
1053
+ "133": {
1054
+ "content": "<|reserved_123|>",
1055
+ "lstrip": false,
1056
+ "normalized": false,
1057
+ "rstrip": false,
1058
+ "single_word": false,
1059
+ "special": true
1060
+ },
1061
+ "134": {
1062
+ "content": "<|reserved_124|>",
1063
+ "lstrip": false,
1064
+ "normalized": false,
1065
+ "rstrip": false,
1066
+ "single_word": false,
1067
+ "special": true
1068
+ },
1069
+ "135": {
1070
+ "content": "<|reserved_125|>",
1071
+ "lstrip": false,
1072
+ "normalized": false,
1073
+ "rstrip": false,
1074
+ "single_word": false,
1075
+ "special": true
1076
+ },
1077
+ "136": {
1078
+ "content": "<|reserved_126|>",
1079
+ "lstrip": false,
1080
+ "normalized": false,
1081
+ "rstrip": false,
1082
+ "single_word": false,
1083
+ "special": true
1084
+ },
1085
+ "137": {
1086
+ "content": "<|reserved_127|>",
1087
+ "lstrip": false,
1088
+ "normalized": false,
1089
+ "rstrip": false,
1090
+ "single_word": false,
1091
+ "special": true
1092
+ },
1093
+ "138": {
1094
+ "content": "<|reserved_128|>",
1095
+ "lstrip": false,
1096
+ "normalized": false,
1097
+ "rstrip": false,
1098
+ "single_word": false,
1099
+ "special": true
1100
+ },
1101
+ "139": {
1102
+ "content": "<|reserved_129|>",
1103
+ "lstrip": false,
1104
+ "normalized": false,
1105
+ "rstrip": false,
1106
+ "single_word": false,
1107
+ "special": true
1108
+ },
1109
+ "140": {
1110
+ "content": "<|reserved_130|>",
1111
+ "lstrip": false,
1112
+ "normalized": false,
1113
+ "rstrip": false,
1114
+ "single_word": false,
1115
+ "special": true
1116
+ },
1117
+ "141": {
1118
+ "content": "<|reserved_131|>",
1119
+ "lstrip": false,
1120
+ "normalized": false,
1121
+ "rstrip": false,
1122
+ "single_word": false,
1123
+ "special": true
1124
+ },
1125
+ "142": {
1126
+ "content": "<|reserved_132|>",
1127
+ "lstrip": false,
1128
+ "normalized": false,
1129
+ "rstrip": false,
1130
+ "single_word": false,
1131
+ "special": true
1132
+ },
1133
+ "143": {
1134
+ "content": "<|reserved_133|>",
1135
+ "lstrip": false,
1136
+ "normalized": false,
1137
+ "rstrip": false,
1138
+ "single_word": false,
1139
+ "special": true
1140
+ },
1141
+ "144": {
1142
+ "content": "<|reserved_134|>",
1143
+ "lstrip": false,
1144
+ "normalized": false,
1145
+ "rstrip": false,
1146
+ "single_word": false,
1147
+ "special": true
1148
+ },
1149
+ "145": {
1150
+ "content": "<|reserved_135|>",
1151
+ "lstrip": false,
1152
+ "normalized": false,
1153
+ "rstrip": false,
1154
+ "single_word": false,
1155
+ "special": true
1156
+ },
1157
+ "146": {
1158
+ "content": "<|reserved_136|>",
1159
+ "lstrip": false,
1160
+ "normalized": false,
1161
+ "rstrip": false,
1162
+ "single_word": false,
1163
+ "special": true
1164
+ },
1165
+ "147": {
1166
+ "content": "<|reserved_137|>",
1167
+ "lstrip": false,
1168
+ "normalized": false,
1169
+ "rstrip": false,
1170
+ "single_word": false,
1171
+ "special": true
1172
+ },
1173
+ "148": {
1174
+ "content": "<|reserved_138|>",
1175
+ "lstrip": false,
1176
+ "normalized": false,
1177
+ "rstrip": false,
1178
+ "single_word": false,
1179
+ "special": true
1180
+ },
1181
+ "149": {
1182
+ "content": "<|reserved_139|>",
1183
+ "lstrip": false,
1184
+ "normalized": false,
1185
+ "rstrip": false,
1186
+ "single_word": false,
1187
+ "special": true
1188
+ },
1189
+ "150": {
1190
+ "content": "<|reserved_140|>",
1191
+ "lstrip": false,
1192
+ "normalized": false,
1193
+ "rstrip": false,
1194
+ "single_word": false,
1195
+ "special": true
1196
+ },
1197
+ "151": {
1198
+ "content": "<|reserved_141|>",
1199
+ "lstrip": false,
1200
+ "normalized": false,
1201
+ "rstrip": false,
1202
+ "single_word": false,
1203
+ "special": true
1204
+ },
1205
+ "152": {
1206
+ "content": "<|reserved_142|>",
1207
+ "lstrip": false,
1208
+ "normalized": false,
1209
+ "rstrip": false,
1210
+ "single_word": false,
1211
+ "special": true
1212
+ },
1213
+ "153": {
1214
+ "content": "<|reserved_143|>",
1215
+ "lstrip": false,
1216
+ "normalized": false,
1217
+ "rstrip": false,
1218
+ "single_word": false,
1219
+ "special": true
1220
+ },
1221
+ "154": {
1222
+ "content": "<|reserved_144|>",
1223
+ "lstrip": false,
1224
+ "normalized": false,
1225
+ "rstrip": false,
1226
+ "single_word": false,
1227
+ "special": true
1228
+ },
1229
+ "155": {
1230
+ "content": "<|reserved_145|>",
1231
+ "lstrip": false,
1232
+ "normalized": false,
1233
+ "rstrip": false,
1234
+ "single_word": false,
1235
+ "special": true
1236
+ },
1237
+ "156": {
1238
+ "content": "<|reserved_146|>",
1239
+ "lstrip": false,
1240
+ "normalized": false,
1241
+ "rstrip": false,
1242
+ "single_word": false,
1243
+ "special": true
1244
+ },
1245
+ "157": {
1246
+ "content": "<|reserved_147|>",
1247
+ "lstrip": false,
1248
+ "normalized": false,
1249
+ "rstrip": false,
1250
+ "single_word": false,
1251
+ "special": true
1252
+ },
1253
+ "158": {
1254
+ "content": "<|reserved_148|>",
1255
+ "lstrip": false,
1256
+ "normalized": false,
1257
+ "rstrip": false,
1258
+ "single_word": false,
1259
+ "special": true
1260
+ },
1261
+ "159": {
1262
+ "content": "<|reserved_149|>",
1263
+ "lstrip": false,
1264
+ "normalized": false,
1265
+ "rstrip": false,
1266
+ "single_word": false,
1267
+ "special": true
1268
+ },
1269
+ "160": {
1270
+ "content": "<|reserved_150|>",
1271
+ "lstrip": false,
1272
+ "normalized": false,
1273
+ "rstrip": false,
1274
+ "single_word": false,
1275
+ "special": true
1276
+ },
1277
+ "161": {
1278
+ "content": "<|reserved_151|>",
1279
+ "lstrip": false,
1280
+ "normalized": false,
1281
+ "rstrip": false,
1282
+ "single_word": false,
1283
+ "special": true
1284
+ },
1285
+ "162": {
1286
+ "content": "<|reserved_152|>",
1287
+ "lstrip": false,
1288
+ "normalized": false,
1289
+ "rstrip": false,
1290
+ "single_word": false,
1291
+ "special": true
1292
+ },
1293
+ "163": {
1294
+ "content": "<|reserved_153|>",
1295
+ "lstrip": false,
1296
+ "normalized": false,
1297
+ "rstrip": false,
1298
+ "single_word": false,
1299
+ "special": true
1300
+ },
1301
+ "164": {
1302
+ "content": "<|reserved_154|>",
1303
+ "lstrip": false,
1304
+ "normalized": false,
1305
+ "rstrip": false,
1306
+ "single_word": false,
1307
+ "special": true
1308
+ },
1309
+ "165": {
1310
+ "content": "<|reserved_155|>",
1311
+ "lstrip": false,
1312
+ "normalized": false,
1313
+ "rstrip": false,
1314
+ "single_word": false,
1315
+ "special": true
1316
+ },
1317
+ "166": {
1318
+ "content": "<|reserved_156|>",
1319
+ "lstrip": false,
1320
+ "normalized": false,
1321
+ "rstrip": false,
1322
+ "single_word": false,
1323
+ "special": true
1324
+ },
1325
+ "167": {
1326
+ "content": "<|reserved_157|>",
1327
+ "lstrip": false,
1328
+ "normalized": false,
1329
+ "rstrip": false,
1330
+ "single_word": false,
1331
+ "special": true
1332
+ },
1333
+ "168": {
1334
+ "content": "<|reserved_158|>",
1335
+ "lstrip": false,
1336
+ "normalized": false,
1337
+ "rstrip": false,
1338
+ "single_word": false,
1339
+ "special": true
1340
+ },
1341
+ "169": {
1342
+ "content": "<|reserved_159|>",
1343
+ "lstrip": false,
1344
+ "normalized": false,
1345
+ "rstrip": false,
1346
+ "single_word": false,
1347
+ "special": true
1348
+ },
1349
+ "170": {
1350
+ "content": "<|reserved_160|>",
1351
+ "lstrip": false,
1352
+ "normalized": false,
1353
+ "rstrip": false,
1354
+ "single_word": false,
1355
+ "special": true
1356
+ },
1357
+ "171": {
1358
+ "content": "<|reserved_161|>",
1359
+ "lstrip": false,
1360
+ "normalized": false,
1361
+ "rstrip": false,
1362
+ "single_word": false,
1363
+ "special": true
1364
+ },
1365
+ "172": {
1366
+ "content": "<|reserved_162|>",
1367
+ "lstrip": false,
1368
+ "normalized": false,
1369
+ "rstrip": false,
1370
+ "single_word": false,
1371
+ "special": true
1372
+ },
1373
+ "173": {
1374
+ "content": "<|reserved_163|>",
1375
+ "lstrip": false,
1376
+ "normalized": false,
1377
+ "rstrip": false,
1378
+ "single_word": false,
1379
+ "special": true
1380
+ },
1381
+ "174": {
1382
+ "content": "<|reserved_164|>",
1383
+ "lstrip": false,
1384
+ "normalized": false,
1385
+ "rstrip": false,
1386
+ "single_word": false,
1387
+ "special": true
1388
+ },
1389
+ "175": {
1390
+ "content": "<|reserved_165|>",
1391
+ "lstrip": false,
1392
+ "normalized": false,
1393
+ "rstrip": false,
1394
+ "single_word": false,
1395
+ "special": true
1396
+ },
1397
+ "176": {
1398
+ "content": "<|reserved_166|>",
1399
+ "lstrip": false,
1400
+ "normalized": false,
1401
+ "rstrip": false,
1402
+ "single_word": false,
1403
+ "special": true
1404
+ },
1405
+ "177": {
1406
+ "content": "<|reserved_167|>",
1407
+ "lstrip": false,
1408
+ "normalized": false,
1409
+ "rstrip": false,
1410
+ "single_word": false,
1411
+ "special": true
1412
+ },
1413
+ "178": {
1414
+ "content": "<|reserved_168|>",
1415
+ "lstrip": false,
1416
+ "normalized": false,
1417
+ "rstrip": false,
1418
+ "single_word": false,
1419
+ "special": true
1420
+ },
1421
+ "179": {
1422
+ "content": "<|reserved_169|>",
1423
+ "lstrip": false,
1424
+ "normalized": false,
1425
+ "rstrip": false,
1426
+ "single_word": false,
1427
+ "special": true
1428
+ },
1429
+ "180": {
1430
+ "content": "<|reserved_170|>",
1431
+ "lstrip": false,
1432
+ "normalized": false,
1433
+ "rstrip": false,
1434
+ "single_word": false,
1435
+ "special": true
1436
+ },
1437
+ "181": {
1438
+ "content": "<|reserved_171|>",
1439
+ "lstrip": false,
1440
+ "normalized": false,
1441
+ "rstrip": false,
1442
+ "single_word": false,
1443
+ "special": true
1444
+ },
1445
+ "182": {
1446
+ "content": "<|reserved_172|>",
1447
+ "lstrip": false,
1448
+ "normalized": false,
1449
+ "rstrip": false,
1450
+ "single_word": false,
1451
+ "special": true
1452
+ },
1453
+ "183": {
1454
+ "content": "<|reserved_173|>",
1455
+ "lstrip": false,
1456
+ "normalized": false,
1457
+ "rstrip": false,
1458
+ "single_word": false,
1459
+ "special": true
1460
+ },
1461
+ "184": {
1462
+ "content": "<|reserved_174|>",
1463
+ "lstrip": false,
1464
+ "normalized": false,
1465
+ "rstrip": false,
1466
+ "single_word": false,
1467
+ "special": true
1468
+ },
1469
+ "185": {
1470
+ "content": "<|reserved_175|>",
1471
+ "lstrip": false,
1472
+ "normalized": false,
1473
+ "rstrip": false,
1474
+ "single_word": false,
1475
+ "special": true
1476
+ },
1477
+ "186": {
1478
+ "content": "<|reserved_176|>",
1479
+ "lstrip": false,
1480
+ "normalized": false,
1481
+ "rstrip": false,
1482
+ "single_word": false,
1483
+ "special": true
1484
+ },
1485
+ "187": {
1486
+ "content": "<|reserved_177|>",
1487
+ "lstrip": false,
1488
+ "normalized": false,
1489
+ "rstrip": false,
1490
+ "single_word": false,
1491
+ "special": true
1492
+ },
1493
+ "188": {
1494
+ "content": "<|reserved_178|>",
1495
+ "lstrip": false,
1496
+ "normalized": false,
1497
+ "rstrip": false,
1498
+ "single_word": false,
1499
+ "special": true
1500
+ },
1501
+ "189": {
1502
+ "content": "<|reserved_179|>",
1503
+ "lstrip": false,
1504
+ "normalized": false,
1505
+ "rstrip": false,
1506
+ "single_word": false,
1507
+ "special": true
1508
+ },
1509
+ "190": {
1510
+ "content": "<|reserved_180|>",
1511
+ "lstrip": false,
1512
+ "normalized": false,
1513
+ "rstrip": false,
1514
+ "single_word": false,
1515
+ "special": true
1516
+ },
1517
+ "191": {
1518
+ "content": "<|reserved_181|>",
1519
+ "lstrip": false,
1520
+ "normalized": false,
1521
+ "rstrip": false,
1522
+ "single_word": false,
1523
+ "special": true
1524
+ },
1525
+ "192": {
1526
+ "content": "<|reserved_182|>",
1527
+ "lstrip": false,
1528
+ "normalized": false,
1529
+ "rstrip": false,
1530
+ "single_word": false,
1531
+ "special": true
1532
+ },
1533
+ "193": {
1534
+ "content": "<|reserved_183|>",
1535
+ "lstrip": false,
1536
+ "normalized": false,
1537
+ "rstrip": false,
1538
+ "single_word": false,
1539
+ "special": true
1540
+ },
1541
+ "194": {
1542
+ "content": "<|reserved_184|>",
1543
+ "lstrip": false,
1544
+ "normalized": false,
1545
+ "rstrip": false,
1546
+ "single_word": false,
1547
+ "special": true
1548
+ },
1549
+ "195": {
1550
+ "content": "<|reserved_185|>",
1551
+ "lstrip": false,
1552
+ "normalized": false,
1553
+ "rstrip": false,
1554
+ "single_word": false,
1555
+ "special": true
1556
+ },
1557
+ "196": {
1558
+ "content": "<|reserved_186|>",
1559
+ "lstrip": false,
1560
+ "normalized": false,
1561
+ "rstrip": false,
1562
+ "single_word": false,
1563
+ "special": true
1564
+ },
1565
+ "197": {
1566
+ "content": "<|reserved_187|>",
1567
+ "lstrip": false,
1568
+ "normalized": false,
1569
+ "rstrip": false,
1570
+ "single_word": false,
1571
+ "special": true
1572
+ },
1573
+ "198": {
1574
+ "content": "<|reserved_188|>",
1575
+ "lstrip": false,
1576
+ "normalized": false,
1577
+ "rstrip": false,
1578
+ "single_word": false,
1579
+ "special": true
1580
+ },
1581
+ "199": {
1582
+ "content": "<|reserved_189|>",
1583
+ "lstrip": false,
1584
+ "normalized": false,
1585
+ "rstrip": false,
1586
+ "single_word": false,
1587
+ "special": true
1588
+ },
1589
+ "200": {
1590
+ "content": "<|reserved_190|>",
1591
+ "lstrip": false,
1592
+ "normalized": false,
1593
+ "rstrip": false,
1594
+ "single_word": false,
1595
+ "special": true
1596
+ },
1597
+ "201": {
1598
+ "content": "<|reserved_191|>",
1599
+ "lstrip": false,
1600
+ "normalized": false,
1601
+ "rstrip": false,
1602
+ "single_word": false,
1603
+ "special": true
1604
+ },
1605
+ "202": {
1606
+ "content": "<|reserved_192|>",
1607
+ "lstrip": false,
1608
+ "normalized": false,
1609
+ "rstrip": false,
1610
+ "single_word": false,
1611
+ "special": true
1612
+ },
1613
+ "203": {
1614
+ "content": "<|reserved_193|>",
1615
+ "lstrip": false,
1616
+ "normalized": false,
1617
+ "rstrip": false,
1618
+ "single_word": false,
1619
+ "special": true
1620
+ },
1621
+ "204": {
1622
+ "content": "<|reserved_194|>",
1623
+ "lstrip": false,
1624
+ "normalized": false,
1625
+ "rstrip": false,
1626
+ "single_word": false,
1627
+ "special": true
1628
+ },
1629
+ "205": {
1630
+ "content": "<|reserved_195|>",
1631
+ "lstrip": false,
1632
+ "normalized": false,
1633
+ "rstrip": false,
1634
+ "single_word": false,
1635
+ "special": true
1636
+ },
1637
+ "206": {
1638
+ "content": "<|reserved_196|>",
1639
+ "lstrip": false,
1640
+ "normalized": false,
1641
+ "rstrip": false,
1642
+ "single_word": false,
1643
+ "special": true
1644
+ },
1645
+ "207": {
1646
+ "content": "<|reserved_197|>",
1647
+ "lstrip": false,
1648
+ "normalized": false,
1649
+ "rstrip": false,
1650
+ "single_word": false,
1651
+ "special": true
1652
+ },
1653
+ "208": {
1654
+ "content": "<|reserved_198|>",
1655
+ "lstrip": false,
1656
+ "normalized": false,
1657
+ "rstrip": false,
1658
+ "single_word": false,
1659
+ "special": true
1660
+ },
1661
+ "209": {
1662
+ "content": "<|reserved_199|>",
1663
+ "lstrip": false,
1664
+ "normalized": false,
1665
+ "rstrip": false,
1666
+ "single_word": false,
1667
+ "special": true
1668
+ },
1669
+ "210": {
1670
+ "content": "<|reserved_200|>",
1671
+ "lstrip": false,
1672
+ "normalized": false,
1673
+ "rstrip": false,
1674
+ "single_word": false,
1675
+ "special": true
1676
+ },
1677
+ "211": {
1678
+ "content": "<|reserved_201|>",
1679
+ "lstrip": false,
1680
+ "normalized": false,
1681
+ "rstrip": false,
1682
+ "single_word": false,
1683
+ "special": true
1684
+ },
1685
+ "212": {
1686
+ "content": "<|reserved_202|>",
1687
+ "lstrip": false,
1688
+ "normalized": false,
1689
+ "rstrip": false,
1690
+ "single_word": false,
1691
+ "special": true
1692
+ },
1693
+ "213": {
1694
+ "content": "<|reserved_203|>",
1695
+ "lstrip": false,
1696
+ "normalized": false,
1697
+ "rstrip": false,
1698
+ "single_word": false,
1699
+ "special": true
1700
+ },
1701
+ "214": {
1702
+ "content": "<|reserved_204|>",
1703
+ "lstrip": false,
1704
+ "normalized": false,
1705
+ "rstrip": false,
1706
+ "single_word": false,
1707
+ "special": true
1708
+ },
1709
+ "215": {
1710
+ "content": "<|reserved_205|>",
1711
+ "lstrip": false,
1712
+ "normalized": false,
1713
+ "rstrip": false,
1714
+ "single_word": false,
1715
+ "special": true
1716
+ },
1717
+ "216": {
1718
+ "content": "<|reserved_206|>",
1719
+ "lstrip": false,
1720
+ "normalized": false,
1721
+ "rstrip": false,
1722
+ "single_word": false,
1723
+ "special": true
1724
+ },
1725
+ "217": {
1726
+ "content": "<|reserved_207|>",
1727
+ "lstrip": false,
1728
+ "normalized": false,
1729
+ "rstrip": false,
1730
+ "single_word": false,
1731
+ "special": true
1732
+ },
1733
+ "218": {
1734
+ "content": "<|reserved_208|>",
1735
+ "lstrip": false,
1736
+ "normalized": false,
1737
+ "rstrip": false,
1738
+ "single_word": false,
1739
+ "special": true
1740
+ },
1741
+ "219": {
1742
+ "content": "<|reserved_209|>",
1743
+ "lstrip": false,
1744
+ "normalized": false,
1745
+ "rstrip": false,
1746
+ "single_word": false,
1747
+ "special": true
1748
+ },
1749
+ "220": {
1750
+ "content": "<|reserved_210|>",
1751
+ "lstrip": false,
1752
+ "normalized": false,
1753
+ "rstrip": false,
1754
+ "single_word": false,
1755
+ "special": true
1756
+ },
1757
+ "221": {
1758
+ "content": "<|reserved_211|>",
1759
+ "lstrip": false,
1760
+ "normalized": false,
1761
+ "rstrip": false,
1762
+ "single_word": false,
1763
+ "special": true
1764
+ },
1765
+ "222": {
1766
+ "content": "<|reserved_212|>",
1767
+ "lstrip": false,
1768
+ "normalized": false,
1769
+ "rstrip": false,
1770
+ "single_word": false,
1771
+ "special": true
1772
+ },
1773
+ "223": {
1774
+ "content": "<|reserved_213|>",
1775
+ "lstrip": false,
1776
+ "normalized": false,
1777
+ "rstrip": false,
1778
+ "single_word": false,
1779
+ "special": true
1780
+ },
1781
+ "224": {
1782
+ "content": "<|reserved_214|>",
1783
+ "lstrip": false,
1784
+ "normalized": false,
1785
+ "rstrip": false,
1786
+ "single_word": false,
1787
+ "special": true
1788
+ },
1789
+ "225": {
1790
+ "content": "<|reserved_215|>",
1791
+ "lstrip": false,
1792
+ "normalized": false,
1793
+ "rstrip": false,
1794
+ "single_word": false,
1795
+ "special": true
1796
+ },
1797
+ "226": {
1798
+ "content": "<|reserved_216|>",
1799
+ "lstrip": false,
1800
+ "normalized": false,
1801
+ "rstrip": false,
1802
+ "single_word": false,
1803
+ "special": true
1804
+ },
1805
+ "227": {
1806
+ "content": "<|reserved_217|>",
1807
+ "lstrip": false,
1808
+ "normalized": false,
1809
+ "rstrip": false,
1810
+ "single_word": false,
1811
+ "special": true
1812
+ },
1813
+ "228": {
1814
+ "content": "<|reserved_218|>",
1815
+ "lstrip": false,
1816
+ "normalized": false,
1817
+ "rstrip": false,
1818
+ "single_word": false,
1819
+ "special": true
1820
+ },
1821
+ "229": {
1822
+ "content": "<|reserved_219|>",
1823
+ "lstrip": false,
1824
+ "normalized": false,
1825
+ "rstrip": false,
1826
+ "single_word": false,
1827
+ "special": true
1828
+ },
1829
+ "230": {
1830
+ "content": "<|reserved_220|>",
1831
+ "lstrip": false,
1832
+ "normalized": false,
1833
+ "rstrip": false,
1834
+ "single_word": false,
1835
+ "special": true
1836
+ },
1837
+ "231": {
1838
+ "content": "<|reserved_221|>",
1839
+ "lstrip": false,
1840
+ "normalized": false,
1841
+ "rstrip": false,
1842
+ "single_word": false,
1843
+ "special": true
1844
+ },
1845
+ "232": {
1846
+ "content": "<|reserved_222|>",
1847
+ "lstrip": false,
1848
+ "normalized": false,
1849
+ "rstrip": false,
1850
+ "single_word": false,
1851
+ "special": true
1852
+ },
1853
+ "233": {
1854
+ "content": "<|reserved_223|>",
1855
+ "lstrip": false,
1856
+ "normalized": false,
1857
+ "rstrip": false,
1858
+ "single_word": false,
1859
+ "special": true
1860
+ },
1861
+ "234": {
1862
+ "content": "<|reserved_224|>",
1863
+ "lstrip": false,
1864
+ "normalized": false,
1865
+ "rstrip": false,
1866
+ "single_word": false,
1867
+ "special": true
1868
+ },
1869
+ "235": {
1870
+ "content": "<|reserved_225|>",
1871
+ "lstrip": false,
1872
+ "normalized": false,
1873
+ "rstrip": false,
1874
+ "single_word": false,
1875
+ "special": true
1876
+ },
1877
+ "236": {
1878
+ "content": "<|reserved_226|>",
1879
+ "lstrip": false,
1880
+ "normalized": false,
1881
+ "rstrip": false,
1882
+ "single_word": false,
1883
+ "special": true
1884
+ },
1885
+ "237": {
1886
+ "content": "<|reserved_227|>",
1887
+ "lstrip": false,
1888
+ "normalized": false,
1889
+ "rstrip": false,
1890
+ "single_word": false,
1891
+ "special": true
1892
+ },
1893
+ "238": {
1894
+ "content": "<|reserved_228|>",
1895
+ "lstrip": false,
1896
+ "normalized": false,
1897
+ "rstrip": false,
1898
+ "single_word": false,
1899
+ "special": true
1900
+ },
1901
+ "239": {
1902
+ "content": "<|reserved_229|>",
1903
+ "lstrip": false,
1904
+ "normalized": false,
1905
+ "rstrip": false,
1906
+ "single_word": false,
1907
+ "special": true
1908
+ },
1909
+ "240": {
1910
+ "content": "<|reserved_230|>",
1911
+ "lstrip": false,
1912
+ "normalized": false,
1913
+ "rstrip": false,
1914
+ "single_word": false,
1915
+ "special": true
1916
+ },
1917
+ "241": {
1918
+ "content": "<|reserved_231|>",
1919
+ "lstrip": false,
1920
+ "normalized": false,
1921
+ "rstrip": false,
1922
+ "single_word": false,
1923
+ "special": true
1924
+ },
1925
+ "242": {
1926
+ "content": "<|reserved_232|>",
1927
+ "lstrip": false,
1928
+ "normalized": false,
1929
+ "rstrip": false,
1930
+ "single_word": false,
1931
+ "special": true
1932
+ },
1933
+ "243": {
1934
+ "content": "<|reserved_233|>",
1935
+ "lstrip": false,
1936
+ "normalized": false,
1937
+ "rstrip": false,
1938
+ "single_word": false,
1939
+ "special": true
1940
+ },
1941
+ "244": {
1942
+ "content": "<|reserved_234|>",
1943
+ "lstrip": false,
1944
+ "normalized": false,
1945
+ "rstrip": false,
1946
+ "single_word": false,
1947
+ "special": true
1948
+ },
1949
+ "245": {
1950
+ "content": "<|reserved_235|>",
1951
+ "lstrip": false,
1952
+ "normalized": false,
1953
+ "rstrip": false,
1954
+ "single_word": false,
1955
+ "special": true
1956
+ },
1957
+ "246": {
1958
+ "content": "<|reserved_236|>",
1959
+ "lstrip": false,
1960
+ "normalized": false,
1961
+ "rstrip": false,
1962
+ "single_word": false,
1963
+ "special": true
1964
+ },
1965
+ "247": {
1966
+ "content": "<|reserved_237|>",
1967
+ "lstrip": false,
1968
+ "normalized": false,
1969
+ "rstrip": false,
1970
+ "single_word": false,
1971
+ "special": true
1972
+ },
1973
+ "248": {
1974
+ "content": "<|reserved_238|>",
1975
+ "lstrip": false,
1976
+ "normalized": false,
1977
+ "rstrip": false,
1978
+ "single_word": false,
1979
+ "special": true
1980
+ },
1981
+ "249": {
1982
+ "content": "<|reserved_239|>",
1983
+ "lstrip": false,
1984
+ "normalized": false,
1985
+ "rstrip": false,
1986
+ "single_word": false,
1987
+ "special": true
1988
+ },
1989
+ "250": {
1990
+ "content": "<|reserved_240|>",
1991
+ "lstrip": false,
1992
+ "normalized": false,
1993
+ "rstrip": false,
1994
+ "single_word": false,
1995
+ "special": true
1996
+ },
1997
+ "251": {
1998
+ "content": "<|reserved_241|>",
1999
+ "lstrip": false,
2000
+ "normalized": false,
2001
+ "rstrip": false,
2002
+ "single_word": false,
2003
+ "special": true
2004
+ },
2005
+ "252": {
2006
+ "content": "<|reserved_242|>",
2007
+ "lstrip": false,
2008
+ "normalized": false,
2009
+ "rstrip": false,
2010
+ "single_word": false,
2011
+ "special": true
2012
+ },
2013
+ "253": {
2014
+ "content": "<|reserved_243|>",
2015
+ "lstrip": false,
2016
+ "normalized": false,
2017
+ "rstrip": false,
2018
+ "single_word": false,
2019
+ "special": true
2020
+ },
2021
+ "254": {
2022
+ "content": "<|reserved_244|>",
2023
+ "lstrip": false,
2024
+ "normalized": false,
2025
+ "rstrip": false,
2026
+ "single_word": false,
2027
+ "special": true
2028
+ },
2029
+ "255": {
2030
+ "content": "<|reserved_245|>",
2031
+ "lstrip": false,
2032
+ "normalized": false,
2033
+ "rstrip": false,
2034
+ "single_word": false,
2035
+ "special": true
2036
+ },
2037
+ "256": {
2038
+ "content": "<|reserved_246|>",
2039
+ "lstrip": false,
2040
+ "normalized": false,
2041
+ "rstrip": false,
2042
+ "single_word": false,
2043
+ "special": true
2044
+ },
2045
+ "257": {
2046
+ "content": "<|reserved_247|>",
2047
+ "lstrip": false,
2048
+ "normalized": false,
2049
+ "rstrip": false,
2050
+ "single_word": false,
2051
+ "special": true
2052
+ },
2053
+ "258": {
2054
+ "content": "<|reserved_248|>",
2055
+ "lstrip": false,
2056
+ "normalized": false,
2057
+ "rstrip": false,
2058
+ "single_word": false,
2059
+ "special": true
2060
+ },
2061
+ "259": {
2062
+ "content": "<|reserved_249|>",
2063
+ "lstrip": false,
2064
+ "normalized": false,
2065
+ "rstrip": false,
2066
+ "single_word": false,
2067
+ "special": true
2068
+ },
2069
+ "260": {
2070
+ "content": "<|reserved_250|>",
2071
+ "lstrip": false,
2072
+ "normalized": false,
2073
+ "rstrip": false,
2074
+ "single_word": false,
2075
+ "special": true
2076
+ },
2077
+ "261": {
2078
+ "content": "<|reserved_251|>",
2079
+ "lstrip": false,
2080
+ "normalized": false,
2081
+ "rstrip": false,
2082
+ "single_word": false,
2083
+ "special": true
2084
+ },
2085
+ "262": {
2086
+ "content": "<|reserved_252|>",
2087
+ "lstrip": false,
2088
+ "normalized": false,
2089
+ "rstrip": false,
2090
+ "single_word": false,
2091
+ "special": true
2092
+ },
2093
+ "263": {
2094
+ "content": "<|reserved_253|>",
2095
+ "lstrip": false,
2096
+ "normalized": false,
2097
+ "rstrip": false,
2098
+ "single_word": false,
2099
+ "special": true
2100
+ },
2101
+ "264": {
2102
+ "content": "<|reserved_254|>",
2103
+ "lstrip": false,
2104
+ "normalized": false,
2105
+ "rstrip": false,
2106
+ "single_word": false,
2107
+ "special": true
2108
+ },
2109
+ "265": {
2110
+ "content": "<|reserved_255|>",
2111
+ "lstrip": false,
2112
+ "normalized": false,
2113
+ "rstrip": false,
2114
+ "single_word": false,
2115
+ "special": true
2116
+ },
2117
+ "266": {
2118
+ "content": "<|reserved_256|>",
2119
+ "lstrip": false,
2120
+ "normalized": false,
2121
+ "rstrip": false,
2122
+ "single_word": false,
2123
+ "special": true
2124
+ },
2125
+ "267": {
2126
+ "content": "<|reserved_257|>",
2127
+ "lstrip": false,
2128
+ "normalized": false,
2129
+ "rstrip": false,
2130
+ "single_word": false,
2131
+ "special": true
2132
+ },
2133
+ "268": {
2134
+ "content": "<|reserved_258|>",
2135
+ "lstrip": false,
2136
+ "normalized": false,
2137
+ "rstrip": false,
2138
+ "single_word": false,
2139
+ "special": true
2140
+ },
2141
+ "269": {
2142
+ "content": "<|reserved_259|>",
2143
+ "lstrip": false,
2144
+ "normalized": false,
2145
+ "rstrip": false,
2146
+ "single_word": false,
2147
+ "special": true
2148
+ },
2149
+ "270": {
2150
+ "content": "<|reserved_260|>",
2151
+ "lstrip": false,
2152
+ "normalized": false,
2153
+ "rstrip": false,
2154
+ "single_word": false,
2155
+ "special": true
2156
+ },
2157
+ "271": {
2158
+ "content": "<|reserved_261|>",
2159
+ "lstrip": false,
2160
+ "normalized": false,
2161
+ "rstrip": false,
2162
+ "single_word": false,
2163
+ "special": true
2164
+ },
2165
+ "272": {
2166
+ "content": "<|reserved_262|>",
2167
+ "lstrip": false,
2168
+ "normalized": false,
2169
+ "rstrip": false,
2170
+ "single_word": false,
2171
+ "special": true
2172
+ },
2173
+ "273": {
2174
+ "content": "<|reserved_263|>",
2175
+ "lstrip": false,
2176
+ "normalized": false,
2177
+ "rstrip": false,
2178
+ "single_word": false,
2179
+ "special": true
2180
+ },
2181
+ "274": {
2182
+ "content": "<|reserved_264|>",
2183
+ "lstrip": false,
2184
+ "normalized": false,
2185
+ "rstrip": false,
2186
+ "single_word": false,
2187
+ "special": true
2188
+ },
2189
+ "275": {
2190
+ "content": "<|reserved_265|>",
2191
+ "lstrip": false,
2192
+ "normalized": false,
2193
+ "rstrip": false,
2194
+ "single_word": false,
2195
+ "special": true
2196
+ },
2197
+ "276": {
2198
+ "content": "<|reserved_266|>",
2199
+ "lstrip": false,
2200
+ "normalized": false,
2201
+ "rstrip": false,
2202
+ "single_word": false,
2203
+ "special": true
2204
+ },
2205
+ "277": {
2206
+ "content": "<|reserved_267|>",
2207
+ "lstrip": false,
2208
+ "normalized": false,
2209
+ "rstrip": false,
2210
+ "single_word": false,
2211
+ "special": true
2212
+ },
2213
+ "278": {
2214
+ "content": "<|reserved_268|>",
2215
+ "lstrip": false,
2216
+ "normalized": false,
2217
+ "rstrip": false,
2218
+ "single_word": false,
2219
+ "special": true
2220
+ },
2221
+ "279": {
2222
+ "content": "<|reserved_269|>",
2223
+ "lstrip": false,
2224
+ "normalized": false,
2225
+ "rstrip": false,
2226
+ "single_word": false,
2227
+ "special": true
2228
+ },
2229
+ "280": {
2230
+ "content": "<|reserved_270|>",
2231
+ "lstrip": false,
2232
+ "normalized": false,
2233
+ "rstrip": false,
2234
+ "single_word": false,
2235
+ "special": true
2236
+ },
2237
+ "281": {
2238
+ "content": "<|reserved_271|>",
2239
+ "lstrip": false,
2240
+ "normalized": false,
2241
+ "rstrip": false,
2242
+ "single_word": false,
2243
+ "special": true
2244
+ },
2245
+ "282": {
2246
+ "content": "<|reserved_272|>",
2247
+ "lstrip": false,
2248
+ "normalized": false,
2249
+ "rstrip": false,
2250
+ "single_word": false,
2251
+ "special": true
2252
+ },
2253
+ "283": {
2254
+ "content": "<|reserved_273|>",
2255
+ "lstrip": false,
2256
+ "normalized": false,
2257
+ "rstrip": false,
2258
+ "single_word": false,
2259
+ "special": true
2260
+ },
2261
+ "284": {
2262
+ "content": "<|reserved_274|>",
2263
+ "lstrip": false,
2264
+ "normalized": false,
2265
+ "rstrip": false,
2266
+ "single_word": false,
2267
+ "special": true
2268
+ },
2269
+ "285": {
2270
+ "content": "<|reserved_275|>",
2271
+ "lstrip": false,
2272
+ "normalized": false,
2273
+ "rstrip": false,
2274
+ "single_word": false,
2275
+ "special": true
2276
+ },
2277
+ "286": {
2278
+ "content": "<|reserved_276|>",
2279
+ "lstrip": false,
2280
+ "normalized": false,
2281
+ "rstrip": false,
2282
+ "single_word": false,
2283
+ "special": true
2284
+ },
2285
+ "287": {
2286
+ "content": "<|reserved_277|>",
2287
+ "lstrip": false,
2288
+ "normalized": false,
2289
+ "rstrip": false,
2290
+ "single_word": false,
2291
+ "special": true
2292
+ },
2293
+ "288": {
2294
+ "content": "<|reserved_278|>",
2295
+ "lstrip": false,
2296
+ "normalized": false,
2297
+ "rstrip": false,
2298
+ "single_word": false,
2299
+ "special": true
2300
+ },
2301
+ "289": {
2302
+ "content": "<|reserved_279|>",
2303
+ "lstrip": false,
2304
+ "normalized": false,
2305
+ "rstrip": false,
2306
+ "single_word": false,
2307
+ "special": true
2308
+ },
2309
+ "290": {
2310
+ "content": "<|reserved_280|>",
2311
+ "lstrip": false,
2312
+ "normalized": false,
2313
+ "rstrip": false,
2314
+ "single_word": false,
2315
+ "special": true
2316
+ },
2317
+ "291": {
2318
+ "content": "<|reserved_281|>",
2319
+ "lstrip": false,
2320
+ "normalized": false,
2321
+ "rstrip": false,
2322
+ "single_word": false,
2323
+ "special": true
2324
+ },
2325
+ "292": {
2326
+ "content": "<|reserved_282|>",
2327
+ "lstrip": false,
2328
+ "normalized": false,
2329
+ "rstrip": false,
2330
+ "single_word": false,
2331
+ "special": true
2332
+ },
2333
+ "293": {
2334
+ "content": "<|reserved_283|>",
2335
+ "lstrip": false,
2336
+ "normalized": false,
2337
+ "rstrip": false,
2338
+ "single_word": false,
2339
+ "special": true
2340
+ },
2341
+ "294": {
2342
+ "content": "<|reserved_284|>",
2343
+ "lstrip": false,
2344
+ "normalized": false,
2345
+ "rstrip": false,
2346
+ "single_word": false,
2347
+ "special": true
2348
+ },
2349
+ "295": {
2350
+ "content": "<|reserved_285|>",
2351
+ "lstrip": false,
2352
+ "normalized": false,
2353
+ "rstrip": false,
2354
+ "single_word": false,
2355
+ "special": true
2356
+ },
2357
+ "296": {
2358
+ "content": "<|reserved_286|>",
2359
+ "lstrip": false,
2360
+ "normalized": false,
2361
+ "rstrip": false,
2362
+ "single_word": false,
2363
+ "special": true
2364
+ },
2365
+ "297": {
2366
+ "content": "<|reserved_287|>",
2367
+ "lstrip": false,
2368
+ "normalized": false,
2369
+ "rstrip": false,
2370
+ "single_word": false,
2371
+ "special": true
2372
+ },
2373
+ "298": {
2374
+ "content": "<|reserved_288|>",
2375
+ "lstrip": false,
2376
+ "normalized": false,
2377
+ "rstrip": false,
2378
+ "single_word": false,
2379
+ "special": true
2380
+ },
2381
+ "299": {
2382
+ "content": "<|reserved_289|>",
2383
+ "lstrip": false,
2384
+ "normalized": false,
2385
+ "rstrip": false,
2386
+ "single_word": false,
2387
+ "special": true
2388
+ },
2389
+ "300": {
2390
+ "content": "<|reserved_290|>",
2391
+ "lstrip": false,
2392
+ "normalized": false,
2393
+ "rstrip": false,
2394
+ "single_word": false,
2395
+ "special": true
2396
+ },
2397
+ "301": {
2398
+ "content": "<|reserved_291|>",
2399
+ "lstrip": false,
2400
+ "normalized": false,
2401
+ "rstrip": false,
2402
+ "single_word": false,
2403
+ "special": true
2404
+ },
2405
+ "302": {
2406
+ "content": "<|reserved_292|>",
2407
+ "lstrip": false,
2408
+ "normalized": false,
2409
+ "rstrip": false,
2410
+ "single_word": false,
2411
+ "special": true
2412
+ },
2413
+ "303": {
2414
+ "content": "<|reserved_293|>",
2415
+ "lstrip": false,
2416
+ "normalized": false,
2417
+ "rstrip": false,
2418
+ "single_word": false,
2419
+ "special": true
2420
+ },
2421
+ "304": {
2422
+ "content": "<|reserved_294|>",
2423
+ "lstrip": false,
2424
+ "normalized": false,
2425
+ "rstrip": false,
2426
+ "single_word": false,
2427
+ "special": true
2428
+ },
2429
+ "305": {
2430
+ "content": "<|reserved_295|>",
2431
+ "lstrip": false,
2432
+ "normalized": false,
2433
+ "rstrip": false,
2434
+ "single_word": false,
2435
+ "special": true
2436
+ },
2437
+ "306": {
2438
+ "content": "<|reserved_296|>",
2439
+ "lstrip": false,
2440
+ "normalized": false,
2441
+ "rstrip": false,
2442
+ "single_word": false,
2443
+ "special": true
2444
+ },
2445
+ "307": {
2446
+ "content": "<|reserved_297|>",
2447
+ "lstrip": false,
2448
+ "normalized": false,
2449
+ "rstrip": false,
2450
+ "single_word": false,
2451
+ "special": true
2452
+ },
2453
+ "308": {
2454
+ "content": "<|reserved_298|>",
2455
+ "lstrip": false,
2456
+ "normalized": false,
2457
+ "rstrip": false,
2458
+ "single_word": false,
2459
+ "special": true
2460
+ },
2461
+ "309": {
2462
+ "content": "<|reserved_299|>",
2463
+ "lstrip": false,
2464
+ "normalized": false,
2465
+ "rstrip": false,
2466
+ "single_word": false,
2467
+ "special": true
2468
+ },
2469
+ "310": {
2470
+ "content": "<|reserved_300|>",
2471
+ "lstrip": false,
2472
+ "normalized": false,
2473
+ "rstrip": false,
2474
+ "single_word": false,
2475
+ "special": true
2476
+ },
2477
+ "311": {
2478
+ "content": "<|reserved_301|>",
2479
+ "lstrip": false,
2480
+ "normalized": false,
2481
+ "rstrip": false,
2482
+ "single_word": false,
2483
+ "special": true
2484
+ },
2485
+ "312": {
2486
+ "content": "<|reserved_302|>",
2487
+ "lstrip": false,
2488
+ "normalized": false,
2489
+ "rstrip": false,
2490
+ "single_word": false,
2491
+ "special": true
2492
+ },
2493
+ "313": {
2494
+ "content": "<|reserved_303|>",
2495
+ "lstrip": false,
2496
+ "normalized": false,
2497
+ "rstrip": false,
2498
+ "single_word": false,
2499
+ "special": true
2500
+ },
2501
+ "314": {
2502
+ "content": "<|reserved_304|>",
2503
+ "lstrip": false,
2504
+ "normalized": false,
2505
+ "rstrip": false,
2506
+ "single_word": false,
2507
+ "special": true
2508
+ },
2509
+ "315": {
2510
+ "content": "<|reserved_305|>",
2511
+ "lstrip": false,
2512
+ "normalized": false,
2513
+ "rstrip": false,
2514
+ "single_word": false,
2515
+ "special": true
2516
+ },
2517
+ "316": {
2518
+ "content": "<|reserved_306|>",
2519
+ "lstrip": false,
2520
+ "normalized": false,
2521
+ "rstrip": false,
2522
+ "single_word": false,
2523
+ "special": true
2524
+ },
2525
+ "317": {
2526
+ "content": "<|reserved_307|>",
2527
+ "lstrip": false,
2528
+ "normalized": false,
2529
+ "rstrip": false,
2530
+ "single_word": false,
2531
+ "special": true
2532
+ },
2533
+ "318": {
2534
+ "content": "<|reserved_308|>",
2535
+ "lstrip": false,
2536
+ "normalized": false,
2537
+ "rstrip": false,
2538
+ "single_word": false,
2539
+ "special": true
2540
+ },
2541
+ "319": {
2542
+ "content": "<|reserved_309|>",
2543
+ "lstrip": false,
2544
+ "normalized": false,
2545
+ "rstrip": false,
2546
+ "single_word": false,
2547
+ "special": true
2548
+ },
2549
+ "320": {
2550
+ "content": "<|reserved_310|>",
2551
+ "lstrip": false,
2552
+ "normalized": false,
2553
+ "rstrip": false,
2554
+ "single_word": false,
2555
+ "special": true
2556
+ },
2557
+ "321": {
2558
+ "content": "<|reserved_311|>",
2559
+ "lstrip": false,
2560
+ "normalized": false,
2561
+ "rstrip": false,
2562
+ "single_word": false,
2563
+ "special": true
2564
+ },
2565
+ "322": {
2566
+ "content": "<|reserved_312|>",
2567
+ "lstrip": false,
2568
+ "normalized": false,
2569
+ "rstrip": false,
2570
+ "single_word": false,
2571
+ "special": true
2572
+ },
2573
+ "323": {
2574
+ "content": "<|reserved_313|>",
2575
+ "lstrip": false,
2576
+ "normalized": false,
2577
+ "rstrip": false,
2578
+ "single_word": false,
2579
+ "special": true
2580
+ },
2581
+ "324": {
2582
+ "content": "<|reserved_314|>",
2583
+ "lstrip": false,
2584
+ "normalized": false,
2585
+ "rstrip": false,
2586
+ "single_word": false,
2587
+ "special": true
2588
+ },
2589
+ "325": {
2590
+ "content": "<|reserved_315|>",
2591
+ "lstrip": false,
2592
+ "normalized": false,
2593
+ "rstrip": false,
2594
+ "single_word": false,
2595
+ "special": true
2596
+ },
2597
+ "326": {
2598
+ "content": "<|reserved_316|>",
2599
+ "lstrip": false,
2600
+ "normalized": false,
2601
+ "rstrip": false,
2602
+ "single_word": false,
2603
+ "special": true
2604
+ },
2605
+ "327": {
2606
+ "content": "<|reserved_317|>",
2607
+ "lstrip": false,
2608
+ "normalized": false,
2609
+ "rstrip": false,
2610
+ "single_word": false,
2611
+ "special": true
2612
+ },
2613
+ "328": {
2614
+ "content": "<|reserved_318|>",
2615
+ "lstrip": false,
2616
+ "normalized": false,
2617
+ "rstrip": false,
2618
+ "single_word": false,
2619
+ "special": true
2620
+ },
2621
+ "329": {
2622
+ "content": "<|reserved_319|>",
2623
+ "lstrip": false,
2624
+ "normalized": false,
2625
+ "rstrip": false,
2626
+ "single_word": false,
2627
+ "special": true
2628
+ },
2629
+ "330": {
2630
+ "content": "<|reserved_320|>",
2631
+ "lstrip": false,
2632
+ "normalized": false,
2633
+ "rstrip": false,
2634
+ "single_word": false,
2635
+ "special": true
2636
+ },
2637
+ "331": {
2638
+ "content": "<|reserved_321|>",
2639
+ "lstrip": false,
2640
+ "normalized": false,
2641
+ "rstrip": false,
2642
+ "single_word": false,
2643
+ "special": true
2644
+ },
2645
+ "332": {
2646
+ "content": "<|reserved_322|>",
2647
+ "lstrip": false,
2648
+ "normalized": false,
2649
+ "rstrip": false,
2650
+ "single_word": false,
2651
+ "special": true
2652
+ },
2653
+ "333": {
2654
+ "content": "<|reserved_323|>",
2655
+ "lstrip": false,
2656
+ "normalized": false,
2657
+ "rstrip": false,
2658
+ "single_word": false,
2659
+ "special": true
2660
+ },
2661
+ "334": {
2662
+ "content": "<|reserved_324|>",
2663
+ "lstrip": false,
2664
+ "normalized": false,
2665
+ "rstrip": false,
2666
+ "single_word": false,
2667
+ "special": true
2668
+ },
2669
+ "335": {
2670
+ "content": "<|reserved_325|>",
2671
+ "lstrip": false,
2672
+ "normalized": false,
2673
+ "rstrip": false,
2674
+ "single_word": false,
2675
+ "special": true
2676
+ },
2677
+ "336": {
2678
+ "content": "<|reserved_326|>",
2679
+ "lstrip": false,
2680
+ "normalized": false,
2681
+ "rstrip": false,
2682
+ "single_word": false,
2683
+ "special": true
2684
+ },
2685
+ "337": {
2686
+ "content": "<|reserved_327|>",
2687
+ "lstrip": false,
2688
+ "normalized": false,
2689
+ "rstrip": false,
2690
+ "single_word": false,
2691
+ "special": true
2692
+ },
2693
+ "338": {
2694
+ "content": "<|reserved_328|>",
2695
+ "lstrip": false,
2696
+ "normalized": false,
2697
+ "rstrip": false,
2698
+ "single_word": false,
2699
+ "special": true
2700
+ },
2701
+ "339": {
2702
+ "content": "<|reserved_329|>",
2703
+ "lstrip": false,
2704
+ "normalized": false,
2705
+ "rstrip": false,
2706
+ "single_word": false,
2707
+ "special": true
2708
+ },
2709
+ "340": {
2710
+ "content": "<|reserved_330|>",
2711
+ "lstrip": false,
2712
+ "normalized": false,
2713
+ "rstrip": false,
2714
+ "single_word": false,
2715
+ "special": true
2716
+ },
2717
+ "341": {
2718
+ "content": "<|reserved_331|>",
2719
+ "lstrip": false,
2720
+ "normalized": false,
2721
+ "rstrip": false,
2722
+ "single_word": false,
2723
+ "special": true
2724
+ },
2725
+ "342": {
2726
+ "content": "<|reserved_332|>",
2727
+ "lstrip": false,
2728
+ "normalized": false,
2729
+ "rstrip": false,
2730
+ "single_word": false,
2731
+ "special": true
2732
+ },
2733
+ "343": {
2734
+ "content": "<|reserved_333|>",
2735
+ "lstrip": false,
2736
+ "normalized": false,
2737
+ "rstrip": false,
2738
+ "single_word": false,
2739
+ "special": true
2740
+ },
2741
+ "344": {
2742
+ "content": "<|reserved_334|>",
2743
+ "lstrip": false,
2744
+ "normalized": false,
2745
+ "rstrip": false,
2746
+ "single_word": false,
2747
+ "special": true
2748
+ },
2749
+ "345": {
2750
+ "content": "<|reserved_335|>",
2751
+ "lstrip": false,
2752
+ "normalized": false,
2753
+ "rstrip": false,
2754
+ "single_word": false,
2755
+ "special": true
2756
+ },
2757
+ "346": {
2758
+ "content": "<|reserved_336|>",
2759
+ "lstrip": false,
2760
+ "normalized": false,
2761
+ "rstrip": false,
2762
+ "single_word": false,
2763
+ "special": true
2764
+ },
2765
+ "347": {
2766
+ "content": "<|reserved_337|>",
2767
+ "lstrip": false,
2768
+ "normalized": false,
2769
+ "rstrip": false,
2770
+ "single_word": false,
2771
+ "special": true
2772
+ },
2773
+ "348": {
2774
+ "content": "<|reserved_338|>",
2775
+ "lstrip": false,
2776
+ "normalized": false,
2777
+ "rstrip": false,
2778
+ "single_word": false,
2779
+ "special": true
2780
+ },
2781
+ "349": {
2782
+ "content": "<|reserved_339|>",
2783
+ "lstrip": false,
2784
+ "normalized": false,
2785
+ "rstrip": false,
2786
+ "single_word": false,
2787
+ "special": true
2788
+ },
2789
+ "350": {
2790
+ "content": "<|reserved_340|>",
2791
+ "lstrip": false,
2792
+ "normalized": false,
2793
+ "rstrip": false,
2794
+ "single_word": false,
2795
+ "special": true
2796
+ },
2797
+ "351": {
2798
+ "content": "<|reserved_341|>",
2799
+ "lstrip": false,
2800
+ "normalized": false,
2801
+ "rstrip": false,
2802
+ "single_word": false,
2803
+ "special": true
2804
+ },
2805
+ "352": {
2806
+ "content": "<|reserved_342|>",
2807
+ "lstrip": false,
2808
+ "normalized": false,
2809
+ "rstrip": false,
2810
+ "single_word": false,
2811
+ "special": true
2812
+ },
2813
+ "353": {
2814
+ "content": "<|reserved_343|>",
2815
+ "lstrip": false,
2816
+ "normalized": false,
2817
+ "rstrip": false,
2818
+ "single_word": false,
2819
+ "special": true
2820
+ },
2821
+ "354": {
2822
+ "content": "<|reserved_344|>",
2823
+ "lstrip": false,
2824
+ "normalized": false,
2825
+ "rstrip": false,
2826
+ "single_word": false,
2827
+ "special": true
2828
+ },
2829
+ "355": {
2830
+ "content": "<|reserved_345|>",
2831
+ "lstrip": false,
2832
+ "normalized": false,
2833
+ "rstrip": false,
2834
+ "single_word": false,
2835
+ "special": true
2836
+ },
2837
+ "356": {
2838
+ "content": "<|reserved_346|>",
2839
+ "lstrip": false,
2840
+ "normalized": false,
2841
+ "rstrip": false,
2842
+ "single_word": false,
2843
+ "special": true
2844
+ },
2845
+ "357": {
2846
+ "content": "<|reserved_347|>",
2847
+ "lstrip": false,
2848
+ "normalized": false,
2849
+ "rstrip": false,
2850
+ "single_word": false,
2851
+ "special": true
2852
+ },
2853
+ "358": {
2854
+ "content": "<|reserved_348|>",
2855
+ "lstrip": false,
2856
+ "normalized": false,
2857
+ "rstrip": false,
2858
+ "single_word": false,
2859
+ "special": true
2860
+ },
2861
+ "359": {
2862
+ "content": "<|reserved_349|>",
2863
+ "lstrip": false,
2864
+ "normalized": false,
2865
+ "rstrip": false,
2866
+ "single_word": false,
2867
+ "special": true
2868
+ },
2869
+ "360": {
2870
+ "content": "<|reserved_350|>",
2871
+ "lstrip": false,
2872
+ "normalized": false,
2873
+ "rstrip": false,
2874
+ "single_word": false,
2875
+ "special": true
2876
+ },
2877
+ "361": {
2878
+ "content": "<|reserved_351|>",
2879
+ "lstrip": false,
2880
+ "normalized": false,
2881
+ "rstrip": false,
2882
+ "single_word": false,
2883
+ "special": true
2884
+ },
2885
+ "362": {
2886
+ "content": "<|reserved_352|>",
2887
+ "lstrip": false,
2888
+ "normalized": false,
2889
+ "rstrip": false,
2890
+ "single_word": false,
2891
+ "special": true
2892
+ },
2893
+ "363": {
2894
+ "content": "<|reserved_353|>",
2895
+ "lstrip": false,
2896
+ "normalized": false,
2897
+ "rstrip": false,
2898
+ "single_word": false,
2899
+ "special": true
2900
+ },
2901
+ "364": {
2902
+ "content": "<|reserved_354|>",
2903
+ "lstrip": false,
2904
+ "normalized": false,
2905
+ "rstrip": false,
2906
+ "single_word": false,
2907
+ "special": true
2908
+ },
2909
+ "365": {
2910
+ "content": "<|reserved_355|>",
2911
+ "lstrip": false,
2912
+ "normalized": false,
2913
+ "rstrip": false,
2914
+ "single_word": false,
2915
+ "special": true
2916
+ },
2917
+ "366": {
2918
+ "content": "<|reserved_356|>",
2919
+ "lstrip": false,
2920
+ "normalized": false,
2921
+ "rstrip": false,
2922
+ "single_word": false,
2923
+ "special": true
2924
+ },
2925
+ "367": {
2926
+ "content": "<|reserved_357|>",
2927
+ "lstrip": false,
2928
+ "normalized": false,
2929
+ "rstrip": false,
2930
+ "single_word": false,
2931
+ "special": true
2932
+ },
2933
+ "368": {
2934
+ "content": "<|reserved_358|>",
2935
+ "lstrip": false,
2936
+ "normalized": false,
2937
+ "rstrip": false,
2938
+ "single_word": false,
2939
+ "special": true
2940
+ },
2941
+ "369": {
2942
+ "content": "<|reserved_359|>",
2943
+ "lstrip": false,
2944
+ "normalized": false,
2945
+ "rstrip": false,
2946
+ "single_word": false,
2947
+ "special": true
2948
+ },
2949
+ "370": {
2950
+ "content": "<|reserved_360|>",
2951
+ "lstrip": false,
2952
+ "normalized": false,
2953
+ "rstrip": false,
2954
+ "single_word": false,
2955
+ "special": true
2956
+ },
2957
+ "371": {
2958
+ "content": "<|reserved_361|>",
2959
+ "lstrip": false,
2960
+ "normalized": false,
2961
+ "rstrip": false,
2962
+ "single_word": false,
2963
+ "special": true
2964
+ },
2965
+ "372": {
2966
+ "content": "<|reserved_362|>",
2967
+ "lstrip": false,
2968
+ "normalized": false,
2969
+ "rstrip": false,
2970
+ "single_word": false,
2971
+ "special": true
2972
+ },
2973
+ "373": {
2974
+ "content": "<|reserved_363|>",
2975
+ "lstrip": false,
2976
+ "normalized": false,
2977
+ "rstrip": false,
2978
+ "single_word": false,
2979
+ "special": true
2980
+ },
2981
+ "374": {
2982
+ "content": "<|reserved_364|>",
2983
+ "lstrip": false,
2984
+ "normalized": false,
2985
+ "rstrip": false,
2986
+ "single_word": false,
2987
+ "special": true
2988
+ },
2989
+ "375": {
2990
+ "content": "<|reserved_365|>",
2991
+ "lstrip": false,
2992
+ "normalized": false,
2993
+ "rstrip": false,
2994
+ "single_word": false,
2995
+ "special": true
2996
+ },
2997
+ "376": {
2998
+ "content": "<|reserved_366|>",
2999
+ "lstrip": false,
3000
+ "normalized": false,
3001
+ "rstrip": false,
3002
+ "single_word": false,
3003
+ "special": true
3004
+ },
3005
+ "377": {
3006
+ "content": "<|reserved_367|>",
3007
+ "lstrip": false,
3008
+ "normalized": false,
3009
+ "rstrip": false,
3010
+ "single_word": false,
3011
+ "special": true
3012
+ },
3013
+ "378": {
3014
+ "content": "<|reserved_368|>",
3015
+ "lstrip": false,
3016
+ "normalized": false,
3017
+ "rstrip": false,
3018
+ "single_word": false,
3019
+ "special": true
3020
+ },
3021
+ "379": {
3022
+ "content": "<|reserved_369|>",
3023
+ "lstrip": false,
3024
+ "normalized": false,
3025
+ "rstrip": false,
3026
+ "single_word": false,
3027
+ "special": true
3028
+ },
3029
+ "380": {
3030
+ "content": "<|reserved_370|>",
3031
+ "lstrip": false,
3032
+ "normalized": false,
3033
+ "rstrip": false,
3034
+ "single_word": false,
3035
+ "special": true
3036
+ },
3037
+ "381": {
3038
+ "content": "<|reserved_371|>",
3039
+ "lstrip": false,
3040
+ "normalized": false,
3041
+ "rstrip": false,
3042
+ "single_word": false,
3043
+ "special": true
3044
+ },
3045
+ "382": {
3046
+ "content": "<|reserved_372|>",
3047
+ "lstrip": false,
3048
+ "normalized": false,
3049
+ "rstrip": false,
3050
+ "single_word": false,
3051
+ "special": true
3052
+ },
3053
+ "383": {
3054
+ "content": "<|reserved_373|>",
3055
+ "lstrip": false,
3056
+ "normalized": false,
3057
+ "rstrip": false,
3058
+ "single_word": false,
3059
+ "special": true
3060
+ },
3061
+ "384": {
3062
+ "content": "<|reserved_374|>",
3063
+ "lstrip": false,
3064
+ "normalized": false,
3065
+ "rstrip": false,
3066
+ "single_word": false,
3067
+ "special": true
3068
+ },
3069
+ "385": {
3070
+ "content": "<|reserved_375|>",
3071
+ "lstrip": false,
3072
+ "normalized": false,
3073
+ "rstrip": false,
3074
+ "single_word": false,
3075
+ "special": true
3076
+ },
3077
+ "386": {
3078
+ "content": "<|reserved_376|>",
3079
+ "lstrip": false,
3080
+ "normalized": false,
3081
+ "rstrip": false,
3082
+ "single_word": false,
3083
+ "special": true
3084
+ },
3085
+ "387": {
3086
+ "content": "<|reserved_377|>",
3087
+ "lstrip": false,
3088
+ "normalized": false,
3089
+ "rstrip": false,
3090
+ "single_word": false,
3091
+ "special": true
3092
+ },
3093
+ "388": {
3094
+ "content": "<|reserved_378|>",
3095
+ "lstrip": false,
3096
+ "normalized": false,
3097
+ "rstrip": false,
3098
+ "single_word": false,
3099
+ "special": true
3100
+ },
3101
+ "389": {
3102
+ "content": "<|reserved_379|>",
3103
+ "lstrip": false,
3104
+ "normalized": false,
3105
+ "rstrip": false,
3106
+ "single_word": false,
3107
+ "special": true
3108
+ },
3109
+ "390": {
3110
+ "content": "<|reserved_380|>",
3111
+ "lstrip": false,
3112
+ "normalized": false,
3113
+ "rstrip": false,
3114
+ "single_word": false,
3115
+ "special": true
3116
+ },
3117
+ "391": {
3118
+ "content": "<|reserved_381|>",
3119
+ "lstrip": false,
3120
+ "normalized": false,
3121
+ "rstrip": false,
3122
+ "single_word": false,
3123
+ "special": true
3124
+ },
3125
+ "392": {
3126
+ "content": "<|reserved_382|>",
3127
+ "lstrip": false,
3128
+ "normalized": false,
3129
+ "rstrip": false,
3130
+ "single_word": false,
3131
+ "special": true
3132
+ },
3133
+ "393": {
3134
+ "content": "<|reserved_383|>",
3135
+ "lstrip": false,
3136
+ "normalized": false,
3137
+ "rstrip": false,
3138
+ "single_word": false,
3139
+ "special": true
3140
+ },
3141
+ "394": {
3142
+ "content": "<|reserved_384|>",
3143
+ "lstrip": false,
3144
+ "normalized": false,
3145
+ "rstrip": false,
3146
+ "single_word": false,
3147
+ "special": true
3148
+ },
3149
+ "395": {
3150
+ "content": "<|reserved_385|>",
3151
+ "lstrip": false,
3152
+ "normalized": false,
3153
+ "rstrip": false,
3154
+ "single_word": false,
3155
+ "special": true
3156
+ },
3157
+ "396": {
3158
+ "content": "<|reserved_386|>",
3159
+ "lstrip": false,
3160
+ "normalized": false,
3161
+ "rstrip": false,
3162
+ "single_word": false,
3163
+ "special": true
3164
+ },
3165
+ "397": {
3166
+ "content": "<|reserved_387|>",
3167
+ "lstrip": false,
3168
+ "normalized": false,
3169
+ "rstrip": false,
3170
+ "single_word": false,
3171
+ "special": true
3172
+ },
3173
+ "398": {
3174
+ "content": "<|reserved_388|>",
3175
+ "lstrip": false,
3176
+ "normalized": false,
3177
+ "rstrip": false,
3178
+ "single_word": false,
3179
+ "special": true
3180
+ },
3181
+ "399": {
3182
+ "content": "<|reserved_389|>",
3183
+ "lstrip": false,
3184
+ "normalized": false,
3185
+ "rstrip": false,
3186
+ "single_word": false,
3187
+ "special": true
3188
+ },
3189
+ "400": {
3190
+ "content": "<|reserved_390|>",
3191
+ "lstrip": false,
3192
+ "normalized": false,
3193
+ "rstrip": false,
3194
+ "single_word": false,
3195
+ "special": true
3196
+ },
3197
+ "401": {
3198
+ "content": "<|reserved_391|>",
3199
+ "lstrip": false,
3200
+ "normalized": false,
3201
+ "rstrip": false,
3202
+ "single_word": false,
3203
+ "special": true
3204
+ },
3205
+ "402": {
3206
+ "content": "<|reserved_392|>",
3207
+ "lstrip": false,
3208
+ "normalized": false,
3209
+ "rstrip": false,
3210
+ "single_word": false,
3211
+ "special": true
3212
+ },
3213
+ "403": {
3214
+ "content": "<|reserved_393|>",
3215
+ "lstrip": false,
3216
+ "normalized": false,
3217
+ "rstrip": false,
3218
+ "single_word": false,
3219
+ "special": true
3220
+ },
3221
+ "404": {
3222
+ "content": "<|reserved_394|>",
3223
+ "lstrip": false,
3224
+ "normalized": false,
3225
+ "rstrip": false,
3226
+ "single_word": false,
3227
+ "special": true
3228
+ },
3229
+ "405": {
3230
+ "content": "<|reserved_395|>",
3231
+ "lstrip": false,
3232
+ "normalized": false,
3233
+ "rstrip": false,
3234
+ "single_word": false,
3235
+ "special": true
3236
+ },
3237
+ "406": {
3238
+ "content": "<|reserved_396|>",
3239
+ "lstrip": false,
3240
+ "normalized": false,
3241
+ "rstrip": false,
3242
+ "single_word": false,
3243
+ "special": true
3244
+ },
3245
+ "407": {
3246
+ "content": "<|reserved_397|>",
3247
+ "lstrip": false,
3248
+ "normalized": false,
3249
+ "rstrip": false,
3250
+ "single_word": false,
3251
+ "special": true
3252
+ },
3253
+ "408": {
3254
+ "content": "<|reserved_398|>",
3255
+ "lstrip": false,
3256
+ "normalized": false,
3257
+ "rstrip": false,
3258
+ "single_word": false,
3259
+ "special": true
3260
+ },
3261
+ "409": {
3262
+ "content": "<|reserved_399|>",
3263
+ "lstrip": false,
3264
+ "normalized": false,
3265
+ "rstrip": false,
3266
+ "single_word": false,
3267
+ "special": true
3268
+ },
3269
+ "410": {
3270
+ "content": "<|reserved_400|>",
3271
+ "lstrip": false,
3272
+ "normalized": false,
3273
+ "rstrip": false,
3274
+ "single_word": false,
3275
+ "special": true
3276
+ },
3277
+ "411": {
3278
+ "content": "<|reserved_401|>",
3279
+ "lstrip": false,
3280
+ "normalized": false,
3281
+ "rstrip": false,
3282
+ "single_word": false,
3283
+ "special": true
3284
+ },
3285
+ "412": {
3286
+ "content": "<|reserved_402|>",
3287
+ "lstrip": false,
3288
+ "normalized": false,
3289
+ "rstrip": false,
3290
+ "single_word": false,
3291
+ "special": true
3292
+ },
3293
+ "413": {
3294
+ "content": "<|reserved_403|>",
3295
+ "lstrip": false,
3296
+ "normalized": false,
3297
+ "rstrip": false,
3298
+ "single_word": false,
3299
+ "special": true
3300
+ },
3301
+ "414": {
3302
+ "content": "<|reserved_404|>",
3303
+ "lstrip": false,
3304
+ "normalized": false,
3305
+ "rstrip": false,
3306
+ "single_word": false,
3307
+ "special": true
3308
+ },
3309
+ "415": {
3310
+ "content": "<|reserved_405|>",
3311
+ "lstrip": false,
3312
+ "normalized": false,
3313
+ "rstrip": false,
3314
+ "single_word": false,
3315
+ "special": true
3316
+ },
3317
+ "416": {
3318
+ "content": "<|reserved_406|>",
3319
+ "lstrip": false,
3320
+ "normalized": false,
3321
+ "rstrip": false,
3322
+ "single_word": false,
3323
+ "special": true
3324
+ },
3325
+ "417": {
3326
+ "content": "<|reserved_407|>",
3327
+ "lstrip": false,
3328
+ "normalized": false,
3329
+ "rstrip": false,
3330
+ "single_word": false,
3331
+ "special": true
3332
+ },
3333
+ "418": {
3334
+ "content": "<|reserved_408|>",
3335
+ "lstrip": false,
3336
+ "normalized": false,
3337
+ "rstrip": false,
3338
+ "single_word": false,
3339
+ "special": true
3340
+ },
3341
+ "419": {
3342
+ "content": "<|reserved_409|>",
3343
+ "lstrip": false,
3344
+ "normalized": false,
3345
+ "rstrip": false,
3346
+ "single_word": false,
3347
+ "special": true
3348
+ },
3349
+ "420": {
3350
+ "content": "<|reserved_410|>",
3351
+ "lstrip": false,
3352
+ "normalized": false,
3353
+ "rstrip": false,
3354
+ "single_word": false,
3355
+ "special": true
3356
+ },
3357
+ "421": {
3358
+ "content": "<|reserved_411|>",
3359
+ "lstrip": false,
3360
+ "normalized": false,
3361
+ "rstrip": false,
3362
+ "single_word": false,
3363
+ "special": true
3364
+ },
3365
+ "422": {
3366
+ "content": "<|reserved_412|>",
3367
+ "lstrip": false,
3368
+ "normalized": false,
3369
+ "rstrip": false,
3370
+ "single_word": false,
3371
+ "special": true
3372
+ },
3373
+ "423": {
3374
+ "content": "<|reserved_413|>",
3375
+ "lstrip": false,
3376
+ "normalized": false,
3377
+ "rstrip": false,
3378
+ "single_word": false,
3379
+ "special": true
3380
+ },
3381
+ "424": {
3382
+ "content": "<|reserved_414|>",
3383
+ "lstrip": false,
3384
+ "normalized": false,
3385
+ "rstrip": false,
3386
+ "single_word": false,
3387
+ "special": true
3388
+ },
3389
+ "425": {
3390
+ "content": "<|reserved_415|>",
3391
+ "lstrip": false,
3392
+ "normalized": false,
3393
+ "rstrip": false,
3394
+ "single_word": false,
3395
+ "special": true
3396
+ },
3397
+ "426": {
3398
+ "content": "<|reserved_416|>",
3399
+ "lstrip": false,
3400
+ "normalized": false,
3401
+ "rstrip": false,
3402
+ "single_word": false,
3403
+ "special": true
3404
+ },
3405
+ "427": {
3406
+ "content": "<|reserved_417|>",
3407
+ "lstrip": false,
3408
+ "normalized": false,
3409
+ "rstrip": false,
3410
+ "single_word": false,
3411
+ "special": true
3412
+ },
3413
+ "428": {
3414
+ "content": "<|reserved_418|>",
3415
+ "lstrip": false,
3416
+ "normalized": false,
3417
+ "rstrip": false,
3418
+ "single_word": false,
3419
+ "special": true
3420
+ },
3421
+ "429": {
3422
+ "content": "<|reserved_419|>",
3423
+ "lstrip": false,
3424
+ "normalized": false,
3425
+ "rstrip": false,
3426
+ "single_word": false,
3427
+ "special": true
3428
+ },
3429
+ "430": {
3430
+ "content": "<|reserved_420|>",
3431
+ "lstrip": false,
3432
+ "normalized": false,
3433
+ "rstrip": false,
3434
+ "single_word": false,
3435
+ "special": true
3436
+ },
3437
+ "431": {
3438
+ "content": "<|reserved_421|>",
3439
+ "lstrip": false,
3440
+ "normalized": false,
3441
+ "rstrip": false,
3442
+ "single_word": false,
3443
+ "special": true
3444
+ },
3445
+ "432": {
3446
+ "content": "<|reserved_422|>",
3447
+ "lstrip": false,
3448
+ "normalized": false,
3449
+ "rstrip": false,
3450
+ "single_word": false,
3451
+ "special": true
3452
+ },
3453
+ "433": {
3454
+ "content": "<|reserved_423|>",
3455
+ "lstrip": false,
3456
+ "normalized": false,
3457
+ "rstrip": false,
3458
+ "single_word": false,
3459
+ "special": true
3460
+ },
3461
+ "434": {
3462
+ "content": "<|reserved_424|>",
3463
+ "lstrip": false,
3464
+ "normalized": false,
3465
+ "rstrip": false,
3466
+ "single_word": false,
3467
+ "special": true
3468
+ },
3469
+ "435": {
3470
+ "content": "<|reserved_425|>",
3471
+ "lstrip": false,
3472
+ "normalized": false,
3473
+ "rstrip": false,
3474
+ "single_word": false,
3475
+ "special": true
3476
+ },
3477
+ "436": {
3478
+ "content": "<|reserved_426|>",
3479
+ "lstrip": false,
3480
+ "normalized": false,
3481
+ "rstrip": false,
3482
+ "single_word": false,
3483
+ "special": true
3484
+ },
3485
+ "437": {
3486
+ "content": "<|reserved_427|>",
3487
+ "lstrip": false,
3488
+ "normalized": false,
3489
+ "rstrip": false,
3490
+ "single_word": false,
3491
+ "special": true
3492
+ },
3493
+ "438": {
3494
+ "content": "<|reserved_428|>",
3495
+ "lstrip": false,
3496
+ "normalized": false,
3497
+ "rstrip": false,
3498
+ "single_word": false,
3499
+ "special": true
3500
+ },
3501
+ "439": {
3502
+ "content": "<|reserved_429|>",
3503
+ "lstrip": false,
3504
+ "normalized": false,
3505
+ "rstrip": false,
3506
+ "single_word": false,
3507
+ "special": true
3508
+ },
3509
+ "440": {
3510
+ "content": "<|reserved_430|>",
3511
+ "lstrip": false,
3512
+ "normalized": false,
3513
+ "rstrip": false,
3514
+ "single_word": false,
3515
+ "special": true
3516
+ },
3517
+ "441": {
3518
+ "content": "<|reserved_431|>",
3519
+ "lstrip": false,
3520
+ "normalized": false,
3521
+ "rstrip": false,
3522
+ "single_word": false,
3523
+ "special": true
3524
+ },
3525
+ "442": {
3526
+ "content": "<|reserved_432|>",
3527
+ "lstrip": false,
3528
+ "normalized": false,
3529
+ "rstrip": false,
3530
+ "single_word": false,
3531
+ "special": true
3532
+ },
3533
+ "443": {
3534
+ "content": "<|reserved_433|>",
3535
+ "lstrip": false,
3536
+ "normalized": false,
3537
+ "rstrip": false,
3538
+ "single_word": false,
3539
+ "special": true
3540
+ },
3541
+ "444": {
3542
+ "content": "<|reserved_434|>",
3543
+ "lstrip": false,
3544
+ "normalized": false,
3545
+ "rstrip": false,
3546
+ "single_word": false,
3547
+ "special": true
3548
+ },
3549
+ "445": {
3550
+ "content": "<|reserved_435|>",
3551
+ "lstrip": false,
3552
+ "normalized": false,
3553
+ "rstrip": false,
3554
+ "single_word": false,
3555
+ "special": true
3556
+ },
3557
+ "446": {
3558
+ "content": "<|reserved_436|>",
3559
+ "lstrip": false,
3560
+ "normalized": false,
3561
+ "rstrip": false,
3562
+ "single_word": false,
3563
+ "special": true
3564
+ },
3565
+ "447": {
3566
+ "content": "<|reserved_437|>",
3567
+ "lstrip": false,
3568
+ "normalized": false,
3569
+ "rstrip": false,
3570
+ "single_word": false,
3571
+ "special": true
3572
+ },
3573
+ "448": {
3574
+ "content": "<|reserved_438|>",
3575
+ "lstrip": false,
3576
+ "normalized": false,
3577
+ "rstrip": false,
3578
+ "single_word": false,
3579
+ "special": true
3580
+ },
3581
+ "449": {
3582
+ "content": "<|reserved_439|>",
3583
+ "lstrip": false,
3584
+ "normalized": false,
3585
+ "rstrip": false,
3586
+ "single_word": false,
3587
+ "special": true
3588
+ },
3589
+ "450": {
3590
+ "content": "<|reserved_440|>",
3591
+ "lstrip": false,
3592
+ "normalized": false,
3593
+ "rstrip": false,
3594
+ "single_word": false,
3595
+ "special": true
3596
+ },
3597
+ "451": {
3598
+ "content": "<|reserved_441|>",
3599
+ "lstrip": false,
3600
+ "normalized": false,
3601
+ "rstrip": false,
3602
+ "single_word": false,
3603
+ "special": true
3604
+ },
3605
+ "452": {
3606
+ "content": "<|reserved_442|>",
3607
+ "lstrip": false,
3608
+ "normalized": false,
3609
+ "rstrip": false,
3610
+ "single_word": false,
3611
+ "special": true
3612
+ },
3613
+ "453": {
3614
+ "content": "<|reserved_443|>",
3615
+ "lstrip": false,
3616
+ "normalized": false,
3617
+ "rstrip": false,
3618
+ "single_word": false,
3619
+ "special": true
3620
+ },
3621
+ "454": {
3622
+ "content": "<|reserved_444|>",
3623
+ "lstrip": false,
3624
+ "normalized": false,
3625
+ "rstrip": false,
3626
+ "single_word": false,
3627
+ "special": true
3628
+ },
3629
+ "455": {
3630
+ "content": "<|reserved_445|>",
3631
+ "lstrip": false,
3632
+ "normalized": false,
3633
+ "rstrip": false,
3634
+ "single_word": false,
3635
+ "special": true
3636
+ },
3637
+ "456": {
3638
+ "content": "<|reserved_446|>",
3639
+ "lstrip": false,
3640
+ "normalized": false,
3641
+ "rstrip": false,
3642
+ "single_word": false,
3643
+ "special": true
3644
+ },
3645
+ "457": {
3646
+ "content": "<|reserved_447|>",
3647
+ "lstrip": false,
3648
+ "normalized": false,
3649
+ "rstrip": false,
3650
+ "single_word": false,
3651
+ "special": true
3652
+ },
3653
+ "458": {
3654
+ "content": "<|reserved_448|>",
3655
+ "lstrip": false,
3656
+ "normalized": false,
3657
+ "rstrip": false,
3658
+ "single_word": false,
3659
+ "special": true
3660
+ },
3661
+ "459": {
3662
+ "content": "<|reserved_449|>",
3663
+ "lstrip": false,
3664
+ "normalized": false,
3665
+ "rstrip": false,
3666
+ "single_word": false,
3667
+ "special": true
3668
+ },
3669
+ "460": {
3670
+ "content": "<|reserved_450|>",
3671
+ "lstrip": false,
3672
+ "normalized": false,
3673
+ "rstrip": false,
3674
+ "single_word": false,
3675
+ "special": true
3676
+ },
3677
+ "461": {
3678
+ "content": "<|reserved_451|>",
3679
+ "lstrip": false,
3680
+ "normalized": false,
3681
+ "rstrip": false,
3682
+ "single_word": false,
3683
+ "special": true
3684
+ },
3685
+ "462": {
3686
+ "content": "<|reserved_452|>",
3687
+ "lstrip": false,
3688
+ "normalized": false,
3689
+ "rstrip": false,
3690
+ "single_word": false,
3691
+ "special": true
3692
+ },
3693
+ "463": {
3694
+ "content": "<|reserved_453|>",
3695
+ "lstrip": false,
3696
+ "normalized": false,
3697
+ "rstrip": false,
3698
+ "single_word": false,
3699
+ "special": true
3700
+ },
3701
+ "464": {
3702
+ "content": "<|reserved_454|>",
3703
+ "lstrip": false,
3704
+ "normalized": false,
3705
+ "rstrip": false,
3706
+ "single_word": false,
3707
+ "special": true
3708
+ },
3709
+ "465": {
3710
+ "content": "<|reserved_455|>",
3711
+ "lstrip": false,
3712
+ "normalized": false,
3713
+ "rstrip": false,
3714
+ "single_word": false,
3715
+ "special": true
3716
+ },
3717
+ "466": {
3718
+ "content": "<|reserved_456|>",
3719
+ "lstrip": false,
3720
+ "normalized": false,
3721
+ "rstrip": false,
3722
+ "single_word": false,
3723
+ "special": true
3724
+ },
3725
+ "467": {
3726
+ "content": "<|reserved_457|>",
3727
+ "lstrip": false,
3728
+ "normalized": false,
3729
+ "rstrip": false,
3730
+ "single_word": false,
3731
+ "special": true
3732
+ },
3733
+ "468": {
3734
+ "content": "<|reserved_458|>",
3735
+ "lstrip": false,
3736
+ "normalized": false,
3737
+ "rstrip": false,
3738
+ "single_word": false,
3739
+ "special": true
3740
+ },
3741
+ "469": {
3742
+ "content": "<|reserved_459|>",
3743
+ "lstrip": false,
3744
+ "normalized": false,
3745
+ "rstrip": false,
3746
+ "single_word": false,
3747
+ "special": true
3748
+ },
3749
+ "470": {
3750
+ "content": "<|reserved_460|>",
3751
+ "lstrip": false,
3752
+ "normalized": false,
3753
+ "rstrip": false,
3754
+ "single_word": false,
3755
+ "special": true
3756
+ },
3757
+ "471": {
3758
+ "content": "<|reserved_461|>",
3759
+ "lstrip": false,
3760
+ "normalized": false,
3761
+ "rstrip": false,
3762
+ "single_word": false,
3763
+ "special": true
3764
+ },
3765
+ "472": {
3766
+ "content": "<|reserved_462|>",
3767
+ "lstrip": false,
3768
+ "normalized": false,
3769
+ "rstrip": false,
3770
+ "single_word": false,
3771
+ "special": true
3772
+ },
3773
+ "473": {
3774
+ "content": "<|reserved_463|>",
3775
+ "lstrip": false,
3776
+ "normalized": false,
3777
+ "rstrip": false,
3778
+ "single_word": false,
3779
+ "special": true
3780
+ },
3781
+ "474": {
3782
+ "content": "<|reserved_464|>",
3783
+ "lstrip": false,
3784
+ "normalized": false,
3785
+ "rstrip": false,
3786
+ "single_word": false,
3787
+ "special": true
3788
+ },
3789
+ "475": {
3790
+ "content": "<|reserved_465|>",
3791
+ "lstrip": false,
3792
+ "normalized": false,
3793
+ "rstrip": false,
3794
+ "single_word": false,
3795
+ "special": true
3796
+ },
3797
+ "476": {
3798
+ "content": "<|reserved_466|>",
3799
+ "lstrip": false,
3800
+ "normalized": false,
3801
+ "rstrip": false,
3802
+ "single_word": false,
3803
+ "special": true
3804
+ },
3805
+ "477": {
3806
+ "content": "<|reserved_467|>",
3807
+ "lstrip": false,
3808
+ "normalized": false,
3809
+ "rstrip": false,
3810
+ "single_word": false,
3811
+ "special": true
3812
+ },
3813
+ "478": {
3814
+ "content": "<|reserved_468|>",
3815
+ "lstrip": false,
3816
+ "normalized": false,
3817
+ "rstrip": false,
3818
+ "single_word": false,
3819
+ "special": true
3820
+ },
3821
+ "479": {
3822
+ "content": "<|reserved_469|>",
3823
+ "lstrip": false,
3824
+ "normalized": false,
3825
+ "rstrip": false,
3826
+ "single_word": false,
3827
+ "special": true
3828
+ },
3829
+ "480": {
3830
+ "content": "<|reserved_470|>",
3831
+ "lstrip": false,
3832
+ "normalized": false,
3833
+ "rstrip": false,
3834
+ "single_word": false,
3835
+ "special": true
3836
+ },
3837
+ "481": {
3838
+ "content": "<|reserved_471|>",
3839
+ "lstrip": false,
3840
+ "normalized": false,
3841
+ "rstrip": false,
3842
+ "single_word": false,
3843
+ "special": true
3844
+ },
3845
+ "482": {
3846
+ "content": "<|reserved_472|>",
3847
+ "lstrip": false,
3848
+ "normalized": false,
3849
+ "rstrip": false,
3850
+ "single_word": false,
3851
+ "special": true
3852
+ },
3853
+ "483": {
3854
+ "content": "<|reserved_473|>",
3855
+ "lstrip": false,
3856
+ "normalized": false,
3857
+ "rstrip": false,
3858
+ "single_word": false,
3859
+ "special": true
3860
+ },
3861
+ "484": {
3862
+ "content": "<|reserved_474|>",
3863
+ "lstrip": false,
3864
+ "normalized": false,
3865
+ "rstrip": false,
3866
+ "single_word": false,
3867
+ "special": true
3868
+ },
3869
+ "485": {
3870
+ "content": "<|reserved_475|>",
3871
+ "lstrip": false,
3872
+ "normalized": false,
3873
+ "rstrip": false,
3874
+ "single_word": false,
3875
+ "special": true
3876
+ },
3877
+ "486": {
3878
+ "content": "<|reserved_476|>",
3879
+ "lstrip": false,
3880
+ "normalized": false,
3881
+ "rstrip": false,
3882
+ "single_word": false,
3883
+ "special": true
3884
+ },
3885
+ "487": {
3886
+ "content": "<|reserved_477|>",
3887
+ "lstrip": false,
3888
+ "normalized": false,
3889
+ "rstrip": false,
3890
+ "single_word": false,
3891
+ "special": true
3892
+ },
3893
+ "488": {
3894
+ "content": "<|reserved_478|>",
3895
+ "lstrip": false,
3896
+ "normalized": false,
3897
+ "rstrip": false,
3898
+ "single_word": false,
3899
+ "special": true
3900
+ },
3901
+ "489": {
3902
+ "content": "<|reserved_479|>",
3903
+ "lstrip": false,
3904
+ "normalized": false,
3905
+ "rstrip": false,
3906
+ "single_word": false,
3907
+ "special": true
3908
+ },
3909
+ "490": {
3910
+ "content": "<|reserved_480|>",
3911
+ "lstrip": false,
3912
+ "normalized": false,
3913
+ "rstrip": false,
3914
+ "single_word": false,
3915
+ "special": true
3916
+ },
3917
+ "491": {
3918
+ "content": "<|reserved_481|>",
3919
+ "lstrip": false,
3920
+ "normalized": false,
3921
+ "rstrip": false,
3922
+ "single_word": false,
3923
+ "special": true
3924
+ },
3925
+ "492": {
3926
+ "content": "<|reserved_482|>",
3927
+ "lstrip": false,
3928
+ "normalized": false,
3929
+ "rstrip": false,
3930
+ "single_word": false,
3931
+ "special": true
3932
+ },
3933
+ "493": {
3934
+ "content": "<|reserved_483|>",
3935
+ "lstrip": false,
3936
+ "normalized": false,
3937
+ "rstrip": false,
3938
+ "single_word": false,
3939
+ "special": true
3940
+ },
3941
+ "494": {
3942
+ "content": "<|reserved_484|>",
3943
+ "lstrip": false,
3944
+ "normalized": false,
3945
+ "rstrip": false,
3946
+ "single_word": false,
3947
+ "special": true
3948
+ },
3949
+ "495": {
3950
+ "content": "<|reserved_485|>",
3951
+ "lstrip": false,
3952
+ "normalized": false,
3953
+ "rstrip": false,
3954
+ "single_word": false,
3955
+ "special": true
3956
+ },
3957
+ "496": {
3958
+ "content": "<|reserved_486|>",
3959
+ "lstrip": false,
3960
+ "normalized": false,
3961
+ "rstrip": false,
3962
+ "single_word": false,
3963
+ "special": true
3964
+ },
3965
+ "497": {
3966
+ "content": "<|reserved_487|>",
3967
+ "lstrip": false,
3968
+ "normalized": false,
3969
+ "rstrip": false,
3970
+ "single_word": false,
3971
+ "special": true
3972
+ },
3973
+ "498": {
3974
+ "content": "<|reserved_488|>",
3975
+ "lstrip": false,
3976
+ "normalized": false,
3977
+ "rstrip": false,
3978
+ "single_word": false,
3979
+ "special": true
3980
+ },
3981
+ "499": {
3982
+ "content": "<|reserved_489|>",
3983
+ "lstrip": false,
3984
+ "normalized": false,
3985
+ "rstrip": false,
3986
+ "single_word": false,
3987
+ "special": true
3988
+ },
3989
+ "500": {
3990
+ "content": "<|reserved_490|>",
3991
+ "lstrip": false,
3992
+ "normalized": false,
3993
+ "rstrip": false,
3994
+ "single_word": false,
3995
+ "special": true
3996
+ },
3997
+ "64011": {
3998
+ "content": "Mathias",
3999
+ "lstrip": false,
4000
+ "normalized": true,
4001
+ "rstrip": false,
4002
+ "single_word": false,
4003
+ "special": false
4004
+ },
4005
+ "64014": {
4006
+ "content": "python",
4007
+ "lstrip": false,
4008
+ "normalized": true,
4009
+ "rstrip": false,
4010
+ "single_word": false,
4011
+ "special": false
4012
+ },
4013
+ "64394": {
4014
+ "content": "<|cot_start|>",
4015
+ "lstrip": false,
4016
+ "normalized": false,
4017
+ "rstrip": false,
4018
+ "single_word": false,
4019
+ "special": true
4020
+ },
4021
+ "64395": {
4022
+ "content": "<|cot_end|>",
4023
+ "lstrip": false,
4024
+ "normalized": false,
4025
+ "rstrip": false,
4026
+ "single_word": false,
4027
+ "special": true
4028
+ },
4029
+ "64396": {
4030
+ "content": "<|review_start|>",
4031
+ "lstrip": false,
4032
+ "normalized": false,
4033
+ "rstrip": false,
4034
+ "single_word": false,
4035
+ "special": true
4036
+ },
4037
+ "64397": {
4038
+ "content": "<|review_end|>",
4039
+ "lstrip": false,
4040
+ "normalized": false,
4041
+ "rstrip": false,
4042
+ "single_word": false,
4043
+ "special": true
4044
+ },
4045
+ "64398": {
4046
+ "content": "<|file_start|>",
4047
+ "lstrip": false,
4048
+ "normalized": false,
4049
+ "rstrip": false,
4050
+ "single_word": false,
4051
+ "special": true
4052
+ },
4053
+ "64399": {
4054
+ "content": "<|file_end|>",
4055
+ "lstrip": false,
4056
+ "normalized": false,
4057
+ "rstrip": false,
4058
+ "single_word": false,
4059
+ "special": true
4060
+ }
4061
+ },
4062
+ "bos_token": "<|startoftext|>",
4063
+ "clean_up_tokenization_spaces": true,
4064
+ "eos_token": "<|im_end|>",
4065
+ "extra_special_tokens": {},
4066
+ "legacy": false,
4067
+ "model_max_length": 1000000000000000019884624838656,
4068
+ "pad_token": "<|pad|>",
4069
+ "sp_model_kwargs": {},
4070
+ "spaces_between_special_tokens": false,
4071
+ "tokenizer_class": "PreTrainedTokenizerFast",
4072
+ "use_default_system_prompt": false,
4073
+ "use_fast": true
4074
+ }