Text Generation
Transformers
Safetensors
English
qwen3
Merge
programming
code generation
code
coding
coder
chat
brainstorm
qwen
qwencoder
brainstorm20x
esper
esper-3
valiant
valiant-labs
qwen-3
qwen-3-14b
14b
reasoning
code-instruct
python
javascript
dev-ops
jenkins
terraform
scripting
powershell
azure
aws
gcp
cloud
problem-solving
architect
engineer
developer
creative
analytical
expert
rationality
conversational
instruct
float32
text-generation-inference
Commit
·
b8dede7
verified
·
0
Parent(s):
Duplicate from DavidAU/Qwen3-Esper3-Reasoning-CODER-Instruct-21B-Brainstorm20x
Browse files- .gitattributes +36 -0
- README.md +300 -0
- config.json +30 -0
- model-00001-of-00018.safetensors +3 -0
- model-00002-of-00018.safetensors +3 -0
- model-00003-of-00018.safetensors +3 -0
- model-00004-of-00018.safetensors +3 -0
- model-00005-of-00018.safetensors +3 -0
- model-00006-of-00018.safetensors +3 -0
- model-00007-of-00018.safetensors +3 -0
- model-00008-of-00018.safetensors +3 -0
- model-00009-of-00018.safetensors +3 -0
- model-00010-of-00018.safetensors +3 -0
- model-00011-of-00018.safetensors +3 -0
- model-00012-of-00018.safetensors +3 -0
- model-00013-of-00018.safetensors +3 -0
- model-00014-of-00018.safetensors +3 -0
- model-00015-of-00018.safetensors +3 -0
- model-00016-of-00018.safetensors +3 -0
- model-00017-of-00018.safetensors +3 -0
- model-00018-of-00018.safetensors +3 -0
- model.safetensors.index.json +1 -0
- special_tokens_map.json +31 -0
- tokenizer.json +3 -0
- tokenizer_config.json +240 -0
.gitattributes
ADDED
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,300 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
base_model:
|
4 |
+
- ValiantLabs/Qwen3-14B-Esper3
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
pipeline_tag: text-generation
|
8 |
+
tags:
|
9 |
+
- merge
|
10 |
+
- programming
|
11 |
+
- code generation
|
12 |
+
- code
|
13 |
+
- coding
|
14 |
+
- coder
|
15 |
+
- chat
|
16 |
+
- code
|
17 |
+
- chat
|
18 |
+
- brainstorm
|
19 |
+
- qwen
|
20 |
+
- qwen3
|
21 |
+
- qwencoder
|
22 |
+
- brainstorm20x
|
23 |
+
- esper
|
24 |
+
- esper-3
|
25 |
+
- valiant
|
26 |
+
- valiant-labs
|
27 |
+
- qwen
|
28 |
+
- qwen-3
|
29 |
+
- qwen-3-14b
|
30 |
+
- 14b
|
31 |
+
- reasoning
|
32 |
+
- code
|
33 |
+
- code-instruct
|
34 |
+
- python
|
35 |
+
- javascript
|
36 |
+
- dev-ops
|
37 |
+
- jenkins
|
38 |
+
- terraform
|
39 |
+
- scripting
|
40 |
+
- powershell
|
41 |
+
- azure
|
42 |
+
- aws
|
43 |
+
- gcp
|
44 |
+
- cloud
|
45 |
+
- problem-solving
|
46 |
+
- architect
|
47 |
+
- engineer
|
48 |
+
- developer
|
49 |
+
- creative
|
50 |
+
- analytical
|
51 |
+
- expert
|
52 |
+
- rationality
|
53 |
+
- conversational
|
54 |
+
- chat
|
55 |
+
- instruct
|
56 |
+
- float32
|
57 |
+
datasets:
|
58 |
+
- sequelbox/Titanium2.1-DeepSeek-R1
|
59 |
+
- sequelbox/Tachibana2-DeepSeek-R1
|
60 |
+
- sequelbox/Raiden-DeepSeek-R1
|
61 |
+
library_name: transformers
|
62 |
+
---
|
63 |
+
|
64 |
+
<h2>Qwen3-Esper3-Reasoning-CODER-Instruct-21B-Brainstorm20x</h2>
|
65 |
+
|
66 |
+
This repo contains the full precision source code, in "safe tensors" format to generate GGUFs, GPTQ, EXL2, AWQ, HQQ and other formats.
|
67 |
+
The source code can also be used directly.
|
68 |
+
|
69 |
+
This source code is in float32, which mirrors the quality of the original model's source.
|
70 |
+
|
71 |
+
This model contains Brainstorm 20x, combined with ValiantLabs's 14B General / Coder (instruct model):
|
72 |
+
|
73 |
+
https://huggingface.co/ValiantLabs/Qwen3-14B-Esper3
|
74 |
+
|
75 |
+
Information on the 14B model below, followed by Brainstorm 20x adapter (by DavidAU) and then a complete help
|
76 |
+
section for running LLM / AI models.
|
77 |
+
|
78 |
+
The Brainstorm adapter improves code generation, and unique code solving abilities.
|
79 |
+
|
80 |
+
This model requires:
|
81 |
+
- Jinja (embedded) or CHATML template
|
82 |
+
- Max context of 40k.
|
83 |
+
|
84 |
+
Settings used for testing (suggested):
|
85 |
+
- Temp .3 to .7
|
86 |
+
- Rep pen 1.05 to 1.1
|
87 |
+
- Topp .8 , minp .05
|
88 |
+
- Topk 20
|
89 |
+
- No system prompt.
|
90 |
+
|
91 |
+
FOR CODING:
|
92 |
+
|
93 |
+
Higher temps: .6 to .9 (even over 1) work better for more complex coding / especially with more restrictions.
|
94 |
+
|
95 |
+
Also, temp .9 with rep pen of 1.05 worked very well with this specific model.
|
96 |
+
|
97 |
+
This model will respond well to both detailed instructions and step by step refinement and additions to code.
|
98 |
+
|
99 |
+
As this is an instruct model, it will also benefit from a detailed system prompt too.
|
100 |
+
|
101 |
+
For simpler coding problems, lower quants will work well; but for complex/multi-step problem solving suggest Q6 or Q8.
|
102 |
+
|
103 |
+
---
|
104 |
+
|
105 |
+
**[Support our open-source dataset and model releases!](https://huggingface.co/spaces/sequelbox/SupportOpenSource)**
|
106 |
+
|
107 |
+
|
108 |
+

|
109 |
+
|
110 |
+
Esper 3: [Qwen3-4B](https://huggingface.co/ValiantLabs/Qwen3-4B-Esper3), [Qwen3-8B](https://huggingface.co/ValiantLabs/Qwen3-8B-Esper3), [Qwen3-14B](https://huggingface.co/ValiantLabs/Qwen3-14B-Esper3)
|
111 |
+
|
112 |
+
|
113 |
+
Esper 3 is a coding, architecture, and DevOps reasoning specialist built on Qwen 3.
|
114 |
+
- Finetuned on our [DevOps and architecture reasoning](https://huggingface.co/datasets/sequelbox/Titanium2.1-DeepSeek-R1) and [code reasoning](https://huggingface.co/datasets/sequelbox/Tachibana2-DeepSeek-R1) data generated with Deepseek R1!
|
115 |
+
- Improved [general and creative reasoning](https://huggingface.co/datasets/sequelbox/Raiden-DeepSeek-R1) to supplement problem-solving and general chat performance.
|
116 |
+
- Small model sizes allow running on local desktop and mobile, plus super-fast server inference!
|
117 |
+
|
118 |
+
|
119 |
+
## Prompting Guide
|
120 |
+
Esper 3 uses the [Qwen 3](https://huggingface.co/Qwen/Qwen3-4B) prompt format.
|
121 |
+
|
122 |
+
Esper 3 is a reasoning finetune; we recommend enable_thinking=True for all chats.
|
123 |
+
|
124 |
+
Example inference script to get started:
|
125 |
+
|
126 |
+
```python
|
127 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
128 |
+
|
129 |
+
model_name = "ValiantLabs/Qwen3-4B-Esper3"
|
130 |
+
|
131 |
+
# load the tokenizer and the model
|
132 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
133 |
+
model = AutoModelForCausalLM.from_pretrained(
|
134 |
+
model_name,
|
135 |
+
torch_dtype="auto",
|
136 |
+
device_map="auto"
|
137 |
+
)
|
138 |
+
|
139 |
+
# prepare the model input
|
140 |
+
prompt = "Write a Terraform configuration that uses the `aws_ami` data source to find the latest Amazon Linux 2 AMI. Then, provision an EC2 instance using this dynamically determined AMI ID."
|
141 |
+
messages = [
|
142 |
+
{"role": "user", "content": prompt}
|
143 |
+
]
|
144 |
+
text = tokenizer.apply_chat_template(
|
145 |
+
messages,
|
146 |
+
tokenize=False,
|
147 |
+
add_generation_prompt=True,
|
148 |
+
enable_thinking=True # Switches between thinking and non-thinking modes. Default is True.
|
149 |
+
)
|
150 |
+
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
|
151 |
+
|
152 |
+
# conduct text completion
|
153 |
+
generated_ids = model.generate(
|
154 |
+
**model_inputs,
|
155 |
+
max_new_tokens=32768
|
156 |
+
)
|
157 |
+
output_ids = generated_ids[0][len(model_inputs.input_ids[0]):].tolist()
|
158 |
+
|
159 |
+
# parsing thinking content
|
160 |
+
try:
|
161 |
+
# rindex finding 151668 (</think>)
|
162 |
+
index = len(output_ids) - output_ids[::-1].index(151668)
|
163 |
+
except ValueError:
|
164 |
+
index = 0
|
165 |
+
|
166 |
+
thinking_content = tokenizer.decode(output_ids[:index], skip_special_tokens=True).strip("\n")
|
167 |
+
content = tokenizer.decode(output_ids[index:], skip_special_tokens=True).strip("\n")
|
168 |
+
|
169 |
+
print("thinking content:", thinking_content)
|
170 |
+
print("content:", content)
|
171 |
+
```
|
172 |
+
|
173 |
+
|
174 |
+

|
175 |
+
|
176 |
+
|
177 |
+
Esper 3 is created by [Valiant Labs.](http://valiantlabs.ca/)
|
178 |
+
|
179 |
+
[Check out our HuggingFace page to see all of our models!](https://huggingface.co/ValiantLabs)
|
180 |
+
|
181 |
+
We care about open source. For everyone to use.
|
182 |
+
|
183 |
+
See more here:
|
184 |
+
|
185 |
+
https://huggingface.co/ValiantLabs/Qwen3-14B-Esper3
|
186 |
+
|
187 |
+
---
|
188 |
+
|
189 |
+
<H2>What is Brainstorm?</H2>
|
190 |
+
|
191 |
+
---
|
192 |
+
|
193 |
+
<B>Brainstorm 20x</B>
|
194 |
+
|
195 |
+
The BRAINSTORM process was developed by David_AU.
|
196 |
+
|
197 |
+
Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
|
198 |
+
scientific paper : Progressive LLaMA with Block Expansion </a>.
|
199 |
+
|
200 |
+
However I went in a completely different direction from what was outlined in this paper.
|
201 |
+
|
202 |
+
What is "Brainstorm" ?
|
203 |
+
|
204 |
+
The reasoning center of an LLM is taken apart, reassembled, and expanded.
|
205 |
+
|
206 |
+
In this case for this model: 20 times
|
207 |
+
|
208 |
+
Then these centers are individually calibrated. These "centers" also interact with each other.
|
209 |
+
This introduces subtle changes into the reasoning process.
|
210 |
+
The calibrations further adjust - dial up or down - these "changes" further.
|
211 |
+
The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
|
212 |
+
|
213 |
+
The core aim of this process is to increase the model's detail, concept and connection to the "world",
|
214 |
+
general concept connections, prose quality and prose length without affecting instruction following.
|
215 |
+
|
216 |
+
This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
|
217 |
+
|
218 |
+
Here are some of the enhancements this process brings to the model's performance:
|
219 |
+
|
220 |
+
- Prose generation seems more focused on the moment to moment.
|
221 |
+
- Sometimes there will be "preamble" and/or foreshadowing present.
|
222 |
+
- Fewer or no "cliches"
|
223 |
+
- Better overall prose and/or more complex / nuanced prose.
|
224 |
+
- A greater sense of nuance on all levels.
|
225 |
+
- Coherence is stronger.
|
226 |
+
- Description is more detailed, and connected closer to the content.
|
227 |
+
- Simile and Metaphors are stronger and better connected to the prose, story, and character.
|
228 |
+
- Sense of "there" / in the moment is enhanced.
|
229 |
+
- Details are more vivid, and there are more of them.
|
230 |
+
- Prose generation length can be long to extreme.
|
231 |
+
- Emotional engagement is stronger.
|
232 |
+
- The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
|
233 |
+
- The MORE instructions and/or details you provide the more strongly the model will respond.
|
234 |
+
- Depending on the model "voice" may be more "human" vs original model's "voice".
|
235 |
+
|
236 |
+
Other "lab" observations:
|
237 |
+
|
238 |
+
- This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
|
239 |
+
- However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
|
240 |
+
- From lab testing it seems to ponder, and consider more carefully roughly speaking.
|
241 |
+
- You could say this process sharpens the model's focus on it's task(s) at a deeper level.
|
242 |
+
|
243 |
+
The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
|
244 |
+
|
245 |
+
---
|
246 |
+
|
247 |
+
For more information / other Qwen/Mistral Coders / additional settings see:
|
248 |
+
|
249 |
+
[ https://huggingface.co/DavidAU/Qwen2.5-MOE-2x-4x-6x-8x__7B__Power-CODER__19B-30B-42B-53B-gguf ]
|
250 |
+
|
251 |
+
---
|
252 |
+
|
253 |
+
<H2>Help, Adjustments, Samplers, Parameters and More</H2>
|
254 |
+
|
255 |
+
---
|
256 |
+
|
257 |
+
<B>CHANGE THE NUMBER OF ACTIVE EXPERTS:</B>
|
258 |
+
|
259 |
+
See this document:
|
260 |
+
|
261 |
+
https://huggingface.co/DavidAU/How-To-Set-and-Manage-MOE-Mix-of-Experts-Model-Activation-of-Experts
|
262 |
+
|
263 |
+
<B>Settings: CHAT / ROLEPLAY and/or SMOOTHER operation of this model:</B>
|
264 |
+
|
265 |
+
In "KoboldCpp" or "oobabooga/text-generation-webui" or "Silly Tavern" ;
|
266 |
+
|
267 |
+
Set the "Smoothing_factor" to 1.5
|
268 |
+
|
269 |
+
: in KoboldCpp -> Settings->Samplers->Advanced-> "Smooth_F"
|
270 |
+
|
271 |
+
: in text-generation-webui -> parameters -> lower right.
|
272 |
+
|
273 |
+
: In Silly Tavern this is called: "Smoothing"
|
274 |
+
|
275 |
+
|
276 |
+
NOTE: For "text-generation-webui"
|
277 |
+
|
278 |
+
-> if using GGUFs you need to use "llama_HF" (which involves downloading some config files from the SOURCE version of this model)
|
279 |
+
|
280 |
+
Source versions (and config files) of my models are here:
|
281 |
+
|
282 |
+
https://huggingface.co/collections/DavidAU/d-au-source-files-for-gguf-exl2-awq-gptq-hqq-etc-etc-66b55cb8ba25f914cbf210be
|
283 |
+
|
284 |
+
OTHER OPTIONS:
|
285 |
+
|
286 |
+
- Increase rep pen to 1.1 to 1.15 (you don't need to do this if you use "smoothing_factor")
|
287 |
+
|
288 |
+
- If the interface/program you are using to run AI MODELS supports "Quadratic Sampling" ("smoothing") just make the adjustment as noted.
|
289 |
+
|
290 |
+
<B>Highest Quality Settings / Optimal Operation Guide / Parameters and Samplers</B>
|
291 |
+
|
292 |
+
This a "Class 1" model:
|
293 |
+
|
294 |
+
For all settings used for this model (including specifics for its "class"), including example generation(s) and for advanced settings guide (which many times addresses any model issue(s)), including methods to improve model performance for all use case(s) as well as chat, roleplay and other use case(s) please see:
|
295 |
+
|
296 |
+
[ https://huggingface.co/DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters ]
|
297 |
+
|
298 |
+
You can see all parameters used for generation, in addition to advanced parameters and samplers to get the most out of this model here:
|
299 |
+
|
300 |
+
[ https://huggingface.co/DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters ]
|
config.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"Qwen3ForCausalLM"
|
4 |
+
],
|
5 |
+
"attention_bias": false,
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"eos_token_id": 151645,
|
9 |
+
"head_dim": 128,
|
10 |
+
"hidden_act": "silu",
|
11 |
+
"hidden_size": 5120,
|
12 |
+
"initializer_range": 0.02,
|
13 |
+
"intermediate_size": 17408,
|
14 |
+
"max_position_embeddings": 40960,
|
15 |
+
"max_window_layers": 40,
|
16 |
+
"model_type": "qwen3",
|
17 |
+
"num_attention_heads": 40,
|
18 |
+
"num_hidden_layers": 59,
|
19 |
+
"num_key_value_heads": 8,
|
20 |
+
"rms_norm_eps": 1e-06,
|
21 |
+
"rope_scaling": null,
|
22 |
+
"rope_theta": 1000000,
|
23 |
+
"sliding_window": null,
|
24 |
+
"tie_word_embeddings": false,
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.52.0.dev0",
|
27 |
+
"use_cache": true,
|
28 |
+
"use_sliding_window": false,
|
29 |
+
"vocab_size": 151936
|
30 |
+
}
|
model-00001-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1c2903d48786e688cb003178c58438de66df39cb5e12be508936ca53465be7a8
|
3 |
+
size 3111649408
|
model-00002-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1ea996f2f489866d4f20882ec85b183a91bb4a089db9f1f164617459dae7ec12
|
3 |
+
size 4789434960
|
model-00003-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eb55bf71d290ef144286f6e9127ec6938728dbc58cae723ca7e4c0cb0130ecc6
|
3 |
+
size 4928480080
|
model-00004-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:369d6b5c2603b435ad683af99c487244340513a2141659742fd0bd393023d573
|
3 |
+
size 4676778968
|
model-00005-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:17508aecfaff90f89203da2807cab372f0100dd2da1b322b815cbfbe1bd2d5a0
|
3 |
+
size 4928480088
|
model-00006-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d96f039e7ba67f4eb291779801f303b75cb2294c9deb2275ebcdcdc33c9ea589
|
3 |
+
size 4928480080
|
model-00007-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7372aac52eb93e2318f8a6e4b6406585cfc0a111af3293d90247e87d1a39332e
|
3 |
+
size 4676778968
|
model-00008-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4b69750120a6411bddfe47114e0af05d763b2a39f5c3643f47893a96784440e8
|
3 |
+
size 4928480088
|
model-00009-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:64c10ffecfe8b0c1217433341835d6c23c418ac1f3d10d47e668dba1be18cdef
|
3 |
+
size 4928480080
|
model-00010-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f337de80c146ebb6f36841ec872bbb8183b73c6a32e3c0ea1eeaf3024107591e
|
3 |
+
size 4676778968
|
model-00011-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:81fdd046edb3e03de68a3cc586c5249e5c5cc464f92ff62d9d621b1242badbb2
|
3 |
+
size 4677170176
|
model-00012-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:124b98839fbe86d1fbec331869557561d3c1bb7c3b0aa5018273e07a0d6b0dfb
|
3 |
+
size 4991223400
|
model-00013-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3d20ca07ba568c1904e2a81bcd26c7ea832750fe796b775ccc92daa1c235c5ba
|
3 |
+
size 4991223400
|
model-00014-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5abfb66fdac557947aefd76634c126dc8172d76baed0ec4c4505eb41108d69f3
|
3 |
+
size 4991223392
|
model-00015-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:95047eca5f3b73c0f7d61a9d1d43fde853d922d5c082c391ef2de4347b367be2
|
3 |
+
size 4991647824
|
model-00016-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:412ffdc0813d37b0cff9bc3d1bf336588fa958f9d5574104b2bd0af67c0ecbf3
|
3 |
+
size 4991243344
|
model-00017-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1c7983ee01348d98054e285a397761cdc5f9b19bbfcb3322e482220c3ecf91dd
|
3 |
+
size 4718742784
|
model-00018-of-00018.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:31a1bae4d77662ba18d32f3458ecff856a30b19f7fd2b16a3e04f59eda9e1276
|
3 |
+
size 3250715080
|
model.safetensors.index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metadata": {"mergekit_version": "0.1.2"}, "weight_map": {"lm_head.weight": "model-00001-of-00018.safetensors", "model.embed_tokens.weight": "model-00002-of-00018.safetensors", "model.layers.0.input_layernorm.weight": "model-00002-of-00018.safetensors", "model.layers.0.mlp.down_proj.weight": "model-00002-of-00018.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00018.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00002-of-00018.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00018.safetensors", "model.layers.0.self_attn.k_norm.weight": "model-00002-of-00018.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00018.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00002-of-00018.safetensors", "model.layers.0.self_attn.q_norm.weight": "model-00002-of-00018.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00018.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00002-of-00018.safetensors", "model.layers.1.input_layernorm.weight": "model-00002-of-00018.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00002-of-00018.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00003-of-00018.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00003-of-00018.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00018.safetensors", "model.layers.1.self_attn.k_norm.weight": "model-00003-of-00018.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00003-of-00018.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00003-of-00018.safetensors", "model.layers.1.self_attn.q_norm.weight": "model-00003-of-00018.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00003-of-00018.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00003-of-00018.safetensors", "model.layers.10.input_layernorm.weight": "model-00003-of-00018.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00003-of-00018.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00003-of-00018.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00003-of-00018.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00003-of-00018.safetensors", "model.layers.10.self_attn.k_norm.weight": "model-00003-of-00018.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00003-of-00018.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00003-of-00018.safetensors", "model.layers.10.self_attn.q_norm.weight": "model-00003-of-00018.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00003-of-00018.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00003-of-00018.safetensors", "model.layers.11.input_layernorm.weight": "model-00003-of-00018.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00003-of-00018.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00003-of-00018.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00003-of-00018.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00003-of-00018.safetensors", "model.layers.11.self_attn.k_norm.weight": "model-00003-of-00018.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00003-of-00018.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00003-of-00018.safetensors", "model.layers.11.self_attn.q_norm.weight": "model-00003-of-00018.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00003-of-00018.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00003-of-00018.safetensors", "model.layers.12.input_layernorm.weight": "model-00003-of-00018.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00003-of-00018.safetensors", "model.layers.12.mlp.gate_proj.weight": "model-00003-of-00018.safetensors", "model.layers.12.mlp.up_proj.weight": "model-00003-of-00018.safetensors", "model.layers.12.post_attention_layernorm.weight": "model-00003-of-00018.safetensors", "model.layers.12.self_attn.k_norm.weight": "model-00003-of-00018.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00003-of-00018.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00003-of-00018.safetensors", "model.layers.12.self_attn.q_norm.weight": "model-00003-of-00018.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00003-of-00018.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00003-of-00018.safetensors", "model.layers.13.input_layernorm.weight": "model-00003-of-00018.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00004-of-00018.safetensors", "model.layers.13.mlp.gate_proj.weight": "model-00004-of-00018.safetensors", "model.layers.13.mlp.up_proj.weight": "model-00004-of-00018.safetensors", "model.layers.13.post_attention_layernorm.weight": "model-00004-of-00018.safetensors", "model.layers.13.self_attn.k_norm.weight": "model-00004-of-00018.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00004-of-00018.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00004-of-00018.safetensors", "model.layers.13.self_attn.q_norm.weight": "model-00004-of-00018.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00004-of-00018.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00004-of-00018.safetensors", "model.layers.14.input_layernorm.weight": "model-00004-of-00018.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00004-of-00018.safetensors", "model.layers.14.mlp.gate_proj.weight": "model-00004-of-00018.safetensors", "model.layers.14.mlp.up_proj.weight": "model-00004-of-00018.safetensors", "model.layers.14.post_attention_layernorm.weight": "model-00004-of-00018.safetensors", "model.layers.14.self_attn.k_norm.weight": "model-00004-of-00018.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00004-of-00018.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00004-of-00018.safetensors", "model.layers.14.self_attn.q_norm.weight": "model-00004-of-00018.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00004-of-00018.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00004-of-00018.safetensors", "model.layers.15.input_layernorm.weight": "model-00004-of-00018.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00004-of-00018.safetensors", "model.layers.15.mlp.gate_proj.weight": "model-00004-of-00018.safetensors", "model.layers.15.mlp.up_proj.weight": "model-00004-of-00018.safetensors", "model.layers.15.post_attention_layernorm.weight": "model-00004-of-00018.safetensors", "model.layers.15.self_attn.k_norm.weight": "model-00004-of-00018.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00004-of-00018.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00004-of-00018.safetensors", "model.layers.15.self_attn.q_norm.weight": "model-00004-of-00018.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00004-of-00018.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00004-of-00018.safetensors", "model.layers.16.input_layernorm.weight": "model-00004-of-00018.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00004-of-00018.safetensors", "model.layers.16.mlp.gate_proj.weight": "model-00004-of-00018.safetensors", "model.layers.16.mlp.up_proj.weight": "model-00005-of-00018.safetensors", "model.layers.16.post_attention_layernorm.weight": "model-00005-of-00018.safetensors", "model.layers.16.self_attn.k_norm.weight": "model-00005-of-00018.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00005-of-00018.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00005-of-00018.safetensors", "model.layers.16.self_attn.q_norm.weight": "model-00005-of-00018.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00005-of-00018.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00005-of-00018.safetensors", "model.layers.17.input_layernorm.weight": "model-00005-of-00018.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00005-of-00018.safetensors", "model.layers.17.mlp.gate_proj.weight": "model-00005-of-00018.safetensors", "model.layers.17.mlp.up_proj.weight": "model-00005-of-00018.safetensors", "model.layers.17.post_attention_layernorm.weight": "model-00005-of-00018.safetensors", "model.layers.17.self_attn.k_norm.weight": "model-00005-of-00018.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00005-of-00018.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00005-of-00018.safetensors", "model.layers.17.self_attn.q_norm.weight": "model-00005-of-00018.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00005-of-00018.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00005-of-00018.safetensors", "model.layers.18.input_layernorm.weight": "model-00005-of-00018.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00005-of-00018.safetensors", "model.layers.18.mlp.gate_proj.weight": "model-00005-of-00018.safetensors", "model.layers.18.mlp.up_proj.weight": "model-00005-of-00018.safetensors", "model.layers.18.post_attention_layernorm.weight": "model-00005-of-00018.safetensors", "model.layers.18.self_attn.k_norm.weight": "model-00005-of-00018.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00005-of-00018.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00005-of-00018.safetensors", "model.layers.18.self_attn.q_norm.weight": "model-00005-of-00018.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00005-of-00018.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00005-of-00018.safetensors", "model.layers.19.input_layernorm.weight": "model-00005-of-00018.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00005-of-00018.safetensors", "model.layers.19.mlp.gate_proj.weight": "model-00005-of-00018.safetensors", "model.layers.19.mlp.up_proj.weight": "model-00005-of-00018.safetensors", "model.layers.19.post_attention_layernorm.weight": "model-00005-of-00018.safetensors", "model.layers.19.self_attn.k_norm.weight": "model-00005-of-00018.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00005-of-00018.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00005-of-00018.safetensors", "model.layers.19.self_attn.q_norm.weight": "model-00005-of-00018.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00005-of-00018.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00005-of-00018.safetensors", "model.layers.2.input_layernorm.weight": "model-00005-of-00018.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00005-of-00018.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00006-of-00018.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00006-of-00018.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00006-of-00018.safetensors", "model.layers.2.self_attn.k_norm.weight": "model-00006-of-00018.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00006-of-00018.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00006-of-00018.safetensors", "model.layers.2.self_attn.q_norm.weight": "model-00006-of-00018.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00006-of-00018.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00006-of-00018.safetensors", "model.layers.20.input_layernorm.weight": "model-00006-of-00018.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00006-of-00018.safetensors", "model.layers.20.mlp.gate_proj.weight": "model-00006-of-00018.safetensors", "model.layers.20.mlp.up_proj.weight": "model-00006-of-00018.safetensors", "model.layers.20.post_attention_layernorm.weight": "model-00006-of-00018.safetensors", "model.layers.20.self_attn.k_norm.weight": "model-00006-of-00018.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00006-of-00018.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00006-of-00018.safetensors", "model.layers.20.self_attn.q_norm.weight": "model-00006-of-00018.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00006-of-00018.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00006-of-00018.safetensors", "model.layers.21.input_layernorm.weight": "model-00006-of-00018.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00006-of-00018.safetensors", "model.layers.21.mlp.gate_proj.weight": "model-00006-of-00018.safetensors", "model.layers.21.mlp.up_proj.weight": "model-00006-of-00018.safetensors", "model.layers.21.post_attention_layernorm.weight": "model-00006-of-00018.safetensors", "model.layers.21.self_attn.k_norm.weight": "model-00006-of-00018.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00006-of-00018.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00006-of-00018.safetensors", "model.layers.21.self_attn.q_norm.weight": "model-00006-of-00018.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00006-of-00018.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00006-of-00018.safetensors", "model.layers.22.input_layernorm.weight": "model-00006-of-00018.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00006-of-00018.safetensors", "model.layers.22.mlp.gate_proj.weight": "model-00006-of-00018.safetensors", "model.layers.22.mlp.up_proj.weight": "model-00006-of-00018.safetensors", "model.layers.22.post_attention_layernorm.weight": "model-00006-of-00018.safetensors", "model.layers.22.self_attn.k_norm.weight": "model-00006-of-00018.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00006-of-00018.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00006-of-00018.safetensors", "model.layers.22.self_attn.q_norm.weight": "model-00006-of-00018.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00006-of-00018.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00006-of-00018.safetensors", "model.layers.23.input_layernorm.weight": "model-00006-of-00018.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00007-of-00018.safetensors", "model.layers.23.mlp.gate_proj.weight": "model-00007-of-00018.safetensors", "model.layers.23.mlp.up_proj.weight": "model-00007-of-00018.safetensors", "model.layers.23.post_attention_layernorm.weight": "model-00007-of-00018.safetensors", "model.layers.23.self_attn.k_norm.weight": "model-00007-of-00018.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00007-of-00018.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00007-of-00018.safetensors", "model.layers.23.self_attn.q_norm.weight": "model-00007-of-00018.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00007-of-00018.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00007-of-00018.safetensors", "model.layers.24.input_layernorm.weight": "model-00007-of-00018.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00007-of-00018.safetensors", "model.layers.24.mlp.gate_proj.weight": "model-00007-of-00018.safetensors", "model.layers.24.mlp.up_proj.weight": "model-00007-of-00018.safetensors", "model.layers.24.post_attention_layernorm.weight": "model-00007-of-00018.safetensors", "model.layers.24.self_attn.k_norm.weight": "model-00007-of-00018.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00007-of-00018.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00007-of-00018.safetensors", "model.layers.24.self_attn.q_norm.weight": "model-00007-of-00018.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00007-of-00018.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00007-of-00018.safetensors", "model.layers.25.input_layernorm.weight": "model-00007-of-00018.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00007-of-00018.safetensors", "model.layers.25.mlp.gate_proj.weight": "model-00007-of-00018.safetensors", "model.layers.25.mlp.up_proj.weight": "model-00007-of-00018.safetensors", "model.layers.25.post_attention_layernorm.weight": "model-00007-of-00018.safetensors", "model.layers.25.self_attn.k_norm.weight": "model-00007-of-00018.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00007-of-00018.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00007-of-00018.safetensors", "model.layers.25.self_attn.q_norm.weight": "model-00007-of-00018.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00007-of-00018.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00007-of-00018.safetensors", "model.layers.26.input_layernorm.weight": "model-00007-of-00018.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00007-of-00018.safetensors", "model.layers.26.mlp.gate_proj.weight": "model-00007-of-00018.safetensors", "model.layers.26.mlp.up_proj.weight": "model-00008-of-00018.safetensors", "model.layers.26.post_attention_layernorm.weight": "model-00008-of-00018.safetensors", "model.layers.26.self_attn.k_norm.weight": "model-00008-of-00018.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00008-of-00018.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00008-of-00018.safetensors", "model.layers.26.self_attn.q_norm.weight": "model-00008-of-00018.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00008-of-00018.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00008-of-00018.safetensors", "model.layers.27.input_layernorm.weight": "model-00008-of-00018.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00008-of-00018.safetensors", "model.layers.27.mlp.gate_proj.weight": "model-00008-of-00018.safetensors", "model.layers.27.mlp.up_proj.weight": "model-00008-of-00018.safetensors", "model.layers.27.post_attention_layernorm.weight": "model-00008-of-00018.safetensors", "model.layers.27.self_attn.k_norm.weight": "model-00008-of-00018.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00008-of-00018.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00008-of-00018.safetensors", "model.layers.27.self_attn.q_norm.weight": "model-00008-of-00018.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00008-of-00018.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00008-of-00018.safetensors", "model.layers.28.input_layernorm.weight": "model-00008-of-00018.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00008-of-00018.safetensors", "model.layers.28.mlp.gate_proj.weight": "model-00008-of-00018.safetensors", "model.layers.28.mlp.up_proj.weight": "model-00008-of-00018.safetensors", "model.layers.28.post_attention_layernorm.weight": "model-00008-of-00018.safetensors", "model.layers.28.self_attn.k_norm.weight": "model-00008-of-00018.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00008-of-00018.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00008-of-00018.safetensors", "model.layers.28.self_attn.q_norm.weight": "model-00008-of-00018.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00008-of-00018.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00008-of-00018.safetensors", "model.layers.29.input_layernorm.weight": "model-00008-of-00018.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00008-of-00018.safetensors", "model.layers.29.mlp.gate_proj.weight": "model-00008-of-00018.safetensors", "model.layers.29.mlp.up_proj.weight": "model-00008-of-00018.safetensors", "model.layers.29.post_attention_layernorm.weight": "model-00008-of-00018.safetensors", "model.layers.29.self_attn.k_norm.weight": "model-00008-of-00018.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00008-of-00018.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00008-of-00018.safetensors", "model.layers.29.self_attn.q_norm.weight": "model-00008-of-00018.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00008-of-00018.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00008-of-00018.safetensors", "model.layers.3.input_layernorm.weight": "model-00008-of-00018.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00008-of-00018.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00009-of-00018.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00009-of-00018.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00009-of-00018.safetensors", "model.layers.3.self_attn.k_norm.weight": "model-00009-of-00018.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00009-of-00018.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00009-of-00018.safetensors", "model.layers.3.self_attn.q_norm.weight": "model-00009-of-00018.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00009-of-00018.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00009-of-00018.safetensors", "model.layers.30.input_layernorm.weight": "model-00009-of-00018.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00009-of-00018.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00009-of-00018.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00009-of-00018.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00009-of-00018.safetensors", "model.layers.30.self_attn.k_norm.weight": "model-00009-of-00018.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00009-of-00018.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00009-of-00018.safetensors", "model.layers.30.self_attn.q_norm.weight": "model-00009-of-00018.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00009-of-00018.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00009-of-00018.safetensors", "model.layers.31.input_layernorm.weight": "model-00009-of-00018.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00009-of-00018.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00009-of-00018.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00009-of-00018.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00009-of-00018.safetensors", "model.layers.31.self_attn.k_norm.weight": "model-00009-of-00018.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00009-of-00018.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00009-of-00018.safetensors", "model.layers.31.self_attn.q_norm.weight": "model-00009-of-00018.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00009-of-00018.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00009-of-00018.safetensors", "model.layers.32.input_layernorm.weight": "model-00009-of-00018.safetensors", "model.layers.32.mlp.down_proj.weight": "model-00009-of-00018.safetensors", "model.layers.32.mlp.gate_proj.weight": "model-00009-of-00018.safetensors", "model.layers.32.mlp.up_proj.weight": "model-00009-of-00018.safetensors", "model.layers.32.post_attention_layernorm.weight": "model-00009-of-00018.safetensors", "model.layers.32.self_attn.k_norm.weight": "model-00009-of-00018.safetensors", "model.layers.32.self_attn.k_proj.weight": "model-00009-of-00018.safetensors", "model.layers.32.self_attn.o_proj.weight": "model-00009-of-00018.safetensors", "model.layers.32.self_attn.q_norm.weight": "model-00009-of-00018.safetensors", "model.layers.32.self_attn.q_proj.weight": "model-00009-of-00018.safetensors", "model.layers.32.self_attn.v_proj.weight": "model-00009-of-00018.safetensors", "model.layers.33.input_layernorm.weight": "model-00009-of-00018.safetensors", "model.layers.33.mlp.down_proj.weight": "model-00010-of-00018.safetensors", "model.layers.33.mlp.gate_proj.weight": "model-00010-of-00018.safetensors", "model.layers.33.mlp.up_proj.weight": "model-00010-of-00018.safetensors", "model.layers.33.post_attention_layernorm.weight": "model-00010-of-00018.safetensors", "model.layers.33.self_attn.k_norm.weight": "model-00010-of-00018.safetensors", "model.layers.33.self_attn.k_proj.weight": "model-00010-of-00018.safetensors", "model.layers.33.self_attn.o_proj.weight": "model-00010-of-00018.safetensors", "model.layers.33.self_attn.q_norm.weight": "model-00010-of-00018.safetensors", "model.layers.33.self_attn.q_proj.weight": "model-00010-of-00018.safetensors", "model.layers.33.self_attn.v_proj.weight": "model-00010-of-00018.safetensors", "model.layers.34.input_layernorm.weight": "model-00010-of-00018.safetensors", "model.layers.34.mlp.down_proj.weight": "model-00010-of-00018.safetensors", "model.layers.34.mlp.gate_proj.weight": "model-00010-of-00018.safetensors", "model.layers.34.mlp.up_proj.weight": "model-00010-of-00018.safetensors", "model.layers.34.post_attention_layernorm.weight": "model-00010-of-00018.safetensors", "model.layers.34.self_attn.k_norm.weight": "model-00010-of-00018.safetensors", "model.layers.34.self_attn.k_proj.weight": "model-00010-of-00018.safetensors", "model.layers.34.self_attn.o_proj.weight": "model-00010-of-00018.safetensors", "model.layers.34.self_attn.q_norm.weight": "model-00010-of-00018.safetensors", "model.layers.34.self_attn.q_proj.weight": "model-00010-of-00018.safetensors", "model.layers.34.self_attn.v_proj.weight": "model-00010-of-00018.safetensors", "model.layers.35.input_layernorm.weight": "model-00010-of-00018.safetensors", "model.layers.35.mlp.down_proj.weight": "model-00010-of-00018.safetensors", "model.layers.35.mlp.gate_proj.weight": "model-00010-of-00018.safetensors", "model.layers.35.mlp.up_proj.weight": "model-00010-of-00018.safetensors", "model.layers.35.post_attention_layernorm.weight": "model-00010-of-00018.safetensors", "model.layers.35.self_attn.k_norm.weight": "model-00010-of-00018.safetensors", "model.layers.35.self_attn.k_proj.weight": "model-00010-of-00018.safetensors", "model.layers.35.self_attn.o_proj.weight": "model-00010-of-00018.safetensors", "model.layers.35.self_attn.q_norm.weight": "model-00010-of-00018.safetensors", "model.layers.35.self_attn.q_proj.weight": "model-00010-of-00018.safetensors", "model.layers.35.self_attn.v_proj.weight": "model-00010-of-00018.safetensors", "model.layers.36.input_layernorm.weight": "model-00010-of-00018.safetensors", "model.layers.36.mlp.down_proj.weight": "model-00010-of-00018.safetensors", "model.layers.36.mlp.gate_proj.weight": "model-00010-of-00018.safetensors", "model.layers.36.mlp.up_proj.weight": "model-00011-of-00018.safetensors", "model.layers.36.post_attention_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.36.self_attn.k_norm.weight": "model-00011-of-00018.safetensors", "model.layers.36.self_attn.k_proj.weight": "model-00011-of-00018.safetensors", "model.layers.36.self_attn.o_proj.weight": "model-00011-of-00018.safetensors", "model.layers.36.self_attn.q_norm.weight": "model-00011-of-00018.safetensors", "model.layers.36.self_attn.q_proj.weight": "model-00011-of-00018.safetensors", "model.layers.36.self_attn.v_proj.weight": "model-00011-of-00018.safetensors", "model.layers.37.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.37.mlp.down_proj.weight": "model-00011-of-00018.safetensors", "model.layers.37.mlp.gate_proj.weight": "model-00011-of-00018.safetensors", "model.layers.37.mlp.up_proj.weight": "model-00011-of-00018.safetensors", "model.layers.37.post_attention_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.37.self_attn.k_norm.weight": "model-00011-of-00018.safetensors", "model.layers.37.self_attn.k_proj.weight": "model-00011-of-00018.safetensors", "model.layers.37.self_attn.o_proj.weight": "model-00011-of-00018.safetensors", "model.layers.37.self_attn.q_norm.weight": "model-00011-of-00018.safetensors", "model.layers.37.self_attn.q_proj.weight": "model-00011-of-00018.safetensors", "model.layers.37.self_attn.v_proj.weight": "model-00011-of-00018.safetensors", "model.layers.38.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.38.mlp.down_proj.weight": "model-00011-of-00018.safetensors", "model.layers.38.mlp.gate_proj.weight": "model-00011-of-00018.safetensors", "model.layers.38.mlp.up_proj.weight": "model-00011-of-00018.safetensors", "model.layers.38.post_attention_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.38.self_attn.k_norm.weight": "model-00011-of-00018.safetensors", "model.layers.38.self_attn.k_proj.weight": "model-00011-of-00018.safetensors", "model.layers.38.self_attn.o_proj.weight": "model-00011-of-00018.safetensors", "model.layers.38.self_attn.q_norm.weight": "model-00011-of-00018.safetensors", "model.layers.38.self_attn.q_proj.weight": "model-00011-of-00018.safetensors", "model.layers.38.self_attn.v_proj.weight": "model-00011-of-00018.safetensors", "model.layers.49.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.51.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.53.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.55.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.57.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.50.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.39.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.40.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.41.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.42.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.43.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.44.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.45.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.46.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.47.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.48.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.52.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.54.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.56.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.58.input_layernorm.weight": "model-00011-of-00018.safetensors", "model.layers.50.mlp.down_proj.weight": "model-00011-of-00018.safetensors", "model.layers.52.mlp.down_proj.weight": "model-00011-of-00018.safetensors", "model.layers.54.mlp.down_proj.weight": "model-00011-of-00018.safetensors", "model.layers.56.mlp.down_proj.weight": "model-00011-of-00018.safetensors", "model.layers.58.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.51.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.39.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.40.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.41.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.42.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.43.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.44.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.45.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.46.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.47.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.48.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.49.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.53.mlp.down_proj.weight": "model-00012-of-00018.safetensors", "model.layers.55.mlp.down_proj.weight": "model-00013-of-00018.safetensors", "model.layers.57.mlp.down_proj.weight": "model-00013-of-00018.safetensors", "model.layers.49.mlp.gate_proj.weight": "model-00013-of-00018.safetensors", "model.layers.51.mlp.gate_proj.weight": "model-00013-of-00018.safetensors", "model.layers.53.mlp.gate_proj.weight": "model-00013-of-00018.safetensors", "model.layers.55.mlp.gate_proj.weight": "model-00013-of-00018.safetensors", "model.layers.57.mlp.gate_proj.weight": "model-00013-of-00018.safetensors", "model.layers.52.mlp.gate_proj.weight": "model-00013-of-00018.safetensors", "model.layers.39.mlp.gate_proj.weight": "model-00013-of-00018.safetensors", "model.layers.40.mlp.gate_proj.weight": "model-00013-of-00018.safetensors", "model.layers.41.mlp.gate_proj.weight": "model-00013-of-00018.safetensors", "model.layers.42.mlp.gate_proj.weight": "model-00013-of-00018.safetensors", "model.layers.43.mlp.gate_proj.weight": "model-00013-of-00018.safetensors", "model.layers.44.mlp.gate_proj.weight": "model-00013-of-00018.safetensors", "model.layers.45.mlp.gate_proj.weight": "model-00014-of-00018.safetensors", "model.layers.46.mlp.gate_proj.weight": "model-00014-of-00018.safetensors", "model.layers.47.mlp.gate_proj.weight": "model-00014-of-00018.safetensors", "model.layers.48.mlp.gate_proj.weight": "model-00014-of-00018.safetensors", "model.layers.50.mlp.gate_proj.weight": "model-00014-of-00018.safetensors", "model.layers.54.mlp.gate_proj.weight": "model-00014-of-00018.safetensors", "model.layers.56.mlp.gate_proj.weight": "model-00014-of-00018.safetensors", "model.layers.58.mlp.gate_proj.weight": "model-00014-of-00018.safetensors", "model.layers.49.mlp.up_proj.weight": "model-00014-of-00018.safetensors", "model.layers.50.mlp.up_proj.weight": "model-00014-of-00018.safetensors", "model.layers.52.mlp.up_proj.weight": "model-00014-of-00018.safetensors", "model.layers.54.mlp.up_proj.weight": "model-00014-of-00018.safetensors", "model.layers.56.mlp.up_proj.weight": "model-00014-of-00018.safetensors", "model.layers.58.mlp.up_proj.weight": "model-00014-of-00018.safetensors", "model.layers.53.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.39.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.40.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.41.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.42.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.43.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.44.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.45.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.46.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.47.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.48.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.51.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.55.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.57.mlp.up_proj.weight": "model-00015-of-00018.safetensors", "model.layers.49.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.51.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.53.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.55.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.57.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.50.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.39.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.40.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.41.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.42.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.43.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.44.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.45.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.46.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.47.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.48.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.52.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.54.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.56.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.58.post_attention_layernorm.weight": "model-00015-of-00018.safetensors", "model.layers.50.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.52.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.54.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.56.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.58.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.51.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.39.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.40.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.41.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.42.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.43.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.44.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.45.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.46.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.47.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.48.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.49.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.53.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.55.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.57.self_attn.k_norm.weight": "model-00015-of-00018.safetensors", "model.layers.49.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.51.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.53.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.55.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.57.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.52.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.39.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.40.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.41.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.42.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.43.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.44.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.45.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.46.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.47.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.48.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.50.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.54.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.56.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.58.self_attn.k_proj.weight": "model-00016-of-00018.safetensors", "model.layers.48.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.49.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.51.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.53.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.55.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.57.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.52.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.39.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.40.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.41.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.42.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.43.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.44.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.45.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.46.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.47.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.50.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.54.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.56.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.58.self_attn.o_proj.weight": "model-00016-of-00018.safetensors", "model.layers.48.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.49.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.50.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.52.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.54.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.56.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.58.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.39.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.40.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.41.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.42.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.43.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.44.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.45.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.46.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.47.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.51.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.53.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.55.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.57.self_attn.q_norm.weight": "model-00016-of-00018.safetensors", "model.layers.49.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.51.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.53.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.55.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.57.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.50.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.39.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.40.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.41.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.42.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.43.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.44.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.45.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.46.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.47.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.48.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.52.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.54.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.56.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.58.self_attn.q_proj.weight": "model-00016-of-00018.safetensors", "model.layers.48.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.50.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.52.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.54.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.56.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.58.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.51.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.39.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.40.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.41.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.42.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.43.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.44.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.45.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.46.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.47.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.49.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.53.self_attn.v_proj.weight": "model-00016-of-00018.safetensors", "model.layers.55.self_attn.v_proj.weight": "model-00017-of-00018.safetensors", "model.layers.57.self_attn.v_proj.weight": "model-00017-of-00018.safetensors", "model.layers.4.input_layernorm.weight": "model-00017-of-00018.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00017-of-00018.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00017-of-00018.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00017-of-00018.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00017-of-00018.safetensors", "model.layers.4.self_attn.k_norm.weight": "model-00017-of-00018.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00017-of-00018.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00017-of-00018.safetensors", "model.layers.4.self_attn.q_norm.weight": "model-00017-of-00018.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00017-of-00018.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00017-of-00018.safetensors", "model.layers.5.input_layernorm.weight": "model-00017-of-00018.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00017-of-00018.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00017-of-00018.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00017-of-00018.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00017-of-00018.safetensors", "model.layers.5.self_attn.k_norm.weight": "model-00017-of-00018.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00017-of-00018.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00017-of-00018.safetensors", "model.layers.5.self_attn.q_norm.weight": "model-00017-of-00018.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00017-of-00018.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00017-of-00018.safetensors", "model.layers.6.input_layernorm.weight": "model-00017-of-00018.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00017-of-00018.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00017-of-00018.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00017-of-00018.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00017-of-00018.safetensors", "model.layers.6.self_attn.k_norm.weight": "model-00017-of-00018.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00017-of-00018.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00017-of-00018.safetensors", "model.layers.6.self_attn.q_norm.weight": "model-00017-of-00018.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00017-of-00018.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00017-of-00018.safetensors", "model.layers.7.input_layernorm.weight": "model-00017-of-00018.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00017-of-00018.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00017-of-00018.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00018-of-00018.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00018-of-00018.safetensors", "model.layers.7.self_attn.k_norm.weight": "model-00018-of-00018.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00018-of-00018.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00018-of-00018.safetensors", "model.layers.7.self_attn.q_norm.weight": "model-00018-of-00018.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00018-of-00018.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00018-of-00018.safetensors", "model.layers.8.input_layernorm.weight": "model-00018-of-00018.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00018-of-00018.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00018-of-00018.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00018-of-00018.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00018-of-00018.safetensors", "model.layers.8.self_attn.k_norm.weight": "model-00018-of-00018.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00018-of-00018.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00018-of-00018.safetensors", "model.layers.8.self_attn.q_norm.weight": "model-00018-of-00018.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00018-of-00018.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00018-of-00018.safetensors", "model.layers.9.input_layernorm.weight": "model-00018-of-00018.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00018-of-00018.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00018-of-00018.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00018-of-00018.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00018-of-00018.safetensors", "model.layers.9.self_attn.k_norm.weight": "model-00018-of-00018.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00018-of-00018.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00018-of-00018.safetensors", "model.layers.9.self_attn.q_norm.weight": "model-00018-of-00018.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00018-of-00018.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00018-of-00018.safetensors", "model.norm.weight": "model-00018-of-00018.safetensors"}}
|
special_tokens_map.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
"<|im_start|>",
|
4 |
+
"<|im_end|>",
|
5 |
+
"<|object_ref_start|>",
|
6 |
+
"<|object_ref_end|>",
|
7 |
+
"<|box_start|>",
|
8 |
+
"<|box_end|>",
|
9 |
+
"<|quad_start|>",
|
10 |
+
"<|quad_end|>",
|
11 |
+
"<|vision_start|>",
|
12 |
+
"<|vision_end|>",
|
13 |
+
"<|vision_pad|>",
|
14 |
+
"<|image_pad|>",
|
15 |
+
"<|video_pad|>"
|
16 |
+
],
|
17 |
+
"eos_token": {
|
18 |
+
"content": "<|im_end|>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": false,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
},
|
24 |
+
"pad_token": {
|
25 |
+
"content": "<|endoftext|>",
|
26 |
+
"lstrip": false,
|
27 |
+
"normalized": false,
|
28 |
+
"rstrip": false,
|
29 |
+
"single_word": false
|
30 |
+
}
|
31 |
+
}
|
tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
|
3 |
+
size 11422654
|
tokenizer_config.json
ADDED
@@ -0,0 +1,240 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": false,
|
3 |
+
"add_prefix_space": false,
|
4 |
+
"added_tokens_decoder": {
|
5 |
+
"151643": {
|
6 |
+
"content": "<|endoftext|>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false,
|
11 |
+
"special": true
|
12 |
+
},
|
13 |
+
"151644": {
|
14 |
+
"content": "<|im_start|>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false,
|
19 |
+
"special": true
|
20 |
+
},
|
21 |
+
"151645": {
|
22 |
+
"content": "<|im_end|>",
|
23 |
+
"lstrip": false,
|
24 |
+
"normalized": false,
|
25 |
+
"rstrip": false,
|
26 |
+
"single_word": false,
|
27 |
+
"special": true
|
28 |
+
},
|
29 |
+
"151646": {
|
30 |
+
"content": "<|object_ref_start|>",
|
31 |
+
"lstrip": false,
|
32 |
+
"normalized": false,
|
33 |
+
"rstrip": false,
|
34 |
+
"single_word": false,
|
35 |
+
"special": true
|
36 |
+
},
|
37 |
+
"151647": {
|
38 |
+
"content": "<|object_ref_end|>",
|
39 |
+
"lstrip": false,
|
40 |
+
"normalized": false,
|
41 |
+
"rstrip": false,
|
42 |
+
"single_word": false,
|
43 |
+
"special": true
|
44 |
+
},
|
45 |
+
"151648": {
|
46 |
+
"content": "<|box_start|>",
|
47 |
+
"lstrip": false,
|
48 |
+
"normalized": false,
|
49 |
+
"rstrip": false,
|
50 |
+
"single_word": false,
|
51 |
+
"special": true
|
52 |
+
},
|
53 |
+
"151649": {
|
54 |
+
"content": "<|box_end|>",
|
55 |
+
"lstrip": false,
|
56 |
+
"normalized": false,
|
57 |
+
"rstrip": false,
|
58 |
+
"single_word": false,
|
59 |
+
"special": true
|
60 |
+
},
|
61 |
+
"151650": {
|
62 |
+
"content": "<|quad_start|>",
|
63 |
+
"lstrip": false,
|
64 |
+
"normalized": false,
|
65 |
+
"rstrip": false,
|
66 |
+
"single_word": false,
|
67 |
+
"special": true
|
68 |
+
},
|
69 |
+
"151651": {
|
70 |
+
"content": "<|quad_end|>",
|
71 |
+
"lstrip": false,
|
72 |
+
"normalized": false,
|
73 |
+
"rstrip": false,
|
74 |
+
"single_word": false,
|
75 |
+
"special": true
|
76 |
+
},
|
77 |
+
"151652": {
|
78 |
+
"content": "<|vision_start|>",
|
79 |
+
"lstrip": false,
|
80 |
+
"normalized": false,
|
81 |
+
"rstrip": false,
|
82 |
+
"single_word": false,
|
83 |
+
"special": true
|
84 |
+
},
|
85 |
+
"151653": {
|
86 |
+
"content": "<|vision_end|>",
|
87 |
+
"lstrip": false,
|
88 |
+
"normalized": false,
|
89 |
+
"rstrip": false,
|
90 |
+
"single_word": false,
|
91 |
+
"special": true
|
92 |
+
},
|
93 |
+
"151654": {
|
94 |
+
"content": "<|vision_pad|>",
|
95 |
+
"lstrip": false,
|
96 |
+
"normalized": false,
|
97 |
+
"rstrip": false,
|
98 |
+
"single_word": false,
|
99 |
+
"special": true
|
100 |
+
},
|
101 |
+
"151655": {
|
102 |
+
"content": "<|image_pad|>",
|
103 |
+
"lstrip": false,
|
104 |
+
"normalized": false,
|
105 |
+
"rstrip": false,
|
106 |
+
"single_word": false,
|
107 |
+
"special": true
|
108 |
+
},
|
109 |
+
"151656": {
|
110 |
+
"content": "<|video_pad|>",
|
111 |
+
"lstrip": false,
|
112 |
+
"normalized": false,
|
113 |
+
"rstrip": false,
|
114 |
+
"single_word": false,
|
115 |
+
"special": true
|
116 |
+
},
|
117 |
+
"151657": {
|
118 |
+
"content": "<tool_call>",
|
119 |
+
"lstrip": false,
|
120 |
+
"normalized": false,
|
121 |
+
"rstrip": false,
|
122 |
+
"single_word": false,
|
123 |
+
"special": false
|
124 |
+
},
|
125 |
+
"151658": {
|
126 |
+
"content": "</tool_call>",
|
127 |
+
"lstrip": false,
|
128 |
+
"normalized": false,
|
129 |
+
"rstrip": false,
|
130 |
+
"single_word": false,
|
131 |
+
"special": false
|
132 |
+
},
|
133 |
+
"151659": {
|
134 |
+
"content": "<|fim_prefix|>",
|
135 |
+
"lstrip": false,
|
136 |
+
"normalized": false,
|
137 |
+
"rstrip": false,
|
138 |
+
"single_word": false,
|
139 |
+
"special": false
|
140 |
+
},
|
141 |
+
"151660": {
|
142 |
+
"content": "<|fim_middle|>",
|
143 |
+
"lstrip": false,
|
144 |
+
"normalized": false,
|
145 |
+
"rstrip": false,
|
146 |
+
"single_word": false,
|
147 |
+
"special": false
|
148 |
+
},
|
149 |
+
"151661": {
|
150 |
+
"content": "<|fim_suffix|>",
|
151 |
+
"lstrip": false,
|
152 |
+
"normalized": false,
|
153 |
+
"rstrip": false,
|
154 |
+
"single_word": false,
|
155 |
+
"special": false
|
156 |
+
},
|
157 |
+
"151662": {
|
158 |
+
"content": "<|fim_pad|>",
|
159 |
+
"lstrip": false,
|
160 |
+
"normalized": false,
|
161 |
+
"rstrip": false,
|
162 |
+
"single_word": false,
|
163 |
+
"special": false
|
164 |
+
},
|
165 |
+
"151663": {
|
166 |
+
"content": "<|repo_name|>",
|
167 |
+
"lstrip": false,
|
168 |
+
"normalized": false,
|
169 |
+
"rstrip": false,
|
170 |
+
"single_word": false,
|
171 |
+
"special": false
|
172 |
+
},
|
173 |
+
"151664": {
|
174 |
+
"content": "<|file_sep|>",
|
175 |
+
"lstrip": false,
|
176 |
+
"normalized": false,
|
177 |
+
"rstrip": false,
|
178 |
+
"single_word": false,
|
179 |
+
"special": false
|
180 |
+
},
|
181 |
+
"151665": {
|
182 |
+
"content": "<tool_response>",
|
183 |
+
"lstrip": false,
|
184 |
+
"normalized": false,
|
185 |
+
"rstrip": false,
|
186 |
+
"single_word": false,
|
187 |
+
"special": false
|
188 |
+
},
|
189 |
+
"151666": {
|
190 |
+
"content": "</tool_response>",
|
191 |
+
"lstrip": false,
|
192 |
+
"normalized": false,
|
193 |
+
"rstrip": false,
|
194 |
+
"single_word": false,
|
195 |
+
"special": false
|
196 |
+
},
|
197 |
+
"151667": {
|
198 |
+
"content": "<think>",
|
199 |
+
"lstrip": false,
|
200 |
+
"normalized": false,
|
201 |
+
"rstrip": false,
|
202 |
+
"single_word": false,
|
203 |
+
"special": false
|
204 |
+
},
|
205 |
+
"151668": {
|
206 |
+
"content": "</think>",
|
207 |
+
"lstrip": false,
|
208 |
+
"normalized": false,
|
209 |
+
"rstrip": false,
|
210 |
+
"single_word": false,
|
211 |
+
"special": false
|
212 |
+
}
|
213 |
+
},
|
214 |
+
"additional_special_tokens": [
|
215 |
+
"<|im_start|>",
|
216 |
+
"<|im_end|>",
|
217 |
+
"<|object_ref_start|>",
|
218 |
+
"<|object_ref_end|>",
|
219 |
+
"<|box_start|>",
|
220 |
+
"<|box_end|>",
|
221 |
+
"<|quad_start|>",
|
222 |
+
"<|quad_end|>",
|
223 |
+
"<|vision_start|>",
|
224 |
+
"<|vision_end|>",
|
225 |
+
"<|vision_pad|>",
|
226 |
+
"<|image_pad|>",
|
227 |
+
"<|video_pad|>"
|
228 |
+
],
|
229 |
+
"bos_token": null,
|
230 |
+
"chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0].role == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0].content + '<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}\n{%- for message in messages[::-1] %}\n {%- set index = (messages|length - 1) - loop.index0 %}\n {%- if ns.multi_step_tool and message.role == \"user\" and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}\n {%- set ns.multi_step_tool = false %}\n {%- set ns.last_query_index = index %}\n {%- endif %}\n{%- endfor %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {%- set content = message.content %}\n {%- set reasoning_content = '' %}\n {%- if message.reasoning_content is defined and message.reasoning_content is not none %}\n {%- set reasoning_content = message.reasoning_content %}\n {%- else %}\n {%- if '</think>' in message.content %}\n {%- set content = message.content.split('</think>')[-1].lstrip('\\n') %}\n {%- set reasoning_content = message.content.split('</think>')[0].rstrip('\\n').split('<think>')[-1].lstrip('\\n') %}\n {%- endif %}\n {%- endif %}\n {%- if loop.index0 > ns.last_query_index %}\n {%- if loop.last or (not loop.last and reasoning_content) %}\n {{- '<|im_start|>' + message.role + '\\n<think>\\n' + reasoning_content.strip('\\n') + '\\n</think>\\n\\n' + content.lstrip('\\n') }}\n {%- else %}\n {{- '<|im_start|>' + message.role + '\\n' + content }}\n {%- endif %}\n {%- else %}\n {{- '<|im_start|>' + message.role + '\\n' + content }}\n {%- endif %}\n {%- if message.tool_calls %}\n {%- for tool_call in message.tool_calls %}\n {%- if (loop.first and content) or (not loop.first) %}\n {{- '\\n' }}\n {%- endif %}\n {%- if tool_call.function %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {%- if tool_call.arguments is string %}\n {{- tool_call.arguments }}\n {%- else %}\n {{- tool_call.arguments | tojson }}\n {%- endif %}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {%- endif %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if loop.first or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n {%- if enable_thinking is defined and enable_thinking is false %}\n {{- '<think>\\n\\n</think>\\n\\n' }}\n {%- endif %}\n{%- endif %}",
|
231 |
+
"clean_up_tokenization_spaces": false,
|
232 |
+
"eos_token": "<|im_end|>",
|
233 |
+
"errors": "replace",
|
234 |
+
"extra_special_tokens": {},
|
235 |
+
"model_max_length": 131072,
|
236 |
+
"pad_token": "<|endoftext|>",
|
237 |
+
"split_special_tokens": false,
|
238 |
+
"tokenizer_class": "Qwen2Tokenizer",
|
239 |
+
"unk_token": null
|
240 |
+
}
|