Commit
·
db4d8ee
0
Parent(s):
initial commit
Browse files- .gitattributes +35 -0
- README.md +255 -0
- config.json +26 -0
- model-00001-of-00007.safetensors +3 -0
- model-00002-of-00007.safetensors +3 -0
- model-00003-of-00007.safetensors +3 -0
- model-00004-of-00007.safetensors +3 -0
- model-00005-of-00007.safetensors +3 -0
- model-00006-of-00007.safetensors +3 -0
- model-00007-of-00007.safetensors +3 -0
- model.safetensors.index.json +1 -0
- tokenizer.json +0 -0
- tokenizer_config.json +0 -0
.gitattributes
ADDED
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,255 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
---
|
4 |
+
|
5 |
+
# Apriel-Nemotron-15b-Thinker
|
6 |
+
|
7 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/63d3095c2727d7888cbb54e2/Lt1t0tOO5emz1X23Azg-E.png" width="120" alt="thumbnail"/> `/ˈɑː.pri.əl/`
|
8 |
+
|
9 |
+
---
|
10 |
+
|
11 |
+
## Table of Contents
|
12 |
+
|
13 |
+
1. [Summary](#summary)
|
14 |
+
2. [Evaluation](#evaluation)
|
15 |
+
3. [Training Details](#training-details)
|
16 |
+
4. [How to Use](#how-to-use)
|
17 |
+
5. [Intended Use](#intended-use)
|
18 |
+
6. [Limitations](#limitations)
|
19 |
+
7. [Security and Responsible Use](#security-and-responsible-use)
|
20 |
+
8. [Software](#software)
|
21 |
+
9. [License](#license)
|
22 |
+
10. [Acknowledgements](#acknowledgements)
|
23 |
+
11. [Citation](#citation)
|
24 |
+
|
25 |
+
|
26 |
+
---
|
27 |
+
|
28 |
+
## Summary
|
29 |
+
|
30 |
+
**Apriel-Nemotron-15b-Thinker** is a 15 billion‑parameter reasoning model in ServiceNow’s Apriel SLM series which achieves competitive performance against similarly sized state-of-the-art models like o1‑mini, QWQ‑32b, and EXAONE‑Deep‑32b, all while maintaining only half the memory footprint of those alternatives. It builds upon the **Apriel‑15b‑base** checkpoint through a three‑stage training pipeline (CPT, SFT and GRPO).
|
31 |
+
|
32 |
+
**Highlights**
|
33 |
+
- Half the size of SOTA models like QWQ-32b and EXAONE-32b and hence **memory efficient**.
|
34 |
+
- On par or outperforms on tasks like - MBPP, BFCL, Enterprise RAG, MT Bench, MixEval, IFEval and Multi-Challenge making it great for **Agentic / Enterprise tasks**.
|
35 |
+
- Competitive performance on academic benchmarks like AIME-24 AIME-25, AMC-23, MATH-500 and GPQA considering model size.
|
36 |
+
|
37 |
+
---
|
38 |
+
|
39 |
+
## Evaluation
|
40 |
+
|
41 |
+
Evaluations were conducted using [lm-eval-harness](https://github.com/EleutherAI/lm-evaluation-harness) and [evalchemy](https://github.com/mlfoundations/evalchemy).
|
42 |
+
|
43 |
+
|
44 |
+

|
45 |
+
|
46 |
+
|
47 |
+
|
48 |
+
|
49 |
+

|
50 |
+
|
51 |
+
---
|
52 |
+
|
53 |
+
|
54 |
+
## Training Details
|
55 |
+
|
56 |
+
**Mid training / Continual Pre‑training**
|
57 |
+
In this stage, the model is trained on 100+ billion tokens of carefully curated examples drawn from mathematical reasoning, coding challenges, scientific discourse and logical puzzles. The objective is to strengthen foundational reasoning capabilities of the model. This stage is super critical for the model to function as a reasoner and provides significant lifts in reasoning benchmarks.
|
58 |
+
|
59 |
+
**Supervised Fine‑Tuning (SFT)**
|
60 |
+
Next, we SFT the model using 200,000 high‑quality demonstrations that cover mathematical and scientific problem‑solving, coding tasks, generic instruction‑following scenarios, API/function invocation use cases etc.
|
61 |
+
|
62 |
+
**Reinforcement Learning**
|
63 |
+
Although the SFT‑tuned checkpoint delivers strong performance on core competencies like mathematics and general knowledge, it exhibits weaknesses in instruction following and coding tasks. To address these gaps, we apply GRPO (with some minor modifications to the objective). The result is significant improvement on benchmarks such as IFEval, Multi Challenge, Enterprise RAG, MBPP and BFCL, while preserving scores on competition‑level math exams like AIME and AMC. GRPO also yields modest gains on GPQA and MixEval.
|
64 |
+
|
65 |
+
Throughout training, intermediate snapshots from both the SFT and GRPO stages are periodically merged, improving generalization and catastrophic forgetting.
|
66 |
+
|
67 |
+
*Technical report with more details - coming soon.*
|
68 |
+
|
69 |
+
---
|
70 |
+
|
71 |
+
|
72 |
+
## How to Use
|
73 |
+
|
74 |
+
```bash
|
75 |
+
pip install transformers
|
76 |
+
```
|
77 |
+
|
78 |
+
---
|
79 |
+
|
80 |
+
### Running the Reasoning model
|
81 |
+
|
82 |
+
|
83 |
+
Here is a code snippet demonstrating the model's usage with the transformers library's generate function:
|
84 |
+
|
85 |
+
```python
|
86 |
+
import re
|
87 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
88 |
+
|
89 |
+
model_name = "ServiceNow-AI/Apriel-Nemotron-15b-Thinker"
|
90 |
+
|
91 |
+
# load the tokenizer and the model
|
92 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
93 |
+
model = AutoModelForCausalLM.from_pretrained(
|
94 |
+
model_name,
|
95 |
+
torch_dtype="auto",
|
96 |
+
device_map="auto"
|
97 |
+
)
|
98 |
+
|
99 |
+
# prepare the model input
|
100 |
+
prompt = "Positive real numbers $x$ and $y$ satisfy $y^3=x^2$ and $(y-x)^2=4y^2$. What is $x+y$?\nMark your solution with \\boxed"
|
101 |
+
messages = [
|
102 |
+
{"role": "user", "content": prompt}
|
103 |
+
]
|
104 |
+
|
105 |
+
tools = []
|
106 |
+
|
107 |
+
text = tokenizer.apply_chat_template(
|
108 |
+
messages,
|
109 |
+
tokenize=False,
|
110 |
+
add_generation_prompt=True,
|
111 |
+
tools=tools
|
112 |
+
)
|
113 |
+
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
|
114 |
+
|
115 |
+
# conduct text completion
|
116 |
+
generated_ids = model.generate(
|
117 |
+
**model_inputs,
|
118 |
+
max_new_tokens=65536
|
119 |
+
)
|
120 |
+
output = tokenizer.decode(generated_ids[0], skip_special_tokens=True)
|
121 |
+
|
122 |
+
# parsing the response
|
123 |
+
response = re.findall(r"\[BEGIN FINAL RESPONSE\](.*?)\[END FINAL RESPONSE\]", output, re.DOTALL)[0].strip()
|
124 |
+
print("output:", output)
|
125 |
+
print("response:", response)
|
126 |
+
```
|
127 |
+
---
|
128 |
+
|
129 |
+
### Chat Template
|
130 |
+
|
131 |
+
|
132 |
+
```
|
133 |
+
<|system|>
|
134 |
+
You are a thoughtful and systematic AI assistant built by ServiceNow Language Models (SLAM) lab. Before providing an answer, analyze the problem carefully and present your reasoning step by step. After explaining your thought process, provide the final solution in the following format: [BEGIN FINAL RESPONSE] ... [END FINAL RESPONSE].
|
135 |
+
<|end|>
|
136 |
+
<|user|>
|
137 |
+
# user message here
|
138 |
+
<|end|>
|
139 |
+
<|assistant|>
|
140 |
+
Here are my reasoning steps:
|
141 |
+
# thoughts here
|
142 |
+
[BEGIN FINAL RESPONSE]
|
143 |
+
# assistant response here
|
144 |
+
[END FINAL RESPONSE]
|
145 |
+
<|end|>
|
146 |
+
```
|
147 |
+
The model will first generate its thinking process and then generate its final response between `[BEGIN FINAL RESPONSE]` and `[END FINAL RESPONSE]`. Here is a code snippet demonstrating the application of the chat template:
|
148 |
+
|
149 |
+
|
150 |
+
|
151 |
+
```python
|
152 |
+
from transformers import AutoTokenizer
|
153 |
+
model_name = "ServiceNow-AI/Apriel-Nemotron-15b-Thinker"
|
154 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
155 |
+
|
156 |
+
# prepare the model input
|
157 |
+
custom_system_prompt = "Answer like a pirate."
|
158 |
+
prompt = "You are an expert assistant in the implementation of customer experience management aspect of retail applications \n \nYou will be using Python as the programming language. \n \nYou will utilize a factory design pattern for the implementation and following the dependency inversion principle \n \nYou will modify the implementation based on user requirements. \n \nUpon user request, you will add, update, and remove the features & enhancements in the implementation provided by you. \n \nYou will ask whether the user wants to refactor the provided code or needs a sample implementation for reference. Upon user confirmation, I will proceed accordingly. \n \n**Guidelines:** \n 1. **User Requirements:** \n - You have to ask users about their requirements, clarify the user expectations, and suggest the best possible solution by providing examples of Python code snippets. \n - Ask users about which type of reports they need to assess the AI model's performance, accuracy, and reliability. \n - After providing the solution, you have to ask the user about the trial of the solution and modify the solution based on the user feedback. \n \n 2. **Libraries/Frameworks:** \n - You will be utilizing Python as a programming language. \n - You will be using Flask framework for REST APIS implementation \n \n 3. **Communication Gesture:** \n - Your conversation with the user should be interactive, supportive, courageous, and professional. \n - You have to break down the complex concepts into sub-concepts and try to explain them to the user. \n - You have to ask the user for the required parameters. If the user refuses to provide in 2 attempts, politely exit the conversation. \n - You have to provide your supported parameters to the user, if the user refuses to accept them then you have to put an apology note and exit the conversation. \n - You have to track the conversation about unasked questions by the user. If some/one of the questions remain then you have to remind the user about these questions and proceed to answer them based on the user's confirmation \n \n 4. **Implementation:** \n - Your code/implementations should be reliable, scaleable, modular, and reusable. \n - You will be providing unit tests for the implementation upon user request. \n - You will be following MVC architecture for the applications \n - Your implementations must be well-commented and readable \n \n \n- Today's date is 23rd August 2024. \n- The default sender email is [email protected].\nHi, I am conducting research on retail customer feedback systems and I need assistance with designing and implementing them. Could you kindly provide me with a list of general customer feedback system modules?"
|
159 |
+
messages = [
|
160 |
+
{"role": "user", "content": custom_system_prompt + "\n\n" + prompt}
|
161 |
+
]
|
162 |
+
# example tools
|
163 |
+
tools = [{"type": "function", "function": {"name": "getRetailFeedbackModules", "description": "Returns the list of modules usually present in the retail industry", "parameters": {"type": "object", "properties": {"page": {"type": "integer", "description": "The current page number.", "default": 1}, "page_size": {"type": "integer", "description": "The number of items per page.", "default": 3}}}}}, {"type": "function", "function": {"name": "verifyImplementation", "description": "Returns the list of modules usually present in the retail industry", "parameters": {"type": "object", "properties": {"coding_language": {"type": "string", "description": "The supported languages for verification of implementation.", "default": "python", "enum": ["python", "java", "php"]}, "code": {"type": "string", "description": "The code which needs verification"}, "design_pattern": {"type": "string", "description": "The design pattern to verify in the implementation", "enum": ["factory", "strategy", "singleton"]}, "verify_best_practices": {"type": "boolean", "description": "The verification of the coding style based on the language selected", "default": true}}}}}]
|
164 |
+
text = tokenizer.apply_chat_template(
|
165 |
+
messages,
|
166 |
+
tokenize=False,
|
167 |
+
add_generation_prompt=True,
|
168 |
+
tools=tools
|
169 |
+
)
|
170 |
+
model_inputs = tokenizer([text], return_tensors="pt")
|
171 |
+
```
|
172 |
+
|
173 |
+
### Usage Guidelines
|
174 |
+
1. Use the model’s default chat template, which already includes a system prompt. We recommend adding all other instructions within the user message.
|
175 |
+
2. We recommend setting temperature to `0.6`.
|
176 |
+
3. We ensure the model starts with `Here are my reasoning steps:\n` during all our evaluations. This is implemented in the default chat template.
|
177 |
+
|
178 |
+
---
|
179 |
+
|
180 |
+
|
181 |
+
## Intended Use
|
182 |
+
|
183 |
+
The Apriel family of models are designed for a variety of general-purpose instruction tasks, including:
|
184 |
+
|
185 |
+
- Code assistance and generation
|
186 |
+
- Logical reasoning and multi-step tasks
|
187 |
+
- Question answering and information retrieval
|
188 |
+
- Function calling, complex instruction following and agent use cases
|
189 |
+
|
190 |
+
They are **not intended** for use in safety-critical applications without human oversight or in scenarios requiring guaranteed factual accuracy.
|
191 |
+
|
192 |
+
---
|
193 |
+
|
194 |
+
## Limitations
|
195 |
+
|
196 |
+
- **Factual accuracy:** May produce incorrect, misleading, or outdated content. Outputs should be verified before use in critical contexts.
|
197 |
+
- **Bias:** May reflect societal, cultural, or systemic biases present in training data.
|
198 |
+
- **Ethics:** Do not use the model to produce harmful, unlawful, or unethical content.
|
199 |
+
- **Language:** Strongest performance is in English. Output quality may degrade in underrepresented languages.
|
200 |
+
- **Critical use:** Not suitable for medical, legal, financial, or other high-risk applications without safeguards.
|
201 |
+
|
202 |
+
---
|
203 |
+
|
204 |
+
## Security and Responsible Use
|
205 |
+
|
206 |
+
**Security Responsibilities:**
|
207 |
+
Deployers and users are strongly encouraged to align their security practices with established frameworks and regulatory guidelines such as the EU AI Act and the NIST AI Risk Management Framework (RMF).
|
208 |
+
|
209 |
+
**Guidelines for Deployers:**
|
210 |
+
|
211 |
+
- Regularly conduct robustness assessments to identify and mitigate adversarial inputs.
|
212 |
+
- Implement validation and filtering processes to prevent harmful or biased outputs.
|
213 |
+
- Continuously perform data privacy checks to guard against unintended data leaks.
|
214 |
+
- Document and communicate the model's limitations, intended usage, and known security risks to all end-users.
|
215 |
+
- Schedule periodic security reviews and updates to address emerging threats and vulnerabilities.
|
216 |
+
|
217 |
+
**Guidelines for Users:**
|
218 |
+
|
219 |
+
- Follow established security policies and usage guidelines provided by deployers.
|
220 |
+
- Protect and manage sensitive information when interacting with the model.
|
221 |
+
- Report anomalies, suspicious behavior, or unsafe outputs to deployers or developers.
|
222 |
+
- Maintain human oversight and apply judgment to mitigate potential security or ethical risks during interactions.
|
223 |
+
|
224 |
+
**Disclaimer:**
|
225 |
+
Users accept responsibility for securely deploying, managing, and using this open-source LLM. The model is provided "as-is," without explicit or implied warranty regarding security or fitness for any specific application or environment.
|
226 |
+
|
227 |
+
---
|
228 |
+
|
229 |
+
## Software
|
230 |
+
|
231 |
+
- **Training stack:** [Fast-LLM](https://github.com/ServiceNow/Fast-LLM)
|
232 |
+
|
233 |
+
---
|
234 |
+
|
235 |
+
## License
|
236 |
+
|
237 |
+
MIT
|
238 |
+
|
239 |
+
---
|
240 |
+
|
241 |
+
## Acknowledgments
|
242 |
+
|
243 |
+
>We thank researchers at Nvidia for sharing detailed insights and data from their work in building reasoners! This greatly accelerated our research and we recognize the same with our model naming convention!
|
244 |
+
|
245 |
+
## Citation
|
246 |
+
|
247 |
+
```bibtex
|
248 |
+
@misc{Apriel-nemotron-15b-thinker,
|
249 |
+
author = {Slam labs team},
|
250 |
+
title = {Apriel Nemotron 15b Thinker},
|
251 |
+
howpublished = {https://huggingface.co/ServiceNow-AI/Apriel-Nemotron-15b-Thinker},
|
252 |
+
publisher = {SLAM - ServiceNow Language Models Lab}
|
253 |
+
year = {2025}
|
254 |
+
}
|
255 |
+
```
|
config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"MistralForCausalLM"
|
4 |
+
],
|
5 |
+
"attention_dropout": 0.0,
|
6 |
+
"bos_token_id": 1,
|
7 |
+
"eos_token_id": 2,
|
8 |
+
"head_dim": 128,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 5120,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 14336,
|
13 |
+
"max_position_embeddings": 65536,
|
14 |
+
"model_type": "mistral",
|
15 |
+
"num_attention_heads": 32,
|
16 |
+
"num_hidden_layers": 50,
|
17 |
+
"num_key_value_heads": 8,
|
18 |
+
"rms_norm_eps": 1e-05,
|
19 |
+
"rope_theta": 1000000.0,
|
20 |
+
"sliding_window": null,
|
21 |
+
"tie_word_embeddings": false,
|
22 |
+
"torch_dtype": "bfloat16",
|
23 |
+
"transformers_version": "4.51.0",
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 131072
|
26 |
+
}
|
model-00001-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:00612ad46e14e7681fe64ff50846531653d6e4a4916a3cef4c7538a73a57edee
|
3 |
+
size 4865489336
|
model-00002-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:44c037d202094da58311a86b15dadd99c8ed199b2804a1871b4b21d55961a424
|
3 |
+
size 4907529456
|
model-00003-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:44f90d5d75ef33e0de005eb764b211095799d16a8a7a0381e1ddf806a0a7dfdc
|
3 |
+
size 4907529464
|
model-00004-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8e19fc4c19cff6a5d0992ddedae3e64c8e2a2c13bd5b00dd307779268f2f4f4f
|
3 |
+
size 4907529456
|
model-00005-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35f07af444a1f301e7288e0092c50098fe4cd2d1f32728d9de552e1629a50ad9
|
3 |
+
size 4907529456
|
model-00006-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2d5bf79826c21358dc18028acb9f5db95b4497cc25c73f543fbcf5294975c82c
|
3 |
+
size 4907529432
|
model-00007-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c56d515592e2a4b7882e12af0a4afff254a7c039af21ab59ce243b1d53f3918d
|
3 |
+
size 545281040
|
model.safetensors.index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metadata": {"mergekit_version": "0.1.0"}, "weight_map": {"lm_head.weight": "model-00001-of-00007.safetensors", "model.embed_tokens.weight": "model-00001-of-00007.safetensors", "model.layers.0.input_layernorm.weight": "model-00001-of-00007.safetensors", "model.layers.0.mlp.down_proj.weight": "model-00001-of-00007.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00007.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00001-of-00007.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00007.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00007.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00007.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00007.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00007.safetensors", "model.layers.1.input_layernorm.weight": "model-00001-of-00007.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00001-of-00007.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00007.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00001-of-00007.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00007.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00007.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00007.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00007.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00007.safetensors", "model.layers.10.input_layernorm.weight": "model-00001-of-00007.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00001-of-00007.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00007.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00001-of-00007.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00007.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00007.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00007.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00007.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00007.safetensors", "model.layers.11.input_layernorm.weight": "model-00001-of-00007.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00001-of-00007.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00007.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00001-of-00007.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00007.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00007.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00007.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00007.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00007.safetensors", "model.layers.12.input_layernorm.weight": "model-00001-of-00007.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00002-of-00007.safetensors", "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.12.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.13.input_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00002-of-00007.safetensors", "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.13.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.input_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.15.input_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00002-of-00007.safetensors", "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.15.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.16.input_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00002-of-00007.safetensors", "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.16.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.17.input_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00002-of-00007.safetensors", "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.17.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.18.input_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00002-of-00007.safetensors", "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.18.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.19.input_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00002-of-00007.safetensors", "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.19.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.input_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.input_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.20.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.20.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.20.post_attention_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.21.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.21.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.21.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.21.post_attention_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.22.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.22.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.23.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.23.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.24.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.24.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.25.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.25.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.26.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.26.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.27.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.27.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.28.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.28.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.28.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.29.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.29.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.29.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.29.post_attention_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.3.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.30.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.32.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.32.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.32.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.32.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.32.post_attention_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.32.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.32.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.32.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.32.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.33.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.33.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.33.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.33.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.33.post_attention_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.33.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.33.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.33.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.33.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.34.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.34.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.34.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.34.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.34.post_attention_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.34.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.34.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.34.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.34.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.35.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.35.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.35.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.35.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.35.post_attention_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.35.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.35.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.35.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.35.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.36.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.36.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.36.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.36.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.36.post_attention_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.36.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.36.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.36.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.36.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.37.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.37.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.37.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.37.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.37.post_attention_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.37.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.37.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.37.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.37.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.38.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.38.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.38.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.38.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.38.post_attention_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.38.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.38.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.38.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.38.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.39.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.39.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.39.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.39.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.39.post_attention_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.39.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.39.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.39.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.39.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.4.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.40.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.40.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.40.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.40.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.40.post_attention_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.40.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.40.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.40.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.40.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.41.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.41.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.41.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.41.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.41.post_attention_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.41.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.41.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.41.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.41.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.42.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.42.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.42.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.42.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.42.post_attention_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.42.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.42.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.42.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.42.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.43.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.43.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.43.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.43.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.43.post_attention_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.43.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.43.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.43.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.43.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.44.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.44.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.44.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.44.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.44.post_attention_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.44.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.44.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.44.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.44.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.45.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.45.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.45.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.45.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.45.post_attention_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.45.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.45.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.45.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.45.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.46.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.46.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.46.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.46.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.46.post_attention_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.46.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.46.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.46.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.46.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.47.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.47.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.47.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.47.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.47.post_attention_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.47.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.47.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.47.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.47.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.48.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.48.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.48.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.48.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.48.post_attention_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.48.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.48.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.48.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.48.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.49.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.49.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.49.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.49.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.49.post_attention_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.49.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.49.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.49.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.49.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.5.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.6.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.7.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.8.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.9.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00007-of-00007.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00007-of-00007.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00007-of-00007.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00007-of-00007.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00007.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00007.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00007.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00007.safetensors", "model.norm.weight": "model-00007-of-00007.safetensors"}}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|