Trained with Unsloth
Browse files
pytorch_model-00001-of-00002.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9db1d62d20b1fb85148fc5a288e47ca4d707283f3479ab9d5561146aaa03d747
|
3 |
+
size 4983825789
|
pytorch_model-00002-of-00002.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:169df8ba9ccb4e93588b382233f8d7daecf5567cbc7f89170f8a307d6142be57
|
3 |
+
size 1278363922
|
pytorch_model.bin.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"model.embed_tokens.weight": "pytorch_model-00001-of-00002.bin",
|
@@ -244,11 +244,11 @@
|
|
244 |
"model.layers.26.self_attn.q_proj.weight": "pytorch_model-00001-of-00002.bin",
|
245 |
"model.layers.26.self_attn.v_proj.bias": "pytorch_model-00001-of-00002.bin",
|
246 |
"model.layers.26.self_attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
|
247 |
-
"model.layers.27.input_layernorm.weight": "pytorch_model-
|
248 |
-
"model.layers.27.mlp.down_proj.weight": "pytorch_model-
|
249 |
"model.layers.27.mlp.gate_proj.weight": "pytorch_model-00001-of-00002.bin",
|
250 |
"model.layers.27.mlp.up_proj.weight": "pytorch_model-00001-of-00002.bin",
|
251 |
-
"model.layers.27.post_attention_layernorm.weight": "pytorch_model-
|
252 |
"model.layers.27.self_attn.k_proj.bias": "pytorch_model-00001-of-00002.bin",
|
253 |
"model.layers.27.self_attn.k_proj.weight": "pytorch_model-00001-of-00002.bin",
|
254 |
"model.layers.27.self_attn.o_proj.weight": "pytorch_model-00001-of-00002.bin",
|
@@ -261,13 +261,13 @@
|
|
261 |
"model.layers.28.mlp.gate_proj.weight": "pytorch_model-00002-of-00002.bin",
|
262 |
"model.layers.28.mlp.up_proj.weight": "pytorch_model-00002-of-00002.bin",
|
263 |
"model.layers.28.post_attention_layernorm.weight": "pytorch_model-00002-of-00002.bin",
|
264 |
-
"model.layers.28.self_attn.k_proj.bias": "pytorch_model-
|
265 |
-
"model.layers.28.self_attn.k_proj.weight": "pytorch_model-
|
266 |
-
"model.layers.28.self_attn.o_proj.weight": "pytorch_model-
|
267 |
-
"model.layers.28.self_attn.q_proj.bias": "pytorch_model-
|
268 |
-
"model.layers.28.self_attn.q_proj.weight": "pytorch_model-
|
269 |
-
"model.layers.28.self_attn.v_proj.bias": "pytorch_model-
|
270 |
-
"model.layers.28.self_attn.v_proj.weight": "pytorch_model-
|
271 |
"model.layers.29.input_layernorm.weight": "pytorch_model-00002-of-00002.bin",
|
272 |
"model.layers.29.mlp.down_proj.weight": "pytorch_model-00002-of-00002.bin",
|
273 |
"model.layers.29.mlp.gate_proj.weight": "pytorch_model-00002-of-00002.bin",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 6262054912
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"model.embed_tokens.weight": "pytorch_model-00001-of-00002.bin",
|
|
|
244 |
"model.layers.26.self_attn.q_proj.weight": "pytorch_model-00001-of-00002.bin",
|
245 |
"model.layers.26.self_attn.v_proj.bias": "pytorch_model-00001-of-00002.bin",
|
246 |
"model.layers.26.self_attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
|
247 |
+
"model.layers.27.input_layernorm.weight": "pytorch_model-00002-of-00002.bin",
|
248 |
+
"model.layers.27.mlp.down_proj.weight": "pytorch_model-00002-of-00002.bin",
|
249 |
"model.layers.27.mlp.gate_proj.weight": "pytorch_model-00001-of-00002.bin",
|
250 |
"model.layers.27.mlp.up_proj.weight": "pytorch_model-00001-of-00002.bin",
|
251 |
+
"model.layers.27.post_attention_layernorm.weight": "pytorch_model-00002-of-00002.bin",
|
252 |
"model.layers.27.self_attn.k_proj.bias": "pytorch_model-00001-of-00002.bin",
|
253 |
"model.layers.27.self_attn.k_proj.weight": "pytorch_model-00001-of-00002.bin",
|
254 |
"model.layers.27.self_attn.o_proj.weight": "pytorch_model-00001-of-00002.bin",
|
|
|
261 |
"model.layers.28.mlp.gate_proj.weight": "pytorch_model-00002-of-00002.bin",
|
262 |
"model.layers.28.mlp.up_proj.weight": "pytorch_model-00002-of-00002.bin",
|
263 |
"model.layers.28.post_attention_layernorm.weight": "pytorch_model-00002-of-00002.bin",
|
264 |
+
"model.layers.28.self_attn.k_proj.bias": "pytorch_model-00002-of-00002.bin",
|
265 |
+
"model.layers.28.self_attn.k_proj.weight": "pytorch_model-00002-of-00002.bin",
|
266 |
+
"model.layers.28.self_attn.o_proj.weight": "pytorch_model-00002-of-00002.bin",
|
267 |
+
"model.layers.28.self_attn.q_proj.bias": "pytorch_model-00002-of-00002.bin",
|
268 |
+
"model.layers.28.self_attn.q_proj.weight": "pytorch_model-00002-of-00002.bin",
|
269 |
+
"model.layers.28.self_attn.v_proj.bias": "pytorch_model-00002-of-00002.bin",
|
270 |
+
"model.layers.28.self_attn.v_proj.weight": "pytorch_model-00002-of-00002.bin",
|
271 |
"model.layers.29.input_layernorm.weight": "pytorch_model-00002-of-00002.bin",
|
272 |
"model.layers.29.mlp.down_proj.weight": "pytorch_model-00002-of-00002.bin",
|
273 |
"model.layers.29.mlp.gate_proj.weight": "pytorch_model-00002-of-00002.bin",
|