Pramodith commited on
Commit
1bb3f7a
·
verified ·
1 Parent(s): f493796

Trained with Unsloth

Browse files
pytorch_model-00001-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5cde958cdf3fadeb39988e8f06757a565e152c947e01ed7ae19b98076336dc66
3
- size 4957627586
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9db1d62d20b1fb85148fc5a288e47ca4d707283f3479ab9d5561146aaa03d747
3
+ size 4983825789
pytorch_model-00002-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:767a875d33de970532c1233474e218fe8344a7a1c986b69e81f1fee7db54e626
3
- size 1214384514
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:169df8ba9ccb4e93588b382233f8d7daecf5567cbc7f89170f8a307d6142be57
3
+ size 1278363922
pytorch_model.bin.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 6171877376
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "pytorch_model-00001-of-00002.bin",
@@ -244,11 +244,11 @@
244
  "model.layers.26.self_attn.q_proj.weight": "pytorch_model-00001-of-00002.bin",
245
  "model.layers.26.self_attn.v_proj.bias": "pytorch_model-00001-of-00002.bin",
246
  "model.layers.26.self_attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
247
- "model.layers.27.input_layernorm.weight": "pytorch_model-00001-of-00002.bin",
248
- "model.layers.27.mlp.down_proj.weight": "pytorch_model-00001-of-00002.bin",
249
  "model.layers.27.mlp.gate_proj.weight": "pytorch_model-00001-of-00002.bin",
250
  "model.layers.27.mlp.up_proj.weight": "pytorch_model-00001-of-00002.bin",
251
- "model.layers.27.post_attention_layernorm.weight": "pytorch_model-00001-of-00002.bin",
252
  "model.layers.27.self_attn.k_proj.bias": "pytorch_model-00001-of-00002.bin",
253
  "model.layers.27.self_attn.k_proj.weight": "pytorch_model-00001-of-00002.bin",
254
  "model.layers.27.self_attn.o_proj.weight": "pytorch_model-00001-of-00002.bin",
@@ -261,13 +261,13 @@
261
  "model.layers.28.mlp.gate_proj.weight": "pytorch_model-00002-of-00002.bin",
262
  "model.layers.28.mlp.up_proj.weight": "pytorch_model-00002-of-00002.bin",
263
  "model.layers.28.post_attention_layernorm.weight": "pytorch_model-00002-of-00002.bin",
264
- "model.layers.28.self_attn.k_proj.bias": "pytorch_model-00001-of-00002.bin",
265
- "model.layers.28.self_attn.k_proj.weight": "pytorch_model-00001-of-00002.bin",
266
- "model.layers.28.self_attn.o_proj.weight": "pytorch_model-00001-of-00002.bin",
267
- "model.layers.28.self_attn.q_proj.bias": "pytorch_model-00001-of-00002.bin",
268
- "model.layers.28.self_attn.q_proj.weight": "pytorch_model-00001-of-00002.bin",
269
- "model.layers.28.self_attn.v_proj.bias": "pytorch_model-00001-of-00002.bin",
270
- "model.layers.28.self_attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
271
  "model.layers.29.input_layernorm.weight": "pytorch_model-00002-of-00002.bin",
272
  "model.layers.29.mlp.down_proj.weight": "pytorch_model-00002-of-00002.bin",
273
  "model.layers.29.mlp.gate_proj.weight": "pytorch_model-00002-of-00002.bin",
 
1
  {
2
  "metadata": {
3
+ "total_size": 6262054912
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "pytorch_model-00001-of-00002.bin",
 
244
  "model.layers.26.self_attn.q_proj.weight": "pytorch_model-00001-of-00002.bin",
245
  "model.layers.26.self_attn.v_proj.bias": "pytorch_model-00001-of-00002.bin",
246
  "model.layers.26.self_attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
247
+ "model.layers.27.input_layernorm.weight": "pytorch_model-00002-of-00002.bin",
248
+ "model.layers.27.mlp.down_proj.weight": "pytorch_model-00002-of-00002.bin",
249
  "model.layers.27.mlp.gate_proj.weight": "pytorch_model-00001-of-00002.bin",
250
  "model.layers.27.mlp.up_proj.weight": "pytorch_model-00001-of-00002.bin",
251
+ "model.layers.27.post_attention_layernorm.weight": "pytorch_model-00002-of-00002.bin",
252
  "model.layers.27.self_attn.k_proj.bias": "pytorch_model-00001-of-00002.bin",
253
  "model.layers.27.self_attn.k_proj.weight": "pytorch_model-00001-of-00002.bin",
254
  "model.layers.27.self_attn.o_proj.weight": "pytorch_model-00001-of-00002.bin",
 
261
  "model.layers.28.mlp.gate_proj.weight": "pytorch_model-00002-of-00002.bin",
262
  "model.layers.28.mlp.up_proj.weight": "pytorch_model-00002-of-00002.bin",
263
  "model.layers.28.post_attention_layernorm.weight": "pytorch_model-00002-of-00002.bin",
264
+ "model.layers.28.self_attn.k_proj.bias": "pytorch_model-00002-of-00002.bin",
265
+ "model.layers.28.self_attn.k_proj.weight": "pytorch_model-00002-of-00002.bin",
266
+ "model.layers.28.self_attn.o_proj.weight": "pytorch_model-00002-of-00002.bin",
267
+ "model.layers.28.self_attn.q_proj.bias": "pytorch_model-00002-of-00002.bin",
268
+ "model.layers.28.self_attn.q_proj.weight": "pytorch_model-00002-of-00002.bin",
269
+ "model.layers.28.self_attn.v_proj.bias": "pytorch_model-00002-of-00002.bin",
270
+ "model.layers.28.self_attn.v_proj.weight": "pytorch_model-00002-of-00002.bin",
271
  "model.layers.29.input_layernorm.weight": "pytorch_model-00002-of-00002.bin",
272
  "model.layers.29.mlp.down_proj.weight": "pytorch_model-00002-of-00002.bin",
273
  "model.layers.29.mlp.gate_proj.weight": "pytorch_model-00002-of-00002.bin",