Upload model trained with Unsloth
Browse filesUpload model trained with Unsloth 2x faster
- adapter_config.json +0 -3
- adapter_model.safetensors +1 -1
adapter_config.json
CHANGED
@@ -23,15 +23,12 @@
|
|
23 |
"megatron_core": "megatron.core",
|
24 |
"modules_to_save": null,
|
25 |
"peft_type": "LORA",
|
26 |
-
"qalora_group_size": 16,
|
27 |
"r": 16,
|
28 |
"rank_pattern": {},
|
29 |
"revision": null,
|
30 |
"target_modules": "(?:.*?(?:language|text).*?(?:self_attn|attention|attn|mlp|feed_forward|ffn|dense).*?(?:in_proj|out_proj|w1|w3|w2|q_proj|k_proj|v_proj).*?)|(?:\\bmodel\\.layers\\.[\\d]{1,}\\.(?:self_attn|attention|attn|mlp|feed_forward|ffn|dense)\\.(?:(?:in_proj|out_proj|w1|w3|w2|q_proj|k_proj|v_proj)))",
|
31 |
-
"target_parameters": null,
|
32 |
"task_type": null,
|
33 |
"trainable_token_indices": null,
|
34 |
"use_dora": false,
|
35 |
-
"use_qalora": false,
|
36 |
"use_rslora": false
|
37 |
}
|
|
|
23 |
"megatron_core": "megatron.core",
|
24 |
"modules_to_save": null,
|
25 |
"peft_type": "LORA",
|
|
|
26 |
"r": 16,
|
27 |
"rank_pattern": {},
|
28 |
"revision": null,
|
29 |
"target_modules": "(?:.*?(?:language|text).*?(?:self_attn|attention|attn|mlp|feed_forward|ffn|dense).*?(?:in_proj|out_proj|w1|w3|w2|q_proj|k_proj|v_proj).*?)|(?:\\bmodel\\.layers\\.[\\d]{1,}\\.(?:self_attn|attention|attn|mlp|feed_forward|ffn|dense)\\.(?:(?:in_proj|out_proj|w1|w3|w2|q_proj|k_proj|v_proj)))",
|
|
|
30 |
"task_type": null,
|
31 |
"trainable_token_indices": null,
|
32 |
"use_dora": false,
|
|
|
33 |
"use_rslora": false
|
34 |
}
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 36588320
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2ae7e10eb8a85a9a363b99e3af2bfeaece066a1586d46e597b5a5f6a9bd8ee49
|
3 |
size 36588320
|