aymanbakiri commited on
Commit
16e4953
verified
1 Parent(s): 8114e3b

Model save

Browse files
README.md CHANGED
@@ -2,12 +2,12 @@
2
  library_name: peft
3
  base_model: AnnaelleMyriam/SFT_M3_model
4
  tags:
 
5
  - mcqa
6
  - question-answering
7
  - sft
8
  - lora
9
  - qwen
10
- - unsloth
11
  - generated_from_trainer
12
  model-index:
13
  - name: MNLP_M3_mcqa_model_adapter
 
2
  library_name: peft
3
  base_model: AnnaelleMyriam/SFT_M3_model
4
  tags:
5
+ - unsloth
6
  - mcqa
7
  - question-answering
8
  - sft
9
  - lora
10
  - qwen
 
11
  - generated_from_trainer
12
  model-index:
13
  - name: MNLP_M3_mcqa_model_adapter
final_model/adapter_config.json CHANGED
@@ -24,14 +24,14 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
 
 
 
27
  "gate_proj",
28
  "o_proj",
29
- "q_proj",
30
- "up_proj",
31
  "lm_head",
32
- "down_proj",
33
- "v_proj",
34
- "k_proj"
35
  ],
36
  "task_type": "CAUSAL_LM",
37
  "trainable_token_indices": null,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "up_proj",
28
+ "q_proj",
29
+ "down_proj",
30
  "gate_proj",
31
  "o_proj",
 
 
32
  "lm_head",
33
+ "k_proj",
34
+ "v_proj"
 
35
  ],
36
  "task_type": "CAUSAL_LM",
37
  "trainable_token_indices": null,
final_model/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:12d761df02e0abda0ea4847d8797711e00275f6db8b32d3d1dcfd5d18973bee0
3
  size 361377080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c0f061612ab369443d8376ec3b6480450b8942cae7c74e6d8c77dbfe843a468
3
  size 361377080