OumaymaELBIACH commited on
Commit
4691cce
·
verified ·
1 Parent(s): 0a71003

OumaymaeELBIACH/Results_deepseek

Browse files
README.md CHANGED
@@ -37,8 +37,8 @@ This model was trained with SFT.
37
 
38
  - TRL: 0.15.2
39
  - Transformers: 4.48.3
40
- - Pytorch: 2.6.0
41
- - Datasets: 3.3.2
42
  - Tokenizers: 0.21.0
43
 
44
  ## Citations
 
37
 
38
  - TRL: 0.15.2
39
  - Transformers: 4.48.3
40
+ - Pytorch: 2.6.0+cu124
41
+ - Datasets: 3.4.0
42
  - Tokenizers: 0.21.0
43
 
44
  ## Citations
adapter_config.json CHANGED
@@ -14,22 +14,22 @@
14
  "loftq_config": {},
15
  "lora_alpha": 16,
16
  "lora_bias": false,
17
- "lora_dropout": 0,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
20
  "modules_to_save": null,
21
  "peft_type": "LORA",
22
- "r": 16,
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
  "gate_proj",
27
  "up_proj",
28
- "o_proj",
29
- "v_proj",
30
  "q_proj",
 
 
31
  "k_proj",
32
- "down_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
14
  "loftq_config": {},
15
  "lora_alpha": 16,
16
  "lora_bias": false,
17
+ "lora_dropout": 0.5,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
20
  "modules_to_save": null,
21
  "peft_type": "LORA",
22
+ "r": 32,
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
  "gate_proj",
27
  "up_proj",
 
 
28
  "q_proj",
29
+ "down_proj",
30
+ "v_proj",
31
  "k_proj",
32
+ "o_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44c39d1541016b2520f5e927237787630be4da5e2d25efb43ce116eed973f512
3
- size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c60767656cc5e970d113dde1e4a5f531ee68e9b22751da91182a3eac375573b0
3
+ size 335604696
runs/Mar15_11-28-08_ca51643881dd/events.out.tfevents.1742038163.ca51643881dd.5562.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7964a2b4e75478f5000e4fbcc920b78572a2e113521383e61a066341e162c372
3
+ size 11951
runs/Mar15_11-28-08_ca51643881dd/events.out.tfevents.1742038727.ca51643881dd.5562.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e71293f3bf5f1ed6dc659845a5b8ddf4cc9875821526c3b288cfb2dbcabe86f
3
+ size 422
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be4d8ea6c277c6d21f16057b4d29bac1a8418e6e87f417c2b88a6fe3e07b6bc4
3
  size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f67e7bfaffe5eacadd01da108aaa6d737a0347ca83024b5f74924b95d9c19def
3
  size 5624