huzaifa11111 commited on
Commit
22e2302
·
verified ·
1 Parent(s): 3c3ebbd

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -4,8 +4,8 @@ library_name: transformers
4
  model_name: llama3.18B-Fine-tuned1
5
  tags:
6
  - generated_from_trainer
7
- - sft
8
  - trl
 
9
  licence: license
10
  ---
11
 
@@ -27,17 +27,17 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/huzaifanaveedd111-fast-nuces/huggingface/runs/eneico7y)
31
 
32
 
33
  This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
- - TRL: 0.19.0
38
- - Transformers: 4.53.0
39
  - Pytorch: 2.6.0+cu124
40
- - Datasets: 3.6.0
41
  - Tokenizers: 0.21.2
42
 
43
  ## Citations
 
4
  model_name: llama3.18B-Fine-tuned1
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
+ - sft
9
  licence: license
10
  ---
11
 
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/huzaifanaveedd111-fast-nuces/huggingface/runs/b4ixch8e)
31
 
32
 
33
  This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
+ - TRL: 0.19.1
38
+ - Transformers: 4.53.1
39
  - Pytorch: 2.6.0+cu124
40
+ - Datasets: 4.0.0
41
  - Tokenizers: 0.21.2
42
 
43
  ## Citations
adapter_config.json CHANGED
@@ -20,6 +20,7 @@
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
 
23
  "r": 8,
24
  "rank_pattern": {},
25
  "revision": null,
@@ -30,5 +31,6 @@
30
  "task_type": "CAUSAL_LM",
31
  "trainable_token_indices": null,
32
  "use_dora": false,
 
33
  "use_rslora": false
34
  }
 
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
  "r": 8,
25
  "rank_pattern": {},
26
  "revision": null,
 
31
  "task_type": "CAUSAL_LM",
32
  "trainable_token_indices": null,
33
  "use_dora": false,
34
+ "use_qalora": false,
35
  "use_rslora": false
36
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a4580fbbb6ffbbe17e853067ff56865c9041f9b70afb1fef4e5c32a88719683
3
  size 13648432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98f98b7a046163c7f21609c2cef5527d3f135b6b411377fdca4e2755b9bdf16f
3
  size 13648432
runs/Jul11_07-31-30_c35d55f5a272/events.out.tfevents.1752219098.c35d55f5a272.337.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c829fc76a175a1cda1cfd64a03e910e6866c238ad1d562e713d0e56cb42a22a
3
+ size 6307
runs/Jul11_07-35-56_c35d55f5a272/events.out.tfevents.1752219386.c35d55f5a272.3801.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8a5c764963cb8e8650765afa2fdf3e7c62adb8c4a3248f6b9be360f4c8e5ebe
3
+ size 6631
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8dc6c7152949a5c0fe1bf012b046359c801fe5cbdc5964a72196215663ee2899
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:954378ab4e08bb7c978ed3e9007b144251709bed68ff5ece28f2d5b53db7bbce
3
  size 5752