pmahdavi commited on
Commit
11b7e83
·
verified ·
1 Parent(s): 05065a3

Upload run root files (non-recursive) - training_config.yaml

Browse files
Files changed (1) hide show
  1. training_config.yaml +32 -0
training_config.yaml ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bf16: true
2
+ cutoff_len: 4096
3
+ dataset: tulu3_mixture_coding
4
+ ddp_timeout: 180000000
5
+ deepspeed: examples/deepspeed/ds_z3_config.json
6
+ do_train: true
7
+ eval_steps: 100
8
+ eval_strategy: steps
9
+ finetuning_type: full
10
+ gradient_accumulation_steps: 32
11
+ learning_rate: 5.0e-06
12
+ logging_steps: 10
13
+ lr_scheduler_kwargs:
14
+ cooldown_ratio: 0.4
15
+ lr_scheduler_type: warmup_stable_decay
16
+ model_name_or_path: meta-llama/Llama-3.1-8B
17
+ num_train_epochs: 1.0
18
+ output_dir: /scratch/pxm5426/runs/lora-exploration/llama-factory/Llama-3.1-8B_tulu3_mixture_coding_full_adamw_ebs128_lr5e-06_wsd-cr0.4
19
+ overwrite_cache: true
20
+ overwrite_output_dir: true
21
+ per_device_eval_batch_size: 2
22
+ per_device_train_batch_size: 2
23
+ plot_loss: true
24
+ preprocessing_num_workers: 16
25
+ report_to: wandb
26
+ save_steps: 1200
27
+ save_strategy: steps
28
+ stage: sft
29
+ template: tulu_v3
30
+ trust_remote_code: true
31
+ val_size: 0.01
32
+ warmup_ratio: 0.0