File size: 703 Bytes
5f7d284
 
 
 
d52e265
5f7d284
 
 
d52e265
5f7d284
 
 
 
d52e265
5f7d284
d52e265
5f7d284
 
 
 
 
 
 
 
d52e265
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
model_name_or_path: "openai-community/gpt2-large"
dataset_name_or_path: "allenai/tulu-3-sft-olmo-2-mixture-0225"
project_name: "scaling-post-training" 
training_args:
  seed: 42
  num_train_epochs: 1
  per_device_train_batch_size: 2
  per_device_eval_batch_size: 2
  gradient_accumulation_steps: 8
  warmup_ratio: 0.05
  weight_decay: 0.01
  logging_steps: 10
  eval_strategy: "steps"
  eval_steps: 50
  report_to: "wandb"
  fp16: true
  learning_rate: 3.0e-5
  lr_scheduler_type: "cosine"
  run_name: "gpt2-large-sft"
  output_dir: "models/gpt2-large/sft"
  save_strategy: "best"
  metric_for_best_model: "eval_loss"
  load_best_model_at_end: true
  save_total_limit: 1
  hub_model_id: "gpt2-large-sft"