RyanYr commited on
Commit
1095e19
·
verified ·
1 Parent(s): 00fb98e

Save model at global step 5

Browse files
config.yaml ADDED
@@ -0,0 +1,177 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data:
2
+ tokenizer: null
3
+ train_files: dapo_ds_train_sample.parquet
4
+ val_files: matheval.parquet
5
+ prompt_key: prompt
6
+ reward_fn_key: data_source
7
+ max_prompt_length: 768
8
+ max_response_length: 13312
9
+ train_batch_size: 1024
10
+ val_batch_size: 640
11
+ return_raw_input_ids: false
12
+ return_raw_chat: false
13
+ shuffle: true
14
+ filter_overlong_prompts: false
15
+ filter_overlong_prompts_workers: 1
16
+ truncation: left
17
+ image_key: images
18
+ custom_cls:
19
+ path: null
20
+ name: null
21
+ actor_rollout_ref:
22
+ hybrid_engine: true
23
+ model:
24
+ path: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
25
+ external_lib: null
26
+ override_config: {}
27
+ enable_gradient_checkpointing: true
28
+ use_remove_padding: false
29
+ use_liger: false
30
+ save_hf_repo_id: RyanYr/brm-dapo-r1qwen1.5B-base-lr2.5e-6-beta0.002
31
+ tokenizer_chat_template: null
32
+ actor:
33
+ brm:
34
+ norm_factor: value
35
+ value_constant: null
36
+ buffer:
37
+ buffer_type: null
38
+ offline_dataset_buffer:
39
+ train_files: null
40
+ response_key: response
41
+ response_truncation: right
42
+ shuffle: true
43
+ update_size: 1024
44
+ strategy: fsdp
45
+ ppo_mini_batch_size: 1024
46
+ ppo_micro_batch_size: null
47
+ ppo_micro_batch_size_per_gpu: 2
48
+ use_dynamic_bsz: false
49
+ ppo_max_token_len_per_gpu: 16384
50
+ grad_clip: 1.0
51
+ use_torch_compile: true
52
+ ppo_epochs: 1
53
+ shuffle: false
54
+ ulysses_sequence_parallel_size: 1
55
+ checkpoint:
56
+ contents:
57
+ - model
58
+ - optimizer
59
+ - extra
60
+ optim:
61
+ lr: 2.5e-06
62
+ lr_warmup_steps: -1
63
+ lr_warmup_steps_ratio: 0
64
+ min_lr_ratio: null
65
+ warmup_style: constant
66
+ total_training_steps: 200
67
+ weight_decay: 0.01
68
+ fsdp_config:
69
+ wrap_policy:
70
+ min_num_params: 0
71
+ param_offload: false
72
+ optimizer_offload: false
73
+ fsdp_size: -1
74
+ report_entropy: false
75
+ ref:
76
+ ref_model_path: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
77
+ strategy: fsdp
78
+ fsdp_config:
79
+ param_offload: false
80
+ wrap_policy:
81
+ min_num_params: 0
82
+ log_prob_micro_batch_size: null
83
+ log_prob_micro_batch_size_per_gpu: 4
84
+ log_prob_use_dynamic_bsz: false
85
+ log_prob_max_token_len_per_gpu: 16384
86
+ ulysses_sequence_parallel_size: 1
87
+ rollout:
88
+ name: vllm
89
+ temperature: 1.0
90
+ top_k: -1
91
+ top_p: 1
92
+ use_fire_sampling: false
93
+ prompt_length: 768
94
+ response_length: 13312
95
+ dtype: bfloat16
96
+ gpu_memory_utilization: 0.8
97
+ ignore_eos: false
98
+ enforce_eager: false
99
+ free_cache_engine: false
100
+ load_format: dummy_dtensor
101
+ tensor_model_parallel_size: 4
102
+ max_num_batched_tokens: 14080
103
+ max_model_len: null
104
+ max_num_seqs: 1024
105
+ log_prob_micro_batch_size: null
106
+ log_prob_micro_batch_size_per_gpu: 2
107
+ log_prob_use_dynamic_bsz: false
108
+ log_prob_max_token_len_per_gpu: 16384
109
+ disable_log_stats: true
110
+ enable_chunked_prefill: true
111
+ do_sample: true
112
+ 'n': 1
113
+ engine_kwargs:
114
+ swap_space: null
115
+ val_kwargs:
116
+ top_k: -1
117
+ top_p: 1.0
118
+ temperature: 0.6
119
+ 'n': 32
120
+ do_sample: true
121
+ reward_model:
122
+ enable: false
123
+ strategy: fsdp
124
+ model:
125
+ input_tokenizer: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
126
+ path: ~/models/FsfairX-LLaMA3-RM-v0.1
127
+ external_lib: null
128
+ use_remove_padding: false
129
+ fsdp_config:
130
+ wrap_policy:
131
+ min_num_params: 0
132
+ param_offload: false
133
+ fsdp_size: -1
134
+ micro_batch_size: null
135
+ micro_batch_size_per_gpu: null
136
+ max_length: null
137
+ ulysses_sequence_parallel_size: 1
138
+ use_dynamic_bsz: false
139
+ forward_max_token_len_per_gpu: 16384
140
+ reward_manager: prime
141
+ reward_kwargs:
142
+ format_reward: 0.0
143
+ format_type: null
144
+ custom_reward_function:
145
+ path: null
146
+ name: compute_score
147
+ trainer:
148
+ balance_batch: true
149
+ total_epochs: 100
150
+ total_training_steps: 200
151
+ project_name: value-LLM
152
+ experiment_name: brm-dapo-r1qwen1.5B-base_lr2.5e-6-beta0.002
153
+ logger:
154
+ - console
155
+ - wandb
156
+ log_val_generations: 0
157
+ nnodes: 1
158
+ n_gpus_per_node: 4
159
+ save_freq: 5
160
+ resume_mode: auto
161
+ resume_from_path: null
162
+ val_before_train: false
163
+ test_freq: -1
164
+ default_hdfs_dir: null
165
+ del_local_ckpt_after_load: false
166
+ default_local_dir: ./BRM
167
+ max_actor_ckpt_to_keep: 1
168
+ ray_wait_register_center_timeout: 300
169
+ hf_token: null
170
+ resume_from_hf:
171
+ enable: false
172
+ hf_repo_id: null
173
+ hf_token: null
174
+ revision: main
175
+ algorithm:
176
+ kl_ctrl:
177
+ kl_coef: 0.002
data.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12408ef6110042eb1cb7500f4879b9e09b1909a2b1fac49f8db18c0a48d137c3
3
+ size 1492
extra_state_world_size_4_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:affb70bc0793fe37cc07d59e93ebc59f6e2c36f455abb03b59c91cfde80781d0
3
+ size 14632
extra_state_world_size_4_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42e5e254daebb5ef8160a836b3f36d0c78e3a4379cc09452cd5138dbec5852cc
3
+ size 14632
extra_state_world_size_4_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:850f52845d8a9470deba24d84081c5192299d77fb4481de8f6f1575d65391daa
3
+ size 14632
extra_state_world_size_4_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d4fe13b8a4ebbe2935ec9356d1b415a673ac30c66b1534ca5e9611fa7c3db3d
3
+ size 14632
model_world_size_4_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5eac1a698e037e6c102c9f1edae64c6ee469561411f752d68bd40d9c1c63c32
3
+ size 1777276538
model_world_size_4_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9559f3288e7b4ea28bff63a4252429f42eb16afaabb3505c129d6e957868a9c
3
+ size 1777276538
model_world_size_4_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d216c7e2ead64713419e32d630f3274433673aa563bbfd94b2f48fe5f9f7d6ea
3
+ size 1777276538
model_world_size_4_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a202bad5d16d43bba2713aecfbe59b9f6012b88d1c7b72bcf82b2e108491f8b
3
+ size 1777276538
optim_world_size_4_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96a63649144a2b754f65678beddcd776f8689b42a23a57e2e3774e2edac5f937
3
+ size 3554202167
optim_world_size_4_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0824ba46aab1e958ea43274a14a82f995f080f66bf002a18931bc3f7086822f4
3
+ size 3554202167
optim_world_size_4_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f462195f2bf19864adc3b78f8c1f0a463365399896825585a464731c7684cf9b
3
+ size 3554202167
optim_world_size_4_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52b2b82802d55189049bb2ee3b8577dd94ec6aa1b3e486b44a7daefcbc8361d3
3
+ size 3554202167