jkminder commited on
Commit
926cbb1
·
verified ·
1 Parent(s): a1b2279

End of training

Browse files
.hydra/config.yaml CHANGED
@@ -32,10 +32,10 @@ finetune:
32
  - o_proj
33
  training:
34
  name: standard
35
- learning_rate: 0.0001
36
  batch_size: 4
37
  gradient_accumulation_steps: 4
38
- num_epochs: 2
39
  warmup_steps: 100
40
  weight_decay: 0.01
41
  max_grad_norm: 1.0
 
32
  - o_proj
33
  training:
34
  name: standard
35
+ learning_rate: 5.0e-05
36
  batch_size: 4
37
  gradient_accumulation_steps: 4
38
+ num_epochs: 1
39
  warmup_steps: 100
40
  weight_decay: 0.01
41
  max_grad_norm: 1.0
.hydra/hydra.yaml CHANGED
@@ -113,11 +113,10 @@ hydra:
113
  - hydra.mode=RUN
114
  task:
115
  - finetune/organism=backdoorA
116
- - finetune.training.learning_rate=0.0001
117
  job:
118
  name: main
119
  chdir: null
120
- override_dirname: finetune.training.learning_rate=0.0001,finetune/organism=backdoorA
121
  id: ???
122
  num: ???
123
  config_name: main
 
113
  - hydra.mode=RUN
114
  task:
115
  - finetune/organism=backdoorA
 
116
  job:
117
  name: main
118
  chdir: null
119
+ override_dirname: finetune/organism=backdoorA
120
  id: ???
121
  num: ???
122
  config_name: main
.hydra/overrides.yaml CHANGED
@@ -1,2 +1 @@
1
  - finetune/organism=backdoorA
2
- - finetune.training.learning_rate=0.0001
 
1
  - finetune/organism=backdoorA
 
README.md CHANGED
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/jkminder/model-organisms-finetune/runs/x285sao1)
31
 
32
 
33
  This model was trained with SFT.
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/jkminder/model-organisms-finetune/runs/oz7d1vyp)
31
 
32
 
33
  This model was trained with SFT.
adapter_config.json CHANGED
@@ -25,12 +25,12 @@
25
  "revision": null,
26
  "target_modules": [
27
  "q_proj",
 
 
28
  "down_proj",
29
  "k_proj",
30
- "up_proj",
31
- "gate_proj",
32
  "v_proj",
33
- "o_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
 
25
  "revision": null,
26
  "target_modules": [
27
  "q_proj",
28
+ "gate_proj",
29
+ "o_proj",
30
  "down_proj",
31
  "k_proj",
 
 
32
  "v_proj",
33
+ "up_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61307050df10b1d2f89cba1fb834baf3588c83986bb80508aeb186afbcc1b5d6
3
  size 26139264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d1ecd5bceb0e74ecd2e579e9d6fcb76fdb85121dfd9567dbb660014295ef563
3
  size 26139264
main.log CHANGED
@@ -18,3 +18,8 @@
18
  [2025-05-31 12:33:35,709][accelerate.utils.modeling][INFO] - We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
19
  [2025-05-31 13:22:04,694][accelerate.utils.modeling][INFO] - We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
20
  [2025-05-31 14:10:58,726][accelerate.utils.modeling][INFO] - We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
 
 
 
 
 
 
18
  [2025-05-31 12:33:35,709][accelerate.utils.modeling][INFO] - We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
19
  [2025-05-31 13:22:04,694][accelerate.utils.modeling][INFO] - We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
20
  [2025-05-31 14:10:58,726][accelerate.utils.modeling][INFO] - We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
21
+ [2025-06-01 15:46:48,163][accelerate.utils.modeling][INFO] - We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
22
+ [2025-06-01 15:49:39,975][accelerate.utils.modeling][INFO] - We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
23
+ [2025-06-01 15:56:38,618][accelerate.utils.modeling][INFO] - We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
24
+ [2025-06-01 16:13:35,827][accelerate.utils.modeling][INFO] - We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
25
+ [2025-06-01 19:21:51,407][accelerate.utils.modeling][INFO] - We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e1ead2847328a60872ab0fad2d63e794c5d72ff695b7eb79da80c9576218d5f7
3
  size 6225
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a2d47124375f37491fe575f4964f355135569dc8a6410307ac39e177eb5cc17
3
  size 6225