End of training
Browse files- README.md +2 -2
- adapter_model.bin +1 -1
README.md
CHANGED
@@ -65,7 +65,7 @@ lora_model_dir: null
|
|
65 |
lora_r: 8
|
66 |
lora_target_linear: true
|
67 |
lr_scheduler: cosine
|
68 |
-
max_steps:
|
69 |
micro_batch_size: 8
|
70 |
mlflow_experiment_name: /tmp/6d1e77ffd717f97a_train_data.json
|
71 |
model_type: AutoModelForCausalLM
|
@@ -90,7 +90,7 @@ wandb_name: d37958d1-2976-4a76-8e1d-0482feba3f3c
|
|
90 |
wandb_project: Gradients-On-Demand
|
91 |
wandb_run: your_name
|
92 |
wandb_runid: d37958d1-2976-4a76-8e1d-0482feba3f3c
|
93 |
-
warmup_steps:
|
94 |
weight_decay: 0.0
|
95 |
xformers_attention: null
|
96 |
|
|
|
65 |
lora_r: 8
|
66 |
lora_target_linear: true
|
67 |
lr_scheduler: cosine
|
68 |
+
max_steps: 50
|
69 |
micro_batch_size: 8
|
70 |
mlflow_experiment_name: /tmp/6d1e77ffd717f97a_train_data.json
|
71 |
model_type: AutoModelForCausalLM
|
|
|
90 |
wandb_project: Gradients-On-Demand
|
91 |
wandb_run: your_name
|
92 |
wandb_runid: d37958d1-2976-4a76-8e1d-0482feba3f3c
|
93 |
+
warmup_steps: 2
|
94 |
weight_decay: 0.0
|
95 |
xformers_attention: null
|
96 |
|
adapter_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 37070634
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0fa4597d927edcefd8cc7f22479dc18ca9180e766b5bc91c04aa3b834a213ce0
|
3 |
size 37070634
|