Sunnylululu commited on
Commit
cbd622e
·
verified ·
1 Parent(s): e5f896c

Model save

Browse files
README.md CHANGED
@@ -1,10 +1,8 @@
1
  ---
2
- datasets: open-r1/OpenR1-Math-220k
3
  library_name: transformers
4
  model_name: Qwen2.5-1.5B-Open-R1-Distill
5
  tags:
6
  - generated_from_trainer
7
- - open-r1
8
  - trl
9
  - sft
10
  licence: license
@@ -12,7 +10,7 @@ licence: license
12
 
13
  # Model Card for Qwen2.5-1.5B-Open-R1-Distill
14
 
15
- This model is a fine-tuned version of [None](https://huggingface.co/None) on the [open-r1/OpenR1-Math-220k](https://huggingface.co/datasets/open-r1/OpenR1-Math-220k) dataset.
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
17
 
18
  ## Quick start
@@ -28,7 +26,7 @@ print(output["generated_text"])
28
 
29
  ## Training procedure
30
 
31
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/648848014/huggingface/runs/94h4qzbt)
32
 
33
 
34
  This model was trained with SFT.
 
1
  ---
 
2
  library_name: transformers
3
  model_name: Qwen2.5-1.5B-Open-R1-Distill
4
  tags:
5
  - generated_from_trainer
 
6
  - trl
7
  - sft
8
  licence: license
 
10
 
11
  # Model Card for Qwen2.5-1.5B-Open-R1-Distill
12
 
13
+ This model is a fine-tuned version of [None](https://huggingface.co/None).
14
  It has been trained using [TRL](https://github.com/huggingface/trl).
15
 
16
  ## Quick start
 
26
 
27
  ## Training procedure
28
 
29
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/648848014/huggingface/runs/3n5lktnf)
30
 
31
 
32
  This model was trained with SFT.
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 2.5375909456379904e+17,
3
- "train_loss": 0.8519784138752864,
4
- "train_runtime": 3243.3529,
5
  "train_samples": 3472,
6
- "train_samples_per_second": 0.607,
7
- "train_steps_per_second": 0.04
8
  }
 
1
  {
2
  "total_flos": 2.5375909456379904e+17,
3
+ "train_loss": 0.0,
4
+ "train_runtime": 2.0428,
5
  "train_samples": 3472,
6
+ "train_samples_per_second": 964.355,
7
+ "train_steps_per_second": 63.638
8
  }
config.json CHANGED
@@ -23,7 +23,7 @@
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.50.0.dev0",
26
- "use_cache": true,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
29
  }
 
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.50.0.dev0",
26
+ "use_cache": false,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
29
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 2.5375909456379904e+17,
3
- "train_loss": 0.8519784138752864,
4
- "train_runtime": 3243.3529,
5
  "train_samples": 3472,
6
- "train_samples_per_second": 0.607,
7
- "train_steps_per_second": 0.04
8
  }
 
1
  {
2
  "total_flos": 2.5375909456379904e+17,
3
+ "train_loss": 0.0,
4
+ "train_runtime": 2.0428,
5
  "train_samples": 3472,
6
+ "train_samples_per_second": 964.355,
7
+ "train_steps_per_second": 63.638
8
  }
trainer_state.json CHANGED
@@ -194,10 +194,10 @@
194
  "epoch": 10.0,
195
  "step": 130,
196
  "total_flos": 2.5375909456379904e+17,
197
- "train_loss": 0.8519784138752864,
198
- "train_runtime": 3243.3529,
199
- "train_samples_per_second": 0.607,
200
- "train_steps_per_second": 0.04
201
  }
202
  ],
203
  "logging_steps": 5,
 
194
  "epoch": 10.0,
195
  "step": 130,
196
  "total_flos": 2.5375909456379904e+17,
197
+ "train_loss": 0.0,
198
+ "train_runtime": 2.0428,
199
+ "train_samples_per_second": 964.355,
200
+ "train_steps_per_second": 63.638
201
  }
202
  ],
203
  "logging_steps": 5,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a17fee46534f3b0ec08271a60c4aa683765916504169cea34626699f7f792ee
3
  size 6264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8059c50bedd31a2ed36ddf1aa09cc4ef20113c535f90c427a63405d23bc104e2
3
  size 6264