schwamaths commited on
Commit
40d0277
·
verified ·
1 Parent(s): c1280b2

Model save

Browse files
README.md CHANGED
@@ -1,11 +1,9 @@
1
  ---
2
  base_model: Qwen/Qwen2.5-1.5B-Instruct
3
- datasets: zwt963/FinReasoner
4
  library_name: transformers
5
  model_name: Qwen2.5-1.5B-Instruct-Open-R1-GRPO
6
  tags:
7
  - generated_from_trainer
8
- - open-r1
9
  - trl
10
  - grpo
11
  licence: license
@@ -13,7 +11,7 @@ licence: license
13
 
14
  # Model Card for Qwen2.5-1.5B-Instruct-Open-R1-GRPO
15
 
16
- This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) on the [zwt963/FinReasoner](https://huggingface.co/datasets/zwt963/FinReasoner) dataset.
17
  It has been trained using [TRL](https://github.com/huggingface/trl).
18
 
19
  ## Quick start
@@ -29,7 +27,7 @@ print(output["generated_text"])
29
 
30
  ## Training procedure
31
 
32
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/zwt963/huggingface/runs/qf6r4ry3)
33
 
34
 
35
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
 
1
  ---
2
  base_model: Qwen/Qwen2.5-1.5B-Instruct
 
3
  library_name: transformers
4
  model_name: Qwen2.5-1.5B-Instruct-Open-R1-GRPO
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
  - grpo
9
  licence: license
 
11
 
12
  # Model Card for Qwen2.5-1.5B-Instruct-Open-R1-GRPO
13
 
14
+ This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/zwt963/huggingface/runs/z82h6cyg)
31
 
32
 
33
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
all_results.json CHANGED
@@ -5,9 +5,9 @@
5
  "eval_samples_per_second": 7.75,
6
  "eval_steps_per_second": 0.278,
7
  "total_flos": 0.0,
8
- "train_loss": 0.007735961908474564,
9
- "train_runtime": 2204.9615,
10
  "train_samples": 891,
11
- "train_samples_per_second": 4.041,
12
- "train_steps_per_second": 0.036
13
  }
 
5
  "eval_samples_per_second": 7.75,
6
  "eval_steps_per_second": 0.278,
7
  "total_flos": 0.0,
8
+ "train_loss": 0.03456688816659152,
9
+ "train_runtime": 2502.5824,
10
  "train_samples": 891,
11
+ "train_samples_per_second": 3.56,
12
+ "train_steps_per_second": 0.032
13
  }
config.json CHANGED
@@ -23,7 +23,7 @@
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.49.0.dev0",
26
- "use_cache": true,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
29
  }
 
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.49.0.dev0",
26
+ "use_cache": false,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
29
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bcd36eb3e7cb0d3f525d6098036ca449aaf11a2be8091ebc57e69010cb9f99ab
3
  size 3554214752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feab802ea5480be54e59770454989982c4a4ed440afd8de60350a365f1119295
3
  size 3554214752
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 0.007735961908474564,
4
- "train_runtime": 2204.9615,
5
  "train_samples": 891,
6
- "train_samples_per_second": 4.041,
7
- "train_steps_per_second": 0.036
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 0.03456688816659152,
4
+ "train_runtime": 2502.5824,
5
  "train_samples": 891,
6
+ "train_samples_per_second": 3.56,
7
+ "train_steps_per_second": 0.032
8
  }
trainer_state.json CHANGED
@@ -9,229 +9,229 @@
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "completion_length": 91.19632097482682,
13
  "epoch": 1.25,
14
- "grad_norm": 1.1969395875930786,
15
- "kl": 0.1442023515701294,
16
  "learning_rate": 1.9961946980917457e-05,
17
- "loss": 0.0211,
18
- "reward": 1.1485491564497352,
19
- "reward_std": 0.3577419674023986,
20
- "rewards/accuracy_reward": 0.3771205507218838,
21
- "rewards/format_reward": 0.7714285938069224,
22
  "step": 10
23
  },
24
  {
25
  "epoch": 1.25,
26
- "eval_completion_length": 41.17280626296997,
27
- "eval_kl": 9.25634765625,
28
- "eval_loss": 0.3707638382911682,
29
- "eval_reward": 1.5239955931901932,
30
- "eval_reward_std": 0.11370187811553478,
31
- "eval_rewards/accuracy_reward": 0.5267857350409031,
32
- "eval_rewards/format_reward": 0.997209832072258,
33
- "eval_runtime": 17.0057,
34
- "eval_samples_per_second": 13.113,
35
- "eval_steps_per_second": 0.47,
36
  "step": 10
37
  },
38
  {
39
- "completion_length": 77.98214609622956,
40
  "epoch": 2.5,
41
- "grad_norm": 0.38055306673049927,
42
- "kl": 0.1487640380859375,
43
  "learning_rate": 1.866025403784439e-05,
44
- "loss": 0.0059,
45
- "reward": 1.5377232797443867,
46
- "reward_std": 0.112967933062464,
47
- "rewards/accuracy_reward": 0.5417410933063366,
48
- "rewards/format_reward": 0.9959821529686451,
49
  "step": 20
50
  },
51
  {
52
  "epoch": 2.5,
53
- "eval_completion_length": 74.35160541534424,
54
- "eval_kl": 79363.1533203125,
55
- "eval_loss": 3201.5810546875,
56
- "eval_reward": 1.4969618618488312,
57
- "eval_reward_std": 0.11330691166222095,
58
- "eval_rewards/accuracy_reward": 0.4997520036995411,
59
- "eval_rewards/format_reward": 0.997209832072258,
60
- "eval_runtime": 16.492,
61
- "eval_samples_per_second": 13.522,
62
- "eval_steps_per_second": 0.485,
63
  "step": 20
64
  },
65
  {
66
- "completion_length": 68.95022611618042,
67
  "epoch": 3.75,
68
- "grad_norm": 0.411288321018219,
69
- "kl": 0.2060882568359375,
70
  "learning_rate": 1.573576436351046e-05,
71
- "loss": 0.0081,
72
- "reward": 1.601116143167019,
73
- "reward_std": 0.09683309164829552,
74
- "rewards/accuracy_reward": 0.6053571676835418,
75
- "rewards/format_reward": 0.9957589406520129,
76
  "step": 30
77
  },
78
  {
79
  "epoch": 3.75,
80
- "eval_completion_length": 80.51895523071289,
81
- "eval_kl": 1.675048828125,
82
- "eval_loss": 0.06678587198257446,
83
- "eval_reward": 1.5466683954000473,
84
- "eval_reward_std": 0.08276130817830563,
85
- "eval_rewards/accuracy_reward": 0.5472263805568218,
86
- "eval_rewards/format_reward": 0.9994419664144516,
87
- "eval_runtime": 18.4204,
88
- "eval_samples_per_second": 12.106,
89
- "eval_steps_per_second": 0.434,
90
  "step": 30
91
  },
92
  {
93
- "completion_length": 86.34196748733521,
94
  "epoch": 5.0,
95
- "grad_norm": 0.30611079931259155,
96
- "kl": 0.15224609375,
97
  "learning_rate": 1.1736481776669307e-05,
98
- "loss": 0.006,
99
- "reward": 1.7011161364614964,
100
- "reward_std": 0.05814310675486922,
101
- "rewards/accuracy_reward": 0.7026785946451127,
102
- "rewards/format_reward": 0.9984375044703484,
103
  "step": 40
104
  },
105
  {
106
  "epoch": 5.0,
107
- "eval_completion_length": 95.69973850250244,
108
- "eval_kl": 1.921630859375,
109
- "eval_loss": 0.07705733180046082,
110
- "eval_reward": 1.5726687014102936,
111
- "eval_reward_std": 0.08731811679899693,
112
- "eval_rewards/accuracy_reward": 0.5726686716079712,
113
- "eval_rewards/format_reward": 1.0,
114
- "eval_runtime": 27.6097,
115
- "eval_samples_per_second": 8.077,
116
- "eval_steps_per_second": 0.29,
117
  "step": 40
118
  },
119
  {
120
- "completion_length": 88.07578473091125,
121
  "epoch": 6.25,
122
- "grad_norm": 0.186005100607872,
123
- "kl": 0.138330078125,
124
  "learning_rate": 7.411809548974792e-06,
125
- "loss": 0.0056,
126
- "reward": 1.7131697207689285,
127
- "reward_std": 0.046526999399065974,
128
- "rewards/accuracy_reward": 0.7152902040630579,
129
- "rewards/format_reward": 0.997879472374916,
130
  "step": 50
131
  },
132
  {
133
  "epoch": 6.25,
134
- "eval_completion_length": 82.68043422698975,
135
- "eval_kl": 0.50927734375,
136
- "eval_loss": 0.020424507558345795,
137
- "eval_reward": 1.5327174961566925,
138
- "eval_reward_std": 0.06632188148796558,
139
- "eval_rewards/accuracy_reward": 0.5343915484845638,
140
- "eval_rewards/format_reward": 0.9983258992433548,
141
- "eval_runtime": 22.0694,
142
- "eval_samples_per_second": 10.104,
143
- "eval_steps_per_second": 0.362,
144
  "step": 50
145
  },
146
  {
147
- "completion_length": 82.71183357238769,
148
  "epoch": 7.5,
149
- "grad_norm": 0.21738964319229126,
150
- "kl": 0.133831787109375,
151
  "learning_rate": 3.5721239031346067e-06,
152
- "loss": 0.0054,
153
- "reward": 1.7395090013742447,
154
- "reward_std": 0.03326362115330994,
155
- "rewards/accuracy_reward": 0.7408482428640127,
156
- "rewards/format_reward": 0.9986607193946838,
157
  "step": 60
158
  },
159
  {
160
  "epoch": 7.5,
161
- "eval_completion_length": 85.38940906524658,
162
- "eval_kl": 0.82177734375,
163
- "eval_loss": 0.033002614974975586,
164
- "eval_reward": 1.538297861814499,
165
- "eval_reward_std": 0.053902728483080864,
166
- "eval_rewards/accuracy_reward": 0.5399719178676605,
167
- "eval_rewards/format_reward": 0.9983258992433548,
168
- "eval_runtime": 24.6306,
169
- "eval_samples_per_second": 9.054,
170
- "eval_steps_per_second": 0.325,
171
  "step": 60
172
  },
173
  {
174
- "completion_length": 87.68672251701355,
175
  "epoch": 8.75,
176
- "grad_norm": 0.16990460455417633,
177
- "kl": 0.12266845703125,
178
  "learning_rate": 9.369221296335007e-07,
179
- "loss": 0.0049,
180
- "reward": 1.7386161454021931,
181
- "reward_std": 0.025158090936020016,
182
- "rewards/accuracy_reward": 0.739843780361116,
183
- "rewards/format_reward": 0.9987723261117936,
184
  "step": 70
185
  },
186
  {
187
  "epoch": 8.75,
188
- "eval_completion_length": 88.4438066482544,
189
- "eval_kl": 1.0899658203125,
190
- "eval_loss": 0.043835245072841644,
191
- "eval_reward": 1.5355076640844345,
192
- "eval_reward_std": 0.05012673186138272,
193
- "eval_rewards/accuracy_reward": 0.5355076231062412,
194
- "eval_rewards/format_reward": 1.0,
195
- "eval_runtime": 23.079,
196
- "eval_samples_per_second": 9.662,
197
- "eval_steps_per_second": 0.347,
198
  "step": 70
199
  },
200
  {
201
- "completion_length": 87.12489142417908,
202
  "epoch": 10.0,
203
- "grad_norm": 0.18827787041664124,
204
- "kl": 0.1228912353515625,
205
  "learning_rate": 0.0,
206
- "loss": 0.0049,
207
- "reward": 1.7418527446687222,
208
- "reward_std": 0.028121343860402704,
209
- "rewards/accuracy_reward": 0.7435268083587289,
210
- "rewards/format_reward": 0.9983258992433548,
211
  "step": 80
212
  },
213
  {
214
  "epoch": 10.0,
215
- "eval_completion_length": 89.6388931274414,
216
- "eval_kl": 0.638916015625,
217
- "eval_loss": 0.025608114898204803,
218
- "eval_reward": 1.533275529742241,
219
- "eval_reward_std": 0.07100645918399096,
220
- "eval_rewards/accuracy_reward": 0.5360656566917896,
221
- "eval_rewards/format_reward": 0.997209832072258,
222
- "eval_runtime": 26.5482,
223
- "eval_samples_per_second": 8.4,
224
- "eval_steps_per_second": 0.301,
225
  "step": 80
226
  },
227
  {
228
  "epoch": 10.0,
229
  "step": 80,
230
  "total_flos": 0.0,
231
- "train_loss": 0.007735961908474564,
232
- "train_runtime": 2204.9615,
233
- "train_samples_per_second": 4.041,
234
- "train_steps_per_second": 0.036
235
  }
236
  ],
237
  "logging_steps": 10,
 
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "completion_length": 105.54286189079285,
13
  "epoch": 1.25,
14
+ "grad_norm": 2.6259748935699463,
15
+ "kl": 0.09056792259216309,
16
  "learning_rate": 1.9961946980917457e-05,
17
+ "loss": 0.0036,
18
+ "reward": 1.1364955766126514,
19
+ "reward_std": 0.41473900550045073,
20
+ "rewards/accuracy_reward": 0.3872768019326031,
21
+ "rewards/format_reward": 0.7492187711410224,
22
  "step": 10
23
  },
24
  {
25
  "epoch": 1.25,
26
+ "eval_completion_length": 63.21034049987793,
27
+ "eval_kl": 32.33154296875,
28
+ "eval_loss": 1.2949256896972656,
29
+ "eval_reward": 1.5097140073776245,
30
+ "eval_reward_std": 0.20717254094779491,
31
+ "eval_rewards/accuracy_reward": 0.5242228880524635,
32
+ "eval_rewards/format_reward": 0.9854911118745804,
33
+ "eval_runtime": 48.0317,
34
+ "eval_samples_per_second": 4.643,
35
+ "eval_steps_per_second": 0.167,
36
  "step": 10
37
  },
38
  {
39
+ "completion_length": 69.01763689517975,
40
  "epoch": 2.5,
41
+ "grad_norm": 0.5660845637321472,
42
+ "kl": 0.1720977783203125,
43
  "learning_rate": 1.866025403784439e-05,
44
+ "loss": 0.0069,
45
+ "reward": 1.5524554237723351,
46
+ "reward_std": 0.1543258734047413,
47
+ "rewards/accuracy_reward": 0.5655134161934257,
48
+ "rewards/format_reward": 0.9869419772177934,
49
  "step": 20
50
  },
51
  {
52
  "epoch": 2.5,
53
+ "eval_completion_length": 94.81211185455322,
54
+ "eval_kl": 2333184.209838867,
55
+ "eval_loss": 93653.515625,
56
+ "eval_reward": 1.4855117946863174,
57
+ "eval_reward_std": 0.1817244552075863,
58
+ "eval_rewards/accuracy_reward": 0.49109210819005966,
59
+ "eval_rewards/format_reward": 0.994419664144516,
60
+ "eval_runtime": 22.2095,
61
+ "eval_samples_per_second": 10.041,
62
+ "eval_steps_per_second": 0.36,
63
  "step": 20
64
  },
65
  {
66
+ "completion_length": 102.01250414848327,
67
  "epoch": 3.75,
68
+ "grad_norm": 0.38660958409309387,
69
+ "kl": 0.16029052734375,
70
  "learning_rate": 1.573576436351046e-05,
71
+ "loss": 0.0064,
72
+ "reward": 1.6341518580913543,
73
+ "reward_std": 0.12228572466410696,
74
+ "rewards/accuracy_reward": 0.6387277023866773,
75
+ "rewards/format_reward": 0.9954241190105677,
76
  "step": 30
77
  },
78
  {
79
  "epoch": 3.75,
80
+ "eval_completion_length": 84.20744895935059,
81
+ "eval_kl": 700088320.0,
82
+ "eval_loss": 28113970.0,
83
+ "eval_reward": 1.5764509439468384,
84
+ "eval_reward_std": 0.08258435316383839,
85
+ "eval_rewards/accuracy_reward": 0.5803571604192257,
86
+ "eval_rewards/format_reward": 0.9960937649011612,
87
+ "eval_runtime": 19.0164,
88
+ "eval_samples_per_second": 11.727,
89
+ "eval_steps_per_second": 0.421,
90
  "step": 30
91
  },
92
  {
93
+ "completion_length": 80.08147611618043,
94
  "epoch": 5.0,
95
+ "grad_norm": 0.600352942943573,
96
+ "kl": 0.1693634033203125,
97
  "learning_rate": 1.1736481776669307e-05,
98
+ "loss": 0.0067,
99
+ "reward": 1.7243304178118706,
100
+ "reward_std": 0.051013148529455064,
101
+ "rewards/accuracy_reward": 0.726674135029316,
102
+ "rewards/format_reward": 0.9976562574505806,
103
  "step": 40
104
  },
105
  {
106
  "epoch": 5.0,
107
+ "eval_completion_length": 77.53683376312256,
108
+ "eval_kl": 341368621.0,
109
+ "eval_loss": 13754997.0,
110
+ "eval_reward": 1.5658482760190964,
111
+ "eval_reward_std": 0.06126835383474827,
112
+ "eval_rewards/accuracy_reward": 0.5691964514553547,
113
+ "eval_rewards/format_reward": 0.9966517984867096,
114
+ "eval_runtime": 24.4064,
115
+ "eval_samples_per_second": 9.137,
116
+ "eval_steps_per_second": 0.328,
117
  "step": 40
118
  },
119
  {
120
+ "completion_length": 80.83471281528473,
121
  "epoch": 6.25,
122
+ "grad_norm": 0.4441774785518646,
123
+ "kl": 0.16822509765625,
124
  "learning_rate": 7.411809548974792e-06,
125
+ "loss": 0.0068,
126
+ "reward": 1.7502232857048512,
127
+ "reward_std": 0.04045138210058212,
128
+ "rewards/accuracy_reward": 0.7536830637603998,
129
+ "rewards/format_reward": 0.9965401917696,
130
  "step": 50
131
  },
132
  {
133
  "epoch": 6.25,
134
+ "eval_completion_length": 81.76986408233643,
135
+ "eval_kl": 230830211.78515625,
136
+ "eval_loss": 9269680.0,
137
+ "eval_reward": 1.5467717200517654,
138
+ "eval_reward_std": 0.10743661969900131,
139
+ "eval_rewards/accuracy_reward": 0.5506779365241528,
140
+ "eval_rewards/format_reward": 0.9960937649011612,
141
+ "eval_runtime": 21.5108,
142
+ "eval_samples_per_second": 10.367,
143
+ "eval_steps_per_second": 0.372,
144
  "step": 50
145
  },
146
  {
147
+ "completion_length": 79.44810581207275,
148
  "epoch": 7.5,
149
+ "grad_norm": 0.18081872165203094,
150
+ "kl": 0.1697906494140625,
151
  "learning_rate": 3.5721239031346067e-06,
152
+ "loss": 0.0068,
153
+ "reward": 1.7613840006291865,
154
+ "reward_std": 0.04446553010493517,
155
+ "rewards/accuracy_reward": 0.7638393169268965,
156
+ "rewards/format_reward": 0.997544652223587,
157
  "step": 60
158
  },
159
  {
160
  "epoch": 7.5,
161
+ "eval_completion_length": 81.85964775085449,
162
+ "eval_kl": 35744152.900390625,
163
+ "eval_loss": 1435289.125,
164
+ "eval_reward": 1.5892857760190964,
165
+ "eval_reward_std": 0.0803372673690319,
166
+ "eval_rewards/accuracy_reward": 0.5931919813156128,
167
+ "eval_rewards/format_reward": 0.9960937649011612,
168
+ "eval_runtime": 45.193,
169
+ "eval_samples_per_second": 4.934,
170
+ "eval_steps_per_second": 0.177,
171
  "step": 60
172
  },
173
  {
174
+ "completion_length": 73.37768139839173,
175
  "epoch": 8.75,
176
+ "grad_norm": 1514.581298828125,
177
+ "kl": 5.606082153320313,
178
  "learning_rate": 9.369221296335007e-07,
179
+ "loss": 0.2239,
180
+ "reward": 1.7611607864499093,
181
+ "reward_std": 0.046963289054110645,
182
+ "rewards/accuracy_reward": 0.7637277090921998,
183
+ "rewards/format_reward": 0.9974330447614193,
184
  "step": 70
185
  },
186
  {
187
  "epoch": 8.75,
188
+ "eval_completion_length": 74.76281642913818,
189
+ "eval_kl": 19875971.58935547,
190
+ "eval_loss": 796283.5625,
191
+ "eval_reward": 1.6049107760190964,
192
+ "eval_reward_std": 0.08538927137851715,
193
+ "eval_rewards/accuracy_reward": 0.6077009104192257,
194
+ "eval_rewards/format_reward": 0.997209832072258,
195
+ "eval_runtime": 24.3263,
196
+ "eval_samples_per_second": 9.167,
197
+ "eval_steps_per_second": 0.329,
198
  "step": 70
199
  },
200
  {
201
+ "completion_length": 72.58817234039307,
202
  "epoch": 10.0,
203
+ "grad_norm": 0.32780376076698303,
204
+ "kl": 0.3870269775390625,
205
  "learning_rate": 0.0,
206
+ "loss": 0.0155,
207
+ "reward": 1.7863839961588384,
208
+ "reward_std": 0.03610877096652985,
209
+ "rewards/accuracy_reward": 0.7883928820490838,
210
+ "rewards/format_reward": 0.9979910783469677,
211
  "step": 80
212
  },
213
  {
214
  "epoch": 10.0,
215
+ "eval_completion_length": 74.61425590515137,
216
+ "eval_kl": 73859272.65722656,
217
+ "eval_loss": 2970804.5,
218
+ "eval_reward": 1.5872809886932373,
219
+ "eval_reward_std": 0.08278952026739717,
220
+ "eval_rewards/accuracy_reward": 0.5935846753418446,
221
+ "eval_rewards/format_reward": 0.9936962723731995,
222
+ "eval_runtime": 19.1672,
223
+ "eval_samples_per_second": 11.634,
224
+ "eval_steps_per_second": 0.417,
225
  "step": 80
226
  },
227
  {
228
  "epoch": 10.0,
229
  "step": 80,
230
  "total_flos": 0.0,
231
+ "train_loss": 0.03456688816659152,
232
+ "train_runtime": 2502.5824,
233
+ "train_samples_per_second": 3.56,
234
+ "train_steps_per_second": 0.032
235
  }
236
  ],
237
  "logging_steps": 10,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7997b82e14676fde33a3d08444edb9eaa791adab6c2582967d3283951b34decf
3
  size 7224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40335af4c1c3bf3a5c088c8646e1eb73a4804afe2a7404ba98ef56e9bef84fc6
3
  size 7224