Woondsc commited on
Commit
8b3ce91
·
verified ·
1 Parent(s): eaa80db

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +14 -0
  2. checkpoint-11826/config.json +56 -0
  3. checkpoint-11826/generation_config.json +16 -0
  4. checkpoint-11826/model.safetensors +3 -0
  5. checkpoint-11826/optimizer.pt +3 -0
  6. checkpoint-11826/rng_state.pth +3 -0
  7. checkpoint-11826/scheduler.pt +3 -0
  8. checkpoint-11826/source.spm +3 -0
  9. checkpoint-11826/special_tokens_map.json +23 -0
  10. checkpoint-11826/target.spm +3 -0
  11. checkpoint-11826/tokenizer_config.json +39 -0
  12. checkpoint-11826/trainer_state.json +284 -0
  13. checkpoint-11826/training_args.bin +3 -0
  14. checkpoint-11826/vocab.json +0 -0
  15. checkpoint-1971/config.json +56 -0
  16. checkpoint-1971/generation_config.json +16 -0
  17. checkpoint-1971/model.safetensors +3 -0
  18. checkpoint-1971/optimizer.pt +3 -0
  19. checkpoint-1971/rng_state.pth +3 -0
  20. checkpoint-1971/scheduler.pt +3 -0
  21. checkpoint-1971/source.spm +3 -0
  22. checkpoint-1971/special_tokens_map.json +23 -0
  23. checkpoint-1971/target.spm +3 -0
  24. checkpoint-1971/tokenizer_config.json +39 -0
  25. checkpoint-1971/trainer_state.json +69 -0
  26. checkpoint-1971/training_args.bin +3 -0
  27. checkpoint-1971/vocab.json +0 -0
  28. checkpoint-3942/config.json +56 -0
  29. checkpoint-3942/generation_config.json +16 -0
  30. checkpoint-3942/model.safetensors +3 -0
  31. checkpoint-3942/optimizer.pt +3 -0
  32. checkpoint-3942/rng_state.pth +3 -0
  33. checkpoint-3942/scheduler.pt +3 -0
  34. checkpoint-3942/source.spm +3 -0
  35. checkpoint-3942/special_tokens_map.json +23 -0
  36. checkpoint-3942/target.spm +3 -0
  37. checkpoint-3942/tokenizer_config.json +39 -0
  38. checkpoint-3942/trainer_state.json +112 -0
  39. checkpoint-3942/training_args.bin +3 -0
  40. checkpoint-3942/vocab.json +0 -0
  41. checkpoint-5913/config.json +56 -0
  42. checkpoint-5913/generation_config.json +16 -0
  43. checkpoint-5913/model.safetensors +3 -0
  44. checkpoint-5913/optimizer.pt +3 -0
  45. checkpoint-5913/rng_state.pth +3 -0
  46. checkpoint-5913/scheduler.pt +3 -0
  47. checkpoint-5913/source.spm +3 -0
  48. checkpoint-5913/special_tokens_map.json +23 -0
  49. checkpoint-5913/target.spm +3 -0
  50. checkpoint-5913/tokenizer_config.json +39 -0
.gitattributes CHANGED
@@ -33,3 +33,17 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ checkpoint-11826/source.spm filter=lfs diff=lfs merge=lfs -text
37
+ checkpoint-11826/target.spm filter=lfs diff=lfs merge=lfs -text
38
+ checkpoint-1971/source.spm filter=lfs diff=lfs merge=lfs -text
39
+ checkpoint-1971/target.spm filter=lfs diff=lfs merge=lfs -text
40
+ checkpoint-3942/source.spm filter=lfs diff=lfs merge=lfs -text
41
+ checkpoint-3942/target.spm filter=lfs diff=lfs merge=lfs -text
42
+ checkpoint-5913/source.spm filter=lfs diff=lfs merge=lfs -text
43
+ checkpoint-5913/target.spm filter=lfs diff=lfs merge=lfs -text
44
+ checkpoint-7884/source.spm filter=lfs diff=lfs merge=lfs -text
45
+ checkpoint-7884/target.spm filter=lfs diff=lfs merge=lfs -text
46
+ checkpoint-9855/source.spm filter=lfs diff=lfs merge=lfs -text
47
+ checkpoint-9855/target.spm filter=lfs diff=lfs merge=lfs -text
48
+ source.spm filter=lfs diff=lfs merge=lfs -text
49
+ target.spm filter=lfs diff=lfs merge=lfs -text
checkpoint-11826/config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Woondsc/opus-mt-ko-en-medterm",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "swish",
5
+ "add_bias_logits": false,
6
+ "add_final_layer_norm": false,
7
+ "architectures": [
8
+ "MarianMTModel"
9
+ ],
10
+ "attention_dropout": 0.0,
11
+ "bos_token_id": 0,
12
+ "classif_dropout": 0.0,
13
+ "classifier_dropout": 0.0,
14
+ "d_model": 512,
15
+ "decoder_attention_heads": 8,
16
+ "decoder_ffn_dim": 2048,
17
+ "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 6,
19
+ "decoder_start_token_id": 65000,
20
+ "decoder_vocab_size": 65001,
21
+ "dropout": 0.1,
22
+ "encoder_attention_heads": 8,
23
+ "encoder_ffn_dim": 2048,
24
+ "encoder_layerdrop": 0.0,
25
+ "encoder_layers": 6,
26
+ "eos_token_id": 0,
27
+ "extra_pos_embeddings": 65001,
28
+ "forced_eos_token_id": 0,
29
+ "id2label": {
30
+ "0": "LABEL_0",
31
+ "1": "LABEL_1",
32
+ "2": "LABEL_2"
33
+ },
34
+ "init_std": 0.02,
35
+ "is_encoder_decoder": true,
36
+ "label2id": {
37
+ "LABEL_0": 0,
38
+ "LABEL_1": 1,
39
+ "LABEL_2": 2
40
+ },
41
+ "max_length": null,
42
+ "max_position_embeddings": 512,
43
+ "model_type": "marian",
44
+ "normalize_before": false,
45
+ "normalize_embedding": false,
46
+ "num_beams": null,
47
+ "num_hidden_layers": 6,
48
+ "pad_token_id": 65000,
49
+ "scale_embedding": true,
50
+ "share_encoder_decoder_embeddings": true,
51
+ "static_position_embeddings": true,
52
+ "torch_dtype": "float32",
53
+ "transformers_version": "4.48.3",
54
+ "use_cache": true,
55
+ "vocab_size": 65001
56
+ }
checkpoint-11826/generation_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bad_words_ids": [
3
+ [
4
+ 65000
5
+ ]
6
+ ],
7
+ "bos_token_id": 0,
8
+ "decoder_start_token_id": 65000,
9
+ "eos_token_id": 0,
10
+ "forced_eos_token_id": 0,
11
+ "max_length": 512,
12
+ "num_beams": 6,
13
+ "pad_token_id": 65000,
14
+ "renormalize_logits": true,
15
+ "transformers_version": "4.48.3"
16
+ }
checkpoint-11826/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fff6e054402a7e94855cc3b3f2207329373797c55de20bfd0f9c15846558425
3
+ size 309965092
checkpoint-11826/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08d7f5a25e2ec74b12d42479efd0ec1bef3abc7dcec7ec8120160239e75db4ac
3
+ size 619563642
checkpoint-11826/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d34f69b47d6327fa2611573784d2b6e3b933e3338ecc5b3e464fb9350d817c6d
3
+ size 14244
checkpoint-11826/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:600c499244df8db233478d843e4a62f93a9b482036853e3f85972f8d94e6c73d
3
+ size 1064
checkpoint-11826/source.spm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9496f7c2be9aecb84c751ae9f35a875915dde8e3892f652a5c76811ab2a0f49
3
+ size 841805
checkpoint-11826/special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eos_token": {
3
+ "content": "</s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "pad_token": {
10
+ "content": "<pad>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
checkpoint-11826/target.spm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1531ac8576fe64267dfca0dc33950a6e9a2d3fd9e05346558ad7ea5ee0e65bf
3
+ size 813126
checkpoint-11826/tokenizer_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "</s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<unk>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "65000": {
20
+ "content": "<pad>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "clean_up_tokenization_spaces": false,
29
+ "eos_token": "</s>",
30
+ "extra_special_tokens": {},
31
+ "model_max_length": 512,
32
+ "pad_token": "<pad>",
33
+ "separate_vocabs": false,
34
+ "source_lang": "kor",
35
+ "sp_model_kwargs": {},
36
+ "target_lang": "eng",
37
+ "tokenizer_class": "MarianTokenizer",
38
+ "unk_token": "<unk>"
39
+ }
checkpoint-11826/trainer_state.json ADDED
@@ -0,0 +1,284 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 6.0,
5
+ "eval_steps": 500,
6
+ "global_step": 11826,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.2536783358701167,
13
+ "grad_norm": 6.285004615783691,
14
+ "learning_rate": 1.9157787924911212e-05,
15
+ "loss": 1.589,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.5073566717402334,
20
+ "grad_norm": 6.245589256286621,
21
+ "learning_rate": 1.8312193472010826e-05,
22
+ "loss": 0.8971,
23
+ "step": 1000
24
+ },
25
+ {
26
+ "epoch": 0.76103500761035,
27
+ "grad_norm": 7.839916706085205,
28
+ "learning_rate": 1.7466599019110436e-05,
29
+ "loss": 0.7412,
30
+ "step": 1500
31
+ },
32
+ {
33
+ "epoch": 1.0,
34
+ "eval_bleu": 77.9302498710783,
35
+ "eval_loss": 0.5189284682273865,
36
+ "eval_rouge": {
37
+ "rouge1": 0.8422585807632581,
38
+ "rouge2": 0.7462100224193878,
39
+ "rougeL": 0.8347423871515361,
40
+ "rougeLsum": 0.8348341110331449
41
+ },
42
+ "eval_runtime": 589.3644,
43
+ "eval_samples_per_second": 14.402,
44
+ "eval_steps_per_second": 0.901,
45
+ "step": 1971
46
+ },
47
+ {
48
+ "epoch": 1.0147133434804667,
49
+ "grad_norm": 4.194991111755371,
50
+ "learning_rate": 1.6621004566210046e-05,
51
+ "loss": 0.6173,
52
+ "step": 2000
53
+ },
54
+ {
55
+ "epoch": 1.2683916793505834,
56
+ "grad_norm": 4.830150127410889,
57
+ "learning_rate": 1.5775410113309656e-05,
58
+ "loss": 0.4964,
59
+ "step": 2500
60
+ },
61
+ {
62
+ "epoch": 1.5220700152207,
63
+ "grad_norm": 6.662743091583252,
64
+ "learning_rate": 1.4929815660409268e-05,
65
+ "loss": 0.4655,
66
+ "step": 3000
67
+ },
68
+ {
69
+ "epoch": 1.775748351090817,
70
+ "grad_norm": 5.814598560333252,
71
+ "learning_rate": 1.408422120750888e-05,
72
+ "loss": 0.4562,
73
+ "step": 3500
74
+ },
75
+ {
76
+ "epoch": 2.0,
77
+ "eval_bleu": 82.4671860062862,
78
+ "eval_loss": 0.4167616665363312,
79
+ "eval_rouge": {
80
+ "rouge1": 0.8726285922102037,
81
+ "rouge2": 0.7929049594785995,
82
+ "rougeL": 0.8662329420974849,
83
+ "rougeLsum": 0.8662523269791588
84
+ },
85
+ "eval_runtime": 558.5398,
86
+ "eval_samples_per_second": 15.197,
87
+ "eval_steps_per_second": 0.951,
88
+ "step": 3942
89
+ },
90
+ {
91
+ "epoch": 2.0294266869609334,
92
+ "grad_norm": 4.13426399230957,
93
+ "learning_rate": 1.323862675460849e-05,
94
+ "loss": 0.4122,
95
+ "step": 4000
96
+ },
97
+ {
98
+ "epoch": 2.2831050228310503,
99
+ "grad_norm": 4.324504375457764,
100
+ "learning_rate": 1.2393032301708102e-05,
101
+ "loss": 0.3451,
102
+ "step": 4500
103
+ },
104
+ {
105
+ "epoch": 2.536783358701167,
106
+ "grad_norm": 1.6161426305770874,
107
+ "learning_rate": 1.1549129037713514e-05,
108
+ "loss": 0.3359,
109
+ "step": 5000
110
+ },
111
+ {
112
+ "epoch": 2.7904616945712837,
113
+ "grad_norm": 3.2323482036590576,
114
+ "learning_rate": 1.0703534584813124e-05,
115
+ "loss": 0.3413,
116
+ "step": 5500
117
+ },
118
+ {
119
+ "epoch": 3.0,
120
+ "eval_bleu": 84.42968434246376,
121
+ "eval_loss": 0.3801760971546173,
122
+ "eval_rouge": {
123
+ "rouge1": 0.8853534128903551,
124
+ "rouge2": 0.8130151085172689,
125
+ "rougeL": 0.8794009703754694,
126
+ "rougeLsum": 0.8793662457667164
127
+ },
128
+ "eval_runtime": 553.0346,
129
+ "eval_samples_per_second": 15.348,
130
+ "eval_steps_per_second": 0.96,
131
+ "step": 5913
132
+ },
133
+ {
134
+ "epoch": 3.0441400304414,
135
+ "grad_norm": 3.7964985370635986,
136
+ "learning_rate": 9.857940131912736e-06,
137
+ "loss": 0.3074,
138
+ "step": 6000
139
+ },
140
+ {
141
+ "epoch": 3.297818366311517,
142
+ "grad_norm": 3.6037538051605225,
143
+ "learning_rate": 9.012345679012346e-06,
144
+ "loss": 0.269,
145
+ "step": 6500
146
+ },
147
+ {
148
+ "epoch": 3.551496702181634,
149
+ "grad_norm": 3.743192672729492,
150
+ "learning_rate": 8.166751226111958e-06,
151
+ "loss": 0.2655,
152
+ "step": 7000
153
+ },
154
+ {
155
+ "epoch": 3.8051750380517504,
156
+ "grad_norm": 1.6885374784469604,
157
+ "learning_rate": 7.321156773211569e-06,
158
+ "loss": 0.275,
159
+ "step": 7500
160
+ },
161
+ {
162
+ "epoch": 4.0,
163
+ "eval_bleu": 85.34047750600152,
164
+ "eval_loss": 0.3641541302204132,
165
+ "eval_rouge": {
166
+ "rouge1": 0.8913775651214185,
167
+ "rouge2": 0.8220615430193459,
168
+ "rougeL": 0.8858292704586206,
169
+ "rougeLsum": 0.8858430238896269
170
+ },
171
+ "eval_runtime": 560.5886,
172
+ "eval_samples_per_second": 15.141,
173
+ "eval_steps_per_second": 0.947,
174
+ "step": 7884
175
+ },
176
+ {
177
+ "epoch": 4.058853373921867,
178
+ "grad_norm": 2.691307783126831,
179
+ "learning_rate": 6.4755623203111796e-06,
180
+ "loss": 0.2498,
181
+ "step": 8000
182
+ },
183
+ {
184
+ "epoch": 4.312531709791983,
185
+ "grad_norm": 4.331101417541504,
186
+ "learning_rate": 5.6299678674107905e-06,
187
+ "loss": 0.2301,
188
+ "step": 8500
189
+ },
190
+ {
191
+ "epoch": 4.566210045662101,
192
+ "grad_norm": 3.2862532138824463,
193
+ "learning_rate": 4.786064603416202e-06,
194
+ "loss": 0.2386,
195
+ "step": 9000
196
+ },
197
+ {
198
+ "epoch": 4.819888381532217,
199
+ "grad_norm": 3.7428629398345947,
200
+ "learning_rate": 3.940470150515813e-06,
201
+ "loss": 0.2326,
202
+ "step": 9500
203
+ },
204
+ {
205
+ "epoch": 5.0,
206
+ "eval_bleu": 85.62474872165251,
207
+ "eval_loss": 0.35635867714881897,
208
+ "eval_rouge": {
209
+ "rouge1": 0.8957946109809499,
210
+ "rouge2": 0.8288231670050729,
211
+ "rougeL": 0.8903628109630211,
212
+ "rougeLsum": 0.8904334470778201
213
+ },
214
+ "eval_runtime": 554.4082,
215
+ "eval_samples_per_second": 15.31,
216
+ "eval_steps_per_second": 0.958,
217
+ "step": 9855
218
+ },
219
+ {
220
+ "epoch": 5.073566717402334,
221
+ "grad_norm": 2.8402657508850098,
222
+ "learning_rate": 3.0948756976154237e-06,
223
+ "loss": 0.22,
224
+ "step": 10000
225
+ },
226
+ {
227
+ "epoch": 5.327245053272451,
228
+ "grad_norm": 3.1032028198242188,
229
+ "learning_rate": 2.2492812447150347e-06,
230
+ "loss": 0.2083,
231
+ "step": 10500
232
+ },
233
+ {
234
+ "epoch": 5.580923389142567,
235
+ "grad_norm": 3.455146312713623,
236
+ "learning_rate": 1.4053779807204465e-06,
237
+ "loss": 0.2027,
238
+ "step": 11000
239
+ },
240
+ {
241
+ "epoch": 5.834601725012684,
242
+ "grad_norm": 2.128119468688965,
243
+ "learning_rate": 5.597835278200575e-07,
244
+ "loss": 0.2143,
245
+ "step": 11500
246
+ },
247
+ {
248
+ "epoch": 6.0,
249
+ "eval_bleu": 85.9174327850833,
250
+ "eval_loss": 0.3543717563152313,
251
+ "eval_rouge": {
252
+ "rouge1": 0.8969832695452472,
253
+ "rouge2": 0.8310275704571024,
254
+ "rougeL": 0.8913461546133479,
255
+ "rougeLsum": 0.8913688423165305
256
+ },
257
+ "eval_runtime": 553.9415,
258
+ "eval_samples_per_second": 15.323,
259
+ "eval_steps_per_second": 0.959,
260
+ "step": 11826
261
+ }
262
+ ],
263
+ "logging_steps": 500,
264
+ "max_steps": 11826,
265
+ "num_input_tokens_seen": 0,
266
+ "num_train_epochs": 6,
267
+ "save_steps": 500,
268
+ "stateful_callbacks": {
269
+ "TrainerControl": {
270
+ "args": {
271
+ "should_epoch_stop": false,
272
+ "should_evaluate": false,
273
+ "should_log": false,
274
+ "should_save": true,
275
+ "should_training_stop": true
276
+ },
277
+ "attributes": {}
278
+ }
279
+ },
280
+ "total_flos": 1596266324361216.0,
281
+ "train_batch_size": 16,
282
+ "trial_name": null,
283
+ "trial_params": null
284
+ }
checkpoint-11826/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d50538077dc9316af4f0e3a52d83f66a602e8f111533468407cd32fad55c01d
3
+ size 5496
checkpoint-11826/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1971/config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Woondsc/opus-mt-ko-en-medterm",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "swish",
5
+ "add_bias_logits": false,
6
+ "add_final_layer_norm": false,
7
+ "architectures": [
8
+ "MarianMTModel"
9
+ ],
10
+ "attention_dropout": 0.0,
11
+ "bos_token_id": 0,
12
+ "classif_dropout": 0.0,
13
+ "classifier_dropout": 0.0,
14
+ "d_model": 512,
15
+ "decoder_attention_heads": 8,
16
+ "decoder_ffn_dim": 2048,
17
+ "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 6,
19
+ "decoder_start_token_id": 65000,
20
+ "decoder_vocab_size": 65001,
21
+ "dropout": 0.1,
22
+ "encoder_attention_heads": 8,
23
+ "encoder_ffn_dim": 2048,
24
+ "encoder_layerdrop": 0.0,
25
+ "encoder_layers": 6,
26
+ "eos_token_id": 0,
27
+ "extra_pos_embeddings": 65001,
28
+ "forced_eos_token_id": 0,
29
+ "id2label": {
30
+ "0": "LABEL_0",
31
+ "1": "LABEL_1",
32
+ "2": "LABEL_2"
33
+ },
34
+ "init_std": 0.02,
35
+ "is_encoder_decoder": true,
36
+ "label2id": {
37
+ "LABEL_0": 0,
38
+ "LABEL_1": 1,
39
+ "LABEL_2": 2
40
+ },
41
+ "max_length": null,
42
+ "max_position_embeddings": 512,
43
+ "model_type": "marian",
44
+ "normalize_before": false,
45
+ "normalize_embedding": false,
46
+ "num_beams": null,
47
+ "num_hidden_layers": 6,
48
+ "pad_token_id": 65000,
49
+ "scale_embedding": true,
50
+ "share_encoder_decoder_embeddings": true,
51
+ "static_position_embeddings": true,
52
+ "torch_dtype": "float32",
53
+ "transformers_version": "4.48.3",
54
+ "use_cache": true,
55
+ "vocab_size": 65001
56
+ }
checkpoint-1971/generation_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bad_words_ids": [
3
+ [
4
+ 65000
5
+ ]
6
+ ],
7
+ "bos_token_id": 0,
8
+ "decoder_start_token_id": 65000,
9
+ "eos_token_id": 0,
10
+ "forced_eos_token_id": 0,
11
+ "max_length": 512,
12
+ "num_beams": 6,
13
+ "pad_token_id": 65000,
14
+ "renormalize_logits": true,
15
+ "transformers_version": "4.48.3"
16
+ }
checkpoint-1971/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:457f802cad9b26588ccbb265d83407431be6d902f1e1991c4ab26391714c43fa
3
+ size 309965092
checkpoint-1971/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b7286a84d2437f1e1c8b390ad47213ebec48ca128e00c611190138025dc744d
3
+ size 619563642
checkpoint-1971/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9db98f0ce5cbfb036b11f8b357434db0be764cee68f5927c4b00e8f348b29786
3
+ size 14244
checkpoint-1971/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be1e0f3c1bf5bfb20687bc226f7646f4a11677f08c1d29cf97f870f3f8ba0f29
3
+ size 1064
checkpoint-1971/source.spm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9496f7c2be9aecb84c751ae9f35a875915dde8e3892f652a5c76811ab2a0f49
3
+ size 841805
checkpoint-1971/special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eos_token": {
3
+ "content": "</s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "pad_token": {
10
+ "content": "<pad>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
checkpoint-1971/target.spm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1531ac8576fe64267dfca0dc33950a6e9a2d3fd9e05346558ad7ea5ee0e65bf
3
+ size 813126
checkpoint-1971/tokenizer_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "</s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<unk>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "65000": {
20
+ "content": "<pad>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "clean_up_tokenization_spaces": false,
29
+ "eos_token": "</s>",
30
+ "extra_special_tokens": {},
31
+ "model_max_length": 512,
32
+ "pad_token": "<pad>",
33
+ "separate_vocabs": false,
34
+ "source_lang": "kor",
35
+ "sp_model_kwargs": {},
36
+ "target_lang": "eng",
37
+ "tokenizer_class": "MarianTokenizer",
38
+ "unk_token": "<unk>"
39
+ }
checkpoint-1971/trainer_state.json ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 1971,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.2536783358701167,
13
+ "grad_norm": 6.285004615783691,
14
+ "learning_rate": 1.9157787924911212e-05,
15
+ "loss": 1.589,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.5073566717402334,
20
+ "grad_norm": 6.245589256286621,
21
+ "learning_rate": 1.8312193472010826e-05,
22
+ "loss": 0.8971,
23
+ "step": 1000
24
+ },
25
+ {
26
+ "epoch": 0.76103500761035,
27
+ "grad_norm": 7.839916706085205,
28
+ "learning_rate": 1.7466599019110436e-05,
29
+ "loss": 0.7412,
30
+ "step": 1500
31
+ },
32
+ {
33
+ "epoch": 1.0,
34
+ "eval_bleu": 77.9302498710783,
35
+ "eval_loss": 0.5189284682273865,
36
+ "eval_rouge": {
37
+ "rouge1": 0.8422585807632581,
38
+ "rouge2": 0.7462100224193878,
39
+ "rougeL": 0.8347423871515361,
40
+ "rougeLsum": 0.8348341110331449
41
+ },
42
+ "eval_runtime": 589.3644,
43
+ "eval_samples_per_second": 14.402,
44
+ "eval_steps_per_second": 0.901,
45
+ "step": 1971
46
+ }
47
+ ],
48
+ "logging_steps": 500,
49
+ "max_steps": 11826,
50
+ "num_input_tokens_seen": 0,
51
+ "num_train_epochs": 6,
52
+ "save_steps": 500,
53
+ "stateful_callbacks": {
54
+ "TrainerControl": {
55
+ "args": {
56
+ "should_epoch_stop": false,
57
+ "should_evaluate": false,
58
+ "should_log": false,
59
+ "should_save": true,
60
+ "should_training_stop": false
61
+ },
62
+ "attributes": {}
63
+ }
64
+ },
65
+ "total_flos": 266574887976960.0,
66
+ "train_batch_size": 16,
67
+ "trial_name": null,
68
+ "trial_params": null
69
+ }
checkpoint-1971/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d50538077dc9316af4f0e3a52d83f66a602e8f111533468407cd32fad55c01d
3
+ size 5496
checkpoint-1971/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-3942/config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Woondsc/opus-mt-ko-en-medterm",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "swish",
5
+ "add_bias_logits": false,
6
+ "add_final_layer_norm": false,
7
+ "architectures": [
8
+ "MarianMTModel"
9
+ ],
10
+ "attention_dropout": 0.0,
11
+ "bos_token_id": 0,
12
+ "classif_dropout": 0.0,
13
+ "classifier_dropout": 0.0,
14
+ "d_model": 512,
15
+ "decoder_attention_heads": 8,
16
+ "decoder_ffn_dim": 2048,
17
+ "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 6,
19
+ "decoder_start_token_id": 65000,
20
+ "decoder_vocab_size": 65001,
21
+ "dropout": 0.1,
22
+ "encoder_attention_heads": 8,
23
+ "encoder_ffn_dim": 2048,
24
+ "encoder_layerdrop": 0.0,
25
+ "encoder_layers": 6,
26
+ "eos_token_id": 0,
27
+ "extra_pos_embeddings": 65001,
28
+ "forced_eos_token_id": 0,
29
+ "id2label": {
30
+ "0": "LABEL_0",
31
+ "1": "LABEL_1",
32
+ "2": "LABEL_2"
33
+ },
34
+ "init_std": 0.02,
35
+ "is_encoder_decoder": true,
36
+ "label2id": {
37
+ "LABEL_0": 0,
38
+ "LABEL_1": 1,
39
+ "LABEL_2": 2
40
+ },
41
+ "max_length": null,
42
+ "max_position_embeddings": 512,
43
+ "model_type": "marian",
44
+ "normalize_before": false,
45
+ "normalize_embedding": false,
46
+ "num_beams": null,
47
+ "num_hidden_layers": 6,
48
+ "pad_token_id": 65000,
49
+ "scale_embedding": true,
50
+ "share_encoder_decoder_embeddings": true,
51
+ "static_position_embeddings": true,
52
+ "torch_dtype": "float32",
53
+ "transformers_version": "4.48.3",
54
+ "use_cache": true,
55
+ "vocab_size": 65001
56
+ }
checkpoint-3942/generation_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bad_words_ids": [
3
+ [
4
+ 65000
5
+ ]
6
+ ],
7
+ "bos_token_id": 0,
8
+ "decoder_start_token_id": 65000,
9
+ "eos_token_id": 0,
10
+ "forced_eos_token_id": 0,
11
+ "max_length": 512,
12
+ "num_beams": 6,
13
+ "pad_token_id": 65000,
14
+ "renormalize_logits": true,
15
+ "transformers_version": "4.48.3"
16
+ }
checkpoint-3942/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:859e460154977ae936caadc7d20d31e8a8d895f4c30a384e5488a1a0ed451bba
3
+ size 309965092
checkpoint-3942/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8537487bb1d383bdfa0e668b5dc53cb58ba6ac936c76d6c1409f1104dda4d501
3
+ size 619563642
checkpoint-3942/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dca42cded49feccc60369b40fc5385380720f862f43024ffe0fd8fe3a7c92d4b
3
+ size 14244
checkpoint-3942/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f0e64f689759835954ae6c93dc848789ea0a051a857e84fab50db368d9ad311
3
+ size 1064
checkpoint-3942/source.spm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9496f7c2be9aecb84c751ae9f35a875915dde8e3892f652a5c76811ab2a0f49
3
+ size 841805
checkpoint-3942/special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eos_token": {
3
+ "content": "</s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "pad_token": {
10
+ "content": "<pad>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
checkpoint-3942/target.spm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1531ac8576fe64267dfca0dc33950a6e9a2d3fd9e05346558ad7ea5ee0e65bf
3
+ size 813126
checkpoint-3942/tokenizer_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "</s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<unk>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "65000": {
20
+ "content": "<pad>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "clean_up_tokenization_spaces": false,
29
+ "eos_token": "</s>",
30
+ "extra_special_tokens": {},
31
+ "model_max_length": 512,
32
+ "pad_token": "<pad>",
33
+ "separate_vocabs": false,
34
+ "source_lang": "kor",
35
+ "sp_model_kwargs": {},
36
+ "target_lang": "eng",
37
+ "tokenizer_class": "MarianTokenizer",
38
+ "unk_token": "<unk>"
39
+ }
checkpoint-3942/trainer_state.json ADDED
@@ -0,0 +1,112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.0,
5
+ "eval_steps": 500,
6
+ "global_step": 3942,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.2536783358701167,
13
+ "grad_norm": 6.285004615783691,
14
+ "learning_rate": 1.9157787924911212e-05,
15
+ "loss": 1.589,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.5073566717402334,
20
+ "grad_norm": 6.245589256286621,
21
+ "learning_rate": 1.8312193472010826e-05,
22
+ "loss": 0.8971,
23
+ "step": 1000
24
+ },
25
+ {
26
+ "epoch": 0.76103500761035,
27
+ "grad_norm": 7.839916706085205,
28
+ "learning_rate": 1.7466599019110436e-05,
29
+ "loss": 0.7412,
30
+ "step": 1500
31
+ },
32
+ {
33
+ "epoch": 1.0,
34
+ "eval_bleu": 77.9302498710783,
35
+ "eval_loss": 0.5189284682273865,
36
+ "eval_rouge": {
37
+ "rouge1": 0.8422585807632581,
38
+ "rouge2": 0.7462100224193878,
39
+ "rougeL": 0.8347423871515361,
40
+ "rougeLsum": 0.8348341110331449
41
+ },
42
+ "eval_runtime": 589.3644,
43
+ "eval_samples_per_second": 14.402,
44
+ "eval_steps_per_second": 0.901,
45
+ "step": 1971
46
+ },
47
+ {
48
+ "epoch": 1.0147133434804667,
49
+ "grad_norm": 4.194991111755371,
50
+ "learning_rate": 1.6621004566210046e-05,
51
+ "loss": 0.6173,
52
+ "step": 2000
53
+ },
54
+ {
55
+ "epoch": 1.2683916793505834,
56
+ "grad_norm": 4.830150127410889,
57
+ "learning_rate": 1.5775410113309656e-05,
58
+ "loss": 0.4964,
59
+ "step": 2500
60
+ },
61
+ {
62
+ "epoch": 1.5220700152207,
63
+ "grad_norm": 6.662743091583252,
64
+ "learning_rate": 1.4929815660409268e-05,
65
+ "loss": 0.4655,
66
+ "step": 3000
67
+ },
68
+ {
69
+ "epoch": 1.775748351090817,
70
+ "grad_norm": 5.814598560333252,
71
+ "learning_rate": 1.408422120750888e-05,
72
+ "loss": 0.4562,
73
+ "step": 3500
74
+ },
75
+ {
76
+ "epoch": 2.0,
77
+ "eval_bleu": 82.4671860062862,
78
+ "eval_loss": 0.4167616665363312,
79
+ "eval_rouge": {
80
+ "rouge1": 0.8726285922102037,
81
+ "rouge2": 0.7929049594785995,
82
+ "rougeL": 0.8662329420974849,
83
+ "rougeLsum": 0.8662523269791588
84
+ },
85
+ "eval_runtime": 558.5398,
86
+ "eval_samples_per_second": 15.197,
87
+ "eval_steps_per_second": 0.951,
88
+ "step": 3942
89
+ }
90
+ ],
91
+ "logging_steps": 500,
92
+ "max_steps": 11826,
93
+ "num_input_tokens_seen": 0,
94
+ "num_train_epochs": 6,
95
+ "save_steps": 500,
96
+ "stateful_callbacks": {
97
+ "TrainerControl": {
98
+ "args": {
99
+ "should_epoch_stop": false,
100
+ "should_evaluate": false,
101
+ "should_log": false,
102
+ "should_save": true,
103
+ "should_training_stop": false
104
+ },
105
+ "attributes": {}
106
+ }
107
+ },
108
+ "total_flos": 531874085142528.0,
109
+ "train_batch_size": 16,
110
+ "trial_name": null,
111
+ "trial_params": null
112
+ }
checkpoint-3942/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d50538077dc9316af4f0e3a52d83f66a602e8f111533468407cd32fad55c01d
3
+ size 5496
checkpoint-3942/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-5913/config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Woondsc/opus-mt-ko-en-medterm",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "swish",
5
+ "add_bias_logits": false,
6
+ "add_final_layer_norm": false,
7
+ "architectures": [
8
+ "MarianMTModel"
9
+ ],
10
+ "attention_dropout": 0.0,
11
+ "bos_token_id": 0,
12
+ "classif_dropout": 0.0,
13
+ "classifier_dropout": 0.0,
14
+ "d_model": 512,
15
+ "decoder_attention_heads": 8,
16
+ "decoder_ffn_dim": 2048,
17
+ "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 6,
19
+ "decoder_start_token_id": 65000,
20
+ "decoder_vocab_size": 65001,
21
+ "dropout": 0.1,
22
+ "encoder_attention_heads": 8,
23
+ "encoder_ffn_dim": 2048,
24
+ "encoder_layerdrop": 0.0,
25
+ "encoder_layers": 6,
26
+ "eos_token_id": 0,
27
+ "extra_pos_embeddings": 65001,
28
+ "forced_eos_token_id": 0,
29
+ "id2label": {
30
+ "0": "LABEL_0",
31
+ "1": "LABEL_1",
32
+ "2": "LABEL_2"
33
+ },
34
+ "init_std": 0.02,
35
+ "is_encoder_decoder": true,
36
+ "label2id": {
37
+ "LABEL_0": 0,
38
+ "LABEL_1": 1,
39
+ "LABEL_2": 2
40
+ },
41
+ "max_length": null,
42
+ "max_position_embeddings": 512,
43
+ "model_type": "marian",
44
+ "normalize_before": false,
45
+ "normalize_embedding": false,
46
+ "num_beams": null,
47
+ "num_hidden_layers": 6,
48
+ "pad_token_id": 65000,
49
+ "scale_embedding": true,
50
+ "share_encoder_decoder_embeddings": true,
51
+ "static_position_embeddings": true,
52
+ "torch_dtype": "float32",
53
+ "transformers_version": "4.48.3",
54
+ "use_cache": true,
55
+ "vocab_size": 65001
56
+ }
checkpoint-5913/generation_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bad_words_ids": [
3
+ [
4
+ 65000
5
+ ]
6
+ ],
7
+ "bos_token_id": 0,
8
+ "decoder_start_token_id": 65000,
9
+ "eos_token_id": 0,
10
+ "forced_eos_token_id": 0,
11
+ "max_length": 512,
12
+ "num_beams": 6,
13
+ "pad_token_id": 65000,
14
+ "renormalize_logits": true,
15
+ "transformers_version": "4.48.3"
16
+ }
checkpoint-5913/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:838388e5434c711a6e41e7d25815812f9d4631f8c658a1eda64cd6f8045e2861
3
+ size 309965092
checkpoint-5913/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd4dd35331b261c8401fcc244258cf208601e1239bb410f6881a92275402cd6b
3
+ size 619563642
checkpoint-5913/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8041b037c9052a7992e9ff13a835adc04189793127123e37f382a3bbdac2c98f
3
+ size 14244
checkpoint-5913/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a893a098cd7e476391b3409db82effe07a8fb4dcd8a40dacf66da774085b4f75
3
+ size 1064
checkpoint-5913/source.spm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9496f7c2be9aecb84c751ae9f35a875915dde8e3892f652a5c76811ab2a0f49
3
+ size 841805
checkpoint-5913/special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eos_token": {
3
+ "content": "</s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "pad_token": {
10
+ "content": "<pad>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
checkpoint-5913/target.spm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1531ac8576fe64267dfca0dc33950a6e9a2d3fd9e05346558ad7ea5ee0e65bf
3
+ size 813126
checkpoint-5913/tokenizer_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "</s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<unk>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "65000": {
20
+ "content": "<pad>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "clean_up_tokenization_spaces": false,
29
+ "eos_token": "</s>",
30
+ "extra_special_tokens": {},
31
+ "model_max_length": 512,
32
+ "pad_token": "<pad>",
33
+ "separate_vocabs": false,
34
+ "source_lang": "kor",
35
+ "sp_model_kwargs": {},
36
+ "target_lang": "eng",
37
+ "tokenizer_class": "MarianTokenizer",
38
+ "unk_token": "<unk>"
39
+ }