Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +14 -0
- checkpoint-1172/config.json +56 -0
- checkpoint-1172/generation_config.json +16 -0
- checkpoint-1172/model.safetensors +3 -0
- checkpoint-1172/optimizer.pt +3 -0
- checkpoint-1172/rng_state.pth +3 -0
- checkpoint-1172/scheduler.pt +3 -0
- checkpoint-1172/source.spm +3 -0
- checkpoint-1172/special_tokens_map.json +23 -0
- checkpoint-1172/target.spm +3 -0
- checkpoint-1172/tokenizer_config.json +39 -0
- checkpoint-1172/trainer_state.json +77 -0
- checkpoint-1172/training_args.bin +3 -0
- checkpoint-1172/vocab.json +0 -0
- checkpoint-1758/config.json +56 -0
- checkpoint-1758/generation_config.json +16 -0
- checkpoint-1758/model.safetensors +3 -0
- checkpoint-1758/optimizer.pt +3 -0
- checkpoint-1758/rng_state.pth +3 -0
- checkpoint-1758/scheduler.pt +3 -0
- checkpoint-1758/source.spm +3 -0
- checkpoint-1758/special_tokens_map.json +23 -0
- checkpoint-1758/target.spm +3 -0
- checkpoint-1758/tokenizer_config.json +39 -0
- checkpoint-1758/trainer_state.json +99 -0
- checkpoint-1758/training_args.bin +3 -0
- checkpoint-1758/vocab.json +0 -0
- checkpoint-2344/config.json +56 -0
- checkpoint-2344/generation_config.json +16 -0
- checkpoint-2344/model.safetensors +3 -0
- checkpoint-2344/optimizer.pt +3 -0
- checkpoint-2344/rng_state.pth +3 -0
- checkpoint-2344/scheduler.pt +3 -0
- checkpoint-2344/source.spm +3 -0
- checkpoint-2344/special_tokens_map.json +23 -0
- checkpoint-2344/target.spm +3 -0
- checkpoint-2344/tokenizer_config.json +39 -0
- checkpoint-2344/trainer_state.json +121 -0
- checkpoint-2344/training_args.bin +3 -0
- checkpoint-2344/vocab.json +0 -0
- checkpoint-2930/config.json +56 -0
- checkpoint-2930/generation_config.json +16 -0
- checkpoint-2930/model.safetensors +3 -0
- checkpoint-2930/optimizer.pt +3 -0
- checkpoint-2930/rng_state.pth +3 -0
- checkpoint-2930/scheduler.pt +3 -0
- checkpoint-2930/source.spm +3 -0
- checkpoint-2930/special_tokens_map.json +23 -0
- checkpoint-2930/target.spm +3 -0
- checkpoint-2930/tokenizer_config.json +39 -0
.gitattributes
CHANGED
@@ -33,3 +33,17 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
checkpoint-1172/source.spm filter=lfs diff=lfs merge=lfs -text
|
37 |
+
checkpoint-1172/target.spm filter=lfs diff=lfs merge=lfs -text
|
38 |
+
checkpoint-1758/source.spm filter=lfs diff=lfs merge=lfs -text
|
39 |
+
checkpoint-1758/target.spm filter=lfs diff=lfs merge=lfs -text
|
40 |
+
checkpoint-2344/source.spm filter=lfs diff=lfs merge=lfs -text
|
41 |
+
checkpoint-2344/target.spm filter=lfs diff=lfs merge=lfs -text
|
42 |
+
checkpoint-2930/source.spm filter=lfs diff=lfs merge=lfs -text
|
43 |
+
checkpoint-2930/target.spm filter=lfs diff=lfs merge=lfs -text
|
44 |
+
checkpoint-3516/source.spm filter=lfs diff=lfs merge=lfs -text
|
45 |
+
checkpoint-3516/target.spm filter=lfs diff=lfs merge=lfs -text
|
46 |
+
checkpoint-586/source.spm filter=lfs diff=lfs merge=lfs -text
|
47 |
+
checkpoint-586/target.spm filter=lfs diff=lfs merge=lfs -text
|
48 |
+
source.spm filter=lfs diff=lfs merge=lfs -text
|
49 |
+
target.spm filter=lfs diff=lfs merge=lfs -text
|
checkpoint-1172/config.json
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "Woondsc/opus-mt-ko-en-medterm",
|
3 |
+
"activation_dropout": 0.0,
|
4 |
+
"activation_function": "swish",
|
5 |
+
"add_bias_logits": false,
|
6 |
+
"add_final_layer_norm": false,
|
7 |
+
"architectures": [
|
8 |
+
"MarianMTModel"
|
9 |
+
],
|
10 |
+
"attention_dropout": 0.0,
|
11 |
+
"bos_token_id": 0,
|
12 |
+
"classif_dropout": 0.0,
|
13 |
+
"classifier_dropout": 0.0,
|
14 |
+
"d_model": 512,
|
15 |
+
"decoder_attention_heads": 8,
|
16 |
+
"decoder_ffn_dim": 2048,
|
17 |
+
"decoder_layerdrop": 0.0,
|
18 |
+
"decoder_layers": 6,
|
19 |
+
"decoder_start_token_id": 65000,
|
20 |
+
"decoder_vocab_size": 65001,
|
21 |
+
"dropout": 0.1,
|
22 |
+
"encoder_attention_heads": 8,
|
23 |
+
"encoder_ffn_dim": 2048,
|
24 |
+
"encoder_layerdrop": 0.0,
|
25 |
+
"encoder_layers": 6,
|
26 |
+
"eos_token_id": 0,
|
27 |
+
"extra_pos_embeddings": 65001,
|
28 |
+
"forced_eos_token_id": 0,
|
29 |
+
"id2label": {
|
30 |
+
"0": "LABEL_0",
|
31 |
+
"1": "LABEL_1",
|
32 |
+
"2": "LABEL_2"
|
33 |
+
},
|
34 |
+
"init_std": 0.02,
|
35 |
+
"is_encoder_decoder": true,
|
36 |
+
"label2id": {
|
37 |
+
"LABEL_0": 0,
|
38 |
+
"LABEL_1": 1,
|
39 |
+
"LABEL_2": 2
|
40 |
+
},
|
41 |
+
"max_length": null,
|
42 |
+
"max_position_embeddings": 512,
|
43 |
+
"model_type": "marian",
|
44 |
+
"normalize_before": false,
|
45 |
+
"normalize_embedding": false,
|
46 |
+
"num_beams": null,
|
47 |
+
"num_hidden_layers": 6,
|
48 |
+
"pad_token_id": 65000,
|
49 |
+
"scale_embedding": true,
|
50 |
+
"share_encoder_decoder_embeddings": true,
|
51 |
+
"static_position_embeddings": true,
|
52 |
+
"torch_dtype": "float32",
|
53 |
+
"transformers_version": "4.48.3",
|
54 |
+
"use_cache": true,
|
55 |
+
"vocab_size": 65001
|
56 |
+
}
|
checkpoint-1172/generation_config.json
ADDED
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bad_words_ids": [
|
3 |
+
[
|
4 |
+
65000
|
5 |
+
]
|
6 |
+
],
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"decoder_start_token_id": 65000,
|
9 |
+
"eos_token_id": 0,
|
10 |
+
"forced_eos_token_id": 0,
|
11 |
+
"max_length": 512,
|
12 |
+
"num_beams": 6,
|
13 |
+
"pad_token_id": 65000,
|
14 |
+
"renormalize_logits": true,
|
15 |
+
"transformers_version": "4.48.3"
|
16 |
+
}
|
checkpoint-1172/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:62b01b7d210ad9320fe8c03d0660f74a52eb35b623d4eac80c65b8d951d61aad
|
3 |
+
size 309965092
|
checkpoint-1172/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a29572d616980a68aae1f305f07112289b2f2c0ea4f9d9f57b191fb969000ae8
|
3 |
+
size 619563642
|
checkpoint-1172/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5629f5bd384eb75bb6cb7e3bcb08a4dfec5da74d7f3b489cee0905e7cafd73e0
|
3 |
+
size 14244
|
checkpoint-1172/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fd443d2422bc9e6ddbaa16713140944279406fd161739a3d951292509af0de6f
|
3 |
+
size 1064
|
checkpoint-1172/source.spm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c9496f7c2be9aecb84c751ae9f35a875915dde8e3892f652a5c76811ab2a0f49
|
3 |
+
size 841805
|
checkpoint-1172/special_tokens_map.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eos_token": {
|
3 |
+
"content": "</s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"pad_token": {
|
10 |
+
"content": "<pad>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"unk_token": {
|
17 |
+
"content": "<unk>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
}
|
23 |
+
}
|
checkpoint-1172/target.spm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1531ac8576fe64267dfca0dc33950a6e9a2d3fd9e05346558ad7ea5ee0e65bf
|
3 |
+
size 813126
|
checkpoint-1172/tokenizer_config.json
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "</s>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<unk>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"65000": {
|
20 |
+
"content": "<pad>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"clean_up_tokenization_spaces": false,
|
29 |
+
"eos_token": "</s>",
|
30 |
+
"extra_special_tokens": {},
|
31 |
+
"model_max_length": 512,
|
32 |
+
"pad_token": "<pad>",
|
33 |
+
"separate_vocabs": false,
|
34 |
+
"source_lang": "kor",
|
35 |
+
"sp_model_kwargs": {},
|
36 |
+
"target_lang": "eng",
|
37 |
+
"tokenizer_class": "MarianTokenizer",
|
38 |
+
"unk_token": "<unk>"
|
39 |
+
}
|
checkpoint-1172/trainer_state.json
ADDED
@@ -0,0 +1,77 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_metric": null,
|
3 |
+
"best_model_checkpoint": null,
|
4 |
+
"epoch": 2.0,
|
5 |
+
"eval_steps": 500,
|
6 |
+
"global_step": 1172,
|
7 |
+
"is_hyper_param_search": false,
|
8 |
+
"is_local_process_zero": true,
|
9 |
+
"is_world_process_zero": true,
|
10 |
+
"log_history": [
|
11 |
+
{
|
12 |
+
"epoch": 0.8532423208191127,
|
13 |
+
"grad_norm": 8.762770652770996,
|
14 |
+
"learning_rate": 1.7167235494880547e-05,
|
15 |
+
"loss": 1.4672,
|
16 |
+
"step": 500
|
17 |
+
},
|
18 |
+
{
|
19 |
+
"epoch": 1.0,
|
20 |
+
"eval_bleu": 71.17004731575985,
|
21 |
+
"eval_loss": 0.7510850429534912,
|
22 |
+
"eval_rouge": {
|
23 |
+
"rouge1": 0.7586685078425267,
|
24 |
+
"rouge2": 0.6115347357798906,
|
25 |
+
"rougeL": 0.7490533887167732,
|
26 |
+
"rougeLsum": 0.7484630926564035
|
27 |
+
},
|
28 |
+
"eval_runtime": 134.1556,
|
29 |
+
"eval_samples_per_second": 18.807,
|
30 |
+
"eval_steps_per_second": 1.178,
|
31 |
+
"step": 586
|
32 |
+
},
|
33 |
+
{
|
34 |
+
"epoch": 1.7064846416382253,
|
35 |
+
"grad_norm": 5.05549430847168,
|
36 |
+
"learning_rate": 1.4323094425483506e-05,
|
37 |
+
"loss": 0.7479,
|
38 |
+
"step": 1000
|
39 |
+
},
|
40 |
+
{
|
41 |
+
"epoch": 2.0,
|
42 |
+
"eval_bleu": 77.98925833953002,
|
43 |
+
"eval_loss": 0.5916641354560852,
|
44 |
+
"eval_rouge": {
|
45 |
+
"rouge1": 0.7980118177508164,
|
46 |
+
"rouge2": 0.6728157741476863,
|
47 |
+
"rougeL": 0.7898294918297277,
|
48 |
+
"rougeLsum": 0.7889319703951503
|
49 |
+
},
|
50 |
+
"eval_runtime": 134.4934,
|
51 |
+
"eval_samples_per_second": 18.759,
|
52 |
+
"eval_steps_per_second": 1.175,
|
53 |
+
"step": 1172
|
54 |
+
}
|
55 |
+
],
|
56 |
+
"logging_steps": 500,
|
57 |
+
"max_steps": 3516,
|
58 |
+
"num_input_tokens_seen": 0,
|
59 |
+
"num_train_epochs": 6,
|
60 |
+
"save_steps": 500,
|
61 |
+
"stateful_callbacks": {
|
62 |
+
"TrainerControl": {
|
63 |
+
"args": {
|
64 |
+
"should_epoch_stop": false,
|
65 |
+
"should_evaluate": false,
|
66 |
+
"should_log": false,
|
67 |
+
"should_save": true,
|
68 |
+
"should_training_stop": false
|
69 |
+
},
|
70 |
+
"attributes": {}
|
71 |
+
}
|
72 |
+
},
|
73 |
+
"total_flos": 118470548127744.0,
|
74 |
+
"train_batch_size": 16,
|
75 |
+
"trial_name": null,
|
76 |
+
"trial_params": null
|
77 |
+
}
|
checkpoint-1172/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f2d28648ba7f6f8e9e5d172a0e9e7e4d52e850fef13b0741a2fb166b63fba202
|
3 |
+
size 5496
|
checkpoint-1172/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-1758/config.json
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "Woondsc/opus-mt-ko-en-medterm",
|
3 |
+
"activation_dropout": 0.0,
|
4 |
+
"activation_function": "swish",
|
5 |
+
"add_bias_logits": false,
|
6 |
+
"add_final_layer_norm": false,
|
7 |
+
"architectures": [
|
8 |
+
"MarianMTModel"
|
9 |
+
],
|
10 |
+
"attention_dropout": 0.0,
|
11 |
+
"bos_token_id": 0,
|
12 |
+
"classif_dropout": 0.0,
|
13 |
+
"classifier_dropout": 0.0,
|
14 |
+
"d_model": 512,
|
15 |
+
"decoder_attention_heads": 8,
|
16 |
+
"decoder_ffn_dim": 2048,
|
17 |
+
"decoder_layerdrop": 0.0,
|
18 |
+
"decoder_layers": 6,
|
19 |
+
"decoder_start_token_id": 65000,
|
20 |
+
"decoder_vocab_size": 65001,
|
21 |
+
"dropout": 0.1,
|
22 |
+
"encoder_attention_heads": 8,
|
23 |
+
"encoder_ffn_dim": 2048,
|
24 |
+
"encoder_layerdrop": 0.0,
|
25 |
+
"encoder_layers": 6,
|
26 |
+
"eos_token_id": 0,
|
27 |
+
"extra_pos_embeddings": 65001,
|
28 |
+
"forced_eos_token_id": 0,
|
29 |
+
"id2label": {
|
30 |
+
"0": "LABEL_0",
|
31 |
+
"1": "LABEL_1",
|
32 |
+
"2": "LABEL_2"
|
33 |
+
},
|
34 |
+
"init_std": 0.02,
|
35 |
+
"is_encoder_decoder": true,
|
36 |
+
"label2id": {
|
37 |
+
"LABEL_0": 0,
|
38 |
+
"LABEL_1": 1,
|
39 |
+
"LABEL_2": 2
|
40 |
+
},
|
41 |
+
"max_length": null,
|
42 |
+
"max_position_embeddings": 512,
|
43 |
+
"model_type": "marian",
|
44 |
+
"normalize_before": false,
|
45 |
+
"normalize_embedding": false,
|
46 |
+
"num_beams": null,
|
47 |
+
"num_hidden_layers": 6,
|
48 |
+
"pad_token_id": 65000,
|
49 |
+
"scale_embedding": true,
|
50 |
+
"share_encoder_decoder_embeddings": true,
|
51 |
+
"static_position_embeddings": true,
|
52 |
+
"torch_dtype": "float32",
|
53 |
+
"transformers_version": "4.48.3",
|
54 |
+
"use_cache": true,
|
55 |
+
"vocab_size": 65001
|
56 |
+
}
|
checkpoint-1758/generation_config.json
ADDED
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bad_words_ids": [
|
3 |
+
[
|
4 |
+
65000
|
5 |
+
]
|
6 |
+
],
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"decoder_start_token_id": 65000,
|
9 |
+
"eos_token_id": 0,
|
10 |
+
"forced_eos_token_id": 0,
|
11 |
+
"max_length": 512,
|
12 |
+
"num_beams": 6,
|
13 |
+
"pad_token_id": 65000,
|
14 |
+
"renormalize_logits": true,
|
15 |
+
"transformers_version": "4.48.3"
|
16 |
+
}
|
checkpoint-1758/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae5dc2e61c1ba0c32fe7c467afa63d2842cb53e2723e12926d00d645e88b08c1
|
3 |
+
size 309965092
|
checkpoint-1758/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f90d9112c78cc4f08915fdca6ec511e7c1417b98d010e6dd115c3f1038ca1ea8
|
3 |
+
size 619563642
|
checkpoint-1758/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:039369bde3b349b36fed7be1fe52bee483eff4a13522093a7a14dced7f6975b6
|
3 |
+
size 14244
|
checkpoint-1758/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1c40abb44e078978f993b40cee2ee76bc70febdb041dfdebb588802807c323b5
|
3 |
+
size 1064
|
checkpoint-1758/source.spm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c9496f7c2be9aecb84c751ae9f35a875915dde8e3892f652a5c76811ab2a0f49
|
3 |
+
size 841805
|
checkpoint-1758/special_tokens_map.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eos_token": {
|
3 |
+
"content": "</s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"pad_token": {
|
10 |
+
"content": "<pad>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"unk_token": {
|
17 |
+
"content": "<unk>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
}
|
23 |
+
}
|
checkpoint-1758/target.spm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1531ac8576fe64267dfca0dc33950a6e9a2d3fd9e05346558ad7ea5ee0e65bf
|
3 |
+
size 813126
|
checkpoint-1758/tokenizer_config.json
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "</s>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<unk>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"65000": {
|
20 |
+
"content": "<pad>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"clean_up_tokenization_spaces": false,
|
29 |
+
"eos_token": "</s>",
|
30 |
+
"extra_special_tokens": {},
|
31 |
+
"model_max_length": 512,
|
32 |
+
"pad_token": "<pad>",
|
33 |
+
"separate_vocabs": false,
|
34 |
+
"source_lang": "kor",
|
35 |
+
"sp_model_kwargs": {},
|
36 |
+
"target_lang": "eng",
|
37 |
+
"tokenizer_class": "MarianTokenizer",
|
38 |
+
"unk_token": "<unk>"
|
39 |
+
}
|
checkpoint-1758/trainer_state.json
ADDED
@@ -0,0 +1,99 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_metric": null,
|
3 |
+
"best_model_checkpoint": null,
|
4 |
+
"epoch": 3.0,
|
5 |
+
"eval_steps": 500,
|
6 |
+
"global_step": 1758,
|
7 |
+
"is_hyper_param_search": false,
|
8 |
+
"is_local_process_zero": true,
|
9 |
+
"is_world_process_zero": true,
|
10 |
+
"log_history": [
|
11 |
+
{
|
12 |
+
"epoch": 0.8532423208191127,
|
13 |
+
"grad_norm": 8.762770652770996,
|
14 |
+
"learning_rate": 1.7167235494880547e-05,
|
15 |
+
"loss": 1.4672,
|
16 |
+
"step": 500
|
17 |
+
},
|
18 |
+
{
|
19 |
+
"epoch": 1.0,
|
20 |
+
"eval_bleu": 71.17004731575985,
|
21 |
+
"eval_loss": 0.7510850429534912,
|
22 |
+
"eval_rouge": {
|
23 |
+
"rouge1": 0.7586685078425267,
|
24 |
+
"rouge2": 0.6115347357798906,
|
25 |
+
"rougeL": 0.7490533887167732,
|
26 |
+
"rougeLsum": 0.7484630926564035
|
27 |
+
},
|
28 |
+
"eval_runtime": 134.1556,
|
29 |
+
"eval_samples_per_second": 18.807,
|
30 |
+
"eval_steps_per_second": 1.178,
|
31 |
+
"step": 586
|
32 |
+
},
|
33 |
+
{
|
34 |
+
"epoch": 1.7064846416382253,
|
35 |
+
"grad_norm": 5.05549430847168,
|
36 |
+
"learning_rate": 1.4323094425483506e-05,
|
37 |
+
"loss": 0.7479,
|
38 |
+
"step": 1000
|
39 |
+
},
|
40 |
+
{
|
41 |
+
"epoch": 2.0,
|
42 |
+
"eval_bleu": 77.98925833953002,
|
43 |
+
"eval_loss": 0.5916641354560852,
|
44 |
+
"eval_rouge": {
|
45 |
+
"rouge1": 0.7980118177508164,
|
46 |
+
"rouge2": 0.6728157741476863,
|
47 |
+
"rougeL": 0.7898294918297277,
|
48 |
+
"rougeLsum": 0.7889319703951503
|
49 |
+
},
|
50 |
+
"eval_runtime": 134.4934,
|
51 |
+
"eval_samples_per_second": 18.759,
|
52 |
+
"eval_steps_per_second": 1.175,
|
53 |
+
"step": 1172
|
54 |
+
},
|
55 |
+
{
|
56 |
+
"epoch": 2.5597269624573378,
|
57 |
+
"grad_norm": 3.831761360168457,
|
58 |
+
"learning_rate": 1.1478953356086462e-05,
|
59 |
+
"loss": 0.556,
|
60 |
+
"step": 1500
|
61 |
+
},
|
62 |
+
{
|
63 |
+
"epoch": 3.0,
|
64 |
+
"eval_bleu": 81.08165889470924,
|
65 |
+
"eval_loss": 0.5316380858421326,
|
66 |
+
"eval_rouge": {
|
67 |
+
"rouge1": 0.8151391371295702,
|
68 |
+
"rouge2": 0.6981852981887253,
|
69 |
+
"rougeL": 0.806655138353247,
|
70 |
+
"rougeLsum": 0.8059563640828541
|
71 |
+
},
|
72 |
+
"eval_runtime": 132.3183,
|
73 |
+
"eval_samples_per_second": 19.068,
|
74 |
+
"eval_steps_per_second": 1.194,
|
75 |
+
"step": 1758
|
76 |
+
}
|
77 |
+
],
|
78 |
+
"logging_steps": 500,
|
79 |
+
"max_steps": 3516,
|
80 |
+
"num_input_tokens_seen": 0,
|
81 |
+
"num_train_epochs": 6,
|
82 |
+
"save_steps": 500,
|
83 |
+
"stateful_callbacks": {
|
84 |
+
"TrainerControl": {
|
85 |
+
"args": {
|
86 |
+
"should_epoch_stop": false,
|
87 |
+
"should_evaluate": false,
|
88 |
+
"should_log": false,
|
89 |
+
"should_save": true,
|
90 |
+
"should_training_stop": false
|
91 |
+
},
|
92 |
+
"attributes": {}
|
93 |
+
}
|
94 |
+
},
|
95 |
+
"total_flos": 178015144771584.0,
|
96 |
+
"train_batch_size": 16,
|
97 |
+
"trial_name": null,
|
98 |
+
"trial_params": null
|
99 |
+
}
|
checkpoint-1758/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f2d28648ba7f6f8e9e5d172a0e9e7e4d52e850fef13b0741a2fb166b63fba202
|
3 |
+
size 5496
|
checkpoint-1758/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-2344/config.json
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "Woondsc/opus-mt-ko-en-medterm",
|
3 |
+
"activation_dropout": 0.0,
|
4 |
+
"activation_function": "swish",
|
5 |
+
"add_bias_logits": false,
|
6 |
+
"add_final_layer_norm": false,
|
7 |
+
"architectures": [
|
8 |
+
"MarianMTModel"
|
9 |
+
],
|
10 |
+
"attention_dropout": 0.0,
|
11 |
+
"bos_token_id": 0,
|
12 |
+
"classif_dropout": 0.0,
|
13 |
+
"classifier_dropout": 0.0,
|
14 |
+
"d_model": 512,
|
15 |
+
"decoder_attention_heads": 8,
|
16 |
+
"decoder_ffn_dim": 2048,
|
17 |
+
"decoder_layerdrop": 0.0,
|
18 |
+
"decoder_layers": 6,
|
19 |
+
"decoder_start_token_id": 65000,
|
20 |
+
"decoder_vocab_size": 65001,
|
21 |
+
"dropout": 0.1,
|
22 |
+
"encoder_attention_heads": 8,
|
23 |
+
"encoder_ffn_dim": 2048,
|
24 |
+
"encoder_layerdrop": 0.0,
|
25 |
+
"encoder_layers": 6,
|
26 |
+
"eos_token_id": 0,
|
27 |
+
"extra_pos_embeddings": 65001,
|
28 |
+
"forced_eos_token_id": 0,
|
29 |
+
"id2label": {
|
30 |
+
"0": "LABEL_0",
|
31 |
+
"1": "LABEL_1",
|
32 |
+
"2": "LABEL_2"
|
33 |
+
},
|
34 |
+
"init_std": 0.02,
|
35 |
+
"is_encoder_decoder": true,
|
36 |
+
"label2id": {
|
37 |
+
"LABEL_0": 0,
|
38 |
+
"LABEL_1": 1,
|
39 |
+
"LABEL_2": 2
|
40 |
+
},
|
41 |
+
"max_length": null,
|
42 |
+
"max_position_embeddings": 512,
|
43 |
+
"model_type": "marian",
|
44 |
+
"normalize_before": false,
|
45 |
+
"normalize_embedding": false,
|
46 |
+
"num_beams": null,
|
47 |
+
"num_hidden_layers": 6,
|
48 |
+
"pad_token_id": 65000,
|
49 |
+
"scale_embedding": true,
|
50 |
+
"share_encoder_decoder_embeddings": true,
|
51 |
+
"static_position_embeddings": true,
|
52 |
+
"torch_dtype": "float32",
|
53 |
+
"transformers_version": "4.48.3",
|
54 |
+
"use_cache": true,
|
55 |
+
"vocab_size": 65001
|
56 |
+
}
|
checkpoint-2344/generation_config.json
ADDED
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bad_words_ids": [
|
3 |
+
[
|
4 |
+
65000
|
5 |
+
]
|
6 |
+
],
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"decoder_start_token_id": 65000,
|
9 |
+
"eos_token_id": 0,
|
10 |
+
"forced_eos_token_id": 0,
|
11 |
+
"max_length": 512,
|
12 |
+
"num_beams": 6,
|
13 |
+
"pad_token_id": 65000,
|
14 |
+
"renormalize_logits": true,
|
15 |
+
"transformers_version": "4.48.3"
|
16 |
+
}
|
checkpoint-2344/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8bf05405650e7fbac4efa7f1bc9dfba93a8d6bb96de577a8cd16f9a6fd8ce347
|
3 |
+
size 309965092
|
checkpoint-2344/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f763d2ee7c7861666001314852bb014ff5488dc3191b3f393ccd9c9123d4dbfd
|
3 |
+
size 619563642
|
checkpoint-2344/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c83e56e2e5ee23f9fecf215ceded761be05ac394f33023930fd7e7d97910686
|
3 |
+
size 14244
|
checkpoint-2344/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c2639021da6e3d648d570927ed265d371063fd545e105458fc72dff7e0c0f84f
|
3 |
+
size 1064
|
checkpoint-2344/source.spm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c9496f7c2be9aecb84c751ae9f35a875915dde8e3892f652a5c76811ab2a0f49
|
3 |
+
size 841805
|
checkpoint-2344/special_tokens_map.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eos_token": {
|
3 |
+
"content": "</s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"pad_token": {
|
10 |
+
"content": "<pad>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"unk_token": {
|
17 |
+
"content": "<unk>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
}
|
23 |
+
}
|
checkpoint-2344/target.spm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1531ac8576fe64267dfca0dc33950a6e9a2d3fd9e05346558ad7ea5ee0e65bf
|
3 |
+
size 813126
|
checkpoint-2344/tokenizer_config.json
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "</s>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<unk>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"65000": {
|
20 |
+
"content": "<pad>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"clean_up_tokenization_spaces": false,
|
29 |
+
"eos_token": "</s>",
|
30 |
+
"extra_special_tokens": {},
|
31 |
+
"model_max_length": 512,
|
32 |
+
"pad_token": "<pad>",
|
33 |
+
"separate_vocabs": false,
|
34 |
+
"source_lang": "kor",
|
35 |
+
"sp_model_kwargs": {},
|
36 |
+
"target_lang": "eng",
|
37 |
+
"tokenizer_class": "MarianTokenizer",
|
38 |
+
"unk_token": "<unk>"
|
39 |
+
}
|
checkpoint-2344/trainer_state.json
ADDED
@@ -0,0 +1,121 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_metric": null,
|
3 |
+
"best_model_checkpoint": null,
|
4 |
+
"epoch": 4.0,
|
5 |
+
"eval_steps": 500,
|
6 |
+
"global_step": 2344,
|
7 |
+
"is_hyper_param_search": false,
|
8 |
+
"is_local_process_zero": true,
|
9 |
+
"is_world_process_zero": true,
|
10 |
+
"log_history": [
|
11 |
+
{
|
12 |
+
"epoch": 0.8532423208191127,
|
13 |
+
"grad_norm": 8.762770652770996,
|
14 |
+
"learning_rate": 1.7167235494880547e-05,
|
15 |
+
"loss": 1.4672,
|
16 |
+
"step": 500
|
17 |
+
},
|
18 |
+
{
|
19 |
+
"epoch": 1.0,
|
20 |
+
"eval_bleu": 71.17004731575985,
|
21 |
+
"eval_loss": 0.7510850429534912,
|
22 |
+
"eval_rouge": {
|
23 |
+
"rouge1": 0.7586685078425267,
|
24 |
+
"rouge2": 0.6115347357798906,
|
25 |
+
"rougeL": 0.7490533887167732,
|
26 |
+
"rougeLsum": 0.7484630926564035
|
27 |
+
},
|
28 |
+
"eval_runtime": 134.1556,
|
29 |
+
"eval_samples_per_second": 18.807,
|
30 |
+
"eval_steps_per_second": 1.178,
|
31 |
+
"step": 586
|
32 |
+
},
|
33 |
+
{
|
34 |
+
"epoch": 1.7064846416382253,
|
35 |
+
"grad_norm": 5.05549430847168,
|
36 |
+
"learning_rate": 1.4323094425483506e-05,
|
37 |
+
"loss": 0.7479,
|
38 |
+
"step": 1000
|
39 |
+
},
|
40 |
+
{
|
41 |
+
"epoch": 2.0,
|
42 |
+
"eval_bleu": 77.98925833953002,
|
43 |
+
"eval_loss": 0.5916641354560852,
|
44 |
+
"eval_rouge": {
|
45 |
+
"rouge1": 0.7980118177508164,
|
46 |
+
"rouge2": 0.6728157741476863,
|
47 |
+
"rougeL": 0.7898294918297277,
|
48 |
+
"rougeLsum": 0.7889319703951503
|
49 |
+
},
|
50 |
+
"eval_runtime": 134.4934,
|
51 |
+
"eval_samples_per_second": 18.759,
|
52 |
+
"eval_steps_per_second": 1.175,
|
53 |
+
"step": 1172
|
54 |
+
},
|
55 |
+
{
|
56 |
+
"epoch": 2.5597269624573378,
|
57 |
+
"grad_norm": 3.831761360168457,
|
58 |
+
"learning_rate": 1.1478953356086462e-05,
|
59 |
+
"loss": 0.556,
|
60 |
+
"step": 1500
|
61 |
+
},
|
62 |
+
{
|
63 |
+
"epoch": 3.0,
|
64 |
+
"eval_bleu": 81.08165889470924,
|
65 |
+
"eval_loss": 0.5316380858421326,
|
66 |
+
"eval_rouge": {
|
67 |
+
"rouge1": 0.8151391371295702,
|
68 |
+
"rouge2": 0.6981852981887253,
|
69 |
+
"rougeL": 0.806655138353247,
|
70 |
+
"rougeLsum": 0.8059563640828541
|
71 |
+
},
|
72 |
+
"eval_runtime": 132.3183,
|
73 |
+
"eval_samples_per_second": 19.068,
|
74 |
+
"eval_steps_per_second": 1.194,
|
75 |
+
"step": 1758
|
76 |
+
},
|
77 |
+
{
|
78 |
+
"epoch": 3.4129692832764507,
|
79 |
+
"grad_norm": 3.7003045082092285,
|
80 |
+
"learning_rate": 8.63481228668942e-06,
|
81 |
+
"loss": 0.4508,
|
82 |
+
"step": 2000
|
83 |
+
},
|
84 |
+
{
|
85 |
+
"epoch": 4.0,
|
86 |
+
"eval_bleu": 82.60133061747088,
|
87 |
+
"eval_loss": 0.4998228847980499,
|
88 |
+
"eval_rouge": {
|
89 |
+
"rouge1": 0.8238729032217965,
|
90 |
+
"rouge2": 0.7162403358709662,
|
91 |
+
"rougeL": 0.8167800536747541,
|
92 |
+
"rougeLsum": 0.8161987903541148
|
93 |
+
},
|
94 |
+
"eval_runtime": 132.5265,
|
95 |
+
"eval_samples_per_second": 19.038,
|
96 |
+
"eval_steps_per_second": 1.192,
|
97 |
+
"step": 2344
|
98 |
+
}
|
99 |
+
],
|
100 |
+
"logging_steps": 500,
|
101 |
+
"max_steps": 3516,
|
102 |
+
"num_input_tokens_seen": 0,
|
103 |
+
"num_train_epochs": 6,
|
104 |
+
"save_steps": 500,
|
105 |
+
"stateful_callbacks": {
|
106 |
+
"TrainerControl": {
|
107 |
+
"args": {
|
108 |
+
"should_epoch_stop": false,
|
109 |
+
"should_evaluate": false,
|
110 |
+
"should_log": false,
|
111 |
+
"should_save": true,
|
112 |
+
"should_training_stop": false
|
113 |
+
},
|
114 |
+
"attributes": {}
|
115 |
+
}
|
116 |
+
},
|
117 |
+
"total_flos": 237695334875136.0,
|
118 |
+
"train_batch_size": 16,
|
119 |
+
"trial_name": null,
|
120 |
+
"trial_params": null
|
121 |
+
}
|
checkpoint-2344/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f2d28648ba7f6f8e9e5d172a0e9e7e4d52e850fef13b0741a2fb166b63fba202
|
3 |
+
size 5496
|
checkpoint-2344/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-2930/config.json
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "Woondsc/opus-mt-ko-en-medterm",
|
3 |
+
"activation_dropout": 0.0,
|
4 |
+
"activation_function": "swish",
|
5 |
+
"add_bias_logits": false,
|
6 |
+
"add_final_layer_norm": false,
|
7 |
+
"architectures": [
|
8 |
+
"MarianMTModel"
|
9 |
+
],
|
10 |
+
"attention_dropout": 0.0,
|
11 |
+
"bos_token_id": 0,
|
12 |
+
"classif_dropout": 0.0,
|
13 |
+
"classifier_dropout": 0.0,
|
14 |
+
"d_model": 512,
|
15 |
+
"decoder_attention_heads": 8,
|
16 |
+
"decoder_ffn_dim": 2048,
|
17 |
+
"decoder_layerdrop": 0.0,
|
18 |
+
"decoder_layers": 6,
|
19 |
+
"decoder_start_token_id": 65000,
|
20 |
+
"decoder_vocab_size": 65001,
|
21 |
+
"dropout": 0.1,
|
22 |
+
"encoder_attention_heads": 8,
|
23 |
+
"encoder_ffn_dim": 2048,
|
24 |
+
"encoder_layerdrop": 0.0,
|
25 |
+
"encoder_layers": 6,
|
26 |
+
"eos_token_id": 0,
|
27 |
+
"extra_pos_embeddings": 65001,
|
28 |
+
"forced_eos_token_id": 0,
|
29 |
+
"id2label": {
|
30 |
+
"0": "LABEL_0",
|
31 |
+
"1": "LABEL_1",
|
32 |
+
"2": "LABEL_2"
|
33 |
+
},
|
34 |
+
"init_std": 0.02,
|
35 |
+
"is_encoder_decoder": true,
|
36 |
+
"label2id": {
|
37 |
+
"LABEL_0": 0,
|
38 |
+
"LABEL_1": 1,
|
39 |
+
"LABEL_2": 2
|
40 |
+
},
|
41 |
+
"max_length": null,
|
42 |
+
"max_position_embeddings": 512,
|
43 |
+
"model_type": "marian",
|
44 |
+
"normalize_before": false,
|
45 |
+
"normalize_embedding": false,
|
46 |
+
"num_beams": null,
|
47 |
+
"num_hidden_layers": 6,
|
48 |
+
"pad_token_id": 65000,
|
49 |
+
"scale_embedding": true,
|
50 |
+
"share_encoder_decoder_embeddings": true,
|
51 |
+
"static_position_embeddings": true,
|
52 |
+
"torch_dtype": "float32",
|
53 |
+
"transformers_version": "4.48.3",
|
54 |
+
"use_cache": true,
|
55 |
+
"vocab_size": 65001
|
56 |
+
}
|
checkpoint-2930/generation_config.json
ADDED
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bad_words_ids": [
|
3 |
+
[
|
4 |
+
65000
|
5 |
+
]
|
6 |
+
],
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"decoder_start_token_id": 65000,
|
9 |
+
"eos_token_id": 0,
|
10 |
+
"forced_eos_token_id": 0,
|
11 |
+
"max_length": 512,
|
12 |
+
"num_beams": 6,
|
13 |
+
"pad_token_id": 65000,
|
14 |
+
"renormalize_logits": true,
|
15 |
+
"transformers_version": "4.48.3"
|
16 |
+
}
|
checkpoint-2930/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bbb078397915a62890ca13d320685590c5734f3c6c33226b8e62a4cae7f2c822
|
3 |
+
size 309965092
|
checkpoint-2930/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1626ad9a4cc60b8eaab30d92fc713533c0768561aab8255dc9ff75916c59e789
|
3 |
+
size 619563642
|
checkpoint-2930/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dcdcb9f230c546ab808091009cf2f89a1cba7f2471e62b54888df416d85c1ec0
|
3 |
+
size 14244
|
checkpoint-2930/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01fbea6b7c88ed9bb73296c972f4033dded5a90b4db9d824aa0d94455317f560
|
3 |
+
size 1064
|
checkpoint-2930/source.spm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c9496f7c2be9aecb84c751ae9f35a875915dde8e3892f652a5c76811ab2a0f49
|
3 |
+
size 841805
|
checkpoint-2930/special_tokens_map.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eos_token": {
|
3 |
+
"content": "</s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"pad_token": {
|
10 |
+
"content": "<pad>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"unk_token": {
|
17 |
+
"content": "<unk>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
}
|
23 |
+
}
|
checkpoint-2930/target.spm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1531ac8576fe64267dfca0dc33950a6e9a2d3fd9e05346558ad7ea5ee0e65bf
|
3 |
+
size 813126
|
checkpoint-2930/tokenizer_config.json
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "</s>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<unk>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"65000": {
|
20 |
+
"content": "<pad>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"clean_up_tokenization_spaces": false,
|
29 |
+
"eos_token": "</s>",
|
30 |
+
"extra_special_tokens": {},
|
31 |
+
"model_max_length": 512,
|
32 |
+
"pad_token": "<pad>",
|
33 |
+
"separate_vocabs": false,
|
34 |
+
"source_lang": "kor",
|
35 |
+
"sp_model_kwargs": {},
|
36 |
+
"target_lang": "eng",
|
37 |
+
"tokenizer_class": "MarianTokenizer",
|
38 |
+
"unk_token": "<unk>"
|
39 |
+
}
|