Sieun Park commited on
Commit
e087677
·
1 Parent(s): 3960360

Upload ./result with huggingface_hub

Browse files
.gitattributes CHANGED
@@ -32,3 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ result/tokenizer.json filter=lfs diff=lfs merge=lfs -text
result/1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
result/README.md ADDED
@@ -0,0 +1,127 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+ - transformers
8
+
9
+ ---
10
+
11
+ # {MODEL_NAME}
12
+
13
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
14
+
15
+ <!--- Describe your model here -->
16
+
17
+ ## Usage (Sentence-Transformers)
18
+
19
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
20
+
21
+ ```
22
+ pip install -U sentence-transformers
23
+ ```
24
+
25
+ Then you can use the model like this:
26
+
27
+ ```python
28
+ from sentence_transformers import SentenceTransformer
29
+ sentences = ["This is an example sentence", "Each sentence is converted"]
30
+
31
+ model = SentenceTransformer('{MODEL_NAME}')
32
+ embeddings = model.encode(sentences)
33
+ print(embeddings)
34
+ ```
35
+
36
+
37
+
38
+ ## Usage (HuggingFace Transformers)
39
+ Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
40
+
41
+ ```python
42
+ from transformers import AutoTokenizer, AutoModel
43
+ import torch
44
+
45
+
46
+ #Mean Pooling - Take attention mask into account for correct averaging
47
+ def mean_pooling(model_output, attention_mask):
48
+ token_embeddings = model_output[0] #First element of model_output contains all token embeddings
49
+ input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
50
+ return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
51
+
52
+
53
+ # Sentences we want sentence embeddings for
54
+ sentences = ['This is an example sentence', 'Each sentence is converted']
55
+
56
+ # Load model from HuggingFace Hub
57
+ tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}')
58
+ model = AutoModel.from_pretrained('{MODEL_NAME}')
59
+
60
+ # Tokenize sentences
61
+ encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
62
+
63
+ # Compute token embeddings
64
+ with torch.no_grad():
65
+ model_output = model(**encoded_input)
66
+
67
+ # Perform pooling. In this case, mean pooling.
68
+ sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
69
+
70
+ print("Sentence embeddings:")
71
+ print(sentence_embeddings)
72
+ ```
73
+
74
+
75
+
76
+ ## Evaluation Results
77
+
78
+ <!--- Describe how your model was evaluated -->
79
+
80
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
81
+
82
+
83
+ ## Training
84
+ The model was trained with the parameters:
85
+
86
+ **DataLoader**:
87
+
88
+ `torch.utils.data.dataloader.DataLoader` of length 11258 with parameters:
89
+ ```
90
+ {'batch_size': 128, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
91
+ ```
92
+
93
+ **Loss**:
94
+
95
+ `sentence_transformers.losses.MSELoss.MSELoss`
96
+
97
+ Parameters of the fit()-Method:
98
+ ```
99
+ {
100
+ "epochs": 5,
101
+ "evaluation_steps": 1000,
102
+ "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator",
103
+ "max_grad_norm": 1,
104
+ "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
105
+ "optimizer_params": {
106
+ "eps": 1e-06,
107
+ "lr": 2e-05
108
+ },
109
+ "scheduler": "WarmupLinear",
110
+ "steps_per_epoch": null,
111
+ "warmup_steps": 10000,
112
+ "weight_decay": 0.01
113
+ }
114
+ ```
115
+
116
+
117
+ ## Full Model Architecture
118
+ ```
119
+ SentenceTransformer(
120
+ (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
121
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
122
+ )
123
+ ```
124
+
125
+ ## Citing & Authors
126
+
127
+ <!--- Describe where people can find more information -->
result/config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "xlm-roberta-base",
3
+ "architectures": [
4
+ "XLMRobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 514,
17
+ "model_type": "xlm-roberta",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 12,
20
+ "output_past": true,
21
+ "pad_token_id": 1,
22
+ "position_embedding_type": "absolute",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.26.0",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 250002
28
+ }
result/config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.2.2",
4
+ "transformers": "4.26.0",
5
+ "pytorch": "1.13.1+cu116"
6
+ }
7
+ }
result/eval/mse_evaluation_TED2020-en-ja-dev.tsv.gz_results.csv ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,MSE
2
+ 0,1000,2.479737065732479
3
+ 0,2000,0.7978680543601513
4
+ 0,3000,0.465529365465045
5
+ 0,4000,0.38496602792292833
6
+ 0,5000,0.3095171647146344
7
+ 0,6000,1.007575262337923
8
+ 0,7000,2.9757926240563393
9
+ 0,8000,0.1680491492152214
10
+ 0,9000,0.15370475593954325
11
+ 0,10000,0.14578887494280934
12
+ 0,11000,44.72245275974274
13
+ 0,-1,38.11988830566406
14
+ 1,1000,34.88417565822601
15
+ 1,2000,40.527331829071045
16
+ 1,3000,34.18007791042328
17
+ 1,4000,36.79800629615784
18
+ 1,5000,18.30635964870453
19
+ 1,6000,19.035635888576508
20
+ 1,7000,10.42214184999466
21
+ 1,8000,21.36164754629135
22
+ 1,9000,13.449710607528687
23
+ 1,10000,26.817166805267334
24
+ 1,11000,27.421659231185913
25
+ 1,-1,27.256697416305542
26
+ 2,1000,19.7702556848526
27
+ 2,2000,23.021529614925385
28
+ 2,3000,22.112300992012024
29
+ 2,4000,18.621844053268433
30
+ 2,5000,15.16069769859314
31
+ 2,6000,18.26171875
32
+ 2,7000,22.869695723056793
33
+ 2,8000,21.816536784172058
34
+ 2,9000,21.53472751379013
35
+ 2,10000,7.719540596008301
36
+ 2,11000,16.72600358724594
37
+ 2,-1,17.56407469511032
38
+ 3,1000,9.689188748598099
39
+ 3,2000,12.322552502155304
40
+ 3,3000,14.71262276172638
41
+ 3,4000,15.390068292617798
42
+ 3,5000,11.171068251132965
43
+ 3,6000,16.961337625980377
44
+ 3,7000,17.228218913078308
45
+ 3,8000,15.841719508171082
46
+ 3,9000,15.5427947640419
47
+ 3,10000,11.266006529331207
48
+ 3,11000,11.172742396593094
49
+ 3,-1,10.830702632665634
50
+ 4,1000,14.59239274263382
51
+ 4,2000,11.720868200063705
52
+ 4,3000,10.161346197128296
53
+ 4,4000,25.874224305152893
54
+ 4,5000,19.239723682403564
55
+ 4,6000,16.438673436641693
56
+ 4,7000,14.188717305660248
57
+ 4,8000,18.68770271539688
58
+ 4,9000,26.96780264377594
59
+ 4,10000,29.64048683643341
60
+ 4,11000,29.76890504360199
61
+ 4,-1,29.792797565460205
result/eval/mse_evaluation_TED2020-en-ko-dev.tsv.gz_results.csv ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,MSE
2
+ 0,1000,3.534962609410286
3
+ 0,2000,0.7726760115474463
4
+ 0,3000,0.45291599817574024
5
+ 0,4000,0.32619184348732233
6
+ 0,5000,0.3069185884669423
7
+ 0,6000,2.836661972105503
8
+ 0,7000,0.4042532294988632
9
+ 0,8000,0.1647747354581952
10
+ 0,9000,0.15147579833865166
11
+ 0,10000,0.14654598198831081
12
+ 0,11000,44.76562440395355
13
+ 0,-1,38.15537393093109
14
+ 1,1000,34.89996790885925
15
+ 1,2000,40.527620911598206
16
+ 1,3000,34.18329954147339
17
+ 1,4000,36.78698241710663
18
+ 1,5000,18.310919404029846
19
+ 1,6000,19.035445153713226
20
+ 1,7000,10.420460999011993
21
+ 1,8000,21.34985774755478
22
+ 1,9000,13.454011082649231
23
+ 1,10000,26.8216073513031
24
+ 1,11000,27.417752146720886
25
+ 1,-1,27.26142108440399
26
+ 2,1000,19.804176688194275
27
+ 2,2000,23.02238494157791
28
+ 2,3000,22.109846770763397
29
+ 2,4000,18.61446350812912
30
+ 2,5000,15.156088769435883
31
+ 2,6000,18.26195865869522
32
+ 2,7000,22.867940366268158
33
+ 2,8000,21.814395487308502
34
+ 2,9000,21.53337448835373
35
+ 2,10000,7.718393951654434
36
+ 2,11000,16.722606122493744
37
+ 2,-1,17.560648918151855
38
+ 3,1000,9.690673649311066
39
+ 3,2000,12.321917712688446
40
+ 3,3000,14.71351683139801
41
+ 3,4000,15.393638610839844
42
+ 3,5000,11.174257099628448
43
+ 3,6000,16.96329414844513
44
+ 3,7000,17.23286509513855
45
+ 3,8000,15.840160846710205
46
+ 3,9000,15.541520714759827
47
+ 3,10000,11.266209185123444
48
+ 3,11000,11.172308027744293
49
+ 3,-1,10.831181704998016
50
+ 4,1000,14.588256180286407
51
+ 4,2000,11.7195263504982
52
+ 4,3000,10.161814093589783
53
+ 4,4000,25.86996555328369
54
+ 4,5000,19.233083724975586
55
+ 4,6000,16.433295607566833
56
+ 4,7000,14.18774127960205
57
+ 4,8000,18.69051158428192
58
+ 4,9000,26.973900198936462
59
+ 4,10000,29.64228391647339
60
+ 4,11000,29.770362377166748
61
+ 4,-1,29.794207215309143
result/eval/similarity_evaluation_STS.en-en.txt_results.csv ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
2
+ 0,1000,0.1303136936057836,0.16109944136542498,0.09244556100421478,0.12335554401547714,0.09157986691726851,0.16468932075590426,0.15038989729175398,0.12561348935648534
3
+ 0,2000,0.10912365007333646,0.16055590464220645,0.04794033150727805,0.056331392577127014,0.12450082278387518,0.1120412159311675,0.18294141836020666,0.1578924209398153
4
+ 0,3000,0.09178811785759904,0.13018627072806063,0.0375661898593943,0.056640063007639366,0.04663888727682823,0.07652489659469304,0.15624140341294881,0.11134392058751942
5
+ 0,4000,0.022261641891210437,0.0917250894138089,-0.0032208746581034514,-0.01755423729789252,0.031819789953348665,0.05254201137664035,0.0984977047210223,0.0569594965415818
6
+ 0,5000,-0.03234222690147609,0.012956854547048487,0.043243785113505864,0.053620628098953765,0.11241124697815322,0.1310273303818104,0.07201460695465578,0.0003974660338104291
7
+ 0,6000,0.10732756373517004,0.09216061070617566,-0.06477532262169992,-0.04413602759775596,-0.11980084415708929,-0.0779194872819892,0.15117625196053378,0.11561802583891032
8
+ 0,7000,-0.015434569368277777,0.09654273138841829,0.006952299232054829,0.10749380473196307,-0.006192276795040167,0.07389331780604101,0.06795022972242805,0.08323415399713717
9
+ 0,8000,0.038855917169002836,0.11389900446250777,0.04410049822020306,0.07170725462008366,0.011684274348693515,0.036586479334664246,0.0956215677502508,0.1471681415614041
10
+ 0,9000,0.06802973499873043,0.0856623870102711,0.01696305375867538,0.011466549118534914,-0.012749684390103305,0.013824053166271241,0.15155220231419234,0.1381621147662933
11
+ 0,10000,-0.034441350183650146,-0.013730260407809204,-0.012714524337353351,-0.01709872738486993,-0.0035184471403324107,-0.0151848169570729,0.03857051958790033,0.07242761375679617
12
+ 0,11000,nan,nan,0.24928280738809921,0.31627571627535195,0.2508603854853241,0.31979063833837185,0.05797329849630246,0.05705381413486732
13
+ 0,-1,0.1360848883523377,0.20132277443030458,0.23908813771295576,0.3030728479356162,0.2519776031046535,0.3163199218787255,0.028848715141106332,-2.7676594189283463e-05
14
+ 1,1000,0.19974244809375838,0.20245347177311732,0.25347067411442387,0.27351275315803203,0.28700116921788826,0.3208008324745941,0.05359383925846385,0.03508662462308365
15
+ 1,2000,0.20437569550743273,0.21554549906475048,0.3608987761626178,0.394714523297842,0.3948207897031582,0.4549367777649391,0.051173556649996156,0.06522848870950099
16
+ 1,3000,0.2631493723275757,0.276235406787417,0.4045803960265086,0.42943437298049136,0.42728551392421416,0.45886646370656786,-0.0023035696166686715,-0.010497113109601252
17
+ 1,4000,0.23542301041102986,0.32665809541073476,0.30511558493590046,0.33477825986474385,0.307419217616696,0.338778290375219,-0.07456120142942858,-0.10527664585225535
18
+ 1,5000,0.22715013854098173,0.22442419297203847,0.28112469497802683,0.3017370699206343,0.31544004947898346,0.3402824340795712,0.07690988331726949,0.10043422772427255
19
+ 1,6000,0.12234388596115063,0.1500378024615469,0.24740733370350537,0.2754516493616489,0.2675203855244018,0.2931135176938421,-0.04981811535406503,-0.064512045244046
20
+ 1,7000,0.13512547997506566,0.21076810559817868,0.20033756137917422,0.22941016806021952,0.24559472730749363,0.2831572626244766,-0.07190949553486115,-0.05545523517203271
21
+ 1,8000,0.17026289109828452,0.21895191252036092,0.19625284091267522,0.20682533309842266,0.22066654420979503,0.24205758338365352,0.07134276863068595,0.056243398163656116
22
+ 1,9000,0.19810861824151754,0.2499129391551433,0.24922470027007046,0.299364190010855,0.24549427070862212,0.2950012891561661,-0.04002332092012268,-0.040637865797701886
23
+ 1,10000,0.18267236143336868,0.2463119449834827,0.23819637269285435,0.2777322740991704,0.23763524569113523,0.27792485677125844,-0.09354891051908217,-0.08436279086484923
24
+ 1,11000,0.10139322839101546,0.10723969129068593,0.29142426202860766,0.3157717723968941,0.2881089293041709,0.3102180110269781,-0.05869636451238082,-0.0338016885732243
25
+ 1,-1,0.24780231171173409,0.2727846024654737,0.2656441993441646,0.26692342747952297,0.2685682084798254,0.2739075284159849,-0.0613956833471493,-0.05814560792062389
26
+ 2,1000,0.2195269237560483,0.24426968445451466,0.24677160591543668,0.24507432751648156,0.2445487644626471,0.24143985812607768,0.0032952509381747822,0.023508539874027477
27
+ 2,2000,0.22450009092448303,0.24070997271959851,0.2459170273362594,0.25004111535290763,0.2462912257634435,0.2508168275929748,-0.03397917581627379,-0.018032265569433326
28
+ 2,3000,0.0636266513903195,0.0601773345712711,0.16956732342008235,0.17083197764206545,0.20597003291615185,0.2448744413099231,-0.06806763504468355,-0.0389467414593623
29
+ 2,4000,0.17759054381465422,0.19689770276424018,0.20337448467202895,0.2068979840465756,0.23045681563655787,0.2476659290638219,-0.13732921863992542,-0.11093020289608005
30
+ 2,5000,0.1104813651449169,0.11192963874049457,0.15213093857364401,0.16750733187182715,0.1788790768350145,0.20123659164235896,-0.026238978840947027,0.0005792862698172362
31
+ 2,6000,0.09345181681029505,0.09444257712328269,0.21865567808058647,0.2254693346186097,0.2568021586578848,0.275207941948268,-0.08778929192120052,-0.1041550562779488
32
+ 2,7000,nan,nan,0.24746828010279465,0.2789304381488678,0.2530309380333205,0.28746928613113953,0.02501830037195193,0.008639221471884102
33
+ 2,8000,nan,nan,0.2651235620836535,0.30092983716337907,0.2736578316013363,0.3077175114622423,0.06578650715311402,0.061964031159856735
34
+ 2,9000,nan,nan,0.2893845580329779,0.35166787552582957,0.2892389701709721,0.35177665974978345,0.022204879676934425,0.038772091227920545
35
+ 2,10000,0.12793884899455713,0.10371639860628296,0.1500750641001991,0.16450557920449094,0.16127890534887113,0.20212031931327792,-0.10435786183417026,-0.12147648023309993
36
+ 2,11000,0.03425146607622923,0.03915151887769445,0.181143122261767,0.2096698675763696,0.20088348297476613,0.23894743088891443,0.12483395821429048,0.1096038415779373
37
+ 2,-1,0.08282333890004473,0.08720482073031116,0.21042172821058988,0.24784275147731594,0.2234104685279609,0.2731102899690577,0.14393432678315235,0.1292149565421456
38
+ 3,1000,0.13412626679697037,0.1397081120101355,0.16492867494474775,0.1868013479598802,0.16160366403736806,0.19398956346496246,-0.10382512755659874,-0.09745744040596176
39
+ 3,2000,0.13061616407905768,0.12899899343891316,0.16478886237940665,0.18717190623511545,0.16057934547739405,0.1818937571923195,-0.08363591654500684,-0.07758593535289267
40
+ 3,3000,0.14727344031272707,0.1471302042040616,0.1476725066189474,0.15609421387389152,0.15256539684106704,0.16697840221754945,-0.11739930251229268,-0.11234339476087372
41
+ 3,4000,0.13904577959889142,0.10463202397122667,0.16026113148538945,0.17045988178062588,0.15847070810311137,0.16783022497472921,-0.06151952554759519,-0.055748305824548254
42
+ 3,5000,0.14600082993828567,0.12008244768152448,0.15449547264105815,0.15601272180506381,0.1544225631538293,0.15906290843783535,-0.06975847640678581,-0.0648064004229899
43
+ 3,6000,0.14049035861868384,0.13813762630180662,0.14291691978395213,0.13391030451486471,0.1429372271620174,0.13444538451395963,-0.07838725447438305,-0.06593834512494824
44
+ 3,7000,0.13229453981796363,0.1317648615187537,0.1375325801487268,0.13436504563478519,0.13749135338958518,0.13535179158139193,-0.09049782554828013,-0.07767388771310883
45
+ 3,8000,nan,nan,0.22036534815282474,0.2605301440420616,0.21755144768039322,0.25690643775508776,0.06428424636861306,0.03922417803782557
46
+ 3,9000,0.017512162961327713,0.013701787902915465,0.17478134821172578,0.19788273173423904,0.17365562886057204,0.1970628138907984,-0.0957667555786681,-0.07321146834649905
47
+ 3,10000,0.12487572514454401,0.14454880553640054,0.16549027191003318,0.1704668009185452,0.16550610732097198,0.17153811410638822,0.07525952817625017,0.06329983953892442
48
+ 3,11000,0.143833546686662,0.1900547192571339,0.1728691736059246,0.18617362839197651,0.17280563986072367,0.1866479737360017,0.09266033001429355,0.07932590867989191
49
+ 3,-1,nan,nan,0.18890767276343418,0.2284276504756746,0.18816471427111228,0.2275308533220191,0.05474655643961899,0.04304837603699167
50
+ 4,1000,0.12046759499768138,0.15750867125363832,0.1597201321461287,0.18639504080539507,0.16034966712713009,0.18760665873438007,0.1038372384437287,0.08500356396362752
51
+ 4,2000,0.12122984854514621,0.14301977070746208,0.18936652326106462,0.2210337828157678,0.1881517671146057,0.21886117350909803,0.10599993434181572,0.08785804683256576
52
+ 4,3000,nan,nan,0.18927856471908644,0.22614318177264228,0.1879346703933802,0.22278739988176702,0.08515038451209997,0.07536425979003279
53
+ 4,4000,0.12188511099942974,0.16307707797714088,0.17442229965066974,0.19948028380050226,0.20153354291941036,0.2344323090000002,0.07590199569770542,0.08015508393508045
54
+ 4,5000,0.15463962628706768,0.17964773156464098,0.18709649999151967,0.21810506549313563,0.17450455991486272,0.20380243862077735,0.09722117190247603,0.08254348848653233
55
+ 4,6000,0.13061992096856206,0.17242607775223368,0.18331727792333924,0.21112058016012258,0.17835350881125064,0.20845671206118035,0.10778480301833271,0.09321791131803152
56
+ 4,7000,0.14628930788768713,0.14418912166846365,0.21709857967530524,0.23932375511241577,0.2236476730859727,0.24908781190624432,0.0442229980092774,0.05443679604525362
57
+ 4,8000,0.14101433834477534,0.17130812357357073,0.17069387710640505,0.18612596321964334,0.1636616129167783,0.17844072297402255,-0.09097595905430293,-0.07712767481304125
58
+ 4,9000,0.14318552507145513,0.1663340317279959,0.16939403915453485,0.18856918769826914,0.15851091114906635,0.18748288304493427,-0.0725133509336175,-0.06538799262809976
59
+ 4,10000,0.13522866862669775,0.15149789709850744,0.18174808727019096,0.20298828672560282,0.20907780837922346,0.23726377399521068,0.03383215519744723,0.02811055697274588
60
+ 4,11000,0.12703543997981906,0.13783225826535966,0.18267901661878108,0.20266001207098566,0.2146341424783029,0.24048040433459722,0.0387828672724821,0.038845962261876726
61
+ 4,-1,0.12619531361769684,0.13869287384449375,0.1830195242390722,0.20605346482386622,0.21563401213261063,0.24139103976409135,0.039992442895886694,0.040008623403086134
result/eval/translation_evaluation_TED2020-en-ja-dev.tsv.gz_results.csv ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,src2trg,trg2src
2
+ 0,1000,0.014,0.022
3
+ 0,2000,0.005,0.012
4
+ 0,3000,0.009,0.013
5
+ 0,4000,0.007,0.009
6
+ 0,5000,0.007,0.008
7
+ 0,6000,0.006,0.004
8
+ 0,7000,0.005,0.005
9
+ 0,8000,0.006,0.007
10
+ 0,9000,0.005,0.005
11
+ 0,10000,0.005,0.006
12
+ 0,11000,0.0,0.001
13
+ 0,-1,0.002,0.002
14
+ 1,1000,0.0,0.001
15
+ 1,2000,0.0,0.0
16
+ 1,3000,0.002,0.002
17
+ 1,4000,0.0,0.002
18
+ 1,5000,0.001,0.001
19
+ 1,6000,0.002,0.0
20
+ 1,7000,0.002,0.004
21
+ 1,8000,0.001,0.002
22
+ 1,9000,0.002,0.003
23
+ 1,10000,0.001,0.001
24
+ 1,11000,0.002,0.0
25
+ 1,-1,0.001,0.001
26
+ 2,1000,0.001,0.001
27
+ 2,2000,0.0,0.001
28
+ 2,3000,0.002,0.001
29
+ 2,4000,0.0,0.001
30
+ 2,5000,0.002,0.0
31
+ 2,6000,0.0,0.002
32
+ 2,7000,0.001,0.001
33
+ 2,8000,0.002,0.0
34
+ 2,9000,0.001,0.001
35
+ 2,10000,0.003,0.004
36
+ 2,11000,0.001,0.002
37
+ 2,-1,0.0,0.0
38
+ 3,1000,0.002,0.002
39
+ 3,2000,0.0,0.004
40
+ 3,3000,0.002,0.001
41
+ 3,4000,0.001,0.004
42
+ 3,5000,0.003,0.002
43
+ 3,6000,0.002,0.002
44
+ 3,7000,0.001,0.001
45
+ 3,8000,0.001,0.002
46
+ 3,9000,0.001,0.002
47
+ 3,10000,0.0,0.002
48
+ 3,11000,0.002,0.001
49
+ 3,-1,0.001,0.0
50
+ 4,1000,0.001,0.002
51
+ 4,2000,0.001,0.002
52
+ 4,3000,0.0,0.002
53
+ 4,4000,0.003,0.003
54
+ 4,5000,0.001,0.003
55
+ 4,6000,0.002,0.001
56
+ 4,7000,0.001,0.001
57
+ 4,8000,0.001,0.001
58
+ 4,9000,0.003,0.002
59
+ 4,10000,0.002,0.002
60
+ 4,11000,0.0,0.002
61
+ 4,-1,0.003,0.001
result/eval/translation_evaluation_TED2020-en-ko-dev.tsv.gz_results.csv ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,src2trg,trg2src
2
+ 0,1000,0.024,0.011
3
+ 0,2000,0.015,0.011
4
+ 0,3000,0.008,0.013
5
+ 0,4000,0.009,0.011
6
+ 0,5000,0.011,0.008
7
+ 0,6000,0.008,0.008
8
+ 0,7000,0.002,0.005
9
+ 0,8000,0.016,0.013
10
+ 0,9000,0.01,0.01
11
+ 0,10000,0.008,0.004
12
+ 0,11000,0.0,0.0
13
+ 0,-1,0.002,0.002
14
+ 1,1000,0.001,0.001
15
+ 1,2000,0.0,0.001
16
+ 1,3000,0.001,0.004
17
+ 1,4000,0.002,0.001
18
+ 1,5000,0.001,0.0
19
+ 1,6000,0.001,0.0
20
+ 1,7000,0.001,0.002
21
+ 1,8000,0.0,0.002
22
+ 1,9000,0.001,0.002
23
+ 1,10000,0.003,0.003
24
+ 1,11000,0.0,0.001
25
+ 1,-1,0.001,0.001
26
+ 2,1000,0.001,0.001
27
+ 2,2000,0.002,0.003
28
+ 2,3000,0.0,0.001
29
+ 2,4000,0.002,0.003
30
+ 2,5000,0.002,0.003
31
+ 2,6000,0.0,0.0
32
+ 2,7000,0.002,0.001
33
+ 2,8000,0.002,0.001
34
+ 2,9000,0.0,0.002
35
+ 2,10000,0.001,0.0
36
+ 2,11000,0.001,0.0
37
+ 2,-1,0.001,0.001
38
+ 3,1000,0.002,0.001
39
+ 3,2000,0.002,0.0
40
+ 3,3000,0.001,0.001
41
+ 3,4000,0.003,0.001
42
+ 3,5000,0.001,0.002
43
+ 3,6000,0.001,0.001
44
+ 3,7000,0.001,0.001
45
+ 3,8000,0.001,0.0
46
+ 3,9000,0.0,0.0
47
+ 3,10000,0.003,0.002
48
+ 3,11000,0.001,0.001
49
+ 3,-1,0.001,0.0
50
+ 4,1000,0.001,0.002
51
+ 4,2000,0.0,0.001
52
+ 4,3000,0.0,0.0
53
+ 4,4000,0.0,0.002
54
+ 4,5000,0.001,0.002
55
+ 4,6000,0.001,0.002
56
+ 4,7000,0.001,0.001
57
+ 4,8000,0.001,0.002
58
+ 4,9000,0.004,0.003
59
+ 4,10000,0.001,0.001
60
+ 4,11000,0.002,0.002
61
+ 4,-1,0.001,0.001
result/modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
result/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96d6864fdb1d944caebd240504a6d774a08361669e406bebe1a93b0c04f5adea
3
+ size 1112245805
result/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 128,
3
+ "do_lower_case": false
4
+ }
result/sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
+ size 5069051
result/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
result/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b60b6b43406a48bf3638526314f3d232d97058bc93472ff2de930d43686fa441
3
+ size 17082913
result/tokenizer_config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "__type": "AddedToken",
7
+ "content": "<mask>",
8
+ "lstrip": true,
9
+ "normalized": true,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "model_max_length": 512,
14
+ "name_or_path": "xlm-roberta-base",
15
+ "pad_token": "<pad>",
16
+ "sep_token": "</s>",
17
+ "special_tokens_map_file": null,
18
+ "tokenizer_class": "XLMRobertaTokenizer",
19
+ "unk_token": "<unk>"
20
+ }