Sieun Park
commited on
Commit
·
e087677
1
Parent(s):
3960360
Upload ./result with huggingface_hub
Browse files- .gitattributes +1 -0
- result/1_Pooling/config.json +7 -0
- result/README.md +127 -0
- result/config.json +28 -0
- result/config_sentence_transformers.json +7 -0
- result/eval/mse_evaluation_TED2020-en-ja-dev.tsv.gz_results.csv +61 -0
- result/eval/mse_evaluation_TED2020-en-ko-dev.tsv.gz_results.csv +61 -0
- result/eval/similarity_evaluation_STS.en-en.txt_results.csv +61 -0
- result/eval/translation_evaluation_TED2020-en-ja-dev.tsv.gz_results.csv +61 -0
- result/eval/translation_evaluation_TED2020-en-ko-dev.tsv.gz_results.csv +61 -0
- result/modules.json +14 -0
- result/pytorch_model.bin +3 -0
- result/sentence_bert_config.json +4 -0
- result/sentencepiece.bpe.model +3 -0
- result/special_tokens_map.json +15 -0
- result/tokenizer.json +3 -0
- result/tokenizer_config.json +20 -0
.gitattributes
CHANGED
@@ -32,3 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
35 |
+
result/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
result/1_Pooling/config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"word_embedding_dimension": 768,
|
3 |
+
"pooling_mode_cls_token": false,
|
4 |
+
"pooling_mode_mean_tokens": true,
|
5 |
+
"pooling_mode_max_tokens": false,
|
6 |
+
"pooling_mode_mean_sqrt_len_tokens": false
|
7 |
+
}
|
result/README.md
ADDED
@@ -0,0 +1,127 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
pipeline_tag: sentence-similarity
|
3 |
+
tags:
|
4 |
+
- sentence-transformers
|
5 |
+
- feature-extraction
|
6 |
+
- sentence-similarity
|
7 |
+
- transformers
|
8 |
+
|
9 |
+
---
|
10 |
+
|
11 |
+
# {MODEL_NAME}
|
12 |
+
|
13 |
+
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
|
14 |
+
|
15 |
+
<!--- Describe your model here -->
|
16 |
+
|
17 |
+
## Usage (Sentence-Transformers)
|
18 |
+
|
19 |
+
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
|
20 |
+
|
21 |
+
```
|
22 |
+
pip install -U sentence-transformers
|
23 |
+
```
|
24 |
+
|
25 |
+
Then you can use the model like this:
|
26 |
+
|
27 |
+
```python
|
28 |
+
from sentence_transformers import SentenceTransformer
|
29 |
+
sentences = ["This is an example sentence", "Each sentence is converted"]
|
30 |
+
|
31 |
+
model = SentenceTransformer('{MODEL_NAME}')
|
32 |
+
embeddings = model.encode(sentences)
|
33 |
+
print(embeddings)
|
34 |
+
```
|
35 |
+
|
36 |
+
|
37 |
+
|
38 |
+
## Usage (HuggingFace Transformers)
|
39 |
+
Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
|
40 |
+
|
41 |
+
```python
|
42 |
+
from transformers import AutoTokenizer, AutoModel
|
43 |
+
import torch
|
44 |
+
|
45 |
+
|
46 |
+
#Mean Pooling - Take attention mask into account for correct averaging
|
47 |
+
def mean_pooling(model_output, attention_mask):
|
48 |
+
token_embeddings = model_output[0] #First element of model_output contains all token embeddings
|
49 |
+
input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
|
50 |
+
return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
|
51 |
+
|
52 |
+
|
53 |
+
# Sentences we want sentence embeddings for
|
54 |
+
sentences = ['This is an example sentence', 'Each sentence is converted']
|
55 |
+
|
56 |
+
# Load model from HuggingFace Hub
|
57 |
+
tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}')
|
58 |
+
model = AutoModel.from_pretrained('{MODEL_NAME}')
|
59 |
+
|
60 |
+
# Tokenize sentences
|
61 |
+
encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
|
62 |
+
|
63 |
+
# Compute token embeddings
|
64 |
+
with torch.no_grad():
|
65 |
+
model_output = model(**encoded_input)
|
66 |
+
|
67 |
+
# Perform pooling. In this case, mean pooling.
|
68 |
+
sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
|
69 |
+
|
70 |
+
print("Sentence embeddings:")
|
71 |
+
print(sentence_embeddings)
|
72 |
+
```
|
73 |
+
|
74 |
+
|
75 |
+
|
76 |
+
## Evaluation Results
|
77 |
+
|
78 |
+
<!--- Describe how your model was evaluated -->
|
79 |
+
|
80 |
+
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
|
81 |
+
|
82 |
+
|
83 |
+
## Training
|
84 |
+
The model was trained with the parameters:
|
85 |
+
|
86 |
+
**DataLoader**:
|
87 |
+
|
88 |
+
`torch.utils.data.dataloader.DataLoader` of length 11258 with parameters:
|
89 |
+
```
|
90 |
+
{'batch_size': 128, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
|
91 |
+
```
|
92 |
+
|
93 |
+
**Loss**:
|
94 |
+
|
95 |
+
`sentence_transformers.losses.MSELoss.MSELoss`
|
96 |
+
|
97 |
+
Parameters of the fit()-Method:
|
98 |
+
```
|
99 |
+
{
|
100 |
+
"epochs": 5,
|
101 |
+
"evaluation_steps": 1000,
|
102 |
+
"evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator",
|
103 |
+
"max_grad_norm": 1,
|
104 |
+
"optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
|
105 |
+
"optimizer_params": {
|
106 |
+
"eps": 1e-06,
|
107 |
+
"lr": 2e-05
|
108 |
+
},
|
109 |
+
"scheduler": "WarmupLinear",
|
110 |
+
"steps_per_epoch": null,
|
111 |
+
"warmup_steps": 10000,
|
112 |
+
"weight_decay": 0.01
|
113 |
+
}
|
114 |
+
```
|
115 |
+
|
116 |
+
|
117 |
+
## Full Model Architecture
|
118 |
+
```
|
119 |
+
SentenceTransformer(
|
120 |
+
(0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
|
121 |
+
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
|
122 |
+
)
|
123 |
+
```
|
124 |
+
|
125 |
+
## Citing & Authors
|
126 |
+
|
127 |
+
<!--- Describe where people can find more information -->
|
result/config.json
ADDED
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "xlm-roberta-base",
|
3 |
+
"architectures": [
|
4 |
+
"XLMRobertaModel"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"classifier_dropout": null,
|
9 |
+
"eos_token_id": 2,
|
10 |
+
"hidden_act": "gelu",
|
11 |
+
"hidden_dropout_prob": 0.1,
|
12 |
+
"hidden_size": 768,
|
13 |
+
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 3072,
|
15 |
+
"layer_norm_eps": 1e-05,
|
16 |
+
"max_position_embeddings": 514,
|
17 |
+
"model_type": "xlm-roberta",
|
18 |
+
"num_attention_heads": 12,
|
19 |
+
"num_hidden_layers": 12,
|
20 |
+
"output_past": true,
|
21 |
+
"pad_token_id": 1,
|
22 |
+
"position_embedding_type": "absolute",
|
23 |
+
"torch_dtype": "float32",
|
24 |
+
"transformers_version": "4.26.0",
|
25 |
+
"type_vocab_size": 1,
|
26 |
+
"use_cache": true,
|
27 |
+
"vocab_size": 250002
|
28 |
+
}
|
result/config_sentence_transformers.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"__version__": {
|
3 |
+
"sentence_transformers": "2.2.2",
|
4 |
+
"transformers": "4.26.0",
|
5 |
+
"pytorch": "1.13.1+cu116"
|
6 |
+
}
|
7 |
+
}
|
result/eval/mse_evaluation_TED2020-en-ja-dev.tsv.gz_results.csv
ADDED
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
epoch,steps,MSE
|
2 |
+
0,1000,2.479737065732479
|
3 |
+
0,2000,0.7978680543601513
|
4 |
+
0,3000,0.465529365465045
|
5 |
+
0,4000,0.38496602792292833
|
6 |
+
0,5000,0.3095171647146344
|
7 |
+
0,6000,1.007575262337923
|
8 |
+
0,7000,2.9757926240563393
|
9 |
+
0,8000,0.1680491492152214
|
10 |
+
0,9000,0.15370475593954325
|
11 |
+
0,10000,0.14578887494280934
|
12 |
+
0,11000,44.72245275974274
|
13 |
+
0,-1,38.11988830566406
|
14 |
+
1,1000,34.88417565822601
|
15 |
+
1,2000,40.527331829071045
|
16 |
+
1,3000,34.18007791042328
|
17 |
+
1,4000,36.79800629615784
|
18 |
+
1,5000,18.30635964870453
|
19 |
+
1,6000,19.035635888576508
|
20 |
+
1,7000,10.42214184999466
|
21 |
+
1,8000,21.36164754629135
|
22 |
+
1,9000,13.449710607528687
|
23 |
+
1,10000,26.817166805267334
|
24 |
+
1,11000,27.421659231185913
|
25 |
+
1,-1,27.256697416305542
|
26 |
+
2,1000,19.7702556848526
|
27 |
+
2,2000,23.021529614925385
|
28 |
+
2,3000,22.112300992012024
|
29 |
+
2,4000,18.621844053268433
|
30 |
+
2,5000,15.16069769859314
|
31 |
+
2,6000,18.26171875
|
32 |
+
2,7000,22.869695723056793
|
33 |
+
2,8000,21.816536784172058
|
34 |
+
2,9000,21.53472751379013
|
35 |
+
2,10000,7.719540596008301
|
36 |
+
2,11000,16.72600358724594
|
37 |
+
2,-1,17.56407469511032
|
38 |
+
3,1000,9.689188748598099
|
39 |
+
3,2000,12.322552502155304
|
40 |
+
3,3000,14.71262276172638
|
41 |
+
3,4000,15.390068292617798
|
42 |
+
3,5000,11.171068251132965
|
43 |
+
3,6000,16.961337625980377
|
44 |
+
3,7000,17.228218913078308
|
45 |
+
3,8000,15.841719508171082
|
46 |
+
3,9000,15.5427947640419
|
47 |
+
3,10000,11.266006529331207
|
48 |
+
3,11000,11.172742396593094
|
49 |
+
3,-1,10.830702632665634
|
50 |
+
4,1000,14.59239274263382
|
51 |
+
4,2000,11.720868200063705
|
52 |
+
4,3000,10.161346197128296
|
53 |
+
4,4000,25.874224305152893
|
54 |
+
4,5000,19.239723682403564
|
55 |
+
4,6000,16.438673436641693
|
56 |
+
4,7000,14.188717305660248
|
57 |
+
4,8000,18.68770271539688
|
58 |
+
4,9000,26.96780264377594
|
59 |
+
4,10000,29.64048683643341
|
60 |
+
4,11000,29.76890504360199
|
61 |
+
4,-1,29.792797565460205
|
result/eval/mse_evaluation_TED2020-en-ko-dev.tsv.gz_results.csv
ADDED
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
epoch,steps,MSE
|
2 |
+
0,1000,3.534962609410286
|
3 |
+
0,2000,0.7726760115474463
|
4 |
+
0,3000,0.45291599817574024
|
5 |
+
0,4000,0.32619184348732233
|
6 |
+
0,5000,0.3069185884669423
|
7 |
+
0,6000,2.836661972105503
|
8 |
+
0,7000,0.4042532294988632
|
9 |
+
0,8000,0.1647747354581952
|
10 |
+
0,9000,0.15147579833865166
|
11 |
+
0,10000,0.14654598198831081
|
12 |
+
0,11000,44.76562440395355
|
13 |
+
0,-1,38.15537393093109
|
14 |
+
1,1000,34.89996790885925
|
15 |
+
1,2000,40.527620911598206
|
16 |
+
1,3000,34.18329954147339
|
17 |
+
1,4000,36.78698241710663
|
18 |
+
1,5000,18.310919404029846
|
19 |
+
1,6000,19.035445153713226
|
20 |
+
1,7000,10.420460999011993
|
21 |
+
1,8000,21.34985774755478
|
22 |
+
1,9000,13.454011082649231
|
23 |
+
1,10000,26.8216073513031
|
24 |
+
1,11000,27.417752146720886
|
25 |
+
1,-1,27.26142108440399
|
26 |
+
2,1000,19.804176688194275
|
27 |
+
2,2000,23.02238494157791
|
28 |
+
2,3000,22.109846770763397
|
29 |
+
2,4000,18.61446350812912
|
30 |
+
2,5000,15.156088769435883
|
31 |
+
2,6000,18.26195865869522
|
32 |
+
2,7000,22.867940366268158
|
33 |
+
2,8000,21.814395487308502
|
34 |
+
2,9000,21.53337448835373
|
35 |
+
2,10000,7.718393951654434
|
36 |
+
2,11000,16.722606122493744
|
37 |
+
2,-1,17.560648918151855
|
38 |
+
3,1000,9.690673649311066
|
39 |
+
3,2000,12.321917712688446
|
40 |
+
3,3000,14.71351683139801
|
41 |
+
3,4000,15.393638610839844
|
42 |
+
3,5000,11.174257099628448
|
43 |
+
3,6000,16.96329414844513
|
44 |
+
3,7000,17.23286509513855
|
45 |
+
3,8000,15.840160846710205
|
46 |
+
3,9000,15.541520714759827
|
47 |
+
3,10000,11.266209185123444
|
48 |
+
3,11000,11.172308027744293
|
49 |
+
3,-1,10.831181704998016
|
50 |
+
4,1000,14.588256180286407
|
51 |
+
4,2000,11.7195263504982
|
52 |
+
4,3000,10.161814093589783
|
53 |
+
4,4000,25.86996555328369
|
54 |
+
4,5000,19.233083724975586
|
55 |
+
4,6000,16.433295607566833
|
56 |
+
4,7000,14.18774127960205
|
57 |
+
4,8000,18.69051158428192
|
58 |
+
4,9000,26.973900198936462
|
59 |
+
4,10000,29.64228391647339
|
60 |
+
4,11000,29.770362377166748
|
61 |
+
4,-1,29.794207215309143
|
result/eval/similarity_evaluation_STS.en-en.txt_results.csv
ADDED
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
|
2 |
+
0,1000,0.1303136936057836,0.16109944136542498,0.09244556100421478,0.12335554401547714,0.09157986691726851,0.16468932075590426,0.15038989729175398,0.12561348935648534
|
3 |
+
0,2000,0.10912365007333646,0.16055590464220645,0.04794033150727805,0.056331392577127014,0.12450082278387518,0.1120412159311675,0.18294141836020666,0.1578924209398153
|
4 |
+
0,3000,0.09178811785759904,0.13018627072806063,0.0375661898593943,0.056640063007639366,0.04663888727682823,0.07652489659469304,0.15624140341294881,0.11134392058751942
|
5 |
+
0,4000,0.022261641891210437,0.0917250894138089,-0.0032208746581034514,-0.01755423729789252,0.031819789953348665,0.05254201137664035,0.0984977047210223,0.0569594965415818
|
6 |
+
0,5000,-0.03234222690147609,0.012956854547048487,0.043243785113505864,0.053620628098953765,0.11241124697815322,0.1310273303818104,0.07201460695465578,0.0003974660338104291
|
7 |
+
0,6000,0.10732756373517004,0.09216061070617566,-0.06477532262169992,-0.04413602759775596,-0.11980084415708929,-0.0779194872819892,0.15117625196053378,0.11561802583891032
|
8 |
+
0,7000,-0.015434569368277777,0.09654273138841829,0.006952299232054829,0.10749380473196307,-0.006192276795040167,0.07389331780604101,0.06795022972242805,0.08323415399713717
|
9 |
+
0,8000,0.038855917169002836,0.11389900446250777,0.04410049822020306,0.07170725462008366,0.011684274348693515,0.036586479334664246,0.0956215677502508,0.1471681415614041
|
10 |
+
0,9000,0.06802973499873043,0.0856623870102711,0.01696305375867538,0.011466549118534914,-0.012749684390103305,0.013824053166271241,0.15155220231419234,0.1381621147662933
|
11 |
+
0,10000,-0.034441350183650146,-0.013730260407809204,-0.012714524337353351,-0.01709872738486993,-0.0035184471403324107,-0.0151848169570729,0.03857051958790033,0.07242761375679617
|
12 |
+
0,11000,nan,nan,0.24928280738809921,0.31627571627535195,0.2508603854853241,0.31979063833837185,0.05797329849630246,0.05705381413486732
|
13 |
+
0,-1,0.1360848883523377,0.20132277443030458,0.23908813771295576,0.3030728479356162,0.2519776031046535,0.3163199218787255,0.028848715141106332,-2.7676594189283463e-05
|
14 |
+
1,1000,0.19974244809375838,0.20245347177311732,0.25347067411442387,0.27351275315803203,0.28700116921788826,0.3208008324745941,0.05359383925846385,0.03508662462308365
|
15 |
+
1,2000,0.20437569550743273,0.21554549906475048,0.3608987761626178,0.394714523297842,0.3948207897031582,0.4549367777649391,0.051173556649996156,0.06522848870950099
|
16 |
+
1,3000,0.2631493723275757,0.276235406787417,0.4045803960265086,0.42943437298049136,0.42728551392421416,0.45886646370656786,-0.0023035696166686715,-0.010497113109601252
|
17 |
+
1,4000,0.23542301041102986,0.32665809541073476,0.30511558493590046,0.33477825986474385,0.307419217616696,0.338778290375219,-0.07456120142942858,-0.10527664585225535
|
18 |
+
1,5000,0.22715013854098173,0.22442419297203847,0.28112469497802683,0.3017370699206343,0.31544004947898346,0.3402824340795712,0.07690988331726949,0.10043422772427255
|
19 |
+
1,6000,0.12234388596115063,0.1500378024615469,0.24740733370350537,0.2754516493616489,0.2675203855244018,0.2931135176938421,-0.04981811535406503,-0.064512045244046
|
20 |
+
1,7000,0.13512547997506566,0.21076810559817868,0.20033756137917422,0.22941016806021952,0.24559472730749363,0.2831572626244766,-0.07190949553486115,-0.05545523517203271
|
21 |
+
1,8000,0.17026289109828452,0.21895191252036092,0.19625284091267522,0.20682533309842266,0.22066654420979503,0.24205758338365352,0.07134276863068595,0.056243398163656116
|
22 |
+
1,9000,0.19810861824151754,0.2499129391551433,0.24922470027007046,0.299364190010855,0.24549427070862212,0.2950012891561661,-0.04002332092012268,-0.040637865797701886
|
23 |
+
1,10000,0.18267236143336868,0.2463119449834827,0.23819637269285435,0.2777322740991704,0.23763524569113523,0.27792485677125844,-0.09354891051908217,-0.08436279086484923
|
24 |
+
1,11000,0.10139322839101546,0.10723969129068593,0.29142426202860766,0.3157717723968941,0.2881089293041709,0.3102180110269781,-0.05869636451238082,-0.0338016885732243
|
25 |
+
1,-1,0.24780231171173409,0.2727846024654737,0.2656441993441646,0.26692342747952297,0.2685682084798254,0.2739075284159849,-0.0613956833471493,-0.05814560792062389
|
26 |
+
2,1000,0.2195269237560483,0.24426968445451466,0.24677160591543668,0.24507432751648156,0.2445487644626471,0.24143985812607768,0.0032952509381747822,0.023508539874027477
|
27 |
+
2,2000,0.22450009092448303,0.24070997271959851,0.2459170273362594,0.25004111535290763,0.2462912257634435,0.2508168275929748,-0.03397917581627379,-0.018032265569433326
|
28 |
+
2,3000,0.0636266513903195,0.0601773345712711,0.16956732342008235,0.17083197764206545,0.20597003291615185,0.2448744413099231,-0.06806763504468355,-0.0389467414593623
|
29 |
+
2,4000,0.17759054381465422,0.19689770276424018,0.20337448467202895,0.2068979840465756,0.23045681563655787,0.2476659290638219,-0.13732921863992542,-0.11093020289608005
|
30 |
+
2,5000,0.1104813651449169,0.11192963874049457,0.15213093857364401,0.16750733187182715,0.1788790768350145,0.20123659164235896,-0.026238978840947027,0.0005792862698172362
|
31 |
+
2,6000,0.09345181681029505,0.09444257712328269,0.21865567808058647,0.2254693346186097,0.2568021586578848,0.275207941948268,-0.08778929192120052,-0.1041550562779488
|
32 |
+
2,7000,nan,nan,0.24746828010279465,0.2789304381488678,0.2530309380333205,0.28746928613113953,0.02501830037195193,0.008639221471884102
|
33 |
+
2,8000,nan,nan,0.2651235620836535,0.30092983716337907,0.2736578316013363,0.3077175114622423,0.06578650715311402,0.061964031159856735
|
34 |
+
2,9000,nan,nan,0.2893845580329779,0.35166787552582957,0.2892389701709721,0.35177665974978345,0.022204879676934425,0.038772091227920545
|
35 |
+
2,10000,0.12793884899455713,0.10371639860628296,0.1500750641001991,0.16450557920449094,0.16127890534887113,0.20212031931327792,-0.10435786183417026,-0.12147648023309993
|
36 |
+
2,11000,0.03425146607622923,0.03915151887769445,0.181143122261767,0.2096698675763696,0.20088348297476613,0.23894743088891443,0.12483395821429048,0.1096038415779373
|
37 |
+
2,-1,0.08282333890004473,0.08720482073031116,0.21042172821058988,0.24784275147731594,0.2234104685279609,0.2731102899690577,0.14393432678315235,0.1292149565421456
|
38 |
+
3,1000,0.13412626679697037,0.1397081120101355,0.16492867494474775,0.1868013479598802,0.16160366403736806,0.19398956346496246,-0.10382512755659874,-0.09745744040596176
|
39 |
+
3,2000,0.13061616407905768,0.12899899343891316,0.16478886237940665,0.18717190623511545,0.16057934547739405,0.1818937571923195,-0.08363591654500684,-0.07758593535289267
|
40 |
+
3,3000,0.14727344031272707,0.1471302042040616,0.1476725066189474,0.15609421387389152,0.15256539684106704,0.16697840221754945,-0.11739930251229268,-0.11234339476087372
|
41 |
+
3,4000,0.13904577959889142,0.10463202397122667,0.16026113148538945,0.17045988178062588,0.15847070810311137,0.16783022497472921,-0.06151952554759519,-0.055748305824548254
|
42 |
+
3,5000,0.14600082993828567,0.12008244768152448,0.15449547264105815,0.15601272180506381,0.1544225631538293,0.15906290843783535,-0.06975847640678581,-0.0648064004229899
|
43 |
+
3,6000,0.14049035861868384,0.13813762630180662,0.14291691978395213,0.13391030451486471,0.1429372271620174,0.13444538451395963,-0.07838725447438305,-0.06593834512494824
|
44 |
+
3,7000,0.13229453981796363,0.1317648615187537,0.1375325801487268,0.13436504563478519,0.13749135338958518,0.13535179158139193,-0.09049782554828013,-0.07767388771310883
|
45 |
+
3,8000,nan,nan,0.22036534815282474,0.2605301440420616,0.21755144768039322,0.25690643775508776,0.06428424636861306,0.03922417803782557
|
46 |
+
3,9000,0.017512162961327713,0.013701787902915465,0.17478134821172578,0.19788273173423904,0.17365562886057204,0.1970628138907984,-0.0957667555786681,-0.07321146834649905
|
47 |
+
3,10000,0.12487572514454401,0.14454880553640054,0.16549027191003318,0.1704668009185452,0.16550610732097198,0.17153811410638822,0.07525952817625017,0.06329983953892442
|
48 |
+
3,11000,0.143833546686662,0.1900547192571339,0.1728691736059246,0.18617362839197651,0.17280563986072367,0.1866479737360017,0.09266033001429355,0.07932590867989191
|
49 |
+
3,-1,nan,nan,0.18890767276343418,0.2284276504756746,0.18816471427111228,0.2275308533220191,0.05474655643961899,0.04304837603699167
|
50 |
+
4,1000,0.12046759499768138,0.15750867125363832,0.1597201321461287,0.18639504080539507,0.16034966712713009,0.18760665873438007,0.1038372384437287,0.08500356396362752
|
51 |
+
4,2000,0.12122984854514621,0.14301977070746208,0.18936652326106462,0.2210337828157678,0.1881517671146057,0.21886117350909803,0.10599993434181572,0.08785804683256576
|
52 |
+
4,3000,nan,nan,0.18927856471908644,0.22614318177264228,0.1879346703933802,0.22278739988176702,0.08515038451209997,0.07536425979003279
|
53 |
+
4,4000,0.12188511099942974,0.16307707797714088,0.17442229965066974,0.19948028380050226,0.20153354291941036,0.2344323090000002,0.07590199569770542,0.08015508393508045
|
54 |
+
4,5000,0.15463962628706768,0.17964773156464098,0.18709649999151967,0.21810506549313563,0.17450455991486272,0.20380243862077735,0.09722117190247603,0.08254348848653233
|
55 |
+
4,6000,0.13061992096856206,0.17242607775223368,0.18331727792333924,0.21112058016012258,0.17835350881125064,0.20845671206118035,0.10778480301833271,0.09321791131803152
|
56 |
+
4,7000,0.14628930788768713,0.14418912166846365,0.21709857967530524,0.23932375511241577,0.2236476730859727,0.24908781190624432,0.0442229980092774,0.05443679604525362
|
57 |
+
4,8000,0.14101433834477534,0.17130812357357073,0.17069387710640505,0.18612596321964334,0.1636616129167783,0.17844072297402255,-0.09097595905430293,-0.07712767481304125
|
58 |
+
4,9000,0.14318552507145513,0.1663340317279959,0.16939403915453485,0.18856918769826914,0.15851091114906635,0.18748288304493427,-0.0725133509336175,-0.06538799262809976
|
59 |
+
4,10000,0.13522866862669775,0.15149789709850744,0.18174808727019096,0.20298828672560282,0.20907780837922346,0.23726377399521068,0.03383215519744723,0.02811055697274588
|
60 |
+
4,11000,0.12703543997981906,0.13783225826535966,0.18267901661878108,0.20266001207098566,0.2146341424783029,0.24048040433459722,0.0387828672724821,0.038845962261876726
|
61 |
+
4,-1,0.12619531361769684,0.13869287384449375,0.1830195242390722,0.20605346482386622,0.21563401213261063,0.24139103976409135,0.039992442895886694,0.040008623403086134
|
result/eval/translation_evaluation_TED2020-en-ja-dev.tsv.gz_results.csv
ADDED
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
epoch,steps,src2trg,trg2src
|
2 |
+
0,1000,0.014,0.022
|
3 |
+
0,2000,0.005,0.012
|
4 |
+
0,3000,0.009,0.013
|
5 |
+
0,4000,0.007,0.009
|
6 |
+
0,5000,0.007,0.008
|
7 |
+
0,6000,0.006,0.004
|
8 |
+
0,7000,0.005,0.005
|
9 |
+
0,8000,0.006,0.007
|
10 |
+
0,9000,0.005,0.005
|
11 |
+
0,10000,0.005,0.006
|
12 |
+
0,11000,0.0,0.001
|
13 |
+
0,-1,0.002,0.002
|
14 |
+
1,1000,0.0,0.001
|
15 |
+
1,2000,0.0,0.0
|
16 |
+
1,3000,0.002,0.002
|
17 |
+
1,4000,0.0,0.002
|
18 |
+
1,5000,0.001,0.001
|
19 |
+
1,6000,0.002,0.0
|
20 |
+
1,7000,0.002,0.004
|
21 |
+
1,8000,0.001,0.002
|
22 |
+
1,9000,0.002,0.003
|
23 |
+
1,10000,0.001,0.001
|
24 |
+
1,11000,0.002,0.0
|
25 |
+
1,-1,0.001,0.001
|
26 |
+
2,1000,0.001,0.001
|
27 |
+
2,2000,0.0,0.001
|
28 |
+
2,3000,0.002,0.001
|
29 |
+
2,4000,0.0,0.001
|
30 |
+
2,5000,0.002,0.0
|
31 |
+
2,6000,0.0,0.002
|
32 |
+
2,7000,0.001,0.001
|
33 |
+
2,8000,0.002,0.0
|
34 |
+
2,9000,0.001,0.001
|
35 |
+
2,10000,0.003,0.004
|
36 |
+
2,11000,0.001,0.002
|
37 |
+
2,-1,0.0,0.0
|
38 |
+
3,1000,0.002,0.002
|
39 |
+
3,2000,0.0,0.004
|
40 |
+
3,3000,0.002,0.001
|
41 |
+
3,4000,0.001,0.004
|
42 |
+
3,5000,0.003,0.002
|
43 |
+
3,6000,0.002,0.002
|
44 |
+
3,7000,0.001,0.001
|
45 |
+
3,8000,0.001,0.002
|
46 |
+
3,9000,0.001,0.002
|
47 |
+
3,10000,0.0,0.002
|
48 |
+
3,11000,0.002,0.001
|
49 |
+
3,-1,0.001,0.0
|
50 |
+
4,1000,0.001,0.002
|
51 |
+
4,2000,0.001,0.002
|
52 |
+
4,3000,0.0,0.002
|
53 |
+
4,4000,0.003,0.003
|
54 |
+
4,5000,0.001,0.003
|
55 |
+
4,6000,0.002,0.001
|
56 |
+
4,7000,0.001,0.001
|
57 |
+
4,8000,0.001,0.001
|
58 |
+
4,9000,0.003,0.002
|
59 |
+
4,10000,0.002,0.002
|
60 |
+
4,11000,0.0,0.002
|
61 |
+
4,-1,0.003,0.001
|
result/eval/translation_evaluation_TED2020-en-ko-dev.tsv.gz_results.csv
ADDED
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
epoch,steps,src2trg,trg2src
|
2 |
+
0,1000,0.024,0.011
|
3 |
+
0,2000,0.015,0.011
|
4 |
+
0,3000,0.008,0.013
|
5 |
+
0,4000,0.009,0.011
|
6 |
+
0,5000,0.011,0.008
|
7 |
+
0,6000,0.008,0.008
|
8 |
+
0,7000,0.002,0.005
|
9 |
+
0,8000,0.016,0.013
|
10 |
+
0,9000,0.01,0.01
|
11 |
+
0,10000,0.008,0.004
|
12 |
+
0,11000,0.0,0.0
|
13 |
+
0,-1,0.002,0.002
|
14 |
+
1,1000,0.001,0.001
|
15 |
+
1,2000,0.0,0.001
|
16 |
+
1,3000,0.001,0.004
|
17 |
+
1,4000,0.002,0.001
|
18 |
+
1,5000,0.001,0.0
|
19 |
+
1,6000,0.001,0.0
|
20 |
+
1,7000,0.001,0.002
|
21 |
+
1,8000,0.0,0.002
|
22 |
+
1,9000,0.001,0.002
|
23 |
+
1,10000,0.003,0.003
|
24 |
+
1,11000,0.0,0.001
|
25 |
+
1,-1,0.001,0.001
|
26 |
+
2,1000,0.001,0.001
|
27 |
+
2,2000,0.002,0.003
|
28 |
+
2,3000,0.0,0.001
|
29 |
+
2,4000,0.002,0.003
|
30 |
+
2,5000,0.002,0.003
|
31 |
+
2,6000,0.0,0.0
|
32 |
+
2,7000,0.002,0.001
|
33 |
+
2,8000,0.002,0.001
|
34 |
+
2,9000,0.0,0.002
|
35 |
+
2,10000,0.001,0.0
|
36 |
+
2,11000,0.001,0.0
|
37 |
+
2,-1,0.001,0.001
|
38 |
+
3,1000,0.002,0.001
|
39 |
+
3,2000,0.002,0.0
|
40 |
+
3,3000,0.001,0.001
|
41 |
+
3,4000,0.003,0.001
|
42 |
+
3,5000,0.001,0.002
|
43 |
+
3,6000,0.001,0.001
|
44 |
+
3,7000,0.001,0.001
|
45 |
+
3,8000,0.001,0.0
|
46 |
+
3,9000,0.0,0.0
|
47 |
+
3,10000,0.003,0.002
|
48 |
+
3,11000,0.001,0.001
|
49 |
+
3,-1,0.001,0.0
|
50 |
+
4,1000,0.001,0.002
|
51 |
+
4,2000,0.0,0.001
|
52 |
+
4,3000,0.0,0.0
|
53 |
+
4,4000,0.0,0.002
|
54 |
+
4,5000,0.001,0.002
|
55 |
+
4,6000,0.001,0.002
|
56 |
+
4,7000,0.001,0.001
|
57 |
+
4,8000,0.001,0.002
|
58 |
+
4,9000,0.004,0.003
|
59 |
+
4,10000,0.001,0.001
|
60 |
+
4,11000,0.002,0.002
|
61 |
+
4,-1,0.001,0.001
|
result/modules.json
ADDED
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[
|
2 |
+
{
|
3 |
+
"idx": 0,
|
4 |
+
"name": "0",
|
5 |
+
"path": "",
|
6 |
+
"type": "sentence_transformers.models.Transformer"
|
7 |
+
},
|
8 |
+
{
|
9 |
+
"idx": 1,
|
10 |
+
"name": "1",
|
11 |
+
"path": "1_Pooling",
|
12 |
+
"type": "sentence_transformers.models.Pooling"
|
13 |
+
}
|
14 |
+
]
|
result/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:96d6864fdb1d944caebd240504a6d774a08361669e406bebe1a93b0c04f5adea
|
3 |
+
size 1112245805
|
result/sentence_bert_config.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"max_seq_length": 128,
|
3 |
+
"do_lower_case": false
|
4 |
+
}
|
result/sentencepiece.bpe.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
|
3 |
+
size 5069051
|
result/special_tokens_map.json
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "<s>",
|
3 |
+
"cls_token": "<s>",
|
4 |
+
"eos_token": "</s>",
|
5 |
+
"mask_token": {
|
6 |
+
"content": "<mask>",
|
7 |
+
"lstrip": true,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false
|
11 |
+
},
|
12 |
+
"pad_token": "<pad>",
|
13 |
+
"sep_token": "</s>",
|
14 |
+
"unk_token": "<unk>"
|
15 |
+
}
|
result/tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b60b6b43406a48bf3638526314f3d232d97058bc93472ff2de930d43686fa441
|
3 |
+
size 17082913
|
result/tokenizer_config.json
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "<s>",
|
3 |
+
"cls_token": "<s>",
|
4 |
+
"eos_token": "</s>",
|
5 |
+
"mask_token": {
|
6 |
+
"__type": "AddedToken",
|
7 |
+
"content": "<mask>",
|
8 |
+
"lstrip": true,
|
9 |
+
"normalized": true,
|
10 |
+
"rstrip": false,
|
11 |
+
"single_word": false
|
12 |
+
},
|
13 |
+
"model_max_length": 512,
|
14 |
+
"name_or_path": "xlm-roberta-base",
|
15 |
+
"pad_token": "<pad>",
|
16 |
+
"sep_token": "</s>",
|
17 |
+
"special_tokens_map_file": null,
|
18 |
+
"tokenizer_class": "XLMRobertaTokenizer",
|
19 |
+
"unk_token": "<unk>"
|
20 |
+
}
|