Upload folder using huggingface_hub
Browse files- 1_Pooling/config.json +1 -1
- README.md +4 -4
- config.json +19 -12
- eval/similarity_evaluation_results.csv +25 -45
- model.safetensors +2 -2
- special_tokens_map.json +5 -49
- tokenizer.json +0 -0
- tokenizer_config.json +17 -25
- vocab.txt +0 -5
1_Pooling/config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"word_embedding_dimension":
|
3 |
"pooling_mode_cls_token": false,
|
4 |
"pooling_mode_mean_tokens": true,
|
5 |
"pooling_mode_max_tokens": false,
|
|
|
1 |
{
|
2 |
+
"word_embedding_dimension": 384,
|
3 |
"pooling_mode_cls_token": false,
|
4 |
"pooling_mode_mean_tokens": true,
|
5 |
"pooling_mode_max_tokens": false,
|
README.md
CHANGED
@@ -10,7 +10,7 @@ tags:
|
|
10 |
|
11 |
# {MODEL_NAME}
|
12 |
|
13 |
-
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a
|
14 |
|
15 |
<!--- Describe your model here -->
|
16 |
|
@@ -98,7 +98,7 @@ Parameters of the fit()-Method:
|
|
98 |
```
|
99 |
{
|
100 |
"epochs": 5,
|
101 |
-
"evaluation_steps":
|
102 |
"evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
|
103 |
"max_grad_norm": 1,
|
104 |
"optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
|
@@ -116,8 +116,8 @@ Parameters of the fit()-Method:
|
|
116 |
## Full Model Architecture
|
117 |
```
|
118 |
SentenceTransformer(
|
119 |
-
(0): Transformer({'max_seq_length': 256, 'do_lower_case': False}) with Transformer model:
|
120 |
-
(1): Pooling({'word_embedding_dimension':
|
121 |
)
|
122 |
```
|
123 |
|
|
|
10 |
|
11 |
# {MODEL_NAME}
|
12 |
|
13 |
+
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.
|
14 |
|
15 |
<!--- Describe your model here -->
|
16 |
|
|
|
98 |
```
|
99 |
{
|
100 |
"epochs": 5,
|
101 |
+
"evaluation_steps": 4000,
|
102 |
"evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
|
103 |
"max_grad_norm": 1,
|
104 |
"optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
|
|
|
116 |
## Full Model Architecture
|
117 |
```
|
118 |
SentenceTransformer(
|
119 |
+
(0): Transformer({'max_seq_length': 256, 'do_lower_case': False}) with Transformer model: BertModel
|
120 |
+
(1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
|
121 |
)
|
122 |
```
|
123 |
|
config.json
CHANGED
@@ -1,24 +1,31 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
-
"
|
5 |
],
|
6 |
"attention_probs_dropout_prob": 0.1,
|
7 |
-
"
|
8 |
-
"eos_token_id": 2,
|
9 |
"hidden_act": "gelu",
|
10 |
"hidden_dropout_prob": 0.1,
|
11 |
-
"hidden_size":
|
|
|
|
|
|
|
12 |
"initializer_range": 0.02,
|
13 |
-
"intermediate_size":
|
14 |
-
"
|
15 |
-
|
16 |
-
|
|
|
|
|
|
|
17 |
"num_attention_heads": 12,
|
18 |
"num_hidden_layers": 12,
|
19 |
-
"pad_token_id":
|
20 |
-
"
|
21 |
"torch_dtype": "float32",
|
22 |
"transformers_version": "4.35.2",
|
23 |
-
"
|
|
|
|
|
24 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "BAAI/bge-small-en-v1.5",
|
3 |
"architectures": [
|
4 |
+
"BertModel"
|
5 |
],
|
6 |
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"classifier_dropout": null,
|
|
|
8 |
"hidden_act": "gelu",
|
9 |
"hidden_dropout_prob": 0.1,
|
10 |
+
"hidden_size": 384,
|
11 |
+
"id2label": {
|
12 |
+
"0": "LABEL_0"
|
13 |
+
},
|
14 |
"initializer_range": 0.02,
|
15 |
+
"intermediate_size": 1536,
|
16 |
+
"label2id": {
|
17 |
+
"LABEL_0": 0
|
18 |
+
},
|
19 |
+
"layer_norm_eps": 1e-12,
|
20 |
+
"max_position_embeddings": 512,
|
21 |
+
"model_type": "bert",
|
22 |
"num_attention_heads": 12,
|
23 |
"num_hidden_layers": 12,
|
24 |
+
"pad_token_id": 0,
|
25 |
+
"position_embedding_type": "absolute",
|
26 |
"torch_dtype": "float32",
|
27 |
"transformers_version": "4.35.2",
|
28 |
+
"type_vocab_size": 2,
|
29 |
+
"use_cache": true,
|
30 |
+
"vocab_size": 30522
|
31 |
}
|
eval/similarity_evaluation_results.csv
CHANGED
@@ -1,46 +1,26 @@
|
|
1 |
epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
|
2 |
-
0,
|
3 |
-
0,
|
4 |
-
0,
|
5 |
-
0,
|
6 |
-
0,
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
1,
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
1,
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
2,16000,0.8544573177596062,0.8580720727688758,0.8037740803189763,0.8373612490337308,0.8039765512515591,0.8362046535013997,0.8535004359962781,0.8582169400534592
|
28 |
-
2,-1,0.8564469620604049,0.8583573088500128,0.8092963756732118,0.8394582862307234,0.8095295258357031,0.8383808437647896,0.8537510545413085,0.8576369640121451
|
29 |
-
3,2000,0.858335492426398,0.8598730511158054,0.8109701818577197,0.8388938096664709,0.8115046245911267,0.8388572580320608,0.857496614165324,0.8603503131250824
|
30 |
-
3,4000,0.8585720362404455,0.8608787200229299,0.8094691553741892,0.8406964015823338,0.8100279877484762,0.8401941017763299,0.8560533825031614,0.8599173187451638
|
31 |
-
3,6000,0.8625065340034024,0.8643877033313818,0.8147635598777612,0.8439370969066217,0.815024717502594,0.8432979741313638,0.8610900947663982,0.8640921848075493
|
32 |
-
3,8000,0.8636424980300432,0.865122090592666,0.8173109962050679,0.845700575720838,0.8182284738793051,0.8459612367985541,0.8613495822247048,0.8641951639601717
|
33 |
-
3,10000,0.8650135805176045,0.8673640283370629,0.816476323494592,0.8466196160872734,0.8169465400233672,0.8462270727279998,0.8634553032248831,0.8670114328838054
|
34 |
-
3,12000,0.8660554514616089,0.8675529876534124,0.820467698385613,0.8491929868057942,0.8206915904174914,0.848456541205093,0.8637250810237758,0.8665606617213012
|
35 |
-
3,14000,0.869624306329827,0.8718296409176024,0.8209503097190902,0.8506831769725806,0.8219506623807201,0.8510081930208988,0.8674419506031797,0.8709930565184041
|
36 |
-
3,16000,0.8704396511213788,0.8735845147650085,0.8213771401822823,0.8531587401746457,0.821977454760251,0.8527795805352297,0.8698618513314226,0.8733708916919545
|
37 |
-
3,-1,0.8700336133151265,0.8726758833679342,0.8204336806274579,0.8520600699024211,0.8215392375394879,0.8524818449200231,0.8686024548768527,0.8720263802394267
|
38 |
-
4,2000,0.87206925200801,0.8739807045524346,0.8261164267655331,0.8544884307137518,0.8271211846423038,0.8547683535320902,0.8705259947562897,0.8734331721011405
|
39 |
-
4,4000,0.8729561300627455,0.8755517209603371,0.8236263307824584,0.8544608040818665,0.8243894920098801,0.8546285432975391,0.8718757908696867,0.8752564718789628
|
40 |
-
4,6000,0.8733050986710387,0.8763342021631335,0.8229918605011646,0.8544769984685667,0.8241693828088917,0.8550171857019131,0.872459812828531,0.8762565503867752
|
41 |
-
4,8000,0.8747918752753512,0.8776918490177567,0.8266396911478837,0.8576142866877868,0.8273975347484257,0.8575592743630189,0.8737816762278383,0.8772673195672135
|
42 |
-
4,10000,0.8756954383154001,0.8785020324188557,0.826701497021203,0.8580297266019579,0.8275847946482253,0.8580523517597078,0.8740733998601564,0.8777094913541181
|
43 |
-
4,12000,0.8766918428682116,0.8791700693786121,0.8287402810383075,0.8591466484870882,0.8295960716735256,0.8592141541249873,0.8749246498622072,0.8780804557469476
|
44 |
-
4,14000,0.877598616280475,0.8800782071546932,0.829796344766758,0.8600729219172725,0.8306199404272812,0.8601873712190522,0.8759751366450554,0.8789839862502832
|
45 |
-
4,16000,0.8776365492614989,0.8802169350974743,0.8292031255117502,0.8601198002110558,0.830114079320727,0.8603507576377725,0.8759709022137859,0.8791947439523577
|
46 |
-
4,-1,0.8776310070951263,0.8802108509348683,0.8291788384293556,0.860103499714235,0.8300926509841282,0.8603385907137475,0.8759596196632796,0.8791872747352656
|
|
|
1 |
epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
|
2 |
+
0,4000,0.6492103513942427,0.6461809089063794,0.6186867187342933,0.6279257077690662,0.6171589154700341,0.6249954915763268,0.6491827341356649,0.6486156848184234
|
3 |
+
0,8000,0.7260485755206463,0.7246891679751163,0.6932865780842696,0.709693419316205,0.6917438681237527,0.7067058888933562,0.7234659680584743,0.7263276839104037
|
4 |
+
0,12000,0.7583795751489272,0.7573109135435032,0.725038775991098,0.7420763443961781,0.723438986499572,0.7391145649787624,0.757461889695176,0.7596567728219538
|
5 |
+
0,16000,0.7812544580714823,0.7803253408644298,0.7478438028046986,0.767227970225704,0.7460679625936297,0.7640873841989747,0.7775656752902136,0.7812982157860213
|
6 |
+
0,-1,0.7809172008215644,0.7812307342801713,0.7456654131604474,0.7701041693637447,0.7438002358737863,0.7668367422590563,0.7770500745147031,0.781040428822645
|
7 |
+
1,4000,0.7955133778715422,0.7952389528260638,0.7591031710579657,0.7813510328272457,0.7573112856005579,0.7780296701014116,0.7911175981195924,0.7955248246571255
|
8 |
+
1,8000,0.8068296582778156,0.8075955109647435,0.7714895994449444,0.7935610617853895,0.7693372797954768,0.7896408198733824,0.8039918570718593,0.8083538825037162
|
9 |
+
1,12000,0.8143849033955202,0.8142774773615219,0.7777244679556524,0.8005788018244105,0.7758861733414191,0.796939410712197,0.8112363014128732,0.8148742901045252
|
10 |
+
1,16000,0.8216469783249275,0.8223234293667148,0.7829877004893209,0.8067297168438895,0.7817642319516476,0.8036011486806538,0.8192870144997304,0.8236616052031184
|
11 |
+
1,-1,0.8232039576074661,0.8239193036009659,0.7889922177258312,0.8115745320531896,0.7872370252018468,0.8078649112367716,0.8201829041372247,0.8234662592135515
|
12 |
+
2,4000,0.8272624725490654,0.8276880756085873,0.7884759709724365,0.8134267545353865,0.7871934987010161,0.8102372217909231,0.8228011267530628,0.8275540950773731
|
13 |
+
2,8000,0.8334669695360342,0.833370132452343,0.7961106465938914,0.8198214000594783,0.7944658688740782,0.8163468351919869,0.8299611851120992,0.8333746894621054
|
14 |
+
2,12000,0.8385052418817065,0.8395632350790249,0.8007618075524696,0.8255042220501557,0.7986791310118732,0.8213920270394287,0.8360037280001015,0.8401393619804501
|
15 |
+
2,16000,0.8438698251969813,0.8449661776546094,0.8059452724221107,0.8304284840595941,0.8040993077695543,0.8266920054765993,0.841134818068614,0.845432716823413
|
16 |
+
2,-1,0.8416046524522067,0.8422627460400357,0.8026235500864117,0.8270961724160747,0.8011903415392386,0.8239186741110563,0.8389866181716672,0.8431939392966612
|
17 |
+
3,4000,0.8454394690566859,0.8465221221328377,0.80500968483825,0.8290282221771081,0.8028380508889708,0.8246808146056436,0.8424489493340934,0.8475153448438146
|
18 |
+
3,8000,0.8480118151128676,0.8485837737398271,0.8077584367843311,0.8328793565435811,0.806367847470503,0.8295274035642068,0.845295491376847,0.848876350683858
|
19 |
+
3,12000,0.851577391805856,0.8533076282253811,0.810407531023261,0.8371000866394858,0.8087163674687277,0.8332922448486583,0.8482931916768949,0.8536316665201856
|
20 |
+
3,16000,0.8526886847009655,0.8539274897752376,0.8119862938997584,0.8377248270177977,0.8103723333079371,0.8340612996896535,0.8494097260716742,0.8541982762125481
|
21 |
+
3,-1,0.852671368793424,0.8540857683864328,0.8116077576016788,0.8378700680247607,0.8099990764449612,0.8342036842439218,0.8492698088111232,0.8543436165293636
|
22 |
+
4,4000,0.8554723329228191,0.8569701076701354,0.8159138243492129,0.8409225397371998,0.8141857324415636,0.8371256524519275,0.8526372409892025,0.8571867730070265
|
23 |
+
4,8000,0.8559374022998232,0.857378667442124,0.8144072114305754,0.8407414375006325,0.8130947387263745,0.8374179377990042,0.8529746996865721,0.8576263132989826
|
24 |
+
4,12000,0.8571570134265576,0.8586257689308819,0.8159794673348664,0.8415337438693111,0.8143748001626152,0.837914216798177,0.8544632177135479,0.8590811074826755
|
25 |
+
4,16000,0.8580448525387094,0.8596755546755923,0.8169314432511898,0.8424529467813953,0.8152350188024375,0.8386913753530236,0.8554719215010795,0.8602197777893075
|
26 |
+
4,-1,0.8580444084725619,0.8596746908465527,0.8169306593712456,0.8424501034905559,0.8152343993486576,0.8386886463373756,0.8554723650635132,0.8602203436786403
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:99deaea72d14f80608eabc015212bc6c0f532e126831c011c93128cd54dafbee
|
3 |
+
size 133462128
|
special_tokens_map.json
CHANGED
@@ -1,51 +1,7 @@
|
|
1 |
{
|
2 |
-
"
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
"single_word": false
|
8 |
-
},
|
9 |
-
"cls_token": {
|
10 |
-
"content": "<s>",
|
11 |
-
"lstrip": false,
|
12 |
-
"normalized": true,
|
13 |
-
"rstrip": false,
|
14 |
-
"single_word": false
|
15 |
-
},
|
16 |
-
"eos_token": {
|
17 |
-
"content": "</s>",
|
18 |
-
"lstrip": false,
|
19 |
-
"normalized": false,
|
20 |
-
"rstrip": false,
|
21 |
-
"single_word": false
|
22 |
-
},
|
23 |
-
"mask_token": {
|
24 |
-
"content": "<mask>",
|
25 |
-
"lstrip": true,
|
26 |
-
"normalized": false,
|
27 |
-
"rstrip": false,
|
28 |
-
"single_word": false
|
29 |
-
},
|
30 |
-
"pad_token": {
|
31 |
-
"content": "<pad>",
|
32 |
-
"lstrip": false,
|
33 |
-
"normalized": false,
|
34 |
-
"rstrip": false,
|
35 |
-
"single_word": false
|
36 |
-
},
|
37 |
-
"sep_token": {
|
38 |
-
"content": "</s>",
|
39 |
-
"lstrip": false,
|
40 |
-
"normalized": true,
|
41 |
-
"rstrip": false,
|
42 |
-
"single_word": false
|
43 |
-
},
|
44 |
-
"unk_token": {
|
45 |
-
"content": "[UNK]",
|
46 |
-
"lstrip": false,
|
47 |
-
"normalized": false,
|
48 |
-
"rstrip": false,
|
49 |
-
"single_word": false
|
50 |
-
}
|
51 |
}
|
|
|
1 |
{
|
2 |
+
"cls_token": "[CLS]",
|
3 |
+
"mask_token": "[MASK]",
|
4 |
+
"pad_token": "[PAD]",
|
5 |
+
"sep_token": "[SEP]",
|
6 |
+
"unk_token": "[UNK]"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
}
|
tokenizer.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
CHANGED
@@ -1,65 +1,57 @@
|
|
1 |
{
|
2 |
"added_tokens_decoder": {
|
3 |
"0": {
|
4 |
-
"content": "
|
5 |
"lstrip": false,
|
6 |
"normalized": false,
|
7 |
"rstrip": false,
|
8 |
"single_word": false,
|
9 |
"special": true
|
10 |
},
|
11 |
-
"
|
12 |
-
"content": "
|
13 |
"lstrip": false,
|
14 |
"normalized": false,
|
15 |
"rstrip": false,
|
16 |
"single_word": false,
|
17 |
"special": true
|
18 |
},
|
19 |
-
"
|
20 |
-
"content": "
|
21 |
"lstrip": false,
|
22 |
"normalized": false,
|
23 |
"rstrip": false,
|
24 |
"single_word": false,
|
25 |
"special": true
|
26 |
},
|
27 |
-
"
|
28 |
-
"content": "
|
29 |
-
"lstrip": false,
|
30 |
-
"normalized": true,
|
31 |
-
"rstrip": false,
|
32 |
-
"single_word": false,
|
33 |
-
"special": true
|
34 |
-
},
|
35 |
-
"104": {
|
36 |
-
"content": "[UNK]",
|
37 |
"lstrip": false,
|
38 |
"normalized": false,
|
39 |
"rstrip": false,
|
40 |
"single_word": false,
|
41 |
"special": true
|
42 |
},
|
43 |
-
"
|
44 |
-
"content": "
|
45 |
-
"lstrip":
|
46 |
"normalized": false,
|
47 |
"rstrip": false,
|
48 |
"single_word": false,
|
49 |
"special": true
|
50 |
}
|
51 |
},
|
52 |
-
"bos_token": "<s>",
|
53 |
"clean_up_tokenization_spaces": true,
|
54 |
-
"cls_token": "
|
|
|
55 |
"do_lower_case": true,
|
56 |
-
"
|
57 |
-
"mask_token": "<mask>",
|
58 |
"model_max_length": 512,
|
59 |
-
"
|
60 |
-
"
|
|
|
61 |
"strip_accents": null,
|
62 |
"tokenize_chinese_chars": true,
|
63 |
-
"tokenizer_class": "
|
64 |
"unk_token": "[UNK]"
|
65 |
}
|
|
|
1 |
{
|
2 |
"added_tokens_decoder": {
|
3 |
"0": {
|
4 |
+
"content": "[PAD]",
|
5 |
"lstrip": false,
|
6 |
"normalized": false,
|
7 |
"rstrip": false,
|
8 |
"single_word": false,
|
9 |
"special": true
|
10 |
},
|
11 |
+
"100": {
|
12 |
+
"content": "[UNK]",
|
13 |
"lstrip": false,
|
14 |
"normalized": false,
|
15 |
"rstrip": false,
|
16 |
"single_word": false,
|
17 |
"special": true
|
18 |
},
|
19 |
+
"101": {
|
20 |
+
"content": "[CLS]",
|
21 |
"lstrip": false,
|
22 |
"normalized": false,
|
23 |
"rstrip": false,
|
24 |
"single_word": false,
|
25 |
"special": true
|
26 |
},
|
27 |
+
"102": {
|
28 |
+
"content": "[SEP]",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
29 |
"lstrip": false,
|
30 |
"normalized": false,
|
31 |
"rstrip": false,
|
32 |
"single_word": false,
|
33 |
"special": true
|
34 |
},
|
35 |
+
"103": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
"normalized": false,
|
39 |
"rstrip": false,
|
40 |
"single_word": false,
|
41 |
"special": true
|
42 |
}
|
43 |
},
|
|
|
44 |
"clean_up_tokenization_spaces": true,
|
45 |
+
"cls_token": "[CLS]",
|
46 |
+
"do_basic_tokenize": true,
|
47 |
"do_lower_case": true,
|
48 |
+
"mask_token": "[MASK]",
|
|
|
49 |
"model_max_length": 512,
|
50 |
+
"never_split": null,
|
51 |
+
"pad_token": "[PAD]",
|
52 |
+
"sep_token": "[SEP]",
|
53 |
"strip_accents": null,
|
54 |
"tokenize_chinese_chars": true,
|
55 |
+
"tokenizer_class": "BertTokenizer",
|
56 |
"unk_token": "[UNK]"
|
57 |
}
|
vocab.txt
CHANGED
@@ -1,7 +1,3 @@
|
|
1 |
-
<s>
|
2 |
-
<pad>
|
3 |
-
</s>
|
4 |
-
<unk>
|
5 |
[PAD]
|
6 |
[unused0]
|
7 |
[unused1]
|
@@ -30524,4 +30520,3 @@ necessitated
|
|
30524 |
##:
|
30525 |
##?
|
30526 |
##~
|
30527 |
-
<mask>
|
|
|
|
|
|
|
|
|
|
|
1 |
[PAD]
|
2 |
[unused0]
|
3 |
[unused1]
|
|
|
30520 |
##:
|
30521 |
##?
|
30522 |
##~
|
|