tartspuppy commited on
Commit
a7f195f
·
verified ·
1 Parent(s): 0bb08c8

Add new SentenceTransformer model

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,756 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sentence-transformers
4
+ - sentence-similarity
5
+ - feature-extraction
6
+ - generated_from_trainer
7
+ - dataset_size:21196
8
+ - loss:DenoisingAutoEncoderLoss
9
+ base_model: google-bert/bert-base-uncased
10
+ widget:
11
+ - source_sentence: Oldham also has Fencing, Marshall Club is a Competitive of its
12
+ competing on stage . They train times old South High.
13
+ sentences:
14
+ - Several leading batsmen opposed the new law , including the professional Herbert
15
+ <unk> , known as an exponent of pad @-@ play , and amateurs Errol Holmes and Bob
16
+ Wyatt . Wisden <unk> ' <unk> noted that these three improved their batting records
17
+ during the 1935 season , but batsmen generally were less successful . There were
18
+ also fewer drawn matches . There was an increase in the number of lbws — out of
19
+ 1 @,@ 560 lbw dismissals in first @-@ class matches in 1935 , 483 were given under
20
+ the amended law . Wisden judged the experiment a success and several of its opponents
21
+ changed their mind by the end of the season ; batsmen soon became accustomed to
22
+ the alteration . Although Australian authorities were less convinced , and did
23
+ not immediately introduce the revision into domestic first @-@ class cricket ,
24
+ in 1937 the new rule became part of the Laws of Cricket .
25
+ - Oldham also has a Fencing Club , Marshall Fencing Club is a Competitive Fencing
26
+ Club with most of its members competing on the national stage . They train three
27
+ times a week at the old South Chadderton High School .
28
+ - While overseeing an initial <unk> of American involvement in the Vietnam War ,
29
+ he subsequently ended U.S. involvement in 1973 , and eliminated the draft . <unk>
30
+ , his administration generally embraced policies that transferred power from Washington
31
+ to the states . Among other things , he initiated wars on cancer and drugs , imposed
32
+ wage and price controls , enforced <unk> of Southern schools and established the
33
+ Environmental Protection Agency . Though he presided over Apollo 11 and the subsequent
34
+ lunar landings , he later scaled back manned space exploration . In 1972 , he
35
+ was reelected by a landslide , the largest to that date . The Watergate scandal
36
+ , which would consume the larger part of his second term , resulted in his ultimate
37
+ resignation on August 4 , 1974 .
38
+ - source_sentence: Randy contributed tracks on the album, was the honest record's
39
+ made . 't care thought of the lyrics . They were only important her "Carey of
40
+ Def Leppard's song "Bringin Heartbreak . During the photo shoot for Charmbracelet,
41
+ Carey happened to Def Leppard's Vault (1995), which contains the song and decided
42
+ to cover In an interview with, Carey said the song is "an of her musical diversity
43
+ . Jackson on "My Saving ", which Carey said describes thoughts about and process
44
+ . While Capri Carey became ill cancer and she returned to New York to spend time
45
+ he after In his, Carey and produced the song Sunflowers Carey song represents
46
+ "his of the family is kind of hard to talk about . "be "for, and she sang it only
47
+ in studio DJ <unk> songs for the, but of them were
48
+ sentences:
49
+ - = = Taxonomy and phylogeny = =
50
+ - 'Outside Japan , Destiny 2 was released in China and South Korea by Sony Computer
51
+ Entertainment in 2003 : the Korean version was released on March 27 , and the
52
+ Chinese version was released on August 14 . The game was going to be part of a
53
+ world tour by Sony Computer Entertainment to promote the next generation of role
54
+ @-@ playing games , but the tensions between America and Iraq at the time and
55
+ the consequent risks of a terrorist attack caused them to cancel the trip . Asked
56
+ at the launch event whether an overseas version of the game was being developed
57
+ , producer Makoto <unk> said he was " not certain " . The PSP port was released
58
+ in South Korea by Namco Bandai Games ''s local branch on March 5 , 2007 . Neither
59
+ the original nor the port has been released in the west , making it one of three
60
+ mainline Tales titles to remain exclusive to Japan .'
61
+ - Randy Jackson contributed to four tracks on the album , and said it was " the
62
+ most real and honest record she 's made . She didn 't care what anyone thought
63
+ of the lyrics . They were only important to her . " Carey included a cover of
64
+ Def Leppard 's song " Bringin ' On the Heartbreak " . During the photo shoot for
65
+ Charmbracelet at Capri , Carey happened to listen to Def Leppard 's album Vault
66
+ ( 1995 ) , which contains the song , and decided to cover it . In an interview
67
+ with Billboard , Carey said that the song is " an example of her musical diversity
68
+ " . Jackson also worked on " My Saving Grace " , which Carey said describes her
69
+ thoughts about the writing , recording and mastering process . While working in
70
+ Capri , Carey 's father became ill with cancer and she returned to New York to
71
+ spend some time with him ; he died soon after . In his memory , Carey wrote and
72
+ produced the song " Sunflowers for Alfred Roy " . Carey said that the song represents
73
+ " his side of the family and is kind of hard to talk about . " The song proved
74
+ to be " very emotional " for Carey , and she sang it only once in the studio .
75
+ DJ <unk> also produced songs for the album , but none of them were included .
76
+ - source_sentence: = = = Allied =
77
+ sentences:
78
+ - = = = Allied planning = = =
79
+ - = = = English colonists = = =
80
+ - Germany advocated quick recognition of Croatia , stating that it wanted to stop
81
+ ongoing violence in Serb @-@ inhabited areas . It was opposed by France , the
82
+ United Kingdom , and the Netherlands , but the countries agreed to pursue a common
83
+ approach and avoid unilateral actions . On 10 October , two days after the Croatian
84
+ Parliament confirmed the declaration of independence , the EEC decided to postpone
85
+ any decision to recognize Croatia for two months , deciding to recognize Croatian
86
+ independence in two months if the war had not ended by then . As the deadline
87
+ expired , Germany presented its decision to recognize Croatia as its policy and
88
+ duty — a position supported by Italy and Denmark . France and the UK attempted
89
+ to prevent the recognition by drafting a United Nations resolution requesting
90
+ no unilateral actions which could <unk> the situation , but backed down during
91
+ the Security Council debate on 14 December , when Germany appeared determined
92
+ to defy the UN resolution . On 17 December , the EEC formally agreed to grant
93
+ Croatia diplomatic recognition on 15 January 1992 , relying on opinion of the
94
+ Badinter <unk> Committee . The Committee ruled that Croatia 's independence should
95
+ not be recognized immediately , because the new Croatian Constitution did not
96
+ provide protection of minorities required by the EEC . In response , the President
97
+ Franjo Tuđman gave written <unk> to Robert Badinter that the deficit would be
98
+ <unk> . The <unk> formally declared its separation from Croatia on 19 December
99
+ , but its statehood and independence were not recognized internationally . On
100
+ 26 December , Yugoslav authorities announced plans for a smaller state , which
101
+ could include the territory captured from Croatia , but the plan was rejected
102
+ by the UN General Assembly .
103
+ - source_sentence: During the night, the German and force Katia and was Oghratina
104
+ when Division to the Zealand Rifles Brigades and 5th Mounted Brigade were Oghratina
105
+ Despite by brigades to the enemy, they were forced to make a on strongly <unk>
106
+ positions which carefully artillery Meanwhile, the divisions Katia Abu Hamra and
107
+ Lawrence moved his forward from Kantara to The 3rd Light Brigade the right towards
108
+ <unk, but could make small progress, positions securely held by German and.
109
+ sentences:
110
+ - State Route 243 ( SR 243 ) , or the Banning @-@ Idyllwild <unk> Highway , is a
111
+ 30 @-@ mile ( 50 kilometer ) two @-@ lane highway that runs from Banning , California
112
+ ( in the north ) to Idyllwild , California ( in the south ) in Riverside County
113
+ , California . The highway is a connector between Interstate 10 ( I @-@ 10 ) and
114
+ SR 74 . Along its route , it provides access to the San <unk> National Forest
115
+ . A road from Banning to Idyllwild was planned around the turn of the twentieth
116
+ century , and was open by 1910 . The road was added to the state highway system
117
+ in 1970 .
118
+ - During the previous night , the German and Ottoman force evacuated Katia and was
119
+ moving towards Oghratina when Chauvel ordered the Anzac Mounted Division to continue
120
+ the attack . The New Zealand Mounted Rifles Brigades and the 5th Mounted Brigade
121
+ were ordered to capture Oghratina . Despite attempts by these two brigades to
122
+ turn the enemy flank , they were forced to make a frontal attack on strongly entrenched
123
+ <unk> in positions which favoured the defenders and which were supported by carefully
124
+ positioned artillery . Meanwhile , the two infantry divisions moved to garrison
125
+ Katia and Abu Hamra and Lawrence moved his headquarters forward from Kantara to
126
+ Romani . The 3rd Light Horse Brigade on the right advanced towards <unk> , but
127
+ could only make small progress , against positions securely held by German and
128
+ Ottoman forces .
129
+ - The current training ground is located at Bodymoor Heath near <unk> in north Warwickshire
130
+ , the site for which was purchased by former chairman Doug Ellis in the early
131
+ 1970s from a local farmer . Although Bodymoor Heath was state @-@ of @-@ the @-@
132
+ art in the 1970s , by the late 1990s the facilities had started to look dated
133
+ . In November 2005 , Ellis and Aston Villa <unk> announced a state of the art
134
+ GB £ 13 million redevelopment of Bodymoor in two phases . However , work on Bodymoor
135
+ was suspended by Ellis due to financial problems , and was left in an unfinished
136
+ state until new owner Randy Lerner made it one of his priorities to make the site
137
+ one of the best in world football . The new training ground was officially unveiled
138
+ on 6 May 2007 , by then manager Martin O 'Neill , then team captain Gareth Barry
139
+ and 1982 European Cup winning team captain Dennis Mortimer , with the Aston Villa
140
+ squad moving in for the 2007 – 08 season .
141
+ - source_sentence: album five @ -, in an with Billboard magazine, said it was previously
142
+ "something I wanted to revisit as been doing a while . "The medley a written whereas
143
+ McCartney had worked the Beatles' was made of "bits we had knocking . "The off
144
+ with Vintage "McCartney sat one to looking back [and looking back . about life
145
+ followed by the bass @ - @ led That Was Me, which is his school days and ",, "from
146
+ there . songs "Feet the Clouds "about the inactivity while is up of ", about the
147
+ life being a celebrity The final song medley, The End of ", written McCartney's
148
+ unk> playing on his, Jim's piano
149
+ sentences:
150
+ - Severe Tropical Storm Domoina in 1984 caused 100 year floods in South Africa and
151
+ record rainfall in Swaziland . The fourth named storm of the season , Domoina
152
+ developed on January 16 off the northeast coast of Madagascar . With a ridge to
153
+ the north , the storm tracked generally westward and later southwestward . On
154
+ January 21 , Domoina struck eastern Madagascar , the third storm in six weeks
155
+ to affect the nation ; collectively , the storms caused 42 deaths and $ 25 million
156
+ in damage ( 1984 USD ) . After crossing the country , Domoina strengthened in
157
+ the Mozambique Channel to peak 10 minute sustained winds of 95 km / h ( 60 mph
158
+ ) . On January 28 , the storm made landfall in southern Mozambique , and slowly
159
+ weakened over land . Domoina crossed into Swaziland and later eastern South Africa
160
+ before dissipating on February 2 .
161
+ - The album features a five song @-@ medley , which in an interview with Billboard
162
+ magazine , McCartney said that it was previously " something I wanted to revisit
163
+ " as " nobody had been doing that for a while . " The medley was a group of intentionally
164
+ written material , whereas McCartney had worked on the Beatles ' Abbey Road which
165
+ , however , was actually made up of " bits we had knocking around . " The medley
166
+ starts off with " Vintage Clothes " , which McCartney " sat down one day " to
167
+ write , that was " looking back , [ and ] looking back . " , about life . It was
168
+ followed by the bass @-@ led " That Was Me " , which is about his " school days
169
+ and teachers " , the medley , as McCartney stated , then " progressed from there
170
+ . " The next songs are " Feet in the Clouds " , about the inactivity while one
171
+ is growing up , and " House of Wax " , about the life of being a celebrity . The
172
+ final song in medley , " The End of the End " , was written at McCartney 's <unk>
173
+ Avenue home while playing on his father , Jim 's , piano .
174
+ - Varanasi grew as an important industrial centre , famous for its muslin and silk
175
+ <unk> , perfumes , ivory works , and sculpture . Buddha is believed to have founded
176
+ Buddhism here around <unk> BC when he gave his first sermon , " The Setting in
177
+ Motion of the Wheel of Dharma " , at nearby <unk> . The city 's religious importance
178
+ continued to grow in the 8th century , when Adi <unk> established the worship
179
+ of Shiva as an official sect of Varanasi . Despite the Muslim rule , Varanasi
180
+ remained the centre of activity for Hindu intellectuals and theologians during
181
+ the Middle Ages , which further contributed to its reputation as a cultural centre
182
+ of religion and education . <unk> Tulsidas wrote his epic poem on Lord Rama 's
183
+ life called Ram <unk> Manas in Varanasi . Several other major figures of the Bhakti
184
+ movement were born in Varanasi , including Kabir and Ravidas . Guru Nanak Dev
185
+ visited Varanasi for <unk> in <unk> , a trip that played a large role in the founding
186
+ of <unk> . In the 16th century , Varanasi experienced a cultural revival under
187
+ the Muslim Mughal emperor <unk> who invested in the city , and built two large
188
+ temples dedicated to Shiva and Vishnu , though much of modern Varanasi was built
189
+ during the 18th century , by the Maratha and <unk> kings . The kingdom of Benares
190
+ was given official status by the <unk> in 1737 , and continued as a dynasty @-@
191
+ governed area until Indian independence in 1947 . The city is governed by the
192
+ Varanasi Nagar Nigam ( Municipal Corporation ) and is represented in the Parliament
193
+ of India by the current Prime Minister of India <unk> <unk> , who won the <unk>
194
+ <unk> elections in 2014 by a huge margin . Silk weaving , carpets and crafts and
195
+ tourism employ a significant number of the local population , as do the <unk>
196
+ <unk> Works and Bharat Heavy <unk> Limited . Varanasi Hospital was established
197
+ in 1964 .
198
+ pipeline_tag: sentence-similarity
199
+ library_name: sentence-transformers
200
+ metrics:
201
+ - pearson_cosine
202
+ - spearman_cosine
203
+ model-index:
204
+ - name: SentenceTransformer based on google-bert/bert-base-uncased
205
+ results:
206
+ - task:
207
+ type: semantic-similarity
208
+ name: Semantic Similarity
209
+ dataset:
210
+ name: sts dev
211
+ type: sts-dev
212
+ metrics:
213
+ - type: pearson_cosine
214
+ value: 0.6552233601802461
215
+ name: Pearson Cosine
216
+ - type: spearman_cosine
217
+ value: 0.6640796604094039
218
+ name: Spearman Cosine
219
+ - task:
220
+ type: semantic-similarity
221
+ name: Semantic Similarity
222
+ dataset:
223
+ name: sts test
224
+ type: sts-test
225
+ metrics:
226
+ - type: pearson_cosine
227
+ value: 0.7355355958065635
228
+ name: Pearson Cosine
229
+ - type: spearman_cosine
230
+ value: 0.7320302276487962
231
+ name: Spearman Cosine
232
+ ---
233
+
234
+ # SentenceTransformer based on google-bert/bert-base-uncased
235
+
236
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
237
+
238
+ ## Model Details
239
+
240
+ ### Model Description
241
+ - **Model Type:** Sentence Transformer
242
+ - **Base model:** [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased) <!-- at revision 86b5e0934494bd15c9632b12f734a8a67f723594 -->
243
+ - **Maximum Sequence Length:** 75 tokens
244
+ - **Output Dimensionality:** 768 dimensions
245
+ - **Similarity Function:** Cosine Similarity
246
+ <!-- - **Training Dataset:** Unknown -->
247
+ <!-- - **Language:** Unknown -->
248
+ <!-- - **License:** Unknown -->
249
+
250
+ ### Model Sources
251
+
252
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
253
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
254
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
255
+
256
+ ### Full Model Architecture
257
+
258
+ ```
259
+ SentenceTransformer(
260
+ (0): Transformer({'max_seq_length': 75, 'do_lower_case': False}) with Transformer model: BertModel
261
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
262
+ )
263
+ ```
264
+
265
+ ## Usage
266
+
267
+ ### Direct Usage (Sentence Transformers)
268
+
269
+ First install the Sentence Transformers library:
270
+
271
+ ```bash
272
+ pip install -U sentence-transformers
273
+ ```
274
+
275
+ Then you can load this model and run inference.
276
+ ```python
277
+ from sentence_transformers import SentenceTransformer
278
+
279
+ # Download from the 🤗 Hub
280
+ model = SentenceTransformer("tartspuppy/bert-base-uncased-tsdae-encoder")
281
+ # Run inference
282
+ sentences = [
283
+ 'album five @ -, in an with Billboard magazine, said it was previously "something I wanted to revisit as been doing a while . "The medley a written whereas McCartney had worked the Beatles\' was made of "bits we had knocking . "The off with Vintage "McCartney sat one to looking back [and looking back . about life followed by the bass @ - @ led That Was Me, which is his school days and ",, "from there . songs "Feet the Clouds "about the inactivity while is up of ", about the life being a celebrity The final song medley, The End of ", written McCartney\'s unk> playing on his, Jim\'s piano',
284
+ 'The album features a five song @-@ medley , which in an interview with Billboard magazine , McCartney said that it was previously " something I wanted to revisit " as " nobody had been doing that for a while . " The medley was a group of intentionally written material , whereas McCartney had worked on the Beatles \' Abbey Road which , however , was actually made up of " bits we had knocking around . " The medley starts off with " Vintage Clothes " , which McCartney " sat down one day " to write , that was " looking back , [ and ] looking back . " , about life . It was followed by the bass @-@ led " That Was Me " , which is about his " school days and teachers " , the medley , as McCartney stated , then " progressed from there . " The next songs are " Feet in the Clouds " , about the inactivity while one is growing up , and " House of Wax " , about the life of being a celebrity . The final song in medley , " The End of the End " , was written at McCartney \'s <unk> Avenue home while playing on his father , Jim \'s , piano .',
285
+ 'Varanasi grew as an important industrial centre , famous for its muslin and silk <unk> , perfumes , ivory works , and sculpture . Buddha is believed to have founded Buddhism here around <unk> BC when he gave his first sermon , " The Setting in Motion of the Wheel of Dharma " , at nearby <unk> . The city \'s religious importance continued to grow in the 8th century , when Adi <unk> established the worship of Shiva as an official sect of Varanasi . Despite the Muslim rule , Varanasi remained the centre of activity for Hindu intellectuals and theologians during the Middle Ages , which further contributed to its reputation as a cultural centre of religion and education . <unk> Tulsidas wrote his epic poem on Lord Rama \'s life called Ram <unk> Manas in Varanasi . Several other major figures of the Bhakti movement were born in Varanasi , including Kabir and Ravidas . Guru Nanak Dev visited Varanasi for <unk> in <unk> , a trip that played a large role in the founding of <unk> . In the 16th century , Varanasi experienced a cultural revival under the Muslim Mughal emperor <unk> who invested in the city , and built two large temples dedicated to Shiva and Vishnu , though much of modern Varanasi was built during the 18th century , by the Maratha and <unk> kings . The kingdom of Benares was given official status by the <unk> in 1737 , and continued as a dynasty @-@ governed area until Indian independence in 1947 . The city is governed by the Varanasi Nagar Nigam ( Municipal Corporation ) and is represented in the Parliament of India by the current Prime Minister of India <unk> <unk> , who won the <unk> <unk> elections in 2014 by a huge margin . Silk weaving , carpets and crafts and tourism employ a significant number of the local population , as do the <unk> <unk> Works and Bharat Heavy <unk> Limited . Varanasi Hospital was established in 1964 .',
286
+ ]
287
+ embeddings = model.encode(sentences)
288
+ print(embeddings.shape)
289
+ # [3, 768]
290
+
291
+ # Get the similarity scores for the embeddings
292
+ similarities = model.similarity(embeddings, embeddings)
293
+ print(similarities.shape)
294
+ # [3, 3]
295
+ ```
296
+
297
+ <!--
298
+ ### Direct Usage (Transformers)
299
+
300
+ <details><summary>Click to see the direct usage in Transformers</summary>
301
+
302
+ </details>
303
+ -->
304
+
305
+ <!--
306
+ ### Downstream Usage (Sentence Transformers)
307
+
308
+ You can finetune this model on your own dataset.
309
+
310
+ <details><summary>Click to expand</summary>
311
+
312
+ </details>
313
+ -->
314
+
315
+ <!--
316
+ ### Out-of-Scope Use
317
+
318
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
319
+ -->
320
+
321
+ ## Evaluation
322
+
323
+ ### Metrics
324
+
325
+ #### Semantic Similarity
326
+
327
+ * Datasets: `sts-dev` and `sts-test`
328
+ * Evaluated with [<code>EmbeddingSimilarityEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator)
329
+
330
+ | Metric | sts-dev | sts-test |
331
+ |:--------------------|:-----------|:----------|
332
+ | pearson_cosine | 0.6552 | 0.7355 |
333
+ | **spearman_cosine** | **0.6641** | **0.732** |
334
+
335
+ <!--
336
+ ## Bias, Risks and Limitations
337
+
338
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
339
+ -->
340
+
341
+ <!--
342
+ ### Recommendations
343
+
344
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
345
+ -->
346
+
347
+ ## Training Details
348
+
349
+ ### Training Dataset
350
+
351
+ #### Unnamed Dataset
352
+
353
+ * Size: 21,196 training samples
354
+ * Columns: <code>text</code>
355
+ * Approximate statistics based on the first 1000 samples:
356
+ | | text |
357
+ |:--------|:----------------------------------------------------------------------------------|
358
+ | type | string |
359
+ | details | <ul><li>min: 6 tokens</li><li>mean: 51.01 tokens</li><li>max: 75 tokens</li></ul> |
360
+ * Samples:
361
+ | text |
362
+ |:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
363
+ | <code>To promote the album , Carey announced a world tour in April 2003 . As of 2003 , " Charmbracelet World Tour : An Intimate Evening with Mariah Carey " was her most extensive tour , lasting over eight months and performing sixty @-@ nine shows in venues worldwide . Before tickets went on sale in the US , venues were switched from large arenas to smaller , more intimate theater shows . According to Carey , the change was made in order to give fans a more intimate show , and something more Broadway @-@ influenced . She said , " It 's much more intimate so you 'll feel like you had an experience . You experience a night with me . " However , while smaller productions were booked for the US leg of the tour , Carey performed at stadia and arenas in Asia and Europe , and performed for a crowd of over 35 @,@ 000 in Manila , 50 @,@ 000 in Malaysia , and to over 70 @,@ 000 people in China . In the UK , it was Carey 's first tour to feature shows outside London ; she performed in Glasgow , Birming...</code> |
364
+ | <code>By 1916 , these raiding forces were causing serious concern in the Admiralty as the proximity of Bruges to the British coast , to the troopship lanes across the English Channel and for the U @-@ boats , to the Western Approaches ; the heaviest shipping lanes in the World at the time . In the late spring of 1915 , Admiral Reginald <unk> had attempted without success to destroy the lock gates at Ostend with monitors . This effort failed , and Bruges became increasingly important in the Atlantic Campaign , which reached its height in 1917 . By early 1918 , the Admiralty was seeking ever more radical solutions to the problems raised by unrestricted submarine warfare , including instructing the " Allied Naval and Marine Forces " department to plan attacks on U @-@ boat bases in Belgium .</code> |
365
+ | <code>PWI International Heavyweight Championship ( 1 time )</code> |
366
+ * Loss: [<code>DenoisingAutoEncoderLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#denoisingautoencoderloss)
367
+
368
+ ### Evaluation Dataset
369
+
370
+ #### Unnamed Dataset
371
+
372
+ * Size: 2,355 evaluation samples
373
+ * Columns: <code>text</code>
374
+ * Approximate statistics based on the first 1000 samples:
375
+ | | text |
376
+ |:--------|:----------------------------------------------------------------------------------|
377
+ | type | string |
378
+ | details | <ul><li>min: 4 tokens</li><li>mean: 51.08 tokens</li><li>max: 75 tokens</li></ul> |
379
+ * Samples:
380
+ | text |
381
+ |:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
382
+ | <code>Wilde 's two final comedies , An Ideal Husband and The Importance of Being Earnest , were still on stage in London at the time of his prosecution , and they were soon closed as the details of his case became public . After two years in prison with hard labour , Wilde went into exile in Paris , sick and depressed , his reputation destroyed in England . In 1898 , when no @-@ one else would , Leonard Smithers agreed with Wilde to publish the two final plays . Wilde proved to be a <unk> <unk> , sending detailed instructions on stage directions , character listings and the presentation of the book , and insisting that a <unk> from the first performance be reproduced inside . Ellmann argues that the proofs show a man " very much in command of himself and of the play " . Wilde 's name did not appear on the cover , it was " By the Author of Lady Windermere 's Fan " . His return to work was brief though , as he refused to write anything else , " I can write , but have lost the joy of writing " ...</code> |
383
+ | <code>= = = = Ely Viaduct = = = =</code> |
384
+ | <code>= = World War I = =</code> |
385
+ * Loss: [<code>DenoisingAutoEncoderLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#denoisingautoencoderloss)
386
+
387
+ ### Training Hyperparameters
388
+ #### Non-Default Hyperparameters
389
+
390
+ - `eval_strategy`: steps
391
+ - `per_device_train_batch_size`: 64
392
+ - `per_device_eval_batch_size`: 64
393
+ - `learning_rate`: 3e-05
394
+ - `num_train_epochs`: 100
395
+ - `warmup_ratio`: 0.1
396
+ - `fp16`: True
397
+ - `dataloader_num_workers`: 2
398
+ - `load_best_model_at_end`: True
399
+
400
+ #### All Hyperparameters
401
+ <details><summary>Click to expand</summary>
402
+
403
+ - `overwrite_output_dir`: False
404
+ - `do_predict`: False
405
+ - `eval_strategy`: steps
406
+ - `prediction_loss_only`: True
407
+ - `per_device_train_batch_size`: 64
408
+ - `per_device_eval_batch_size`: 64
409
+ - `per_gpu_train_batch_size`: None
410
+ - `per_gpu_eval_batch_size`: None
411
+ - `gradient_accumulation_steps`: 1
412
+ - `eval_accumulation_steps`: None
413
+ - `torch_empty_cache_steps`: None
414
+ - `learning_rate`: 3e-05
415
+ - `weight_decay`: 0.0
416
+ - `adam_beta1`: 0.9
417
+ - `adam_beta2`: 0.999
418
+ - `adam_epsilon`: 1e-08
419
+ - `max_grad_norm`: 1.0
420
+ - `num_train_epochs`: 100
421
+ - `max_steps`: -1
422
+ - `lr_scheduler_type`: linear
423
+ - `lr_scheduler_kwargs`: {}
424
+ - `warmup_ratio`: 0.1
425
+ - `warmup_steps`: 0
426
+ - `log_level`: passive
427
+ - `log_level_replica`: warning
428
+ - `log_on_each_node`: True
429
+ - `logging_nan_inf_filter`: True
430
+ - `save_safetensors`: True
431
+ - `save_on_each_node`: False
432
+ - `save_only_model`: False
433
+ - `restore_callback_states_from_checkpoint`: False
434
+ - `no_cuda`: False
435
+ - `use_cpu`: False
436
+ - `use_mps_device`: False
437
+ - `seed`: 42
438
+ - `data_seed`: None
439
+ - `jit_mode_eval`: False
440
+ - `use_ipex`: False
441
+ - `bf16`: False
442
+ - `fp16`: True
443
+ - `fp16_opt_level`: O1
444
+ - `half_precision_backend`: auto
445
+ - `bf16_full_eval`: False
446
+ - `fp16_full_eval`: False
447
+ - `tf32`: None
448
+ - `local_rank`: 0
449
+ - `ddp_backend`: None
450
+ - `tpu_num_cores`: None
451
+ - `tpu_metrics_debug`: False
452
+ - `debug`: []
453
+ - `dataloader_drop_last`: False
454
+ - `dataloader_num_workers`: 2
455
+ - `dataloader_prefetch_factor`: None
456
+ - `past_index`: -1
457
+ - `disable_tqdm`: False
458
+ - `remove_unused_columns`: True
459
+ - `label_names`: None
460
+ - `load_best_model_at_end`: True
461
+ - `ignore_data_skip`: False
462
+ - `fsdp`: []
463
+ - `fsdp_min_num_params`: 0
464
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
465
+ - `tp_size`: 0
466
+ - `fsdp_transformer_layer_cls_to_wrap`: None
467
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
468
+ - `deepspeed`: None
469
+ - `label_smoothing_factor`: 0.0
470
+ - `optim`: adamw_torch
471
+ - `optim_args`: None
472
+ - `adafactor`: False
473
+ - `group_by_length`: False
474
+ - `length_column_name`: length
475
+ - `ddp_find_unused_parameters`: None
476
+ - `ddp_bucket_cap_mb`: None
477
+ - `ddp_broadcast_buffers`: False
478
+ - `dataloader_pin_memory`: True
479
+ - `dataloader_persistent_workers`: False
480
+ - `skip_memory_metrics`: True
481
+ - `use_legacy_prediction_loop`: False
482
+ - `push_to_hub`: False
483
+ - `resume_from_checkpoint`: None
484
+ - `hub_model_id`: None
485
+ - `hub_strategy`: every_save
486
+ - `hub_private_repo`: None
487
+ - `hub_always_push`: False
488
+ - `gradient_checkpointing`: False
489
+ - `gradient_checkpointing_kwargs`: None
490
+ - `include_inputs_for_metrics`: False
491
+ - `include_for_metrics`: []
492
+ - `eval_do_concat_batches`: True
493
+ - `fp16_backend`: auto
494
+ - `push_to_hub_model_id`: None
495
+ - `push_to_hub_organization`: None
496
+ - `mp_parameters`:
497
+ - `auto_find_batch_size`: False
498
+ - `full_determinism`: False
499
+ - `torchdynamo`: None
500
+ - `ray_scope`: last
501
+ - `ddp_timeout`: 1800
502
+ - `torch_compile`: False
503
+ - `torch_compile_backend`: None
504
+ - `torch_compile_mode`: None
505
+ - `dispatch_batches`: None
506
+ - `split_batches`: None
507
+ - `include_tokens_per_second`: False
508
+ - `include_num_input_tokens_seen`: False
509
+ - `neftune_noise_alpha`: None
510
+ - `optim_target_modules`: None
511
+ - `batch_eval_metrics`: False
512
+ - `eval_on_start`: False
513
+ - `use_liger_kernel`: False
514
+ - `eval_use_gather_object`: False
515
+ - `average_tokens_across_devices`: False
516
+ - `prompts`: None
517
+ - `batch_sampler`: batch_sampler
518
+ - `multi_dataset_batch_sampler`: proportional
519
+
520
+ </details>
521
+
522
+ ### Training Logs
523
+ <details><summary>Click to expand</summary>
524
+
525
+ | Epoch | Step | Training Loss | Validation Loss | sts-dev_spearman_cosine | sts-test_spearman_cosine |
526
+ |:-----------:|:--------:|:-------------:|:---------------:|:-----------------------:|:------------------------:|
527
+ | -1 | -1 | - | - | 0.3173 | - |
528
+ | 0.6024 | 100 | 8.2676 | - | - | - |
529
+ | 1.2048 | 200 | 6.0396 | - | - | - |
530
+ | 1.8072 | 300 | 4.7794 | - | - | - |
531
+ | 2.4096 | 400 | 4.2732 | - | - | - |
532
+ | 3.0120 | 500 | 3.9759 | - | - | - |
533
+ | 3.6145 | 600 | 3.7263 | - | - | - |
534
+ | 4.2169 | 700 | 3.5471 | - | - | - |
535
+ | 4.8193 | 800 | 3.4097 | - | - | - |
536
+ | 5.4217 | 900 | 3.2513 | - | - | - |
537
+ | 6.0241 | 1000 | 3.1646 | 3.3052 | 0.7232 | - |
538
+ | 6.6265 | 1100 | 3.0129 | - | - | - |
539
+ | 7.2289 | 1200 | 2.9307 | - | - | - |
540
+ | 7.8313 | 1300 | 2.8372 | - | - | - |
541
+ | 8.4337 | 1400 | 2.7232 | - | - | - |
542
+ | 9.0361 | 1500 | 2.6845 | - | - | - |
543
+ | 9.6386 | 1600 | 2.546 | - | - | - |
544
+ | 10.2410 | 1700 | 2.4931 | - | - | - |
545
+ | 10.8434 | 1800 | 2.4064 | - | - | - |
546
+ | 11.4458 | 1900 | 2.3145 | - | - | - |
547
+ | 12.0482 | 2000 | 2.2715 | 3.1490 | 0.7177 | - |
548
+ | 12.6506 | 2100 | 2.1495 | - | - | - |
549
+ | 13.2530 | 2200 | 2.1164 | - | - | - |
550
+ | 13.8554 | 2300 | 2.0398 | - | - | - |
551
+ | 14.4578 | 2400 | 1.9538 | - | - | - |
552
+ | 15.0602 | 2500 | 1.9311 | - | - | - |
553
+ | 15.6627 | 2600 | 1.8264 | - | - | - |
554
+ | 16.2651 | 2700 | 1.7786 | - | - | - |
555
+ | 16.8675 | 2800 | 1.7256 | - | - | - |
556
+ | 17.4699 | 2900 | 1.6395 | - | - | - |
557
+ | 18.0723 | 3000 | 1.6082 | 3.4656 | 0.6894 | - |
558
+ | 18.6747 | 3100 | 1.5152 | - | - | - |
559
+ | 19.2771 | 3200 | 1.4678 | - | - | - |
560
+ | 19.8795 | 3300 | 1.425 | - | - | - |
561
+ | 20.4819 | 3400 | 1.3395 | - | - | - |
562
+ | 21.0843 | 3500 | 1.3203 | - | - | - |
563
+ | 21.6867 | 3600 | 1.2275 | - | - | - |
564
+ | 22.2892 | 3700 | 1.1955 | - | - | - |
565
+ | 22.8916 | 3800 | 1.1612 | - | - | - |
566
+ | 23.4940 | 3900 | 1.0792 | - | - | - |
567
+ | 24.0964 | 4000 | 1.0557 | 3.9473 | 0.6822 | - |
568
+ | 24.6988 | 4100 | 0.9793 | - | - | - |
569
+ | 25.3012 | 4200 | 0.9516 | - | - | - |
570
+ | 25.9036 | 4300 | 0.9095 | - | - | - |
571
+ | 26.5060 | 4400 | 0.8408 | - | - | - |
572
+ | 27.1084 | 4500 | 0.8338 | - | - | - |
573
+ | 27.7108 | 4600 | 0.7713 | - | - | - |
574
+ | 28.3133 | 4700 | 0.8312 | - | - | - |
575
+ | 28.9157 | 4800 | 0.8437 | - | - | - |
576
+ | 29.5181 | 4900 | 0.6952 | - | - | - |
577
+ | 30.1205 | 5000 | 0.6825 | 4.3702 | 0.6671 | - |
578
+ | 30.7229 | 5100 | 1.7624 | - | - | - |
579
+ | 31.3253 | 5200 | 6.9439 | - | - | - |
580
+ | 31.9277 | 5300 | 6.2218 | - | - | - |
581
+ | 32.5301 | 5400 | 5.9866 | - | - | - |
582
+ | 33.1325 | 5500 | 5.8608 | - | - | - |
583
+ | 33.7349 | 5600 | 5.7661 | - | - | - |
584
+ | 34.3373 | 5700 | 5.7114 | - | - | - |
585
+ | 34.9398 | 5800 | 5.6526 | - | - | - |
586
+ | 35.5422 | 5900 | 5.5982 | - | - | - |
587
+ | **36.1446** | **6000** | **5.5632** | **5.6696** | **0.7876** | **-** |
588
+ | 36.7470 | 6100 | 5.5455 | - | - | - |
589
+ | 37.3494 | 6200 | 5.4853 | - | - | - |
590
+ | 37.9518 | 6300 | 5.4709 | - | - | - |
591
+ | 38.5542 | 6400 | 5.4372 | - | - | - |
592
+ | 39.1566 | 6500 | 5.405 | - | - | - |
593
+ | 39.7590 | 6600 | 5.4011 | - | - | - |
594
+ | 40.3614 | 6700 | 5.3779 | - | - | - |
595
+ | 40.9639 | 6800 | 5.3684 | - | - | - |
596
+ | 41.5663 | 6900 | 5.3462 | - | - | - |
597
+ | 42.1687 | 7000 | 5.335 | 5.5090 | 0.7515 | - |
598
+ | 42.7711 | 7100 | 5.3273 | - | - | - |
599
+ | 43.3735 | 7200 | 5.3078 | - | - | - |
600
+ | 43.9759 | 7300 | 5.3005 | - | - | - |
601
+ | 44.5783 | 7400 | 5.2836 | - | - | - |
602
+ | 45.1807 | 7500 | 5.2732 | - | - | - |
603
+ | 45.7831 | 7600 | 5.2707 | - | - | - |
604
+ | 46.3855 | 7700 | 5.2525 | - | - | - |
605
+ | 46.9880 | 7800 | 5.2439 | - | - | - |
606
+ | 47.5904 | 7900 | 5.2316 | - | - | - |
607
+ | 48.1928 | 8000 | 5.2121 | 5.4451 | 0.7316 | - |
608
+ | 48.7952 | 8100 | 5.2142 | - | - | - |
609
+ | 49.3976 | 8200 | 5.1939 | - | - | - |
610
+ | 50.0 | 8300 | 5.186 | - | - | - |
611
+ | 50.6024 | 8400 | 5.166 | - | - | - |
612
+ | 51.2048 | 8500 | 5.1727 | - | - | - |
613
+ | 51.8072 | 8600 | 5.1555 | - | - | - |
614
+ | 52.4096 | 8700 | 5.1538 | - | - | - |
615
+ | 53.0120 | 8800 | 5.1413 | - | - | - |
616
+ | 53.6145 | 8900 | 5.1343 | - | - | - |
617
+ | 54.2169 | 9000 | 5.1257 | 5.3939 | 0.7142 | - |
618
+ | 54.8193 | 9100 | 5.1183 | - | - | - |
619
+ | 55.4217 | 9200 | 5.116 | - | - | - |
620
+ | 56.0241 | 9300 | 5.0999 | - | - | - |
621
+ | 56.6265 | 9400 | 5.0922 | - | - | - |
622
+ | 57.2289 | 9500 | 5.0756 | - | - | - |
623
+ | 57.8313 | 9600 | 5.0792 | - | - | - |
624
+ | 58.4337 | 9700 | 5.061 | - | - | - |
625
+ | 59.0361 | 9800 | 5.0663 | - | - | - |
626
+ | 59.6386 | 9900 | 5.0493 | - | - | - |
627
+ | 60.2410 | 10000 | 5.0487 | 5.3613 | 0.7019 | - |
628
+ | 60.8434 | 10100 | 5.0462 | - | - | - |
629
+ | 61.4458 | 10200 | 5.0356 | - | - | - |
630
+ | 62.0482 | 10300 | 5.0379 | - | - | - |
631
+ | 62.6506 | 10400 | 5.0243 | - | - | - |
632
+ | 63.2530 | 10500 | 5.0091 | - | - | - |
633
+ | 63.8554 | 10600 | 5.0128 | - | - | - |
634
+ | 64.4578 | 10700 | 5.0099 | - | - | - |
635
+ | 65.0602 | 10800 | 5.0078 | - | - | - |
636
+ | 65.6627 | 10900 | 4.9965 | - | - | - |
637
+ | 66.2651 | 11000 | 4.9907 | 5.3310 | 0.6963 | - |
638
+ | 66.8675 | 11100 | 4.9918 | - | - | - |
639
+ | 67.4699 | 11200 | 4.9724 | - | - | - |
640
+ | 68.0723 | 11300 | 4.984 | - | - | - |
641
+ | 68.6747 | 11400 | 4.9689 | - | - | - |
642
+ | 69.2771 | 11500 | 4.9636 | - | - | - |
643
+ | 69.8795 | 11600 | 4.9622 | - | - | - |
644
+ | 70.4819 | 11700 | 4.9547 | - | - | - |
645
+ | 71.0843 | 11800 | 4.9527 | - | - | - |
646
+ | 71.6867 | 11900 | 4.9467 | - | - | - |
647
+ | 72.2892 | 12000 | 4.9397 | 5.3186 | 0.6832 | - |
648
+ | 72.8916 | 12100 | 4.9387 | - | - | - |
649
+ | 73.4940 | 12200 | 4.9299 | - | - | - |
650
+ | 74.0964 | 12300 | 4.9454 | - | - | - |
651
+ | 74.6988 | 12400 | 4.9267 | - | - | - |
652
+ | 75.3012 | 12500 | 4.9258 | - | - | - |
653
+ | 75.9036 | 12600 | 4.9244 | - | - | - |
654
+ | 76.5060 | 12700 | 4.9214 | - | - | - |
655
+ | 77.1084 | 12800 | 4.9125 | - | - | - |
656
+ | 77.7108 | 12900 | 4.9122 | - | - | - |
657
+ | 78.3133 | 13000 | 4.9108 | 5.3026 | 0.6840 | - |
658
+ | 78.9157 | 13100 | 4.9073 | - | - | - |
659
+ | 79.5181 | 13200 | 4.8944 | - | - | - |
660
+ | 80.1205 | 13300 | 4.8987 | - | - | - |
661
+ | 80.7229 | 13400 | 4.9013 | - | - | - |
662
+ | 81.3253 | 13500 | 4.8915 | - | - | - |
663
+ | 81.9277 | 13600 | 4.8883 | - | - | - |
664
+ | 82.5301 | 13700 | 4.8861 | - | - | - |
665
+ | 83.1325 | 13800 | 4.882 | - | - | - |
666
+ | 83.7349 | 13900 | 4.8812 | - | - | - |
667
+ | 84.3373 | 14000 | 4.8805 | 5.2968 | 0.6695 | - |
668
+ | 84.9398 | 14100 | 4.8839 | - | - | - |
669
+ | 85.5422 | 14200 | 4.8747 | - | - | - |
670
+ | 86.1446 | 14300 | 4.8652 | - | - | - |
671
+ | 86.7470 | 14400 | 4.8734 | - | - | - |
672
+ | 87.3494 | 14500 | 4.872 | - | - | - |
673
+ | 87.9518 | 14600 | 4.8621 | - | - | - |
674
+ | 88.5542 | 14700 | 4.8599 | - | - | - |
675
+ | 89.1566 | 14800 | 4.8649 | - | - | - |
676
+ | 89.7590 | 14900 | 4.8621 | - | - | - |
677
+ | 90.3614 | 15000 | 4.8483 | 5.2860 | 0.6694 | - |
678
+ | 90.9639 | 15100 | 4.8538 | - | - | - |
679
+ | 91.5663 | 15200 | 4.86 | - | - | - |
680
+ | 92.1687 | 15300 | 4.8463 | - | - | - |
681
+ | 92.7711 | 15400 | 4.8582 | - | - | - |
682
+ | 93.3735 | 15500 | 4.8444 | - | - | - |
683
+ | 93.9759 | 15600 | 4.8482 | - | - | - |
684
+ | 94.5783 | 15700 | 4.848 | - | - | - |
685
+ | 95.1807 | 15800 | 4.8489 | - | - | - |
686
+ | 95.7831 | 15900 | 4.8403 | - | - | - |
687
+ | 96.3855 | 16000 | 4.8425 | 5.2828 | 0.6641 | - |
688
+ | 96.9880 | 16100 | 4.8423 | - | - | - |
689
+ | 97.5904 | 16200 | 4.8377 | - | - | - |
690
+ | 98.1928 | 16300 | 4.8448 | - | - | - |
691
+ | 98.7952 | 16400 | 4.8384 | - | - | - |
692
+ | 99.3976 | 16500 | 4.8381 | - | - | - |
693
+ | 100.0 | 16600 | 4.8389 | - | - | - |
694
+ | -1 | -1 | - | - | - | 0.7320 |
695
+
696
+ * The bold row denotes the saved checkpoint.
697
+ </details>
698
+
699
+ ### Framework Versions
700
+ - Python: 3.12.9
701
+ - Sentence Transformers: 4.0.1
702
+ - Transformers: 4.50.1
703
+ - PyTorch: 2.6.0+cu124
704
+ - Accelerate: 1.5.2
705
+ - Datasets: 3.4.1
706
+ - Tokenizers: 0.21.1
707
+
708
+ ## Citation
709
+
710
+ ### BibTeX
711
+
712
+ #### Sentence Transformers
713
+ ```bibtex
714
+ @inproceedings{reimers-2019-sentence-bert,
715
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
716
+ author = "Reimers, Nils and Gurevych, Iryna",
717
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
718
+ month = "11",
719
+ year = "2019",
720
+ publisher = "Association for Computational Linguistics",
721
+ url = "https://arxiv.org/abs/1908.10084",
722
+ }
723
+ ```
724
+
725
+ #### DenoisingAutoEncoderLoss
726
+ ```bibtex
727
+ @inproceedings{wang-2021-TSDAE,
728
+ title = "TSDAE: Using Transformer-based Sequential Denoising Auto-Encoderfor Unsupervised Sentence Embedding Learning",
729
+ author = "Wang, Kexin and Reimers, Nils and Gurevych, Iryna",
730
+ booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021",
731
+ month = nov,
732
+ year = "2021",
733
+ address = "Punta Cana, Dominican Republic",
734
+ publisher = "Association for Computational Linguistics",
735
+ pages = "671--688",
736
+ url = "https://arxiv.org/abs/2104.06979",
737
+ }
738
+ ```
739
+
740
+ <!--
741
+ ## Glossary
742
+
743
+ *Clearly define terms in order to be accessible across audiences.*
744
+ -->
745
+
746
+ <!--
747
+ ## Model Card Authors
748
+
749
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
750
+ -->
751
+
752
+ <!--
753
+ ## Model Card Contact
754
+
755
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
756
+ -->
config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertModel"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "gradient_checkpointing": false,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 3072,
13
+ "layer_norm_eps": 1e-12,
14
+ "max_position_embeddings": 512,
15
+ "model_type": "bert",
16
+ "num_attention_heads": 12,
17
+ "num_hidden_layers": 12,
18
+ "pad_token_id": 0,
19
+ "position_embedding_type": "absolute",
20
+ "torch_dtype": "float32",
21
+ "transformers_version": "4.50.1",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 30522
25
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "4.0.1",
4
+ "transformers": "4.50.1",
5
+ "pytorch": "2.6.0+cu124"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": "cosine"
10
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:165c6f8fd2098d69eb4a036ce3e57998451cab370bef38e038f7e1fc0927180b
3
+ size 437951328
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 75,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": false,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 75,
50
+ "pad_token": "[PAD]",
51
+ "sep_token": "[SEP]",
52
+ "strip_accents": null,
53
+ "tokenize_chinese_chars": true,
54
+ "tokenizer_class": "BertTokenizer",
55
+ "unk_token": "[UNK]"
56
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff