Jrinky commited on
Commit
bb9e7a6
·
verified ·
1 Parent(s): f4bcb01

Add new SentenceTransformer model

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 1024,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,533 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sentence-transformers
4
+ - sentence-similarity
5
+ - feature-extraction
6
+ - generated_from_trainer
7
+ - dataset_size:11808
8
+ - loss:Infonce
9
+ base_model: BAAI/bge-m3
10
+ widget:
11
+ - source_sentence: Who are some notable individuals named Roger Mason
12
+ sentences:
13
+ - "Rav Kook's writings are extensive, and he is considered one of the most celebrated\
14
+ \ and influential rabbis of the 20th century. Some rabbis recommend that students\
15
+ \ of his begin studying his writings with Ein Ayah. References\n\nExternal links\n\
16
+ \ Ayin Ayah (full text), Hebrew Wikisource\n * Ayn Aya Classes in English\n\n\
17
+ Talmud\nAggadic Midrashim\nAbraham Isaac Kook\nHebrew-language religious books"
18
+ - 'Roger Mason may refer to:
19
+
20
+
21
+ Roger Mason (baseball) (born 1958), American baseball player
22
+
23
+ Roger Mason (geologist) (born 1941), discoverer of Ediacaran fossils
24
+
25
+ Roger Mason Jr. (born 1980), American basketball player
26
+
27
+ Roger Mason (musician), Australian keyboardist
28
+
29
+ L. Roger Mason, Jr., former assistant director of National Intelligence for Systems
30
+ and Resource Analyses'
31
+ - 'Timetabled passenger services on both lines had ceased by the end of February
32
+ 1959. Shipping
33
+
34
+ The Bourne-Morton Canal or Bourne Old Eau connected the town to the sea in Roman
35
+ times. Until the mid-19th century, the present Bourne Eau was capable of carrying
36
+ commercial boat traffic from the Wash coast and Spalding. This resulted from the
37
+ investment following the Bourne Navigation Act of 1780. Passage became impossible
38
+ once the junction of the Eau and the River Glen was converted from gates to a
39
+ sluice in 1860. Media
40
+
41
+ Local news and television programmes are provided by BBC Yorkshire and Lincolnshire
42
+ and ITV Yorkshire. Television signals are received from the Belmont TV transmitter,
43
+ the Waltham TV transmitter can also be received which broadcast BBC East Midlands
44
+ and ITV Central programmes. Local radio stations are BBC Radio Lincolnshire, Greatest
45
+ Hits Radio Lincolnshire and Lincs FM. The town''s local newspapers are Bourne
46
+ Local and Stamford Mercury. Sport
47
+
48
+ Bourne Town Football Club plays football in the United Counties Football League,
49
+ whilst Bourne Cricket Club plays in the Lincolnshire ECB Premier League. These
50
+ teams play their home games at the Abbey Lawn, a recreation ground privately owned
51
+ by the Bourne United Charities. Motor sports
52
+
53
+
54
+ The racing-car marques English Racing Automobiles (ERA) and British Racing Motors
55
+ (BRM) were both founded in Bourne by Raymond Mays, an international racing driver
56
+ and designer who lived in Bourne. The former ERA and BRM workshops in Spalding
57
+ Road are adjacent to Eastgate House, the Mays'' family home in the town''s Eastgate.
58
+ Landmarks
59
+
60
+
61
+ There are currently 71 listed buildings in the parish of Bourne, the most important
62
+ being Bourne Abbey and the Parish Church of St Peter and St Paul (1138), which
63
+ is the only one scheduled Grade I. Notable people
64
+
65
+ Bourne is reputedly the birthplace of Hereward the Wake (in about 1035), although
66
+ the 12th-century source of this information, De Gestis Herwardi Saxonis, refers
67
+ only to his father as being "of Bourne" and to the father''s house and retainers
68
+ there. Robert Mannyng (1264–1340) is credited with putting the speech of the ordinary
69
+ people of his time into recognisable form. He is better known as Robert de Brunne
70
+ because of his long period of residence as a canon at Bourne Abbey. There he completed
71
+ his life''s work of popularising religious and historical material in a Middle
72
+ English dialect that was easily understood at that time. William Cecil (1520–1598)
73
+ became the first Lord Burghley after serving Queen Elizabeth I. He was born at
74
+ a house in the centre of Bourne that is now the Burghley Arms. Dr William Dodd
75
+ (1729–1777), was an Anglican clergyman, man of letters and forger. He was prosecuted,
76
+ sentenced to death and publicly hanged at Tyburn in 1777. Charles Frederick Worth
77
+ (1825–1895), son of a solicitor, lived at Wake House in North Street. He moved
78
+ to Paris and became a renowned designer of women''s fashion and the founder of
79
+ haute couture. The French government awarded him the Légion d''honneur. Sir George
80
+ White (1840-1912), MP for North West Norfolk, a seat he held for twelve years
81
+ until he died in 1912. He was knighted for public service in 1907.'
82
+ - source_sentence: What football team does the Japanese player play for
83
+ sentences:
84
+ - After the meeting, Box summons up the courage to ask Lorraine (Sue Holderness)
85
+ on the date. The act ends with Robert's coat getting on fire because of the cigarette,
86
+ with "Smoke Gets in Your Eyes" on the background.
87
+ - is a Japanese football player. He plays for Honda Lock.
88
+ - As followers on Twitter and FB probably well know I’ve been up to more than a
89
+ spot of preserving of late. It’s my latest addiction, as if I need any more of
90
+ those. My Dad’s the King of Jams, Chutneys and Pickles and I have a feeling he’s
91
+ passed his enthusiastic genes for it on to me!. Which is great, but time consuming.
92
+ Many an evening has been spent peeling, dicing, de-stoning, chopping, stirring,
93
+ testing, sterilising and jarring. And then obviously the tasting. And all the
94
+ crackers, bread and cheese to go with it!. I rarely get to bed much before midnight
95
+ on my chutneying nights. And to be honest my cupboards are now fit to bursting
96
+ with so many goodies, but at least I have christmas presents totally nailed this
97
+ year. My Dad’s been making Hedgerow Chutney for years, and it happens to be everyone’s
98
+ favourite of all his chutney recipes (and he makes quite a number!). Each autumn
99
+ he takes a long walk around the field at the back of his house in Herefordshire
100
+ picking all the freebie hedgerow goodies he can find and transforms them into
101
+ this marvellously fruitful chutney. There’s always plenty of damsons, bullaces,
102
+ sloes, blackberries and a few elderberries. Plus pears or apples for smoothing
103
+ and bulking out. We don’t have quite the same fruit in our hedgerows in France
104
+ but I thought I’d make my own French version picking the fruit from our garden
105
+ and nearby tracks and lanes, managing to find plenty of figs, greengages, plums,
106
+ pears, blackberries and sloes just before the season finished a couple of weeks
107
+ ago. We’ve elderberries here too but they were way past their best by the time
108
+ I got into full chutney mode. There’s no escaping how time consuming and labourious
109
+ chutney making can be, especially when using so much fruit that needs hefty preparatory
110
+ work. I realise now why it’s a hobby generally taken up by retired folk. But the
111
+ results are so worth it, if you can spare it set aside a whole evening in the
112
+ kitchen and wile away the hours getting lost in music or the radio or even catching
113
+ up on a few programmes on You Tube.
114
+ - source_sentence: What is the purpose of Business Intelligence
115
+ sentences:
116
+ - 'College career
117
+
118
+ Proctor played as a defensive lineman for the North Carolina Central Eagles from
119
+ 2008 to 2012. He was redshirted in 2008.'
120
+ - The purpose of Business Intelligence is the transformation of raw data into meaningful
121
+ information which can be used to make better business decisions. Business Intelligence
122
+ grew out of Decision Support systems and is all about collecting data from disparate
123
+ sources, conforming and integrating that data into central repositories which
124
+ support reporting and analysis activities.
125
+ - You have to show the police courtesy, they are only human. No one even WANTS for
126
+ the judicial system to work. They are too lazy.
127
+ - source_sentence: How does the speaker feel about Battle Symphony
128
+ sentences:
129
+ - It's a symptomless prearranged fact that when you afford your babe a infant work
130
+ you motivate the status system, bolster the infant's stressed system, eat up colic,
131
+ and harden your in bondage next to your kid. Now, how satisfying is that
132
+ - Piquet passed Laffite to become the race's fifth different leader. Senna reached
133
+ second just 1.7 seconds behind Piquet by passing Laffite, who then pitted for
134
+ tires. With the two of them in front on their own, and Piquet leading by up to
135
+ 3.5 seconds, Senna was content for the time being to follow his countryman. After
136
+ eight laps in the lead, Piquet pitted for tires. Senna regained first place and
137
+ then also pitted. Piquet's 18.4 second stop was even slower than teammate Mansell's
138
+ had been, but when he returned to the track, the two-time champion got the bit
139
+ between his teeth. Running second behind Senna, Piquet set the fastest lap of
140
+ the race on lap 41, but with a pit stop ten seconds quicker than Piquet's, Senna
141
+ was able to retain the lead. On the very next lap, the 42nd, Piquet pushed a bit
142
+ too much, and crashed hard at the left-hand corner before the last chicane. He
143
+ ended up in the tire barrier, unhurt, but with his car in a very precarious position.
144
+ The crane, present for just that reason, was unable to move the car. Arnoux, now
145
+ 16.6 seconds behind in second, took a second a lap off Senna's lead for five laps
146
+ while a yellow was displayed in the corner where Piquet had crashed. As soon as
147
+ the yellow flag was gone, Arnoux went wide and hit Piquet's abandoned Williams!
148
+ The Frenchman decided that his car was not damaged, and attempted to rejoin the
149
+ field, but did so right in front of Thierry Boutsen's Arrows-BMW, sidelining both
150
+ cars. Very uncharacteristic of a street race, these three – Piquet, Arnoux and
151
+ Boutsen – were the only drivers all afternoon to retire due to accidents.
152
+ - Like Battle Symphony, it's not bad. It's just extremely boring.
153
+ - source_sentence: When did he migrate to New South Wales
154
+ sentences:
155
+ - 'predict ministry in a sales and special floor being Job to the vulnerability
156
+ diver. team: This research will work last for either, often, and also obtaining
157
+ spreadsheets in the funny wedding power of the usability time. Physical Demands:
158
+ The exclusive transitions was temporarily need perfect of those that must share
159
+ developed by an position to badly do the animal objectives of this source. necessary
160
+ terabytes may pay acted to increase streets with hearts to address the professional
161
+ items. solely, the job will distract, Coordinate and be inbox security fun interdisciplinary
162
+ operations that might read in back of 20 updates The service will properly be
163
+ to like the detection throughout the use: logging, including, killing, teaching,
164
+ leading, preparing, operating, and using.'
165
+ - "Shizuka Shirakawa, Scholar of Chinese-language literature. Horin Fukuoji, Nihonga\
166
+ \ painter. 2005\n Mitsuko Mori. Actress. Makoto Saitō (1921–2008). Political scientist,\
167
+ \ specializing in American diplomatic and political history. Ryuzan Aoki, Ceramic\
168
+ \ artist. Toshio Sawada, Civil engineer. Shigeaki Hinohara, Doctor. 2006\n Yoshiaki\
169
+ \ Arata. A pioneer of nuclear fusion research. Jakuchō Setouchi. Writer/Buddhist\
170
+ \ nun. Hidekazu Yoshida. Music critic. Chusaku Oyama, Nihonga painter. Miyohei\
171
+ \ Shinohara, Economist. 2007\n Akira Mikazuki. Former justice minister and professor\
172
+ \ emeritus. Shinya Nakamura. Sculptor. Kōji Nakanishi. Organic chemist. Tokindo\
173
+ \ Okada, Developmental biologist. Shigeyama Sensaku, Kyogen performer. 2008\n\
174
+ \ Hironoshin Furuhashi (1928–2009). Sportsman and sports bureaucrat. Kiyoshi Itō.\
175
+ \ A mathematician whose work is now called Itō calculus. Donald Keene."
176
+ - He attended Derby Grammar School and Beaufort House in London, and migrated to
177
+ New South Wales in 1883. He settled in Newcastle, where he worked as a shipping
178
+ agent, eventually partnering with his brothers in a firm. On 6 May 1893 he married
179
+ Gertrude Mary Saddington, with whom he had five children.
180
+ pipeline_tag: sentence-similarity
181
+ library_name: sentence-transformers
182
+ ---
183
+
184
+ # SentenceTransformer based on BAAI/bge-m3
185
+
186
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [BAAI/bge-m3](https://huggingface.co/BAAI/bge-m3). It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
187
+
188
+ ## Model Details
189
+
190
+ ### Model Description
191
+ - **Model Type:** Sentence Transformer
192
+ - **Base model:** [BAAI/bge-m3](https://huggingface.co/BAAI/bge-m3) <!-- at revision 5617a9f61b028005a4858fdac845db406aefb181 -->
193
+ - **Maximum Sequence Length:** 1024 tokens
194
+ - **Output Dimensionality:** 1024 dimensions
195
+ - **Similarity Function:** Cosine Similarity
196
+ <!-- - **Training Dataset:** Unknown -->
197
+ <!-- - **Language:** Unknown -->
198
+ <!-- - **License:** Unknown -->
199
+
200
+ ### Model Sources
201
+
202
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
203
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
204
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
205
+
206
+ ### Full Model Architecture
207
+
208
+ ```
209
+ SentenceTransformer(
210
+ (0): Transformer({'max_seq_length': 1024, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
211
+ (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
212
+ (2): Normalize()
213
+ )
214
+ ```
215
+
216
+ ## Usage
217
+
218
+ ### Direct Usage (Sentence Transformers)
219
+
220
+ First install the Sentence Transformers library:
221
+
222
+ ```bash
223
+ pip install -U sentence-transformers
224
+ ```
225
+
226
+ Then you can load this model and run inference.
227
+ ```python
228
+ from sentence_transformers import SentenceTransformer
229
+
230
+ # Download from the 🤗 Hub
231
+ model = SentenceTransformer("Jrinky/model3")
232
+ # Run inference
233
+ sentences = [
234
+ 'When did he migrate to New South Wales',
235
+ 'He attended Derby Grammar School and Beaufort House in London, and migrated to New South Wales in 1883. He settled in Newcastle, where he worked as a shipping agent, eventually partnering with his brothers in a firm. On 6 May 1893 he married Gertrude Mary Saddington, with whom he had five children.',
236
+ 'Shizuka Shirakawa, Scholar of Chinese-language literature. Horin Fukuoji, Nihonga painter. 2005\n Mitsuko Mori. Actress. Makoto Saitō (1921–2008). Political scientist, specializing in American diplomatic and political history. Ryuzan Aoki, Ceramic artist. Toshio Sawada, Civil engineer. Shigeaki Hinohara, Doctor. 2006\n Yoshiaki Arata. A pioneer of nuclear fusion research. Jakuchō Setouchi. Writer/Buddhist nun. Hidekazu Yoshida. Music critic. Chusaku Oyama, Nihonga painter. Miyohei Shinohara, Economist. 2007\n Akira Mikazuki. Former justice minister and professor emeritus. Shinya Nakamura. Sculptor. Kōji Nakanishi. Organic chemist. Tokindo Okada, Developmental biologist. Shigeyama Sensaku, Kyogen performer. 2008\n Hironoshin Furuhashi (1928–2009). Sportsman and sports bureaucrat. Kiyoshi Itō. A mathematician whose work is now called Itō calculus. Donald Keene.',
237
+ ]
238
+ embeddings = model.encode(sentences)
239
+ print(embeddings.shape)
240
+ # [3, 1024]
241
+
242
+ # Get the similarity scores for the embeddings
243
+ similarities = model.similarity(embeddings, embeddings)
244
+ print(similarities.shape)
245
+ # [3, 3]
246
+ ```
247
+
248
+ <!--
249
+ ### Direct Usage (Transformers)
250
+
251
+ <details><summary>Click to see the direct usage in Transformers</summary>
252
+
253
+ </details>
254
+ -->
255
+
256
+ <!--
257
+ ### Downstream Usage (Sentence Transformers)
258
+
259
+ You can finetune this model on your own dataset.
260
+
261
+ <details><summary>Click to expand</summary>
262
+
263
+ </details>
264
+ -->
265
+
266
+ <!--
267
+ ### Out-of-Scope Use
268
+
269
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
270
+ -->
271
+
272
+ <!--
273
+ ## Bias, Risks and Limitations
274
+
275
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
276
+ -->
277
+
278
+ <!--
279
+ ### Recommendations
280
+
281
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
282
+ -->
283
+
284
+ ## Training Details
285
+
286
+ ### Training Dataset
287
+
288
+ #### Unnamed Dataset
289
+
290
+ * Size: 11,808 training samples
291
+ * Columns: <code>anchor</code> and <code>positive</code>
292
+ * Approximate statistics based on the first 1000 samples:
293
+ | | anchor | positive |
294
+ |:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
295
+ | type | string | string |
296
+ | details | <ul><li>min: 6 tokens</li><li>mean: 17.85 tokens</li><li>max: 48 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 186.46 tokens</li><li>max: 1024 tokens</li></ul> |
297
+ * Samples:
298
+ | anchor | positive |
299
+ |:-----------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
300
+ | <code>What type of tournament structure was used in this freestyle wrestling competition</code> | <code>This freestyle wrestling competition consisted of a single-elimination tournament, with a repechage used to determine the winners of two bronze medals. Results<br>Legend<br>F — Won by fall<br><br>Final<br><br>Top half<br><br>Bottom half<br><br>Repechage<br><br>References<br>Official website<br><br>Women's freestyle 58 kg<br>World</code> |
301
+ | <code>What was the status of Josip Broz Tito under the 1974 Constitution of Yugoslavia regarding his presidency</code> | <code>1 Wednesday, 22 April 1998. 2 (8.30 a.m.). 3 JUDGE CASSESE: Good morning. May I ask the<br>4 Registrar to call out the case number, please. 5 THE REGISTRAR: Case number IT-95-13a-T,<br>6 Prosecutor versus Slavko Dokmanovic. 7 MR. NIEMANN: My name is Niemann. I appear<br>8 with my colleagues, Mr. Williamson, Mr. Waespi and<br>9 Mr. Vos. 10 MR. FILA: My name is Mr. Toma Fila and<br>11 I appear with Ms. Lopicic and Mr. Petrovic in Defence of<br>12 my client, Mr. Slavko Dokmanovic. 13 JUDGE CASSESE: Mr. Dokmanovic, can you<br>14 follow me? Before we call the witness, may I ask you<br>15 whether you agree to this note from the Registrar about<br>16 the two documents which we discussed yesterday -- you<br>17 have probably received the English translation of the<br>18 bibliography of our witness, plus the missing pages of<br>19 the other document, so I think it is agreed that they<br>20 can be admitted into evidence. 21 MR. NIEMANN: Yes. 22 JUDGE CASSESE: Shall we proceed with the<br>24 MR. FILA: Your Honour, before we continue<br>25 wi...</code> |
302
+ | <code>How quickly can you get loan approval and funds transferred with Crawfort</code> | <code>Then click on the submit button, and it’s done. Make your dream come true with Crawfort<br>When you all submit the loan form, then the agency takes a few hours to process and for approval of the loan. Not only that, you can get your loan amount in your account within a day after getting approval. Many money lenders all take more time in processing things and to credit the amount as well. So, for all that, a customer suffers more as they can’t get the money immediately. But here all these things are not done, and the staff here always make sure to provide you best and fast services. For all these things, you can get the best loan services from here without any doubt.</code> |
303
+ * Loss: <code>selfloss.Infonce</code> with these parameters:
304
+ ```json
305
+ {
306
+ "scale": 20.0,
307
+ "similarity_fct": "cos_sim"
308
+ }
309
+ ```
310
+
311
+ ### Evaluation Dataset
312
+
313
+ #### Unnamed Dataset
314
+
315
+ * Size: 1,476 evaluation samples
316
+ * Columns: <code>anchor</code> and <code>positive</code>
317
+ * Approximate statistics based on the first 1000 samples:
318
+ | | anchor | positive |
319
+ |:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
320
+ | type | string | string |
321
+ | details | <ul><li>min: 6 tokens</li><li>mean: 17.61 tokens</li><li>max: 47 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 171.81 tokens</li><li>max: 1024 tokens</li></ul> |
322
+ * Samples:
323
+ | anchor | positive |
324
+ |:--------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
325
+ | <code>What is Hector Guimard best known for</code> | <code>Hector Guimard (, 10 March 1867 – 20 May 1942) was a French architect and designer, and a prominent figure of the Art Nouveau style. He achieved early fame with his design for the Castel Beranger, the first Art Nouveau apartment building in Paris, which was selected in an 1899 competition as one of the best new building facades in the city. He is best known for the glass and iron edicules or canopies, with ornamental Art Nouveau curves, which he designed to cover the entrances of the first stations of the Paris Metro. Between 1890 and 1930, Guimard designed and built some fifty buildings, in addition to one hundred and forty-one subway entrances for Paris Metro, as well as numerous pieces of furniture and other decorative works. However, in the 1910s Art Nouveau went out of fashion and by the 1960s most of his works had been demolished, and only two of his original Metro edicules were still in place. Guimard's critical reputation revived in the 1960s, in part due to subsequent acquisit...</code> |
326
+ | <code>What does Mark Kantrowitz say about the inclusion of loans in financial aid packages</code> | <code>"They don't always understand that part of the financial aid package includes loans," he says. But loans "don't really reduce your costs," explains Mark Kantrowitz, founder of the financial aid website FinAid.org and publisher of Edvisors Network. "They simply spread them out over time. ... A loan is a loan.</code> |
327
+ | <code>How can Ayurveda support women's health during menopause</code> | <code>Especially as we journey towards menopause, Ayurveda is there to support us with its millenary wisdom. These are some easy routines to incorporate for the daily care of the vulva and vagina, our most delicate flower. Sesame oil: our best allied against dryness, it cannot be missing in our diet.</code> |
328
+ * Loss: <code>selfloss.Infonce</code> with these parameters:
329
+ ```json
330
+ {
331
+ "scale": 20.0,
332
+ "similarity_fct": "cos_sim"
333
+ }
334
+ ```
335
+
336
+ ### Training Hyperparameters
337
+ #### Non-Default Hyperparameters
338
+
339
+ - `eval_strategy`: steps
340
+ - `per_device_train_batch_size`: 2
341
+ - `per_device_eval_batch_size`: 2
342
+ - `learning_rate`: 2e-05
343
+ - `num_train_epochs`: 5
344
+ - `warmup_ratio`: 0.1
345
+ - `fp16`: True
346
+ - `batch_sampler`: no_duplicates
347
+
348
+ #### All Hyperparameters
349
+ <details><summary>Click to expand</summary>
350
+
351
+ - `overwrite_output_dir`: False
352
+ - `do_predict`: False
353
+ - `eval_strategy`: steps
354
+ - `prediction_loss_only`: True
355
+ - `per_device_train_batch_size`: 2
356
+ - `per_device_eval_batch_size`: 2
357
+ - `per_gpu_train_batch_size`: None
358
+ - `per_gpu_eval_batch_size`: None
359
+ - `gradient_accumulation_steps`: 1
360
+ - `eval_accumulation_steps`: None
361
+ - `learning_rate`: 2e-05
362
+ - `weight_decay`: 0.0
363
+ - `adam_beta1`: 0.9
364
+ - `adam_beta2`: 0.999
365
+ - `adam_epsilon`: 1e-08
366
+ - `max_grad_norm`: 1.0
367
+ - `num_train_epochs`: 5
368
+ - `max_steps`: -1
369
+ - `lr_scheduler_type`: linear
370
+ - `lr_scheduler_kwargs`: {}
371
+ - `warmup_ratio`: 0.1
372
+ - `warmup_steps`: 0
373
+ - `log_level`: passive
374
+ - `log_level_replica`: warning
375
+ - `log_on_each_node`: True
376
+ - `logging_nan_inf_filter`: True
377
+ - `save_safetensors`: True
378
+ - `save_on_each_node`: False
379
+ - `save_only_model`: False
380
+ - `restore_callback_states_from_checkpoint`: False
381
+ - `no_cuda`: False
382
+ - `use_cpu`: False
383
+ - `use_mps_device`: False
384
+ - `seed`: 42
385
+ - `data_seed`: None
386
+ - `jit_mode_eval`: False
387
+ - `use_ipex`: False
388
+ - `bf16`: False
389
+ - `fp16`: True
390
+ - `fp16_opt_level`: O1
391
+ - `half_precision_backend`: auto
392
+ - `bf16_full_eval`: False
393
+ - `fp16_full_eval`: False
394
+ - `tf32`: None
395
+ - `local_rank`: 0
396
+ - `ddp_backend`: None
397
+ - `tpu_num_cores`: None
398
+ - `tpu_metrics_debug`: False
399
+ - `debug`: []
400
+ - `dataloader_drop_last`: False
401
+ - `dataloader_num_workers`: 0
402
+ - `dataloader_prefetch_factor`: None
403
+ - `past_index`: -1
404
+ - `disable_tqdm`: False
405
+ - `remove_unused_columns`: True
406
+ - `label_names`: None
407
+ - `load_best_model_at_end`: False
408
+ - `ignore_data_skip`: False
409
+ - `fsdp`: []
410
+ - `fsdp_min_num_params`: 0
411
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
412
+ - `fsdp_transformer_layer_cls_to_wrap`: None
413
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
414
+ - `deepspeed`: None
415
+ - `label_smoothing_factor`: 0.0
416
+ - `optim`: adamw_torch
417
+ - `optim_args`: None
418
+ - `adafactor`: False
419
+ - `group_by_length`: False
420
+ - `length_column_name`: length
421
+ - `ddp_find_unused_parameters`: None
422
+ - `ddp_bucket_cap_mb`: None
423
+ - `ddp_broadcast_buffers`: False
424
+ - `dataloader_pin_memory`: True
425
+ - `dataloader_persistent_workers`: False
426
+ - `skip_memory_metrics`: True
427
+ - `use_legacy_prediction_loop`: False
428
+ - `push_to_hub`: False
429
+ - `resume_from_checkpoint`: None
430
+ - `hub_model_id`: None
431
+ - `hub_strategy`: every_save
432
+ - `hub_private_repo`: False
433
+ - `hub_always_push`: False
434
+ - `gradient_checkpointing`: False
435
+ - `gradient_checkpointing_kwargs`: None
436
+ - `include_inputs_for_metrics`: False
437
+ - `eval_do_concat_batches`: True
438
+ - `fp16_backend`: auto
439
+ - `push_to_hub_model_id`: None
440
+ - `push_to_hub_organization`: None
441
+ - `mp_parameters`:
442
+ - `auto_find_batch_size`: False
443
+ - `full_determinism`: False
444
+ - `torchdynamo`: None
445
+ - `ray_scope`: last
446
+ - `ddp_timeout`: 1800
447
+ - `torch_compile`: False
448
+ - `torch_compile_backend`: None
449
+ - `torch_compile_mode`: None
450
+ - `dispatch_batches`: None
451
+ - `split_batches`: None
452
+ - `include_tokens_per_second`: False
453
+ - `include_num_input_tokens_seen`: False
454
+ - `neftune_noise_alpha`: None
455
+ - `optim_target_modules`: None
456
+ - `batch_eval_metrics`: False
457
+ - `eval_on_start`: False
458
+ - `prompts`: None
459
+ - `batch_sampler`: no_duplicates
460
+ - `multi_dataset_batch_sampler`: proportional
461
+
462
+ </details>
463
+
464
+ ### Training Logs
465
+ | Epoch | Step | Training Loss | Validation Loss |
466
+ |:------:|:----:|:-------------:|:---------------:|
467
+ | 0.2033 | 100 | 0.2694 | 0.0690 |
468
+ | 0.4065 | 200 | 0.0822 | 0.0528 |
469
+ | 0.6098 | 300 | 0.0689 | 0.0497 |
470
+ | 0.8130 | 400 | 0.0644 | 0.0469 |
471
+ | 1.0163 | 500 | 0.0643 | 0.0443 |
472
+ | 1.2195 | 600 | 0.0378 | 0.0473 |
473
+ | 1.4228 | 700 | 0.04 | 0.0479 |
474
+ | 1.6260 | 800 | 0.0358 | 0.0461 |
475
+ | 1.8293 | 900 | 0.0332 | 0.0507 |
476
+ | 2.0325 | 1000 | 0.0283 | 0.0538 |
477
+
478
+
479
+ ### Framework Versions
480
+ - Python: 3.12.3
481
+ - Sentence Transformers: 3.4.0
482
+ - Transformers: 4.42.4
483
+ - PyTorch: 2.2.0+cu121
484
+ - Accelerate: 1.3.0
485
+ - Datasets: 3.2.0
486
+ - Tokenizers: 0.19.1
487
+
488
+ ## Citation
489
+
490
+ ### BibTeX
491
+
492
+ #### Sentence Transformers
493
+ ```bibtex
494
+ @inproceedings{reimers-2019-sentence-bert,
495
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
496
+ author = "Reimers, Nils and Gurevych, Iryna",
497
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
498
+ month = "11",
499
+ year = "2019",
500
+ publisher = "Association for Computational Linguistics",
501
+ url = "https://arxiv.org/abs/1908.10084",
502
+ }
503
+ ```
504
+
505
+ #### Infonce
506
+ ```bibtex
507
+ @misc{henderson2017efficient,
508
+ title={Efficient Natural Language Response Suggestion for Smart Reply},
509
+ author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
510
+ year={2017},
511
+ eprint={1705.00652},
512
+ archivePrefix={arXiv},
513
+ primaryClass={cs.CL}
514
+ }
515
+ ```
516
+
517
+ <!--
518
+ ## Glossary
519
+
520
+ *Clearly define terms in order to be accessible across audiences.*
521
+ -->
522
+
523
+ <!--
524
+ ## Model Card Authors
525
+
526
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
527
+ -->
528
+
529
+ <!--
530
+ ## Model Card Contact
531
+
532
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
533
+ -->
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/root/autodl-tmp/work_dir/models/model3/checkpoint-1000",
3
+ "architectures": [
4
+ "XLMRobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1024,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 4096,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 8194,
17
+ "model_type": "xlm-roberta",
18
+ "num_attention_heads": 16,
19
+ "num_hidden_layers": 24,
20
+ "output_past": true,
21
+ "pad_token_id": 1,
22
+ "position_embedding_type": "absolute",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.42.4",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 250002
28
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.4.0",
4
+ "transformers": "4.42.4",
5
+ "pytorch": "2.2.0+cu121"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": "cosine"
10
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13c915093581542d29183fcb909932efa101abdfcade9e3e184b70c2adb9bf36
3
+ size 2271064456
modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 1024,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e3b8957de04e3a4ed42b1a11381556f9adad8d0d502b9dd071c75f626b28f40
3
+ size 17083053
tokenizer_config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "250001": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "<s>",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "<s>",
47
+ "eos_token": "</s>",
48
+ "mask_token": "<mask>",
49
+ "max_length": 1024,
50
+ "model_max_length": 1024,
51
+ "pad_to_multiple_of": null,
52
+ "pad_token": "<pad>",
53
+ "pad_token_type_id": 0,
54
+ "padding_side": "right",
55
+ "sep_token": "</s>",
56
+ "sp_model_kwargs": {},
57
+ "stride": 0,
58
+ "tokenizer_class": "XLMRobertaTokenizer",
59
+ "truncation_side": "right",
60
+ "truncation_strategy": "longest_first",
61
+ "unk_token": "<unk>"
62
+ }