File size: 41,899 Bytes
4efe6e2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
---
tags:
- sentence-transformers
- sentence-similarity
- feature-extraction
- generated_from_trainer
- dataset_size:98112
- loss:MultipleNegativesRankingLoss
base_model: thenlper/gte-small
widget:
- source_sentence: How does a photocell control outdoor lighting?
  sentences:
  - 'To solve this problem, we can use the binomial probability formula:


    P(X = k) = C(n, k) * p^k * (1-p)^(n-k)


    where:

    - P(X = k) is the probability of exactly k successes (faulty keyboards) in n trials
    (laptops produced)

    - C(n, k) is the number of combinations of n items taken k at a time (n! / (k!(n-k)!))

    - p is the probability of success (5% or 0.05)

    - n is the number of trials (400 laptops)

    - k is the number of successes (20 faulty keyboards)


    However, we want to find the probability of at least 20 faulty keyboards, so we
    need to find the sum of probabilities for k = 20, 21, 22, ..., 400.


    P(X >= 20) = 1 - P(X < 20) = 1 - Σ P(X = k) for k = 0 to 19


    Now, we can calculate the probabilities for each value of k and sum them up:


    P(X >= 20) = 1 - Σ C(400, k) * 0.05^k * 0.95^(400-k) for k = 0 to 19


    Using a calculator or software to compute the sum, we get:


    P(X >= 20) ≈ 1 - 0.0184 = 0.9816


    So, the probability that at least 20 laptops will have a faulty keyboard is approximately
    98.16%.'
  - A photocell controls outdoor lighting by detecting the level of ambient light.
    It automatically turns the lights on when it becomes dark and off when it becomes
    light, functioning as a light-dependent switch for energy efficiency and convenience.
  - 'Glycosylation with β-N-acetylglucosamine (O-GlcNAcylation) is one of the most
    complex post-translational modifications. The cycling of O-GlcNAc is controlled
    by two enzymes: UDP-NAc transferase (OGT) and O-GlcNAcase (OGA). We recently reported
    that endothelin-1 (ET-1) augments vascular levels of O-GlcNAcylated proteins.
    Here we tested the hypothesis that O-GlcNAcylation contributes to the vascular
    effects of ET-1 via activation of the RhoA/Rho-kinase pathway. Incubation of vascular
    smooth muscle cells (VSMCs) with ET-1 (0.1 μM) produces a time-dependent increase
    in O-GlcNAc levels. ET-1-induced O-GlcNAcylation is not observed when VSMCs are
    previously transfected with OGT siRNA, treated with ST045849 (OGT inhibitor) or
    atrasentan (ET(A) antagonist). ET-1 as well as PugNAc (OGA inhibitor) augmented
    contractions to phenylephrine in endothelium-denuded rat aortas, an effect that
    was abolished by the Rho kinase inhibitor Y-27632. Incubation of VSMCs with ET-1
    increased expression of the phosphorylated forms of myosin phosphatase target
    subunit 1 (MYPT-1), protein kinase C-potentiated protein phosphatase 1 inhibitor
    protein (protein kinase C-potentiated phosphatase inhibitor-17), and myosin light
    chain (MLC) and RhoA expression and activity, and this effect was abolished by
    both OGT siRNA transfection or OGT inhibition and atrasentan. ET-1 also augmented
    expression of PDZ-Rho GEF (guanine nucleotide exchange factor) and p115-Rho GEF
    in VSMCs and this was prevented by OGT siRNA, ST045849, and atrasentan.'
- source_sentence: A torus has a major radius of 5 cm and a minor radius of 3 cm.
    Find the volume of the torus.
  sentences:
  - 'To find the Hausdorff dimension of the Koch curve, we can use the formula:


    Hausdorff dimension (D) = log(N) / log(1/s)


    where N is the number of self-similar pieces and s is the scaling factor.


    For the Koch curve, each line segment is divided into four segments, each of which
    is 1/3 the length of the original segment. Therefore, N = 4 and s = 1/3.


    Now, we can plug these values into the formula:


    D = log(4) / log(1/3)


    D ≈ 1.2619


    So, the Hausdorff dimension of the Koch curve is approximately 1.2619.'
  - 'To find the volume of a torus, we can use the formula:


    Volume = (π * minor_radius^2) * (2 * π * major_radius)


    where minor_radius is the minor radius of the torus and major_radius is the major
    radius of the torus.


    Given that the major radius is 5 cm and the minor radius is 3 cm, we can plug
    these values into the formula:


    Volume = (π * 3^2) * (2 * π * 5)


    Volume = (π * 9) * (10 * π)


    Volume = 90 * π^2


    The volume of the torus is approximately 282.74 cubic centimeters.'
  - The purpose of the present study was to elucidate the mechanisms of action mediating
    enhancement of basal glucose uptake in skeletal muscle cells by seven medicinal
    plant products recently identified from the pharmacopeia of native Canadian populations
    (Spoor et al., 2006). Activity of the major signaling pathways that regulate glucose
    uptake was assessed by western immunoblot in C2C12 muscle cells treated with extracts
    from these plant species. Effects of extracts on mitochondrial function were assessed
    by respirometry in isolated rat liver mitochondria. Metabolic stress induced by
    extracts was assessed by measuring ATP concentration and rate of cell medium acidification
    in C2C12 myotubes and H4IIE hepatocytes. Extracts were applied at a dose of 15-100
    microg/ml. The effect of all seven products was achieved through a common mechanism
    mediated not by the insulin signaling pathway but rather by the AMP-activated
    protein kinase (AMPK) pathway in response to the disruption of mitochondrial function
    and ensuing metabolic stress. Disruption of mitochondrial function occurred in
    the form of uncoupling of oxidative phosphorylation and/or inhibition of ATPsynthase.
    Activity of the AMPK pathway, in some instances comparable to that stimulated
    by 4mM of the AMP-mimetic AICAR, was in several cases sustained for at least 18h
    post-treatment. Duration of metabolic stress, however, was in most cases in the
    order of 1h.
- source_sentence: Consider the elliptic curve given by the equation $y^2=x^3-2x+5$
    over the field of rational numbers $\mathbb{Q}$. Let $P=(1,2)$ and $Q=(-1,2)$
    be two points on the curve. Find the equation of the line passing through $P$
    and $Q$ and show that it intersects the curve at another point $R$. Then, find
    the coordinates of the point $R$.
  sentences:
  - Fifteen novel derivatives of D-DIBOA, including aromatic ring modifications and
    the addition of side chains in positions C-2 and N-4, had previously been synthesised
    and their phytotoxicity on standard target species (STS) evaluated. This strategy
    combined steric, electronic, solubility and lipophilicity requirements to achieve
    the maximum phytotoxic activity. An evaluation of the bioactivity of these compounds
    on the systems Oryza sativa-Echinochloa crus-galli and Triticum aestivum-Avena
    fatua is reported here. All compounds showed inhibition profiles on the two species
    Echinochloa crus-galli (L.) Beauv. and Avena fatua L. The most marked effects
    were caused by 6F-4Pr-D-DIBOA, 6F-4Val-D-DIBOA, 6Cl-4Pr-D-DIBOA and 6Cl-4Val-D-DIBOA.
    The IC(50) values for the systems Echinochloa crus-galli-Oryza sativa and Avena
    fatua-Triticum aestivum for all compounds were compared. The compound that showed
    the greatest selectivity for the system Echinochloa crus-galli-Oryza sativa was
    8Cl-4Pr-D-DIBOA, which was 15 times more selective than the commercial herbicide
    propanil (Cotanil-35). With regard to the system Avena fatua-Triticum aestivum,
    the compounds that showed the highest selectivities were 8Cl-4Val-D-DIBOA and
    6F-4Pr-D-DIBOA. The results obtained for 6F-4Pr-D-DIBOA are of great interest
    because of the high phytotoxicity to Avena fatua (IC(50) = 6 µM, r(2) = 0.9616).
  - 'To find the equation of the line passing through points $P=(1,2)$ and $Q=(-1,2)$,
    we first find the slope of the line. Since the y-coordinates of both points are
    the same, the slope is 0. Therefore, the line is horizontal and its equation is
    given by:


    $y = 2$


    Now, we want to find the point $R$ where this line intersects the elliptic curve
    $y^2 = x^3 - 2x + 5$. Since we know that $y=2$, we can substitute this value into
    the equation of the curve:


    $(2)^2 = x^3 - 2x + 5$


    Simplifying, we get:


    $4 = x^3 - 2x + 5$


    Rearranging the terms, we have:


    $x^3 - 2x + 1 = 0$


    We know that $x=1$ and $x=-1$ are solutions to this equation since they correspond
    to the points $P$ and $Q$. To find the third solution, we can use synthetic division
    or factor the polynomial. Factoring, we get:


    $(x-1)(x+1)(x-1) = 0$


    So, the third solution is $x=1$. Substituting this value back into the equation
    of the line, we find the corresponding y-coordinate:


    $y = 2$


    Thus, the third point of intersection is $R=(1,2)$. However, in the context of
    elliptic curves, we should take the "sum" of the points $P$ and $Q$ as the negative
    of the third intersection point. Since $R=(1,2)$, the negative of this point is
    given by $-R=(1,-2)$. Therefore, the "sum" of the points $P$ and $Q$ on the elliptic
    curve is:


    $P + Q = -R = (1,-2)$.'
  - The use of geospatial analysis may be subject to regulatory compliance depending
    on the specific application and the jurisdiction in which it is used. For example,
    the use of geospatial data for marketing purposes may be subject to privacy regulations,
    and the use of geospatial data for land use planning may be subject to environmental
    regulations. It is important to consult with legal counsel to ensure compliance
    with all applicable laws and regulations.
- source_sentence: Does sLEDAI-2K Conceal Worsening in a Particular System When There
    Is Overall Improvement?
  sentences:
  - To determine whether the Systemic Lupus Erythematosus Disease Activity Index 2000
    (SLEDAI-2K) is valid in identifying patients who had a clinically important overall
    improvement with no worsening in other descriptors/systems. Consecutive patients
    with systemic lupus erythematosus with active disease who attended the Lupus Clinic
    between 2000 and 2012 were studied. Based on the change in the total SLEDAI-2K
    scores on last visit, patients were grouped as improved, flared/worsened, and
    unchanged. Patients showing improvement were evaluated for the presence of new
    active descriptors at last visit compared with baseline visit. Of the 158 patients
    studied, 109 patients had improved, 38 remained unchanged, and 11 flared/worsened
    at last visit. In the improved group, 11 patients had a new laboratory descriptor
    that was not present at baseline visit. In those 11 patients, this new laboratory
    descriptor was not clinically significant and did not require a change in disease
    management.
  - 'To find the dot product of two vectors using their magnitudes, angle between
    them, and trigonometry, we can use the formula:


    Dot product = |A| * |B| * cos(θ)


    where |A| and |B| are the magnitudes of the vectors, and θ is the angle between
    them.


    In this case, |A| = 5 units, |B| = 8 units, and θ = 60 degrees.


    First, we need to convert the angle from degrees to radians:


    θ = 60 * (π / 180) = π / 3 radians


    Now, we can find the dot product:


    Dot product = 5 * 8 * cos(π / 3)

    Dot product = 40 * (1/2)

    Dot product = 20


    So, the dot product of the two vectors is 20.'
  - To determine if hospitals that routinely discharge patients early after lobectomy
    have increased readmissions. Hospitals are increasingly motivated to reduce length
    of stay (LOS) after lung cancer surgery, yet it is unclear if a routine of early
    discharge is associated with increased readmissions. The relationship between
    hospital discharge practices and readmission rates is therefore of tremendous
    clinical and financial importance. The National Cancer Database was queried for
    patients undergoing lobectomy for lung cancer from 2004 to 2013 at Commission
    on Cancer-accredited hospitals, which performed at least 25 lobectomies in a 2-year
    period. Facility discharge practices were characterized by a facility's median
    LOS relative to the median LOS for all patients in that same time period. In all,
    59,734 patients met inclusion criteria; 2687 (4.5%) experienced an unplanned readmission.
    In a hierarchical logistic regression model, a routine of early discharge (defined
    as a facility's tendency to discharge patients faster than the population median
    in the same time period) was not associated with increased risk of readmission
    (odds ratio 1.12, 95% confidence interval 0.97-1.28, P = 0.12). In a risk-adjusted
    hospital readmission rate analysis, hospitals that discharged patients early did
    not experience more readmissions (P = 0.39). The lack of effect of early discharge
    practices on readmission rates was observed for both minimally invasive and thoracotomy
    approaches.
- source_sentence: Does systemic administration of urocortin after intracerebral hemorrhage
    reduce neurological deficits and neuroinflammation in rats?
  sentences:
  - Intracerebral hemorrhage (ICH) remains a serious clinical problem lacking effective
    treatment. Urocortin (UCN), a novel anti-inflammatory neuropeptide, protects injured
    cardiomyocytes and dopaminergic neurons. Our preliminary studies indicate UCN
    alleviates ICH-induced brain injury when administered intracerebroventricularly
    (ICV). The present study examines the therapeutic effect of UCN on ICH-induced
    neurological deficits and neuroinflammation when administered by the more convenient
    intraperitoneal (i.p.) route. ICH was induced in male Sprague-Dawley rats by intrastriatal
    infusion of bacterial collagenase VII-S or autologous blood. UCN (2.5 or 25 μg/kg)
    was administered i.p. at 60 minutes post-ICH. Penetration of i.p. administered
    fluorescently labeled UCN into the striatum was examined by fluorescence microscopy.
    Neurological deficits were evaluated by modified neurological severity score (mNSS).
    Brain edema was assessed using the dry/wet method. Blood-brain barrier (BBB) disruption
    was assessed using the Evans blue assay. Hemorrhagic volume and lesion volume
    were assessed by Drabkin's method and morphometric assay, respectively. Pro-inflammatory
    cytokine (TNF-α, IL-1β, and IL-6) expression was evaluated by enzyme-linked immunosorbent
    assay (ELISA). Microglial activation and neuronal loss were evaluated by immunohistochemistry.
    Administration of UCN reduced neurological deficits from 1 to 7 days post-ICH.
    Surprisingly, although a higher dose (25 μg/kg, i.p.) also reduced the functional
    deficits associated with ICH, it is significantly less effective than the lower
    dose (2.5 μg/kg, i.p.). Beneficial results with the low dose of UCN included a
    reduction in neurological deficits from 1 to 7 days post-ICH, as well as a reduction
    in brain edema, BBB disruption, lesion volume, microglial activation and neuronal
    loss 3 days post-ICH, and suppression of TNF-α, IL-1β, and IL-6 production 1,
    3 and 7 days post-ICH.
  - 'A perfect number is a positive integer that is equal to the sum of its proper
    divisors (excluding itself). The first perfect numbers are 6, 28, 496, and 8128.
    Perfect numbers can be generated using the formula 2^(p-1) * (2^p - 1), where
    p and 2^p - 1 are both prime numbers.


    The first five (p, 2^p - 1) pairs are:

    (2, 3) - 6

    (3, 7) - 28

    (5, 31) - 496

    (7, 127) - 8128

    (13, 8191) - 33,550,336


    To find the 6th perfect number, we need to find the next prime number p such that
    2^p - 1 is also prime. The next such pair is (17, 131071). Using the formula:


    2^(17-1) * (2^17 - 1) = 2^16 * 131071 = 65,536 * 131071 = 8,589,869,056


    So, the 6th perfect number is 8,589,869,056.'
  - 'In type theory, the successor function $S$ is used to represent the next number
    in the sequence. When you apply the successor function $S$ three times to the
    number 0, you get:


    1. $S(0)$, which represents 1.

    2. $S(S(0))$, which represents 2.

    3. $S(S(S(0)))$, which represents 3.


    So, the result of applying the successor function $S$ three times to the number
    0 in type theory is 3.'
pipeline_tag: sentence-similarity
library_name: sentence-transformers
metrics:
- cosine_accuracy@1
- cosine_accuracy@3
- cosine_accuracy@5
- cosine_accuracy@10
- cosine_precision@1
- cosine_precision@3
- cosine_precision@5
- cosine_recall@1
- cosine_recall@3
- cosine_recall@5
- cosine_ndcg@10
- cosine_mrr@10
- cosine_map@100
model-index:
- name: SentenceTransformer based on thenlper/gte-small
  results:
  - task:
      type: logging
      name: Logging
    dataset:
      name: ir eval
      type: ir-eval
    metrics:
    - type: cosine_accuracy@1
      value: 0.9291020819957809
      name: Cosine Accuracy@1
    - type: cosine_accuracy@3
      value: 0.9819315784646427
      name: Cosine Accuracy@3
    - type: cosine_accuracy@5
      value: 0.9933963129413923
      name: Cosine Accuracy@5
    - type: cosine_accuracy@10
      value: 0.9984407961111621
      name: Cosine Accuracy@10
    - type: cosine_precision@1
      value: 0.9291020819957809
      name: Cosine Precision@1
    - type: cosine_precision@3
      value: 0.32731052615488093
      name: Cosine Precision@3
    - type: cosine_precision@5
      value: 0.19867926258827848
      name: Cosine Precision@5
    - type: cosine_recall@1
      value: 0.9291020819957809
      name: Cosine Recall@1
    - type: cosine_recall@3
      value: 0.9819315784646427
      name: Cosine Recall@3
    - type: cosine_recall@5
      value: 0.9933963129413923
      name: Cosine Recall@5
    - type: cosine_ndcg@10
      value: 0.9670096227619588
      name: Cosine Ndcg@10
    - type: cosine_mrr@10
      value: 0.9565327512887825
      name: Cosine Mrr@10
    - type: cosine_map@100
      value: 0.9565967419425125
      name: Cosine Map@100
---

# SentenceTransformer based on thenlper/gte-small

This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [thenlper/gte-small](https://huggingface.co/thenlper/gte-small). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.

## Model Details

### Model Description
- **Model Type:** Sentence Transformer
- **Base model:** [thenlper/gte-small](https://huggingface.co/thenlper/gte-small) <!-- at revision 17e1f347d17fe144873b1201da91788898c639cd -->
- **Maximum Sequence Length:** 512 tokens
- **Output Dimensionality:** 384 dimensions
- **Similarity Function:** Cosine Similarity
<!-- - **Training Dataset:** Unknown -->
<!-- - **Language:** Unknown -->
<!-- - **License:** Unknown -->

### Model Sources

- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)

### Full Model Architecture

```
SentenceTransformer(
  (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel 
  (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
  (2): Normalize()
)
```

## Usage

### Direct Usage (Sentence Transformers)

First install the Sentence Transformers library:

```bash
pip install -U sentence-transformers
```

Then you can load this model and run inference.
```python
from sentence_transformers import SentenceTransformer

# Download from the 🤗 Hub
model = SentenceTransformer("sucharush/gte_MNR")
# Run inference
sentences = [
    'Does systemic administration of urocortin after intracerebral hemorrhage reduce neurological deficits and neuroinflammation in rats?',
    "Intracerebral hemorrhage (ICH) remains a serious clinical problem lacking effective treatment. Urocortin (UCN), a novel anti-inflammatory neuropeptide, protects injured cardiomyocytes and dopaminergic neurons. Our preliminary studies indicate UCN alleviates ICH-induced brain injury when administered intracerebroventricularly (ICV). The present study examines the therapeutic effect of UCN on ICH-induced neurological deficits and neuroinflammation when administered by the more convenient intraperitoneal (i.p.) route. ICH was induced in male Sprague-Dawley rats by intrastriatal infusion of bacterial collagenase VII-S or autologous blood. UCN (2.5 or 25 μg/kg) was administered i.p. at 60 minutes post-ICH. Penetration of i.p. administered fluorescently labeled UCN into the striatum was examined by fluorescence microscopy. Neurological deficits were evaluated by modified neurological severity score (mNSS). Brain edema was assessed using the dry/wet method. Blood-brain barrier (BBB) disruption was assessed using the Evans blue assay. Hemorrhagic volume and lesion volume were assessed by Drabkin's method and morphometric assay, respectively. Pro-inflammatory cytokine (TNF-α, IL-1β, and IL-6) expression was evaluated by enzyme-linked immunosorbent assay (ELISA). Microglial activation and neuronal loss were evaluated by immunohistochemistry. Administration of UCN reduced neurological deficits from 1 to 7 days post-ICH. Surprisingly, although a higher dose (25 μg/kg, i.p.) also reduced the functional deficits associated with ICH, it is significantly less effective than the lower dose (2.5 μg/kg, i.p.). Beneficial results with the low dose of UCN included a reduction in neurological deficits from 1 to 7 days post-ICH, as well as a reduction in brain edema, BBB disruption, lesion volume, microglial activation and neuronal loss 3 days post-ICH, and suppression of TNF-α, IL-1β, and IL-6 production 1, 3 and 7 days post-ICH.",
    'In type theory, the successor function $S$ is used to represent the next number in the sequence. When you apply the successor function $S$ three times to the number 0, you get:\n\n1. $S(0)$, which represents 1.\n2. $S(S(0))$, which represents 2.\n3. $S(S(S(0)))$, which represents 3.\n\nSo, the result of applying the successor function $S$ three times to the number 0 in type theory is 3.',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 384]

# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]
```

<!--
### Direct Usage (Transformers)

<details><summary>Click to see the direct usage in Transformers</summary>

</details>
-->

<!--
### Downstream Usage (Sentence Transformers)

You can finetune this model on your own dataset.

<details><summary>Click to expand</summary>

</details>
-->

<!--
### Out-of-Scope Use

*List how the model may foreseeably be misused and address what users ought not to do with the model.*
-->

## Evaluation

### Metrics

#### Logging

* Dataset: `ir-eval`
* Evaluated with <code>__main__.LoggingEvaluator</code>

| Metric             | Value     |
|:-------------------|:----------|
| cosine_accuracy@1  | 0.9291    |
| cosine_accuracy@3  | 0.9819    |
| cosine_accuracy@5  | 0.9934    |
| cosine_accuracy@10 | 0.9984    |
| cosine_precision@1 | 0.9291    |
| cosine_precision@3 | 0.3273    |
| cosine_precision@5 | 0.1987    |
| cosine_recall@1    | 0.9291    |
| cosine_recall@3    | 0.9819    |
| cosine_recall@5    | 0.9934    |
| **cosine_ndcg@10** | **0.967** |
| cosine_mrr@10      | 0.9565    |
| cosine_map@100     | 0.9566    |

<!--
## Bias, Risks and Limitations

*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
-->

<!--
### Recommendations

*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
-->

## Training Details

### Training Dataset

#### Unnamed Dataset

* Size: 98,112 training samples
* Columns: <code>sentence_0</code> and <code>sentence_1</code>
* Approximate statistics based on the first 1000 samples:
  |         | sentence_0                                                                         | sentence_1                                                                          |
  |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
  | type    | string                                                                             | string                                                                              |
  | details | <ul><li>min: 6 tokens</li><li>mean: 44.14 tokens</li><li>max: 512 tokens</li></ul> | <ul><li>min: 12 tokens</li><li>mean: 321.5 tokens</li><li>max: 512 tokens</li></ul> |
* Samples:
  | sentence_0                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                 | sentence_1                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           |
  |:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
  | <code>Are transcobalamin II receptor polymorphisms associated with increased risk for neural tube defects?</code>                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          | <code>Women who have low cobalamin (vitamin B(12)) levels are at increased risk for having children with neural tube defects (NTDs). The transcobalamin II receptor (TCblR) mediates uptake of cobalamin into cells. Inherited variants in the TCblR gene as NTD risk factors were evaluated. Case-control and family-based tests of association were used to screen common variation in TCblR as genetic risk factors for NTDs in a large Irish group. A confirmatory group of NTD triads was used to test positive findings. 2 tightly linked variants associated with NTDs in a recessive model were found: TCblR rs2336573 (G220R; p(corr)=0.0080, corrected for multiple hypothesis testing) and TCblR rs9426 (p(corr)=0.0279). These variants were also associated with NTDs in a family-based test before multiple test correction (log-linear analysis of a recessive model: rs2336573 (G220R; RR=6.59, p=0.0037) and rs9426 (RR=6.71, p=0.0035)). A copy number variant distal to TCblR and two previously unreported exonic insertio...</code>                                                             |
  | <code>A company produces three products: Product A, B, and C. The monthly sales figures and marketing expenses (in thousands of dollars) for each product for the last six months are given below:<br><br>| Product | Sales1 | Sales2 | Sales3 | Sales4 | Sales5 | Sales6 | Marketing Expense1 | Marketing Expense2 | Marketing Expense3 | Marketing Expense4 | Marketing Expense5 | Marketing Expense6 |<br>|---------|--------|--------|--------|--------|--------|--------|--------------------|--------------------|--------------------|--------------------|--------------------|--------------------|<br>| A       | 50     | 45     | 55     | 52     | 48     | 56     | 20                  | 18                  | 25                  | 22                  | 19                  | 30                  |<br>| B       | 40     | 48     | 35     | 37     | 45     | 38     | 12                  | 15                  | 10                  | 14                  | 17                  | 11                  |<br>| C       | 60     | 65     | ...</code> | <code>To calculate the covariance between the sales of Product A and Product B, we first need to find the mean sales for both products. Then, we will calculate the deviations from the mean for each month's sales and multiply these deviations for both products. Finally, we will sum these products and divide by the number of months minus 1.<br><br>Mean sales for Product A:<br>(50 + 45 + 55 + 52 + 48 + 56) / 6 = 306 / 6 = 51<br><br>Mean sales for Product B:<br>(40 + 48 + 35 + 37 + 45 + 38) / 6 = 243 / 6 = 40.5<br><br>Deviations from the mean for Product A:<br>(50 - 51, 45 - 51, 55 - 51, 52 - 51, 48 - 51, 56 - 51) = (-1, -6, 4, 1, -3, 5)<br><br>Deviations from the mean for Product B:<br>(40 - 40.5, 48 - 40.5, 35 - 40.5, 37 - 40.5, 45 - 40.5, 38 - 40.5) = (-0.5, 7.5, -5.5, -3.5, 4.5, -2.5)<br><br>Multiply the deviations for both products:<br>(-1 * -0.5, -6 * 7.5, 4 * -5.5, 1 * -3.5, -3 * 4.5, 5 * -2.5) = (0.5, 45, -22, -3.5, -13.5, -12.5)<br><br>Sum of the products of deviations:<br>0.5 + 45 - 22 - 3.5 - 13.5 - 12.5 = -6<br><br>Covariance between the sale...</code> |
  | <code>Consider a basketball player who has a free-throw shooting percentage of 80%. The player attempts 10 free throws in a game. <br><br>If the player makes a free throw, there is an 80% chance that they will make their next free throw attempt. If they miss a free throw, there's a 60% chance that they will make their next free throw attempt.<br><br>What is the probability that the player makes exactly 7 out of their 10 free throw attempts?</code>                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        | <code>To solve this problem, we can use the concept of conditional probability and the binomial theorem. Let's denote the probability of making a free throw after a successful attempt as P(S) = 0.8 and the probability of making a free throw after a missed attempt as P(M) = 0.6.<br><br>We need to find the probability of making exactly 7 out of 10 free throw attempts. There are multiple ways this can happen, and we need to consider all possible sequences of 7 successes (S) and 3 misses (M). We can represent these sequences as a string of S and M, for example, SSSSSSSMMM.<br><br>There are C(10, 7) = 10! / (7! * 3!) = 120 ways to arrange 7 successes and 3 misses in a sequence of 10 attempts. For each of these sequences, we can calculate the probability of that specific sequence occurring and then sum up the probabilities of all sequences.<br><br>Let's calculate the probability of a specific sequence. For example, consider the sequence SSSSSSSMMM. The probability of this sequence occurring is:<br><br>P(SSSSSSSMMM) = P(S...</code>                                     |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
  ```json
  {
      "scale": 20.0,
      "similarity_fct": "cos_sim"
  }
  ```

### Training Hyperparameters
#### Non-Default Hyperparameters

- `eval_strategy`: steps
- `per_device_train_batch_size`: 32
- `per_device_eval_batch_size`: 32
- `num_train_epochs`: 1
- `batch_sampler`: no_duplicates
- `multi_dataset_batch_sampler`: round_robin

#### All Hyperparameters
<details><summary>Click to expand</summary>

- `overwrite_output_dir`: False
- `do_predict`: False
- `eval_strategy`: steps
- `prediction_loss_only`: True
- `per_device_train_batch_size`: 32
- `per_device_eval_batch_size`: 32
- `per_gpu_train_batch_size`: None
- `per_gpu_eval_batch_size`: None
- `gradient_accumulation_steps`: 1
- `eval_accumulation_steps`: None
- `torch_empty_cache_steps`: None
- `learning_rate`: 5e-05
- `weight_decay`: 0.0
- `adam_beta1`: 0.9
- `adam_beta2`: 0.999
- `adam_epsilon`: 1e-08
- `max_grad_norm`: 1
- `num_train_epochs`: 1
- `max_steps`: -1
- `lr_scheduler_type`: linear
- `lr_scheduler_kwargs`: {}
- `warmup_ratio`: 0.0
- `warmup_steps`: 0
- `log_level`: passive
- `log_level_replica`: warning
- `log_on_each_node`: True
- `logging_nan_inf_filter`: True
- `save_safetensors`: True
- `save_on_each_node`: False
- `save_only_model`: False
- `restore_callback_states_from_checkpoint`: False
- `no_cuda`: False
- `use_cpu`: False
- `use_mps_device`: False
- `seed`: 42
- `data_seed`: None
- `jit_mode_eval`: False
- `use_ipex`: False
- `bf16`: False
- `fp16`: False
- `fp16_opt_level`: O1
- `half_precision_backend`: auto
- `bf16_full_eval`: False
- `fp16_full_eval`: False
- `tf32`: None
- `local_rank`: 0
- `ddp_backend`: None
- `tpu_num_cores`: None
- `tpu_metrics_debug`: False
- `debug`: []
- `dataloader_drop_last`: False
- `dataloader_num_workers`: 0
- `dataloader_prefetch_factor`: None
- `past_index`: -1
- `disable_tqdm`: False
- `remove_unused_columns`: True
- `label_names`: None
- `load_best_model_at_end`: False
- `ignore_data_skip`: False
- `fsdp`: []
- `fsdp_min_num_params`: 0
- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
- `tp_size`: 0
- `fsdp_transformer_layer_cls_to_wrap`: None
- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
- `deepspeed`: None
- `label_smoothing_factor`: 0.0
- `optim`: adamw_torch
- `optim_args`: None
- `adafactor`: False
- `group_by_length`: False
- `length_column_name`: length
- `ddp_find_unused_parameters`: None
- `ddp_bucket_cap_mb`: None
- `ddp_broadcast_buffers`: False
- `dataloader_pin_memory`: True
- `dataloader_persistent_workers`: False
- `skip_memory_metrics`: True
- `use_legacy_prediction_loop`: False
- `push_to_hub`: False
- `resume_from_checkpoint`: None
- `hub_model_id`: None
- `hub_strategy`: every_save
- `hub_private_repo`: None
- `hub_always_push`: False
- `gradient_checkpointing`: False
- `gradient_checkpointing_kwargs`: None
- `include_inputs_for_metrics`: False
- `include_for_metrics`: []
- `eval_do_concat_batches`: True
- `fp16_backend`: auto
- `push_to_hub_model_id`: None
- `push_to_hub_organization`: None
- `mp_parameters`: 
- `auto_find_batch_size`: False
- `full_determinism`: False
- `torchdynamo`: None
- `ray_scope`: last
- `ddp_timeout`: 1800
- `torch_compile`: False
- `torch_compile_backend`: None
- `torch_compile_mode`: None
- `include_tokens_per_second`: False
- `include_num_input_tokens_seen`: False
- `neftune_noise_alpha`: None
- `optim_target_modules`: None
- `batch_eval_metrics`: False
- `eval_on_start`: False
- `use_liger_kernel`: False
- `eval_use_gather_object`: False
- `average_tokens_across_devices`: False
- `prompts`: None
- `batch_sampler`: no_duplicates
- `multi_dataset_batch_sampler`: round_robin

</details>

### Training Logs
| Epoch  | Step | Training Loss | ir-eval_cosine_ndcg@10 |
|:------:|:----:|:-------------:|:----------------------:|
| 0.1631 | 500  | 0.0634        | 0.9563                 |
| 0.3262 | 1000 | 0.005         | 0.9627                 |
| 0.4892 | 1500 | 0.0037        | 0.9631                 |
| 0.6523 | 2000 | 0.0029        | 0.9660                 |
| 0.8154 | 2500 | 0.0033        | 0.9663                 |
| 0.9785 | 3000 | 0.0027        | 0.9670                 |
| 1.0    | 3066 | -             | 0.9670                 |


### Framework Versions
- Python: 3.12.8
- Sentence Transformers: 3.4.1
- Transformers: 4.51.3
- PyTorch: 2.5.1+cu124
- Accelerate: 1.3.0
- Datasets: 3.2.0
- Tokenizers: 0.21.0

## Citation

### BibTeX

#### Sentence Transformers
```bibtex
@inproceedings{reimers-2019-sentence-bert,
    title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
    author = "Reimers, Nils and Gurevych, Iryna",
    booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
    month = "11",
    year = "2019",
    publisher = "Association for Computational Linguistics",
    url = "https://arxiv.org/abs/1908.10084",
}
```

#### MultipleNegativesRankingLoss
```bibtex
@misc{henderson2017efficient,
    title={Efficient Natural Language Response Suggestion for Smart Reply},
    author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
    year={2017},
    eprint={1705.00652},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}
```

<!--
## Glossary

*Clearly define terms in order to be accessible across audiences.*
-->

<!--
## Model Card Authors

*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
-->

<!--
## Model Card Contact

*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
-->