BartekSadlej
commited on
Commit
•
e974ff3
1
Parent(s):
372602d
Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -12,7 +12,7 @@ language:
|
|
12 |
- pl
|
13 |
---
|
14 |
model-index:
|
15 |
-
- name: 2024-06-
|
16 |
results:
|
17 |
- dataset:
|
18 |
config: pl
|
@@ -22,7 +22,7 @@ model-index:
|
|
22 |
type: mteb/amazon_massive_intent
|
23 |
metrics:
|
24 |
- type: accuracy
|
25 |
-
value: 0.
|
26 |
task:
|
27 |
type: Classification
|
28 |
- dataset:
|
@@ -33,7 +33,7 @@ model-index:
|
|
33 |
type: mteb/amazon_massive_intent
|
34 |
metrics:
|
35 |
- type: accuracy
|
36 |
-
value: 0.
|
37 |
task:
|
38 |
type: Classification
|
39 |
- dataset:
|
@@ -44,7 +44,7 @@ model-index:
|
|
44 |
type: mteb/amazon_massive_scenario
|
45 |
metrics:
|
46 |
- type: accuracy
|
47 |
-
value: 0.
|
48 |
task:
|
49 |
type: Classification
|
50 |
- dataset:
|
@@ -55,7 +55,7 @@ model-index:
|
|
55 |
type: mteb/amazon_massive_scenario
|
56 |
metrics:
|
57 |
- type: accuracy
|
58 |
-
value: 0.
|
59 |
task:
|
60 |
type: Classification
|
61 |
- dataset:
|
@@ -66,7 +66,7 @@ model-index:
|
|
66 |
type: PL-MTEB/cbd
|
67 |
metrics:
|
68 |
- type: accuracy
|
69 |
-
value: 0.
|
70 |
task:
|
71 |
type: Classification
|
72 |
- dataset:
|
@@ -77,7 +77,7 @@ model-index:
|
|
77 |
type: PL-MTEB/polemo2_in
|
78 |
metrics:
|
79 |
- type: accuracy
|
80 |
-
value: 0.
|
81 |
task:
|
82 |
type: Classification
|
83 |
- dataset:
|
@@ -88,7 +88,7 @@ model-index:
|
|
88 |
type: PL-MTEB/polemo2_out
|
89 |
metrics:
|
90 |
- type: accuracy
|
91 |
-
value: 0.
|
92 |
task:
|
93 |
type: Classification
|
94 |
- dataset:
|
@@ -99,7 +99,7 @@ model-index:
|
|
99 |
type: PL-MTEB/allegro-reviews
|
100 |
metrics:
|
101 |
- type: accuracy
|
102 |
-
value: 0.
|
103 |
task:
|
104 |
type: Classification
|
105 |
- dataset:
|
@@ -110,7 +110,7 @@ model-index:
|
|
110 |
type: laugustyniak/abusive-clauses-pl
|
111 |
metrics:
|
112 |
- type: accuracy
|
113 |
-
value: 0.
|
114 |
task:
|
115 |
type: Classification
|
116 |
- dataset:
|
@@ -121,7 +121,7 @@ model-index:
|
|
121 |
type: PL-MTEB/8tags-clustering
|
122 |
metrics:
|
123 |
- type: v_measure
|
124 |
-
value: 0.
|
125 |
task:
|
126 |
type: Clustering
|
127 |
- dataset:
|
@@ -132,7 +132,7 @@ model-index:
|
|
132 |
type: PL-MTEB/sicke-pl-pairclassification
|
133 |
metrics:
|
134 |
- type: ap
|
135 |
-
value: 0.
|
136 |
task:
|
137 |
type: PairClassification
|
138 |
- dataset:
|
@@ -143,7 +143,7 @@ model-index:
|
|
143 |
type: PL-MTEB/cdsce-pairclassification
|
144 |
metrics:
|
145 |
- type: ap
|
146 |
-
value: 0.
|
147 |
task:
|
148 |
type: PairClassification
|
149 |
- dataset:
|
@@ -154,7 +154,7 @@ model-index:
|
|
154 |
type: PL-MTEB/psc-pairclassification
|
155 |
metrics:
|
156 |
- type: ap
|
157 |
-
value: 0.
|
158 |
task:
|
159 |
type: PairClassification
|
160 |
- dataset:
|
@@ -165,7 +165,7 @@ model-index:
|
|
165 |
type: mteb/sts22-crosslingual-sts
|
166 |
metrics:
|
167 |
- type: cosine_spearman
|
168 |
-
value: -0.
|
169 |
task:
|
170 |
type: STS
|
171 |
- dataset:
|
@@ -176,7 +176,7 @@ model-index:
|
|
176 |
type: mteb/stsb_multi_mt
|
177 |
metrics:
|
178 |
- type: cosine_spearman
|
179 |
-
value: 0.
|
180 |
task:
|
181 |
type: STS
|
182 |
- dataset:
|
@@ -187,7 +187,7 @@ model-index:
|
|
187 |
type: mteb/stsb_multi_mt
|
188 |
metrics:
|
189 |
- type: cosine_spearman
|
190 |
-
value: 0.
|
191 |
task:
|
192 |
type: STS
|
193 |
- dataset:
|
@@ -198,7 +198,7 @@ model-index:
|
|
198 |
type: PL-MTEB/sickr-pl-sts
|
199 |
metrics:
|
200 |
- type: cosine_spearman
|
201 |
-
value: 0.
|
202 |
task:
|
203 |
type: STS
|
204 |
- dataset:
|
@@ -209,7 +209,7 @@ model-index:
|
|
209 |
type: PL-MTEB/cdscr-sts
|
210 |
metrics:
|
211 |
- type: cosine_spearman
|
212 |
-
value: 0.
|
213 |
task:
|
214 |
type: STS
|
215 |
- dataset:
|
@@ -220,7 +220,7 @@ model-index:
|
|
220 |
type: PL-MTEB/plsc-clustering-s2s
|
221 |
metrics:
|
222 |
- type: v_measure
|
223 |
-
value: 0.
|
224 |
task:
|
225 |
type: Clustering
|
226 |
- dataset:
|
@@ -231,7 +231,7 @@ model-index:
|
|
231 |
type: PL-MTEB/plsc-clustering-p2p
|
232 |
metrics:
|
233 |
- type: v_measure
|
234 |
-
value: 0.
|
235 |
task:
|
236 |
type: Clustering
|
237 |
tags:
|
|
|
12 |
- pl
|
13 |
---
|
14 |
model-index:
|
15 |
+
- name: 2024-06-19_08-22-22
|
16 |
results:
|
17 |
- dataset:
|
18 |
config: pl
|
|
|
22 |
type: mteb/amazon_massive_intent
|
23 |
metrics:
|
24 |
- type: accuracy
|
25 |
+
value: 0.10346334902488233
|
26 |
task:
|
27 |
type: Classification
|
28 |
- dataset:
|
|
|
33 |
type: mteb/amazon_massive_intent
|
34 |
metrics:
|
35 |
- type: accuracy
|
36 |
+
value: 0.10678799803246433
|
37 |
task:
|
38 |
type: Classification
|
39 |
- dataset:
|
|
|
44 |
type: mteb/amazon_massive_scenario
|
45 |
metrics:
|
46 |
- type: accuracy
|
47 |
+
value: 0.1761936785474109
|
48 |
task:
|
49 |
type: Classification
|
50 |
- dataset:
|
|
|
55 |
type: mteb/amazon_massive_scenario
|
56 |
metrics:
|
57 |
- type: accuracy
|
58 |
+
value: 0.17629119527791443
|
59 |
task:
|
60 |
type: Classification
|
61 |
- dataset:
|
|
|
66 |
type: PL-MTEB/cbd
|
67 |
metrics:
|
68 |
- type: accuracy
|
69 |
+
value: 0.5073
|
70 |
task:
|
71 |
type: Classification
|
72 |
- dataset:
|
|
|
77 |
type: PL-MTEB/polemo2_in
|
78 |
metrics:
|
79 |
- type: accuracy
|
80 |
+
value: 0.3889196675900277
|
81 |
task:
|
82 |
type: Classification
|
83 |
- dataset:
|
|
|
88 |
type: PL-MTEB/polemo2_out
|
89 |
metrics:
|
90 |
- type: accuracy
|
91 |
+
value: 0.2937246963562753
|
92 |
task:
|
93 |
type: Classification
|
94 |
- dataset:
|
|
|
99 |
type: PL-MTEB/allegro-reviews
|
100 |
metrics:
|
101 |
- type: accuracy
|
102 |
+
value: 0.22037773359840954
|
103 |
task:
|
104 |
type: Classification
|
105 |
- dataset:
|
|
|
110 |
type: laugustyniak/abusive-clauses-pl
|
111 |
metrics:
|
112 |
- type: accuracy
|
113 |
+
value: 0.5505068056762236
|
114 |
task:
|
115 |
type: Classification
|
116 |
- dataset:
|
|
|
121 |
type: PL-MTEB/8tags-clustering
|
122 |
metrics:
|
123 |
- type: v_measure
|
124 |
+
value: 0.09774455155716733
|
125 |
task:
|
126 |
type: Clustering
|
127 |
- dataset:
|
|
|
132 |
type: PL-MTEB/sicke-pl-pairclassification
|
133 |
metrics:
|
134 |
- type: ap
|
135 |
+
value: 0.5140817433152867
|
136 |
task:
|
137 |
type: PairClassification
|
138 |
- dataset:
|
|
|
143 |
type: PL-MTEB/cdsce-pairclassification
|
144 |
metrics:
|
145 |
- type: ap
|
146 |
+
value: 0.46484295974270995
|
147 |
task:
|
148 |
type: PairClassification
|
149 |
- dataset:
|
|
|
154 |
type: PL-MTEB/psc-pairclassification
|
155 |
metrics:
|
156 |
- type: ap
|
157 |
+
value: 0.4051269007928542
|
158 |
task:
|
159 |
type: PairClassification
|
160 |
- dataset:
|
|
|
165 |
type: mteb/sts22-crosslingual-sts
|
166 |
metrics:
|
167 |
- type: cosine_spearman
|
168 |
+
value: -0.08590905486427534
|
169 |
task:
|
170 |
type: STS
|
171 |
- dataset:
|
|
|
176 |
type: mteb/stsb_multi_mt
|
177 |
metrics:
|
178 |
- type: cosine_spearman
|
179 |
+
value: 0.43179596275645044
|
180 |
task:
|
181 |
type: STS
|
182 |
- dataset:
|
|
|
187 |
type: mteb/stsb_multi_mt
|
188 |
metrics:
|
189 |
- type: cosine_spearman
|
190 |
+
value: 0.3366352638628963
|
191 |
task:
|
192 |
type: STS
|
193 |
- dataset:
|
|
|
198 |
type: PL-MTEB/sickr-pl-sts
|
199 |
metrics:
|
200 |
- type: cosine_spearman
|
201 |
+
value: 0.39053635596458536
|
202 |
task:
|
203 |
type: STS
|
204 |
- dataset:
|
|
|
209 |
type: PL-MTEB/cdscr-sts
|
210 |
metrics:
|
211 |
- type: cosine_spearman
|
212 |
+
value: 0.5696239277348915
|
213 |
task:
|
214 |
type: STS
|
215 |
- dataset:
|
|
|
220 |
type: PL-MTEB/plsc-clustering-s2s
|
221 |
metrics:
|
222 |
- type: v_measure
|
223 |
+
value: 0.2411184934551013
|
224 |
task:
|
225 |
type: Clustering
|
226 |
- dataset:
|
|
|
231 |
type: PL-MTEB/plsc-clustering-p2p
|
232 |
metrics:
|
233 |
- type: v_measure
|
234 |
+
value: 0.2120976462600393
|
235 |
task:
|
236 |
type: Clustering
|
237 |
tags:
|