lv12 commited on
Commit
b9d6070
1 Parent(s): 66df9c3

Add new SentenceTransformer model.

Browse files
Files changed (3) hide show
  1. README.md +117 -177
  2. config.json +1 -1
  3. model.safetensors +1 -1
README.md CHANGED
@@ -5,67 +5,36 @@ tags:
5
  - sentence-transformers
6
  - sentence-similarity
7
  - feature-extraction
8
- - dataset_size:1K<n<10K
9
  - loss:CachedGISTEmbedLoss
10
  base_model: nomic-ai/nomic-embed-text-v1.5
11
- metrics:
12
- - cosine_accuracy
13
- - dot_accuracy
14
- - manhattan_accuracy
15
- - euclidean_accuracy
16
- - max_accuracy
17
  widget:
18
- - source_sentence: Pilot
19
  sentences:
20
- - Episode Two
21
- - dog dinosaur bone
22
- - 10' x 12' gazebo
23
- - source_sentence: skull
24
  sentences:
25
- - cool head s
26
- - trunk bike rack 4
27
- - bread without gluten
28
- - source_sentence: pipes
29
  sentences:
30
- - chillum pipe
31
- - Deckle Edge Ruler
32
- - dog collar for boxer
33
- - source_sentence: ddj400
34
  sentences:
35
- - lc27h711qenxza
36
- - bed frame for full
37
- - chicago bears gifts
38
- - source_sentence: primes
39
  sentences:
40
- - Newton
41
- - big boys sneakers
42
- - large dog clothes
43
  pipeline_tag: sentence-similarity
44
- model-index:
45
- - name: SentenceTransformer based on nomic-ai/nomic-embed-text-v1.5
46
- results:
47
- - task:
48
- type: triplet
49
- name: Triplet
50
- dataset:
51
- name: esci dev
52
- type: esci-dev
53
- metrics:
54
- - type: cosine_accuracy
55
- value: 0.6414052697616061
56
- name: Cosine Accuracy
57
- - type: dot_accuracy
58
- value: 0.36637390213299875
59
- name: Dot Accuracy
60
- - type: manhattan_accuracy
61
- value: 0.6404015056461732
62
- name: Manhattan Accuracy
63
- - type: euclidean_accuracy
64
- value: 0.6406524466750314
65
- name: Euclidean Accuracy
66
- - type: max_accuracy
67
- value: 0.6414052697616061
68
- name: Max Accuracy
69
  ---
70
 
71
  # SentenceTransformer based on nomic-ai/nomic-embed-text-v1.5
@@ -117,9 +86,9 @@ from sentence_transformers import SentenceTransformer
117
  model = SentenceTransformer("sentence_transformers_model_id")
118
  # Run inference
119
  sentences = [
120
- 'primes',
121
- 'Newton',
122
- 'big boys sneakers',
123
  ]
124
  embeddings = model.encode(sentences)
125
  print(embeddings.shape)
@@ -155,22 +124,6 @@ You can finetune this model on your own dataset.
155
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
156
  -->
157
 
158
- ## Evaluation
159
-
160
- ### Metrics
161
-
162
- #### Triplet
163
- * Dataset: `esci-dev`
164
- * Evaluated with [<code>TripletEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.TripletEvaluator)
165
-
166
- | Metric | Value |
167
- |:-------------------|:-----------|
168
- | cosine_accuracy | 0.6414 |
169
- | dot_accuracy | 0.3664 |
170
- | manhattan_accuracy | 0.6404 |
171
- | euclidean_accuracy | 0.6407 |
172
- | **max_accuracy** | **0.6414** |
173
-
174
  <!--
175
  ## Bias, Risks and Limitations
176
 
@@ -190,46 +143,19 @@ You can finetune this model on your own dataset.
190
  #### Unnamed Dataset
191
 
192
 
193
- * Size: 9,090 training samples
194
- * Columns: <code>query</code>, <code>pos</code>, and <code>neg</code>
195
- * Approximate statistics based on the first 1000 samples:
196
- | | query | pos | neg |
197
- |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|
198
- | type | string | string | string |
199
- | details | <ul><li>min: 3 tokens</li><li>mean: 7.42 tokens</li><li>max: 30 tokens</li></ul> | <ul><li>min: 3 tokens</li><li>mean: 29.27 tokens</li><li>max: 87 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 29.8 tokens</li><li>max: 82 tokens</li></ul> |
200
- * Samples:
201
- | query | pos | neg |
202
- |:--------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
203
- | <code>1 3/4 inch tooled belt strap without belt buckle</code> | <code>BS3501 Solid Brass Leaf Belt Buckle Fits 1-3/4"(45mm) Wide Belt</code> | <code>Nocona Men's Hired Brown Floral Eagle, 40</code> |
204
- | <code>7edge phone case peacock</code> | <code>Galaxy S7 Edge Case for Girls Women Clear with Flowers Design Shockproof Protective Cell Phone Cases for Samsung Galaxy S7 Edge 5.5 Inch Cute Floral Pattern Print Flexible Slim Fit Bumper Rubber Cover</code> | <code>Galaxy S7 Case, Galaxy S7 Phone Case with HD Screen Protector for Girls Women, Gritup Cute Clear Gradient Glitter Liquid TPU Slim Phone Case for Samsung Galaxy S7 Teal/Purple</code> |
205
- | <code>girls white shoes</code> | <code>adidas Women's Coast Star Shoes, ftwr White/Silver Met./ core Black, 6 M US</code> | <code>Converse Optical White M7650 - HI TOP Size 6 M US Women / 4 M US Men</code> |
206
- * Loss: [<code>CachedGISTEmbedLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedgistembedloss) with these parameters:
207
- ```json
208
- {'guide': SentenceTransformer(
209
- (0): Transformer({'max_seq_length': 256, 'do_lower_case': False}) with Transformer model: BertModel
210
- (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
211
- (2): Normalize()
212
- ), 'temperature': 0.01}
213
- ```
214
-
215
- ### Evaluation Dataset
216
-
217
- #### Unnamed Dataset
218
-
219
-
220
- * Size: 3,985 evaluation samples
221
- * Columns: <code>query</code>, <code>pos</code>, and <code>neg</code>
222
  * Approximate statistics based on the first 1000 samples:
223
- | | query | pos | neg |
224
- |:--------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
225
- | type | string | string | string |
226
- | details | <ul><li>min: 3 tokens</li><li>mean: 7.28 tokens</li><li>max: 25 tokens</li></ul> | <ul><li>min: 3 tokens</li><li>mean: 28.58 tokens</li><li>max: 116 tokens</li></ul> | <ul><li>min: 3 tokens</li><li>mean: 29.26 tokens</li><li>max: 79 tokens</li></ul> |
227
  * Samples:
228
- | query | pos | neg |
229
- |:--------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
230
- | <code>colors for dining room</code> | <code>AOOS CUSTOM Dimmable LED Neon Signs for Home Bedroom Salon Dining Room Wall Decor (Customization: Texts, Designs, Logos, Languages, Colors, Sizes, Fonts, Color-Changing) (24" / 1 Line Text)</code> | <code>Jetec 5 Pieces EAT Sign Kitchen Wood Rustic Sign Arrow Wall Decor EAT Farmhouse Decoration Hanging Arrow Wooden Sign for Kitchen Wall Home Dining Room (Charming Color)</code> |
231
- | <code>mix no 6 heels for women</code> | <code>DREAM PAIRS Women's Hi-Chunk Gold Glitter High Heel Pump Sandals - 6 M US</code> | <code>Fashare Womens High Heels Pointed Toe Bowtie Back Ankle Buckle Strap Wedding Evening Party Dress Pumps Shoes</code> |
232
- | <code>goxlrmini</code> | <code>Singing Machine SMM-205 Unidirectional Dynamic Microphone with 10 Ft. Cord,Black, one size</code> | <code>Behringer U-Phoria Studio Pro Complete Recording Bundle with UMC202HD USB Audio Interface - With 20' 6mm Rubber XLR Microphone Cable, On-Stage MBS5000 Broadcast/Webcast Boom Arm with XLR Cable</code> |
233
  * Loss: [<code>CachedGISTEmbedLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedgistembedloss) with these parameters:
234
  ```json
235
  {'guide': SentenceTransformer(
@@ -242,9 +168,10 @@ You can finetune this model on your own dataset.
242
  ### Training Hyperparameters
243
  #### Non-Default Hyperparameters
244
 
245
- - `per_device_train_batch_size`: 16
246
- - `per_device_eval_batch_size`: 16
247
- - `num_train_epochs`: 10
 
248
  - `warmup_ratio`: 0.1
249
  - `fp16`: True
250
  - `batch_sampler`: no_duplicates
@@ -255,21 +182,21 @@ You can finetune this model on your own dataset.
255
  - `overwrite_output_dir`: False
256
  - `do_predict`: False
257
  - `prediction_loss_only`: True
258
- - `per_device_train_batch_size`: 16
259
- - `per_device_eval_batch_size`: 16
260
  - `per_gpu_train_batch_size`: None
261
  - `per_gpu_eval_batch_size`: None
262
- - `gradient_accumulation_steps`: 1
263
  - `eval_accumulation_steps`: None
264
- - `learning_rate`: 5e-05
265
  - `weight_decay`: 0.0
266
  - `adam_beta1`: 0.9
267
  - `adam_beta2`: 0.999
268
  - `adam_epsilon`: 1e-08
269
  - `max_grad_norm`: 1.0
270
- - `num_train_epochs`: 10
271
  - `max_steps`: -1
272
- - `lr_scheduler_type`: linear
273
  - `lr_scheduler_kwargs`: {}
274
  - `warmup_ratio`: 0.1
275
  - `warmup_steps`: 0
@@ -359,65 +286,78 @@ You can finetune this model on your own dataset.
359
  </details>
360
 
361
  ### Training Logs
362
- | Epoch | Step | Training Loss | esci-dev_max_accuracy |
363
- |:------:|:----:|:-------------:|:---------------------:|
364
- | 0 | 0 | - | 0.6414 |
365
- | 0.1757 | 100 | 0.8875 | - |
366
- | 0.3515 | 200 | 0.5281 | - |
367
- | 0.5272 | 300 | 0.4621 | - |
368
- | 0.7030 | 400 | 0.4669 | - |
369
- | 0.8787 | 500 | 0.4501 | - |
370
- | 1.0545 | 600 | 0.5379 | - |
371
- | 1.2302 | 700 | 0.4288 | - |
372
- | 1.4060 | 800 | 0.2112 | - |
373
- | 1.5817 | 900 | 0.1508 | - |
374
- | 1.7575 | 1000 | 0.1133 | - |
375
- | 1.9332 | 1100 | 0.1312 | - |
376
- | 2.1090 | 1200 | 0.0784 | - |
377
- | 2.2847 | 1300 | 0.0983 | - |
378
- | 2.4605 | 1400 | 0.106 | - |
379
- | 2.6362 | 1500 | 0.1058 | - |
380
- | 2.8120 | 1600 | 0.0673 | - |
381
- | 2.9877 | 1700 | 0.0355 | - |
382
- | 3.1634 | 1800 | 0.0175 | - |
383
- | 3.3392 | 1900 | 0.0366 | - |
384
- | 3.5149 | 2000 | 0.0332 | - |
385
- | 3.6907 | 2100 | 0.0682 | - |
386
- | 3.8664 | 2200 | 0.0378 | - |
387
- | 4.0422 | 2300 | 0.0239 | - |
388
- | 4.2179 | 2400 | 0.0282 | - |
389
- | 4.3937 | 2500 | 0.0401 | - |
390
- | 4.5694 | 2600 | 0.0268 | - |
391
- | 4.7452 | 2700 | 0.0208 | - |
392
- | 4.9209 | 2800 | 0.0117 | - |
393
- | 5.0967 | 2900 | 0.0045 | - |
394
- | 5.2724 | 3000 | 0.0145 | - |
395
- | 5.4482 | 3100 | 0.029 | - |
396
- | 5.6239 | 3200 | 0.0009 | - |
397
- | 5.7996 | 3300 | 0.0033 | - |
398
- | 5.9754 | 3400 | 0.0088 | - |
399
- | 6.1511 | 3500 | 0.0014 | - |
400
- | 6.3269 | 3600 | 0.0027 | - |
401
- | 6.5026 | 3700 | 0.0021 | - |
402
- | 6.6784 | 3800 | 0.0001 | - |
403
- | 6.8541 | 3900 | 0.0025 | - |
404
- | 7.0299 | 4000 | 0.0059 | - |
405
- | 7.2056 | 4100 | 0.0025 | - |
406
- | 7.3814 | 4200 | 0.0029 | - |
407
- | 7.5571 | 4300 | 0.0007 | - |
408
- | 7.7329 | 4400 | 0.0018 | - |
409
- | 7.9086 | 4500 | 0.0032 | - |
410
- | 8.0844 | 4600 | 0.0007 | - |
411
- | 8.2601 | 4700 | 0.0027 | - |
412
- | 8.4359 | 4800 | 0.0027 | - |
413
- | 8.6116 | 4900 | 0.0 | - |
414
- | 8.7873 | 5000 | 0.0025 | - |
415
- | 8.9631 | 5100 | 0.0025 | - |
416
- | 9.1388 | 5200 | 0.0014 | - |
417
- | 9.3146 | 5300 | 0.0027 | - |
418
- | 9.4903 | 5400 | 0.0021 | - |
419
- | 9.6661 | 5500 | 0.0 | - |
420
- | 9.8418 | 5600 | 0.0025 | - |
 
 
 
 
 
 
 
 
 
 
 
 
 
421
 
422
 
423
  ### Framework Versions
 
5
  - sentence-transformers
6
  - sentence-similarity
7
  - feature-extraction
8
+ - dataset_size:1M<n<10M
9
  - loss:CachedGISTEmbedLoss
10
  base_model: nomic-ai/nomic-embed-text-v1.5
 
 
 
 
 
 
11
  widget:
12
+ - source_sentence: 'search_query: 楢崎壮太'
13
  sentences:
14
+ - 'search_query: 野崎萌香'
15
+ - 'search_query: ps4 slim 1tb'
16
+ - 'search_query: toy story 4 on dvd'
17
+ - source_sentence: 'search_query: テプラ'
18
  sentences:
19
+ - 'search_query: 携帯デコシール'
20
+ - 'search_query: womens boots'
21
+ - 'search_query: nfl gift'
22
+ - source_sentence: 'search_query: 扇子 布'
23
  sentences:
24
+ - 'search_query: 天気の子'
25
+ - 'search_query: 登山ぐつ メンズ 紐なし'
26
+ - 'search_query: 10gbe switch'
27
+ - source_sentence: 'search_query: リング棒'
28
  sentences:
29
+ - 'search_query: ライトショアジギング'
30
+ - 'search_query: auvハンガーすべらない'
31
+ - 'search_query: plastic drum lid'
32
+ - source_sentence: 'search_query: 聖 龍人'
33
  sentences:
34
+ - 'search_query: 越前かに職人甲羅組'
35
+ - 'search_query: tea tree oil'
36
+ - 'search_query: lift storage bed'
37
  pipeline_tag: sentence-similarity
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
  ---
39
 
40
  # SentenceTransformer based on nomic-ai/nomic-embed-text-v1.5
 
86
  model = SentenceTransformer("sentence_transformers_model_id")
87
  # Run inference
88
  sentences = [
89
+ 'search_query: 聖 龍人',
90
+ 'search_query: 越前かに職人甲羅組',
91
+ 'search_query: tea tree oil',
92
  ]
93
  embeddings = model.encode(sentences)
94
  print(embeddings.shape)
 
124
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
125
  -->
126
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
127
  <!--
128
  ## Bias, Risks and Limitations
129
 
 
143
  #### Unnamed Dataset
144
 
145
 
146
+ * Size: 1,767,572 training samples
147
+ * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
148
  * Approximate statistics based on the first 1000 samples:
149
+ | | anchor | positive | negative |
150
+ |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
151
+ | type | string | string | string |
152
+ | details | <ul><li>min: 7 tokens</li><li>mean: 12.26 tokens</li><li>max: 59 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 31.93 tokens</li><li>max: 140 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 34.3 tokens</li><li>max: 157 tokens</li></ul> |
153
  * Samples:
154
+ | anchor | positive | negative |
155
+ |:---------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------|
156
+ | <code>search_query: plus tops for women</code> | <code>search_document:Just My Size Women's Plus-Size Graphic Short Sleeve V-Neck T-Shirt, White-Y07188, 5X, JUST MY SIZE, White-y07188</code> | <code>search_document:Calvin Klein Women's Regular Modern Cotton Bralette, Nymph's Thigh, S, Calvin Klein, Nymph's Thigh</code> |
157
+ | <code>search_query: mens black wallet trifold</code> | <code>search_document:Stealth Mode Trifold RFID Blocking Leather Wallet for Men (Black), Stealth Mode, Black</code> | <code>search_document:RFID Trifold Canvas Outdoor Sports Wallet for Kids - Front Pocket Wallet with Magic Sticker (Black), AI-DEE, Black</code> |
158
+ | <code>search_query: ipad pro reacondicionado 12,9</code> | <code>search_document:Apple iPad Pro (12.9 Pouces, Wi-FI, 64Go) 2018 - Gray (Renewed), Apple, Gris Espacial</code> | <code>search_document:Apple iPad Pro 3rd Generation (11-Inch, Wi-FI Only 64GB) - Space Gray (Renewed), Apple, Gris Espacial</code> |
159
  * Loss: [<code>CachedGISTEmbedLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedgistembedloss) with these parameters:
160
  ```json
161
  {'guide': SentenceTransformer(
 
168
  ### Training Hyperparameters
169
  #### Non-Default Hyperparameters
170
 
171
+ - `per_device_eval_batch_size`: 2
172
+ - `gradient_accumulation_steps`: 2
173
+ - `learning_rate`: 1e-05
174
+ - `lr_scheduler_type`: cosine_with_restarts
175
  - `warmup_ratio`: 0.1
176
  - `fp16`: True
177
  - `batch_sampler`: no_duplicates
 
182
  - `overwrite_output_dir`: False
183
  - `do_predict`: False
184
  - `prediction_loss_only`: True
185
+ - `per_device_train_batch_size`: 8
186
+ - `per_device_eval_batch_size`: 2
187
  - `per_gpu_train_batch_size`: None
188
  - `per_gpu_eval_batch_size`: None
189
+ - `gradient_accumulation_steps`: 2
190
  - `eval_accumulation_steps`: None
191
+ - `learning_rate`: 1e-05
192
  - `weight_decay`: 0.0
193
  - `adam_beta1`: 0.9
194
  - `adam_beta2`: 0.999
195
  - `adam_epsilon`: 1e-08
196
  - `max_grad_norm`: 1.0
197
+ - `num_train_epochs`: 3
198
  - `max_steps`: -1
199
+ - `lr_scheduler_type`: cosine_with_restarts
200
  - `lr_scheduler_kwargs`: {}
201
  - `warmup_ratio`: 0.1
202
  - `warmup_steps`: 0
 
286
  </details>
287
 
288
  ### Training Logs
289
+ | Epoch | Step | Training Loss |
290
+ |:------:|:----:|:-------------:|
291
+ | 0.0009 | 100 | 3.7009 |
292
+ | 0.0018 | 200 | 3.3796 |
293
+ | 0.0027 | 300 | 2.8348 |
294
+ | 0.0036 | 400 | 2.1803 |
295
+ | 0.0045 | 500 | 1.8272 |
296
+ | 0.0054 | 600 | 1.4715 |
297
+ | 0.0063 | 700 | 1.0056 |
298
+ | 0.0072 | 800 | 0.7984 |
299
+ | 0.0081 | 900 | 0.6925 |
300
+ | 0.0091 | 1000 | 0.6552 |
301
+ | 0.0100 | 1100 | 0.6054 |
302
+ | 0.0109 | 1200 | 0.5874 |
303
+ | 0.0118 | 1300 | 0.5641 |
304
+ | 0.0127 | 1400 | 0.528 |
305
+ | 0.0136 | 1500 | 0.5285 |
306
+ | 0.0145 | 1600 | 0.5032 |
307
+ | 0.0154 | 1700 | 0.5238 |
308
+ | 0.0163 | 1800 | 0.4565 |
309
+ | 0.0172 | 1900 | 0.4739 |
310
+ | 0.0181 | 2000 | 0.4614 |
311
+ | 0.0190 | 2100 | 0.4334 |
312
+ | 0.0199 | 2200 | 0.4217 |
313
+ | 0.0208 | 2300 | 0.3931 |
314
+ | 0.0217 | 2400 | 0.4102 |
315
+ | 0.0226 | 2500 | 0.3936 |
316
+ | 0.0235 | 2600 | 0.415 |
317
+ | 0.0244 | 2700 | 0.4462 |
318
+ | 0.0253 | 2800 | 0.3886 |
319
+ | 0.0263 | 2900 | 0.3887 |
320
+ | 0.0272 | 3000 | 0.3629 |
321
+ | 0.0281 | 3100 | 0.37 |
322
+ | 0.0290 | 3200 | 0.3861 |
323
+ | 0.0299 | 3300 | 0.3813 |
324
+ | 0.0308 | 3400 | 0.3348 |
325
+ | 0.0317 | 3500 | 0.3566 |
326
+ | 0.0326 | 3600 | 0.3674 |
327
+ | 0.0335 | 3700 | 0.3421 |
328
+ | 0.0344 | 3800 | 0.3225 |
329
+ | 0.0353 | 3900 | 0.406 |
330
+ | 0.0362 | 4000 | 0.3975 |
331
+ | 0.0371 | 4100 | 0.368 |
332
+ | 0.0380 | 4200 | 0.3481 |
333
+ | 0.0389 | 4300 | 0.3405 |
334
+ | 0.0398 | 4400 | 0.3529 |
335
+ | 0.0407 | 4500 | 0.3968 |
336
+ | 0.0416 | 4600 | 0.3634 |
337
+ | 0.0425 | 4700 | 0.3518 |
338
+ | 0.0434 | 4800 | 0.383 |
339
+ | 0.0444 | 4900 | 0.3261 |
340
+ | 0.0453 | 5000 | 0.323 |
341
+ | 0.0462 | 5100 | 0.3372 |
342
+ | 0.0471 | 5200 | 0.358 |
343
+ | 0.0480 | 5300 | 0.3207 |
344
+ | 0.0489 | 5400 | 0.341 |
345
+ | 0.0498 | 5500 | 0.3146 |
346
+ | 0.0507 | 5600 | 0.3065 |
347
+ | 0.0516 | 5700 | 0.3597 |
348
+ | 0.0525 | 5800 | 0.3352 |
349
+ | 0.0534 | 5900 | 0.3212 |
350
+ | 0.0543 | 6000 | 0.316 |
351
+ | 0.0552 | 6100 | 0.3405 |
352
+ | 0.0561 | 6200 | 0.3416 |
353
+ | 0.0570 | 6300 | 0.3124 |
354
+ | 0.0579 | 6400 | 0.3146 |
355
+ | 0.0588 | 6500 | 0.3043 |
356
+ | 0.0597 | 6600 | 0.3687 |
357
+ | 0.0606 | 6700 | 0.3359 |
358
+ | 0.0616 | 6800 | 0.3414 |
359
+ | 0.0625 | 6900 | 0.3161 |
360
+ | 0.0634 | 7000 | 0.3266 |
361
 
362
 
363
  ### Framework Versions
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "models/nomic-embed-text-esci/checkpoint-5600",
3
  "activation_function": "swiglu",
4
  "architectures": [
5
  "NomicBertModel"
 
1
  {
2
+ "_name_or_path": "models/nomic-embed-text-esci/checkpoint-7000",
3
  "activation_function": "swiglu",
4
  "architectures": [
5
  "NomicBertModel"
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85c0ee2699043dfafce03c1ee919a4bf7ef793421919aeb9892d0f41a6b7de62
3
  size 546938168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d84393e7281f9fa1749b614ba8833b120c31a65d73fe00f693417d03f68231c
3
  size 546938168