LiYuan commited on
Commit
dbad4e3
1 Parent(s): f26777c
Files changed (1) hide show
  1. README.md +13 -5
README.md CHANGED
@@ -5,27 +5,35 @@ tags:
5
  metrics:
6
  - accuracy
7
  model-index:
8
- - name: distilbert-base-uncased-finetuned-mnli
9
  results: []
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
- # distilbert-base-uncased-finetuned-mnli
16
 
17
- This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
 
 
18
  It achieves the following results on the evaluation set:
19
  - Loss: 0.8244
20
  - Accuracy: 0.6617
21
 
22
  ## Model description
23
 
24
- More information needed
 
 
 
25
 
26
  ## Intended uses & limitations
27
 
28
- More information needed
 
 
 
29
 
30
  ## Training and evaluation data
31
 
 
5
  metrics:
6
  - accuracy
7
  model-index:
8
+ - name: distilbert-base-uncased-finetuned-mnli-amazon-query-shopping
9
  results: []
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
+ # distilbert-base-uncased-finetuned-mnli-amazon-query-shopping
16
 
17
+ This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an [Amazon shopping query dataset](https://www.aicrowd.com/challenges/esci-challenge-for-improving-product-search). The code for the fine-tuning process can be found
18
+ [here](https://github.com/vanderbilt-data-science/sna). This model is uncased: it does
19
+ not make a difference between english and English.
20
  It achieves the following results on the evaluation set:
21
  - Loss: 0.8244
22
  - Accuracy: 0.6617
23
 
24
  ## Model description
25
 
26
+ DistilBERT is a transformers model, smaller and faster than BERT, which was pretrained on the same corpus in a
27
+ self-supervised fashion, using the BERT base model as a teacher. This means it was pretrained on the raw texts only,
28
+ with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic
29
+ process to generate inputs and labels from those texts using the BERT base model. We replaced its head with our shopping relevance category to fine-tune it on 571,223 rows of training set while validate it on 142,806 rows of dev set. Finally, we evaluated our model performance on a held-out test set: 79,337 rows.
30
 
31
  ## Intended uses & limitations
32
 
33
+ DistilBERT is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)
34
+ to make decisions, such as sequence classification, token classification, or question answering. This fine-tuned version of DistilBERT is used to predict the relevance between one query and one product description. It also can be used to rerank the relevance order of products given one query for the amazon platform or other e-commerce platforms.
35
+
36
+ ## How to use
37
 
38
  ## Training and evaluation data
39