LiYuan commited on
Commit
5882ec4
1 Parent(s): dbad4e3
Files changed (1) hide show
  1. README.md +17 -1
README.md CHANGED
@@ -33,11 +33,27 @@ process to generate inputs and labels from those texts using the BERT base model
33
  DistilBERT is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)
34
  to make decisions, such as sequence classification, token classification, or question answering. This fine-tuned version of DistilBERT is used to predict the relevance between one query and one product description. It also can be used to rerank the relevance order of products given one query for the amazon platform or other e-commerce platforms.
35
 
 
 
36
  ## How to use
37
 
 
 
 
 
 
 
 
 
 
 
38
  ## Training and evaluation data
39
 
40
- More information needed
 
 
 
 
41
 
42
  ## Training procedure
43
 
 
33
  DistilBERT is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)
34
  to make decisions, such as sequence classification, token classification, or question answering. This fine-tuned version of DistilBERT is used to predict the relevance between one query and one product description. It also can be used to rerank the relevance order of products given one query for the amazon platform or other e-commerce platforms.
35
 
36
+ The limitations are this trained model is focusing on queries and products on Amazon. If you apply this model to other domains, it may perform poorly.
37
+
38
  ## How to use
39
 
40
+ You can use this model directly by downloading the trained weights and configurations like the below code snippet:
41
+
42
+ ```python
43
+ from transformers import AutoTokenizer, AutoModelForSequenceClassification
44
+
45
+ tokenizer = AutoTokenizer.from_pretrained("LiYuan/amazon-query-product-ranking")
46
+
47
+ model = AutoModelForSequenceClassification.from_pretrained("LiYuan/amazon-query-product-ranking")
48
+ ```
49
+
50
  ## Training and evaluation data
51
 
52
+ Download all the raw [dataset](https://www.aicrowd.com/challenges/esci-challenge-for-improving-product-search/dataset_files) from the Amazon KDD Cup website.
53
+
54
+ The three dataset we used like below:
55
+
56
+
57
 
58
  ## Training procedure
59