mohammadhia commited on
Commit
75f6c1d
1 Parent(s): 6dfb2d8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -4
README.md CHANGED
@@ -18,7 +18,19 @@ should probably proofread and complete it, then remove this comment. -->
18
  # t5_recommendation_sports_equipment_english
19
 
20
  This model is a fine-tuned version of [t5-large](https://huggingface.co/t5-large) on a custom dataset, consisting of sports equipment customers have purchased, and items to recommended next.
21
- It achieves the following results on the evaluation set:
 
 
 
 
 
 
 
 
 
 
 
 
22
  - Loss: 0.4554
23
  - Rouge1: 57.1429
24
  - Rouge2: 47.6190
@@ -28,18 +40,24 @@ It achieves the following results on the evaluation set:
28
 
29
  ## Model description
30
 
31
- More information needed
 
 
 
 
32
 
33
  ## Intended uses & limitations
34
 
35
- More information needed
36
 
37
  ## Training and evaluation data
38
 
39
- More information needed
40
 
41
  ## Training procedure
42
 
 
 
43
  ### Training hyperparameters
44
 
45
  The following hyperparameters were used during training:
 
18
  # t5_recommendation_sports_equipment_english
19
 
20
  This model is a fine-tuned version of [t5-large](https://huggingface.co/t5-large) on a custom dataset, consisting of sports equipment customers have purchased, and items to recommended next.
21
+
22
+ This is based on the paper ["Recommendation as Language Processing (RLP): A Unified Pretrain, Personalized Prompt & Predict Paradigm (P5)"](https://arxiv.org/pdf/2203.13366.pdf), where the researchers use a language model as a recommendation system.
23
+ - LLMs can "understand" relationships between words/terms via embeddings produced by the transformer architecture. This allows for relationships to be taken into account.
24
+ - By feeding an LLM a history of items purchased as the input and the next item purchased as the output, the model can learn what to recommend based on the semantics of the product's name.
25
+ - Taking multiple examples of different users' purchase history into account, the LLM can also learn what genres of products go with what.
26
+ - This essentially replicates collaboritve filtering
27
+ - Benefits include:
28
+ - Getting past the cold-start problem with ease (when new items are introduced, the model will be able to understand what's similar to it from the name alone).
29
+ - Avoiding tedious, manual feature engineering (using natural language, the LLM will automatically learn).
30
+
31
+ The github repository for fine-tuning this model can be viewed [here](https://github.com/Mohammadhia/t5_p5_recommendation_system).
32
+
33
+ The fine-tuned T5 model achieves the following results on the evaluation set:
34
  - Loss: 0.4554
35
  - Rouge1: 57.1429
36
  - Rouge2: 47.6190
 
40
 
41
  ## Model description
42
 
43
+ T5 is an open-source sequence-to-sequence model released by Google in 2020, from which several variants have been developed. This fine-tuned version is an attempt to replicate what was presented in the [P5 paper](https://arxiv.org/pdf/2203.13366.pdf), with a custom dataset (based on sports equipment).
44
+
45
+ More about this model (T5) can be viewed [here](https://huggingface.co/docs/transformers/model_doc/t5).
46
+
47
+ The P5 models from the paper can be viewed on the [Hugging Face Hub](https://huggingface.co/makitanikaze/P5) as well as in this [repository](https://github.com/jeykigung/P5).
48
 
49
  ## Intended uses & limitations
50
 
51
+ Can be used as you please, but is limited to the sports equipment dataset it was fine-tuned on. Your mileage may vary.
52
 
53
  ## Training and evaluation data
54
 
55
+ Please see this [repository](https://github.com/Mohammadhia/t5_p5_recommendation_system) for training and evaluation data.
56
 
57
  ## Training procedure
58
 
59
+ Please see this [repository](https://github.com/Mohammadhia/t5_p5_recommendation_system) for training and evaluation data.
60
+
61
  ### Training hyperparameters
62
 
63
  The following hyperparameters were used during training: