kawine commited on
Commit
64f3c4d
·
1 Parent(s): 05a9d3a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -27,7 +27,7 @@ How is SHP different from [Anthropic's HH-RLHF dataset](https://huggingface.co/d
27
 
28
  | Dataset | Input | Output | No. Domains | Data Format |
29
  | -------------------- | -------------------------- | ---------------------------- | ------------------------- | ------------------------------------- |
30
- | SHP | Reddit post and comments | Aggregate Preference Label | 18 (cooking, cars, ...) | Question/Answer + Assertion/Response |
31
  | Anthropic/HH-RLHF | Dialogue history with LLM | Individual Preference Label | 2 (harmful, helpful) | Multi-turn Dialogue |
32
 
33
 
@@ -169,7 +169,9 @@ If you want to finetune a model to predict human preferences (e.g., for NLG eval
169
  ### Evaluating
170
 
171
  Since it is easier to predict stronger preferences than weaker ones (e.g., preferences with a big difference in comment score), we recommend reporting a performance curve instead of a single number.
172
- For example, here is the accuracy curve for a FLAN-T5-xl model trained using the suggestions above, on only preferences with a 2+ score ratio and using no more than 5 preferences from each post to prevent overfitting:
 
 
173
 
174
 
175
 
 
27
 
28
  | Dataset | Input | Output | No. Domains | Data Format |
29
  | -------------------- | -------------------------- | ---------------------------- | ------------------------- | ------------------------------------- |
30
+ | SHP | Reddit post and comments | Aggregate Preference Label with Scores | 18 (cooking, cars, ...) | Question/Answer + Assertion/Response |
31
  | Anthropic/HH-RLHF | Dialogue history with LLM | Individual Preference Label | 2 (harmful, helpful) | Multi-turn Dialogue |
32
 
33
 
 
169
  ### Evaluating
170
 
171
  Since it is easier to predict stronger preferences than weaker ones (e.g., preferences with a big difference in comment score), we recommend reporting a performance curve instead of a single number.
172
+ For example, here is the accuracy curve for a FLAN-T5-xl model trained on the askculinary ddata using the suggestions above.
173
+
174
+ The orange line is without filtering the training data and the blue line is with training only on preferences with a 2+ score ratio and using no more than 5 preferences from each post to prevent overfitting:
175
 
176
 
177