pszemraj commited on
Commit
7286290
1 Parent(s): 8326fe1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -6
README.md CHANGED
@@ -69,6 +69,8 @@ widget:
69
 
70
  # flan-t5-small-instructiongen
71
 
 
 
72
  This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on the None dataset.
73
  It achieves the following results on the evaluation set:
74
  - Loss: 1.3401
@@ -78,17 +80,17 @@ It achieves the following results on the evaluation set:
78
  - Rougelsum: 50.338
79
  - Gen Len: 14.0450
80
 
81
- ## Model description
82
-
83
- More information needed
84
-
85
  ## Intended uses & limitations
86
 
87
- More information needed
 
 
88
 
89
  ## Training and evaluation data
90
 
91
- More information needed
 
 
92
 
93
  ## Training procedure
94
 
 
69
 
70
  # flan-t5-small-instructiongen
71
 
72
+ Instead of generating questions from text, generate instructions for LLMs!
73
+
74
  This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on the None dataset.
75
  It achieves the following results on the evaluation set:
76
  - Loss: 1.3401
 
80
  - Rougelsum: 50.338
81
  - Gen Len: 14.0450
82
 
 
 
 
 
83
  ## Intended uses & limitations
84
 
85
+ This is just a **small** model/example. There is likely to be even better performance with larger models (ex [pszemraj/bart-base-instructiongen)](https://huggingface.co/pszemraj/bart-base-instructiongen) generalizes better)
86
+
87
+ Additionally, this was trained on a dataset of **only** instructions+outputs, with the `inputs` filtered out. This means that text of *1) cookies and cream 2) chocolate chip 3) mint chip 4) oreo* will **not** get you *"Rank the following ice cream flavors: oreo, mint chip, chocolate chip, cookies and cream"*.
88
 
89
  ## Training and evaluation data
90
 
91
+ See the linked dataset `pszemraj/fleece2instructions` - it is a filtered/formatted version of `tatsu-lab/alpaca` to generate instructions for arbitrary text.
92
+
93
+ - Some of the API examples are intentionally weird to demonstrate the generalizability of the model.
94
 
95
  ## Training procedure
96