pszemraj commited on
Commit
48cccc0
1 Parent(s): 97a57f9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -3
README.md CHANGED
@@ -87,9 +87,15 @@ inference:
87
 
88
  # bart-base-instructiongen
89
 
90
- Instead of generating questions from text, generate instructions for LLMs! Demo on Spaces [is here](https://huggingface.co/spaces/pszemraj/generate-instructions).
 
 
 
 
 
 
 
91
 
92
- This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on the pszemraj/fleece2instructions dataset.
93
  It achieves the following results on the evaluation set:
94
  - Loss: 1.0034
95
  - Rouge1: 61.7209
@@ -100,7 +106,7 @@ It achieves the following results on the evaluation set:
100
 
101
  ## Intended uses & limitations
102
 
103
- This is just a base model/example. There is likely to be even better performance with larger models.
104
 
105
  Additionally, this was trained on a dataset of **only** instructions+outputs, with the `inputs` filtered out. This means that text of *1) cookies and cream 2) chocolate chip 3) mint chip 4) oreo* will **not** get you *"Rank the following ice cream flavors: oreo, mint chip, chocolate chip, cookies and cream"*.
106
 
 
87
 
88
  # bart-base-instructiongen
89
 
90
+ Instead of generating questions from text, generate instructions for LLMs!
91
+
92
+ Check out a [basic demo on Spaces](https://huggingface.co/spaces/pszemraj/generate-instructions). You can find other models fine-tuned for instruction generation by [searching for the instructiongen tag](https://huggingface.co/models?other=instructiongen).
93
+
94
+
95
+ ## About
96
+
97
+ This model is a fine-tuned version of the [facebook/bart-base](https://huggingface.co/facebook/bart-base) model, fine-tuned using the `pszemraj/fleece2instructions` dataset. The concept is to apply text-to-text models to unlabeled, domain-specific text to generate appropriate LLM instructions. Consequently, this facilitates domain adaptation of instruction-tuned LLMs, making them more versatile in their respective domains.
98
 
 
99
  It achieves the following results on the evaluation set:
100
  - Loss: 1.0034
101
  - Rouge1: 61.7209
 
106
 
107
  ## Intended uses & limitations
108
 
109
+ This is just a base model/example. There is likely to be even better performance with larger models (click [here to see other checkpoints](https://huggingface.co/models?other=instructiongen))
110
 
111
  Additionally, this was trained on a dataset of **only** instructions+outputs, with the `inputs` filtered out. This means that text of *1) cookies and cream 2) chocolate chip 3) mint chip 4) oreo* will **not** get you *"Rank the following ice cream flavors: oreo, mint chip, chocolate chip, cookies and cream"*.
112