SummerSigh commited on
Commit
380cef1
1 Parent(s): 1f690c1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -0
README.md CHANGED
@@ -8,6 +8,10 @@ This is EleutherAI/pythia-410m finetuned on OpenAssistant/oasst_top1_2023-08-25
8
  # Why
9
  Plain and simple. Im experimenting with making instruction LLMs under 1B params. I think we can still squeeze out better performance out of these models.
10
 
 
 
 
 
11
  # Usage
12
  ```
13
  from transformers import pipeline
 
8
  # Why
9
  Plain and simple. Im experimenting with making instruction LLMs under 1B params. I think we can still squeeze out better performance out of these models.
10
 
11
+ # Random Notes
12
+ - Only using OpenAssistant data gives fantastic results becuase of its high quality. I like the top1 dataset becuase of it's lack of prompt refusals.
13
+ - Prompt refusals have been shown to damage the performance of instruction LLMs. My theory is that the model "spends" parameters learning how to refuse prompts rather than learning actually useful information. Adding to this, I think that unlike other tasks, learning prompt refusals most likely has no other value in terms of transfer learning.
14
+
15
  # Usage
16
  ```
17
  from transformers import pipeline