LDJnr commited on
Commit
02a0ae3
1 Parent(s): 54b9dce

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -16,6 +16,8 @@ datasets:
16
 
17
  **This is trained on the Yi-34B model with 200K context length, for 3 epochs on the Capybara dataset!**
18
 
 
 
19
  The Capybara series is the first Nous collection of models made by fine-tuning mostly on data created by Nous in-house.
20
 
21
  We leverage our novel data synthesis technique called Amplify-instruct (Paper coming soon), the seed distribution and synthesis method are comprised of a synergistic combination of top performing existing data synthesis techniques and distributions used for SOTA models such as Airoboros, Evol-Instruct(WizardLM), Orca, Vicuna, Know_Logic, Lamini, FLASK and others, all into one lean holistically formed methodology for the dataset and model. The seed instructions used for the start of synthesized conversations are largely based on highly datasets like Airoboros, Know logic, EverythingLM, GPTeacher and even entirely new seed instructions derived from posts on the website LessWrong, as well as being supplemented with certain in-house multi-turn datasets like Dove(A successor to Puffin).
 
16
 
17
  **This is trained on the Yi-34B model with 200K context length, for 3 epochs on the Capybara dataset!**
18
 
19
+ **First 34B Nous model and first 200K context length Nous model!**
20
+
21
  The Capybara series is the first Nous collection of models made by fine-tuning mostly on data created by Nous in-house.
22
 
23
  We leverage our novel data synthesis technique called Amplify-instruct (Paper coming soon), the seed distribution and synthesis method are comprised of a synergistic combination of top performing existing data synthesis techniques and distributions used for SOTA models such as Airoboros, Evol-Instruct(WizardLM), Orca, Vicuna, Know_Logic, Lamini, FLASK and others, all into one lean holistically formed methodology for the dataset and model. The seed instructions used for the start of synthesized conversations are largely based on highly datasets like Airoboros, Know logic, EverythingLM, GPTeacher and even entirely new seed instructions derived from posts on the website LessWrong, as well as being supplemented with certain in-house multi-turn datasets like Dove(A successor to Puffin).