winglian commited on
Commit
d543ee8
1 Parent(s): cfa9881

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -8
README.md CHANGED
@@ -17,6 +17,7 @@ datasets:
17
  - camel-ai/biology
18
  - camel-ai/physics
19
  - camel-ai/chemistry
 
20
 
21
  inference: false
22
  ---
@@ -25,7 +26,7 @@ inference: false
25
 
26
  # Minotaur MPT 7B
27
 
28
- Minotaur 7B is an instruct fine-tuned model on top of MPT-7B.
29
 
30
  Questions, comments, feedback, looking to donate, or want to help? Reach out on our [Discord](https://discord.gg/PugNNHAF5r) or email [wing@openaccessaicollective.org](mailto:wing@openaccessaicollective.org)
31
 
@@ -36,25 +37,25 @@ Chat only style prompts using `USER:`,`ASSISTANT:`.
36
 
37
  Minotaur 7B model is fine-tuned on the following datasets:
38
 
39
- - [riddle_sense](https://huggingface.co/datasets/riddle_sense) - instruct augmented
40
- - hellaswag, updated for detailed explanations w 30K+ rows
41
- - [gsm8k](https://huggingface.co/datasets/gsm8k) - instruct augmented
42
  - [WizardLM](https://huggingface.co/datasets/ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered)
43
  - [subset of QingyiSi/Alpaca-CoT for roleplay and CoT](https://huggingface.co/QingyiSi/Alpaca-CoT)
44
  - [GPTeacher-General-Instruct](https://huggingface.co/datasets/teknium/GPTeacher-General-Instruct)
45
- - ARC-Easy & ARC-Challenge - instruct augmented for detailed responses, derived from the `train` split
46
- - [hellaswag](https://huggingface.co/datasets/hellaswag) - 5K row subset of instruct augmented for concise responses, derived from the `train` split
47
  - [metaeval/ScienceQA_text_only](https://huggingface.co/datasets/metaeval/ScienceQA_text_only) - instruct for concise responses
48
  - [openai/summarize_from_feedback](https://huggingface.co/datasets/openai/summarize_from_feedback) - instruct augmented tl;dr summarization
49
  - [camel-ai/math](https://huggingface.co/datasets/camel-ai/math)
50
  - [camel-ai/physics](https://huggingface.co/datasets/camel-ai/physics)
51
  - [camel-ai/chemistry](https://huggingface.co/datasets/camel-ai/chemistry)
52
  - [camel-ai/biology](https://huggingface.co/datasets/camel-ai/biology)
53
- - custom sysnthetic datasets around misconceptions, in-context qa, jokes, N-tasks problems, and context-insensitivity
 
 
 
 
 
54
 
55
  # Shoutouts
56
 
57
- Special thanks to Nanobit for helping with Axolotl, TheBloke for quantizing these models are more accessible to all, ehartford for cleaned datasets, and 0x000011b for the RP dataset.
58
 
59
  # Demo
60
 
 
17
  - camel-ai/biology
18
  - camel-ai/physics
19
  - camel-ai/chemistry
20
+ - winglian/evals
21
 
22
  inference: false
23
  ---
 
26
 
27
  # Minotaur MPT 7B
28
 
29
+ Minotaur 7B is an instruct fine-tuned model on top of MPT-7B. Minotaur 7B is fine-tuned on only completely open datasets making this model reproducible by anyone.
30
 
31
  Questions, comments, feedback, looking to donate, or want to help? Reach out on our [Discord](https://discord.gg/PugNNHAF5r) or email [wing@openaccessaicollective.org](mailto:wing@openaccessaicollective.org)
32
 
 
37
 
38
  Minotaur 7B model is fine-tuned on the following datasets:
39
 
 
 
 
40
  - [WizardLM](https://huggingface.co/datasets/ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered)
41
  - [subset of QingyiSi/Alpaca-CoT for roleplay and CoT](https://huggingface.co/QingyiSi/Alpaca-CoT)
42
  - [GPTeacher-General-Instruct](https://huggingface.co/datasets/teknium/GPTeacher-General-Instruct)
 
 
43
  - [metaeval/ScienceQA_text_only](https://huggingface.co/datasets/metaeval/ScienceQA_text_only) - instruct for concise responses
44
  - [openai/summarize_from_feedback](https://huggingface.co/datasets/openai/summarize_from_feedback) - instruct augmented tl;dr summarization
45
  - [camel-ai/math](https://huggingface.co/datasets/camel-ai/math)
46
  - [camel-ai/physics](https://huggingface.co/datasets/camel-ai/physics)
47
  - [camel-ai/chemistry](https://huggingface.co/datasets/camel-ai/chemistry)
48
  - [camel-ai/biology](https://huggingface.co/datasets/camel-ai/biology)
49
+ - [winglian/evals](https://huggingface.co/datasets/winglian/evals)
50
+ - custom sysnthetic datasets around misconceptions, in-context qa, jokes, N-tasks problems, and context-insensitivity
51
+ - ARC-Easy & ARC-Challenge - instruct augmented for detailed responses, derived from the `train` split
52
+ - [hellaswag](https://huggingface.co/datasets/hellaswag) - 30K+ rows of instruct augmented for detailed explanations w 30K+ rows, derived from the `train` split
53
+ - [riddle_sense](https://huggingface.co/datasets/riddle_sense) - instruct augmented
54
+ - [gsm8k](https://huggingface.co/datasets/gsm8k) - instruct augmented
55
 
56
  # Shoutouts
57
 
58
+ Special thanks to Nanobit for helping with Axolotl and TheBloke for quantizing these models are more accessible to all.
59
 
60
  # Demo
61