BioMistral-v0.2 / README.md
Ba2han's picture
Update README.md
0eeda98 verified
---
license: cc-by-4.0
datasets:
- camel-ai/biology
---
limit: None, provide_description: False, num_fewshot: 5, batch_size: None
| Task |Version| Metric |Value | |Stderr|
|-----------------------------------|------:|--------|-----:|---|-----:|
|hendrycksTest-college_chemistry | 1|acc |0.4600|± |0.0501|
| | |acc_norm|**0.4600**|± |0.0501|
|hendrycksTest-high_school_chemistry| 1|acc |0.5222|± |0.0351|
| | |acc_norm|**0.5222**|± |0.0351|
|hendrycksTest-college_biology | 1|acc |0.7222|± |0.0375|
| | |acc_norm|**0.7222**|± |0.0375|
|hendrycksTest-high_school_biology | 1|acc |0.7355|± |0.0251|
| | |acc_norm|**0.7355**|± |0.0251|
|winogrande | 0|acc |**0.7758**|± |0.0117|
This model was trained from base Mistral-7B-Instruct-v0.2 on 710 examples, 200 of which comes from camel-ai/biology set. The rest were scraped personally and consists of very long scientific articles and text books.
It beats Mistral-7B-Instruct-v0.2 in MMLU chemistry and biology. It should be able to generate mostly factual, basic and lengthy scientific text. I guess it could be "we have cosmopedia at home" for people who want to create cheap pretraining datasets from scratch.
Template:
[Context]
You are a helpful assistant. Read the instruction and write a response accordingly.
[User]
{prompt}
[Assistant]
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6324eabf05bd8a54c6eb1650/ywxKzcQra_1g8EWtMeZ8Q.png)