|
--- |
|
license: cc-by-4.0 |
|
datasets: |
|
- camel-ai/biology |
|
--- |
|
|
|
limit: None, provide_description: False, num_fewshot: 5, batch_size: None |
|
| Task |Version| Metric |Value | |Stderr| |
|
|-----------------------------------|------:|--------|-----:|---|-----:| |
|
|hendrycksTest-college_chemistry | 1|acc |0.4600|± |0.0501| |
|
| | |acc_norm|**0.4600**|± |0.0501| |
|
|hendrycksTest-high_school_chemistry| 1|acc |0.5222|± |0.0351| |
|
| | |acc_norm|**0.5222**|± |0.0351| |
|
|hendrycksTest-college_biology | 1|acc |0.7222|± |0.0375| |
|
| | |acc_norm|**0.7222**|± |0.0375| |
|
|hendrycksTest-high_school_biology | 1|acc |0.7355|± |0.0251| |
|
| | |acc_norm|**0.7355**|± |0.0251| |
|
|winogrande | 0|acc |**0.7758**|± |0.0117| |
|
|
|
This model was trained from base Mistral-7B-Instruct-v0.2 on 710 examples, 200 of which comes from camel-ai/biology set. The rest were scraped personally and consists of very long scientific articles and text books. |
|
|
|
It beats Mistral-7B-Instruct-v0.2 in MMLU chemistry and biology. It should be able to generate mostly factual, basic and lengthy scientific text. I guess it could be "we have cosmopedia at home" for people who want to create cheap pretraining datasets from scratch. |
|
|
|
Template: |
|
|
|
[Context] |
|
You are a helpful assistant. Read the instruction and write a response accordingly. |
|
|
|
[User] |
|
{prompt} |
|
|
|
[Assistant] |
|
|
|
|
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6324eabf05bd8a54c6eb1650/ywxKzcQra_1g8EWtMeZ8Q.png) |
|
|
|
|