Edit model card

Breeze-Petro-7B-Instruct-v1

Model Description

This is a model fine-tuned based on Breeze-7B-Instruct-v1_0. The training set is mainly based on chemical knowledge and procedural knowledge. Supplement knowledge about the petroleum industry.

  • Developed by: RebeccaChou
  • License: apache-2.0
  • Finetuned from model : [MediaTek-Research/Breeze-7B-Instruct-v1_0
  • Language(s) (NLP): [English.繁體中文]

πŸ“– Table of Contents

1.Open LLM Leaderboard - ARC - HellaSwag - MMLU - TruthfulQA - Winogrande - GSM8K 3. EvalPlus Leaderboard - HumanEval - HumanEval_Plus - MBPP - MBPP_Plus 4. Prompt Format 5. Quantized Models 6. Gratitude

πŸ† Open LLM Leaderboard

WestSeverus-7B-DPO-v2 is one of the top 7B model in Open LLM Leaderboard and it outperforms on TruthfulQA and GSM8K.

Metric Value
Avg. 59.32
AI2 Reasoning Challenge (25-Shot) 58.87
HellaSwag (10-Shot) 79.17
MMLU (5-Shot) 56.62
TruthfulQA (0-shot) 46.36
Winogrande (5-shot) 73.64
GSM8k (5-shot) 41.24

Detailed results can be found here

⚑ EvalPlus Leaderboard

Model HumanEval HumanEval_Plus MBPP MBPP_Plus
phi-2-2.7B 48.2 43.3 61.9 51.4
SOLAR-10.7B-Instruct-v1.0 42.1 34.3 42.9 34.6
CodeLlama-7B 37.8 34.1 57.6 45.4

πŸ› οΈ Quantized Models

Training Details

Training Data

  • Dataset: Rebecca19990101/petro-dataset-v2
Downloads last month
2,739
Safetensors
Model size
7.49B params
Tensor type
F32
Β·