Edit model card

Breeze-Petro-7B-Instruct-v1

Model Description

This is a model fine-tuned based on Breeze-7B-Instruct-v1_0. The training set is mainly based on chemical knowledge and procedural knowledge. Supplement knowledge about the petroleum industry.

  • Developed by: RebeccaChou
  • License: apache-2.0
  • Finetuned from model : [MediaTek-Research/Breeze-7B-Instruct-v1_0
  • Language(s) (NLP): [English.繁體中文]

πŸ“– Table of Contents

1.Open LLM Leaderboard - ARC - HellaSwag - MMLU - TruthfulQA - Winogrande - GSM8K 3. EvalPlus Leaderboard - HumanEval - HumanEval_Plus - MBPP - MBPP_Plus 4. Prompt Format 5. Quantized Models 6. Gratitude

πŸ† Open LLM Leaderboard

WestSeverus-7B-DPO-v2 is one of the top 7B model in Open LLM Leaderboard and it outperforms on TruthfulQA and GSM8K.

Metric Value
Avg. 59.32
AI2 Reasoning Challenge (25-Shot) 58.87
HellaSwag (10-Shot) 79.17
MMLU (5-Shot) 56.62
TruthfulQA (0-shot) 46.36
Winogrande (5-shot) 73.64
GSM8k (5-shot) 41.24

Detailed results can be found here

⚑ EvalPlus Leaderboard

Model HumanEval HumanEval_Plus MBPP MBPP_Plus
phi-2-2.7B 48.2 43.3 61.9 51.4
SOLAR-10.7B-Instruct-v1.0 42.1 34.3 42.9 34.6
CodeLlama-7B 37.8 34.1 57.6 45.4

πŸ› οΈ Quantized Models

Training Details

Training Data

  • Dataset: Rebecca19990101/petro-dataset-v2
Downloads last month
9
Safetensors
Model size
7.49B params
Tensor type
F32
Β·
Inference API
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.