metadata
license: other
language:
- en
- zh
library_name: transformers
pipeline_tag: text-generation
inference: false
tags:
- baichuan
- llama2
- baichuan2
This is the LLaMAfied version of Baichuan2-7B-Base model by Baichuan Inc.
This model is converted with https://github.com/hiyouga/LLaMA-Factory/blob/main/tests/llamafy_baichuan2.py
You may use this model for fine-tuning in downstream tasks, we recommend using our efficient fine-tuning toolkit. https://github.com/hiyouga/LLaMA-Factory
- Developed by: Baichuan Inc.
- Language(s) (NLP): Chinese/English
- License: Baichuan2 License
Usage:
from transformers import AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("hiyouga/Baichuan2-7B-Base-LLaMAfied", use_fast=False)
model = AutoModelForCausalLM.from_pretrained("hiyouga/Baichuan2-7B-Base-LLaMAfied").cuda()
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 42.83 |
ARC (25-shot) | 49.57 |
HellaSwag (10-shot) | 73.45 |
MMLU (5-shot) | 54.86 |
TruthfulQA (0-shot) | 37.54 |
Winogrande (5-shot) | 70.72 |
GSM8K (5-shot) | 7.81 |
DROP (3-shot) | 5.85 |