Edit model card

🦙 Qwen-72B-Llama

This is the 🦙 llamafied version of Qwen/Qwen-72B.

🛠️ Reproduction

I used this script to convert the weights:

LLaMA-Factory/tests/llamafy_qwen.py

🔠 Tokenizer

After I converted the weights, I took the tokenizer from KnutJaegersberg/Qwen-14B-Llamafied and uploaded it to this repository.

📊 Eval Scores Compared to Original Model

Here are some of the evaluation score comparisons based on the Open LLM Leaderboard.

Metric Qwen-72B Qwen-72B-Llama
Avg. 73.6 69.53
ARC (25-shot) 65.19 64.85
HellaSwag (10-shot) 85.94 83.27
MMLU (5-shot) 77.37 73.66
TruthfulQA (0-shot) 60.19 57.6
Winogrande (5-shot) 82.48 81.53
GSM8K (5-shot) 70.43 56.25

image/png

Downloads last month
734
Safetensors
Model size
72.3B params
Tensor type
BF16
·
Inference API
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.