🦙 Qwen-72B-Llama
This is the 🦙 llamafied version of Qwen/Qwen-72B.
🛠️ Reproduction
I used this script to convert the weights:
LLaMA-Factory/tests/llamafy_qwen.py
🔠 Tokenizer
After I converted the weights, I took the tokenizer from KnutJaegersberg/Qwen-14B-Llamafied and uploaded it to this repository.
📊 Eval Scores Compared to Original Model
Here are some of the evaluation score comparisons based on the Open LLM Leaderboard.
Metric | Qwen-72B | Qwen-72B-Llama |
---|---|---|
Avg. | 73.6 | 69.53 |
ARC (25-shot) | 65.19 | 64.85 |
HellaSwag (10-shot) | 85.94 | 83.27 |
MMLU (5-shot) | 77.37 | 73.66 |
TruthfulQA (0-shot) | 60.19 | 57.6 |
Winogrande (5-shot) | 82.48 | 81.53 |
GSM8K (5-shot) | 70.43 | 56.25 |
- Downloads last month
- 82
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.