|
--- |
|
license: cc-by-nc-4.0 |
|
datasets: |
|
- meta-math/MetaMathQA |
|
language: |
|
- en |
|
pipeline_tag: text-generation |
|
tags: |
|
- Math |
|
- exl2 |
|
--- |
|
# Merged-AGI-7B |
|
|
|
- Model creator: [Talha Rüzgar Akkuş](https://huggingface.co/Q-bert) |
|
- Original model: [Merged AGI 7B](https://huggingface.co/Q-bert/Merged-AGI-7B) |
|
|
|
## EXL2 Quants |
|
|
|
- [3.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/3.0bpw) |
|
- [4.0bpw (main)](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/main) |
|
- [6.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/6.0bpw) |
|
- [8.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/8.0bpw) |
|
|
|
Zipped Quantization (if you want to download a single file) |
|
- [3.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/3.0bpw-zip) |
|
- [4.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/4.0bpw-zip) |
|
- [6.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/6.0bpw-zip) |
|
- [8.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/8.0bpw-zip) |
|
|
|
## Calibration Dataset |
|
[wikitext-103-v1](https://huggingface.co/datasets/wikitext/blob/9a9e482b5987f9d25b3a9b2883fc6cc9fd8071b3/wikitext-103-v1/wikitext-test.parquet) |
|
|
|
## Memory Usage |
|
Measured using ExLlamaV2 and 4096 max_seq_len with [Oobabooga's Text Generation WebUI](https://github.com/oobabooga/text-generation-webui/tree/main). |
|
| Branch | BPW | VRAM Usage | Description | |
|
| ------ | --- | ---------- | ----------- | |
|
[3.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/3.0bpw)|3.0|3.7 GB|For >=6GB VRAM cards |
|
[4.0bpw (main)](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/main)|4.0|4.4 GB|For >=6GB VRAM cards |
|
[6.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/6.0bpw)|6.0|6.1 GB|For >=8GB VRAM cards |
|
[8.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/8.0bpw)|8.0|7.7 GB|For >=10GB VRAM cards |
|
|
|
## Prompt template: ChatML |
|
|
|
``` |
|
<|im_start|>system |
|
{system_message}<|im_end|> |
|
<|im_start|>user |
|
{prompt}<|im_end|> |
|
<|im_start|>assistant |
|
|
|
``` |
|
|
|
# Original Info |
|
|
|
Merge [Q-bert/MetaMath-Cybertron-Starling](https://huggingface.co/Q-bert/MetaMath-Cybertron-Starling) and [fblgit/juanako-7b-UNA](https://huggingface.co/fblgit/juanako-7b-UNA) using slerp merge. |
|
|
|
You can use ChatML format. |
|
|
|
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) |
|
Detailed results can be found [Coming soon]() |
|
|
|
| Metric | Value | |
|
|-----------------------|---------------------------| |
|
| Avg. | Coming soon | |
|
| ARC (25-shot) | Coming soon | |
|
| HellaSwag (10-shot) | Coming soon | |
|
| MMLU (5-shot) | Coming soon | |
|
| TruthfulQA (0-shot) | Coming soon | |
|
| Winogrande (5-shot) | Coming soon | |
|
| GSM8K (5-shot) | Coming soon | |