Merged-AGI-7B-EXL2 / README.md
hgloow's picture
Update README.md
169a850
|
raw
history blame
2.85 kB
---
license: cc-by-nc-4.0
datasets:
- meta-math/MetaMathQA
language:
- en
pipeline_tag: text-generation
tags:
- Math
- exl2
---
# Merged-AGI-7B
- Model creator: [Talha Rüzgar Akkuş](https://huggingface.co/Q-bert)
- Original model: [Merged AGI 7B](https://huggingface.co/Q-bert/Merged-AGI-7B)
## EXL2 Quants
- [3.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/3.0bpw)
- [4.0bpw (main)](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/main)
- [6.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/6.0bpw)
- [8.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/8.0bpw)
Zipped Quantization (if you want to download a single file)
- [3.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/3.0bpw-zip)
- [4.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/4.0bpw-zip)
- [6.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/6.0bpw-zip)
- [8.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/8.0bpw-zip)
## Calibration Dataset
[wikitext-103-v1](https://huggingface.co/datasets/wikitext/blob/9a9e482b5987f9d25b3a9b2883fc6cc9fd8071b3/wikitext-103-v1/wikitext-test.parquet)
## Memory Usage
Measured using ExLlamaV2 and 4096 max_seq_len with [Oobabooga's Text Generation WebUI](https://github.com/oobabooga/text-generation-webui/tree/main).
| Branch | BPW | VRAM Usage | Description |
| ------ | --- | ---------- | ----------- |
[3.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/3.0bpw)|3.0|3.7 GB|For >=6GB VRAM cards
[4.0bpw (main)](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/main)|4.0|4.4 GB|For >=6GB VRAM cards
[6.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/6.0bpw)|6.0|6.1 GB|For >=8GB VRAM cards
[8.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/8.0bpw)|8.0|7.7 GB|For >=10GB VRAM cards
## Prompt template: ChatML
```
<|im_start|>system
{system_message}<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant
```
# Original Info
Merge [Q-bert/MetaMath-Cybertron-Starling](https://huggingface.co/Q-bert/MetaMath-Cybertron-Starling) and [fblgit/juanako-7b-UNA](https://huggingface.co/fblgit/juanako-7b-UNA) using slerp merge.
You can use ChatML format.
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
Detailed results can be found [Coming soon]()
| Metric | Value |
|-----------------------|---------------------------|
| Avg. | Coming soon |
| ARC (25-shot) | Coming soon |
| HellaSwag (10-shot) | Coming soon |
| MMLU (5-shot) | Coming soon |
| TruthfulQA (0-shot) | Coming soon |
| Winogrande (5-shot) | Coming soon |
| GSM8K (5-shot) | Coming soon |