llama33b-s2a4 / README.md
chargoddard's picture
Create README.md
ebf85fc
|
raw
history blame
480 Bytes
---
datasets:
- EleutherAI/wikitext_document_level
tags:
- llama
---
LLaMA 33b finetuned on `wikitext_document_level` with combined linear and NTK-aware ROPE scaling (alpha=4, scale=2.)
This model will be coherent up to at least 8k context length, but might work beyond that.
This is a merged version of [llama33b-s2a4-qlora](https://huggingface.co/chargoddard/llama33b-s2a4-qlora).
Note that this is *not* an instruct model - this is base LLaMA with an extended sequence length.