andrijdavid
commited on
Commit
•
ab293e3
1
Parent(s):
eaa9b85
Upload folder using huggingface_hub
Browse files
README.md
CHANGED
@@ -208,13 +208,20 @@ Here are guides on using llama-cpp-python and ctransformers with LangChain:
|
|
208 |
<a href="https://www.gradient.ai" target="_blank"><img src="https://cdn-uploads.huggingface.co/production/uploads/655bb613e8a8971e89944f3e/TSa3V8YpoVagnTYgxiLaO.png" width="200"/></a>
|
209 |
|
210 |
# Llama-3 8B Gradient Instruct 1048k
|
|
|
|
|
|
|
211 |
Gradient incorporates your data to deploy autonomous assistants that power critical operations across your business. If you're looking to build custom AI models or agents, email us a message contact@gradient.ai.
|
212 |
|
213 |
For more info see our [End-to-end development service for custom LLMs and AI systems](https://gradient.ai/development-lab)
|
214 |
|
|
|
|
|
215 |
This model extends LLama-3 8B's context length from 8k to > 1040K, developed by Gradient, sponsored by compute from [Crusoe Energy](https://huggingface.co/crusoeai). It demonstrates that SOTA LLMs can learn to operate on long context with minimal training by appropriately adjusting RoPE theta. We trained on 830M tokens for this stage, and 1.4B tokens total for all stages, which is < 0.01% of Llama-3's original pre-training data.
|
216 |
|
217 |
-
|
|
|
|
|
218 |
|
219 |
**Approach:**
|
220 |
|
@@ -230,7 +237,7 @@ Notably, we layered parallelism on top of Ring Attention with a custom network t
|
|
230 |
|
231 |
**Data:**
|
232 |
|
233 |
-
For training data, we generate long contexts by augmenting [SlimPajama](https://huggingface.co/datasets/cerebras/SlimPajama-627B).
|
234 |
|
235 |
**Progressive Training Details:**
|
236 |
|
|
|
208 |
<a href="https://www.gradient.ai" target="_blank"><img src="https://cdn-uploads.huggingface.co/production/uploads/655bb613e8a8971e89944f3e/TSa3V8YpoVagnTYgxiLaO.png" width="200"/></a>
|
209 |
|
210 |
# Llama-3 8B Gradient Instruct 1048k
|
211 |
+
|
212 |
+
Join our custom agent and long context (262k-1M+) waitlist: https://forms.gle/L6TDY7dozx8TuoUv7
|
213 |
+
|
214 |
Gradient incorporates your data to deploy autonomous assistants that power critical operations across your business. If you're looking to build custom AI models or agents, email us a message contact@gradient.ai.
|
215 |
|
216 |
For more info see our [End-to-end development service for custom LLMs and AI systems](https://gradient.ai/development-lab)
|
217 |
|
218 |
+
[Join our Discord](https://discord.com/invite/2QVy2qt2mf)
|
219 |
+
|
220 |
This model extends LLama-3 8B's context length from 8k to > 1040K, developed by Gradient, sponsored by compute from [Crusoe Energy](https://huggingface.co/crusoeai). It demonstrates that SOTA LLMs can learn to operate on long context with minimal training by appropriately adjusting RoPE theta. We trained on 830M tokens for this stage, and 1.4B tokens total for all stages, which is < 0.01% of Llama-3's original pre-training data.
|
221 |
|
222 |
+
**Update (5/3): We further fine-tuned our model to strengthen its assistant-like chat ability as well. The NIAH result is updated.**
|
223 |
+
|
224 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6585dc9be92bc5f258156bd6/-qaI__83ksClzoJzlqZjq.png)
|
225 |
|
226 |
**Approach:**
|
227 |
|
|
|
237 |
|
238 |
**Data:**
|
239 |
|
240 |
+
For training data, we generate long contexts by augmenting [SlimPajama](https://huggingface.co/datasets/cerebras/SlimPajama-627B). We also fine-tune on a chat dataset based on UltraChat [4], following a similar recipe for data augmentation to [2].
|
241 |
|
242 |
**Progressive Training Details:**
|
243 |
|