andrijdavid commited on
Commit
ab293e3
1 Parent(s): eaa9b85

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +9 -2
README.md CHANGED
@@ -208,13 +208,20 @@ Here are guides on using llama-cpp-python and ctransformers with LangChain:
208
  <a href="https://www.gradient.ai" target="_blank"><img src="https://cdn-uploads.huggingface.co/production/uploads/655bb613e8a8971e89944f3e/TSa3V8YpoVagnTYgxiLaO.png" width="200"/></a>
209
 
210
  # Llama-3 8B Gradient Instruct 1048k
 
 
 
211
  Gradient incorporates your data to deploy autonomous assistants that power critical operations across your business. If you're looking to build custom AI models or agents, email us a message contact@gradient.ai.
212
 
213
  For more info see our [End-to-end development service for custom LLMs and AI systems](https://gradient.ai/development-lab)
214
 
 
 
215
  This model extends LLama-3 8B's context length from 8k to > 1040K, developed by Gradient, sponsored by compute from [Crusoe Energy](https://huggingface.co/crusoeai). It demonstrates that SOTA LLMs can learn to operate on long context with minimal training by appropriately adjusting RoPE theta. We trained on 830M tokens for this stage, and 1.4B tokens total for all stages, which is < 0.01% of Llama-3's original pre-training data.
216
 
217
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6585dc9be92bc5f258156bd6/6MKLoX2ruLIaREiyb6coO.png)
 
 
218
 
219
  **Approach:**
220
 
@@ -230,7 +237,7 @@ Notably, we layered parallelism on top of Ring Attention with a custom network t
230
 
231
  **Data:**
232
 
233
- For training data, we generate long contexts by augmenting [SlimPajama](https://huggingface.co/datasets/cerebras/SlimPajama-627B).
234
 
235
  **Progressive Training Details:**
236
 
 
208
  <a href="https://www.gradient.ai" target="_blank"><img src="https://cdn-uploads.huggingface.co/production/uploads/655bb613e8a8971e89944f3e/TSa3V8YpoVagnTYgxiLaO.png" width="200"/></a>
209
 
210
  # Llama-3 8B Gradient Instruct 1048k
211
+
212
+ Join our custom agent and long context (262k-1M+) waitlist: https://forms.gle/L6TDY7dozx8TuoUv7
213
+
214
  Gradient incorporates your data to deploy autonomous assistants that power critical operations across your business. If you're looking to build custom AI models or agents, email us a message contact@gradient.ai.
215
 
216
  For more info see our [End-to-end development service for custom LLMs and AI systems](https://gradient.ai/development-lab)
217
 
218
+ [Join our Discord](https://discord.com/invite/2QVy2qt2mf)
219
+
220
  This model extends LLama-3 8B's context length from 8k to > 1040K, developed by Gradient, sponsored by compute from [Crusoe Energy](https://huggingface.co/crusoeai). It demonstrates that SOTA LLMs can learn to operate on long context with minimal training by appropriately adjusting RoPE theta. We trained on 830M tokens for this stage, and 1.4B tokens total for all stages, which is < 0.01% of Llama-3's original pre-training data.
221
 
222
+ **Update (5/3): We further fine-tuned our model to strengthen its assistant-like chat ability as well. The NIAH result is updated.**
223
+
224
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6585dc9be92bc5f258156bd6/-qaI__83ksClzoJzlqZjq.png)
225
 
226
  **Approach:**
227
 
 
237
 
238
  **Data:**
239
 
240
+ For training data, we generate long contexts by augmenting [SlimPajama](https://huggingface.co/datasets/cerebras/SlimPajama-627B). We also fine-tune on a chat dataset based on UltraChat [4], following a similar recipe for data augmentation to [2].
241
 
242
  **Progressive Training Details:**
243