PY007 commited on
Commit
bed7637
1 Parent(s): b451586

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -11,3 +11,5 @@
11
 
12
  Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
13
 
 
 
 
11
 
12
  Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
13
 
14
+ This model is finetuned from Llama-2-7B-hf with EasyContext on context length 512K and generalized to 1M tokens. Note that I keep max_position_embeddings in config.json to 4096 because HF llama will create 2D causal mask during initialization. If it is set to 1M GPU will just OOM. You can surely use this model with context length longer than 4096.
15
+