--- license: apache-2.0 language: - en tags: - llama2 - 100k - 7b --- Anima LLM supporting 100K input token length. It's trained based on Llama2 7B, so the license support commercial use! We carefully curated long QA training dataset from 30k to 100k length to train this model. We also made a lot of memory optimizations to make it scale to 100k tokens. ## How to train/infer? #### install dependencies ```bash # Please update the path of `CUDA_HOME` export CUDA_HOME=/usr/local/cuda-11.8 pip install transformers==4.31.0 pip install sentencepiece pip install ninja pip install flash-attn --no-build-isolation pip install git+https://github.com/HazyResearch/flash-attention.git#subdirectory=csrc/rotary pip install git+https://github.com/HazyResearch/flash-attention.git#subdirectory=csrc/xentropy pip install evaluate pip install git+https://github.com/huggingface/peft.git@v0.4.0 pip install wandb ``` #### inference ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch base_model = "lyogavin/Anima-7B-100K" tokenizer = AutoTokenizer.from_pretrained(base_model) model = AutoModelForCausalLM.from_pretrained( base_model, torch_dtype=torch.float16, trust_remote_code=True, device_map="auto", ) model.eval() prompt = "Where is the capital of US?" inputs = tokenizer(prompt, return_tensors="pt") inputs['input_ids'] = inputs['input_ids'].cuda() inputs['attention_mask'] = inputs['attention_mask'].cuda() # Generate generate_ids = model.generate(**inputs, max_new_tokens=30, only_last_logit=True, # to save memory use_cache=False, # when run into OOM, enable this can save memory xentropy=True) output = tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0] ``` #### Training ```bash ./run_longer_training.sh ``` ## Evaluations There's almost none evaluation dataset designed for 100k tokens. So we designed/curated some dataset for this model. We compared this model and several other public/private models. #### 1. longchat topic retrieval | Model | Accuracy | |-------------------|---------| | Claude2 | 0.9 | | together llama2 32k | 0.15 | | longchat 32k 1.5 | 0.05 | | Anima 100K | 0.5 | #### 2. longchat number retrieval | Model | Accuracy | |-------------------|---------| | Claude2 | 0.85 | | together llama2 32k | 0.2 | | longchat 32k 1.5 | 0.05 | | Anima 100K | 0.45 | #### 3. Narrative QA in zeroscore | Model | F1 | |-------------------|---------| | Claude2 | 0.6187 | | together llama2 32k | 0.3833 | | longchat 32k 1.5 | 0.2416 | | Anima 100K | 0.4919 | ## Github Github repo is [here](https://github.com/lyogavin/Anima/tree/main/anima_100k)