Safetensors
xiuyul commited on
Commit
434b166
1 Parent(s): e2e506e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +40 -3
README.md CHANGED
@@ -1,3 +1,40 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - allenai/qasper
5
+ metrics:
6
+ - f1
7
+ ---
8
+
9
+ # LLoCO: Learning Long Contexts Offline
10
+ [**Paper**](https://arxiv.org/abs/2404.07979) | [**Code**](https://github.com/jeffreysijuntan/lloco)
11
+
12
+ Lloco-7b-qasper is the LoRA adaptor checkpoint finetuned from [AutoCompressor-Llama-2-7b-6k](https://huggingface.co/princeton-nlp/AutoCompressor-Llama-2-7b-6k/) and [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf)
13
+ using the **LLoCO** method in [LLoCO: Learning Long Contexts Offline](https://arxiv.org/abs/2404.07979). It is instruction-tuned on the Qasper training set.
14
+
15
+ **LLoCO** enables LLMs to process long-context efficiently by learning contexts offline through context compression and in-domain parameter-efficient finetuning with LoRA. This approach extends the effective context window of a 4k token LLaMA2-7B model to handle up to 128k tokens, while using
16
+ 30x fewer tokens and achieving up to 7.62x inference speed-up.
17
+
18
+ ## Released LoRA Checkpoint
19
+ | Model | LoRA Rank | Dataset | Link |
20
+ |:----------------|-----------|-------------|--------------------------------------------------------|
21
+ | Lloco-7b-quality| 8 | QuALITY | [link](https://huggingface.co/xiuyul/Lloco-7b-quality/)|
22
+ | Lloco-7b-qasper | 8 | Qasper | [link](https://huggingface.co/xiuyul/Lloco-7b-qasper/) |
23
+ | Lloco-7b-qmsum | 8 | QMSum | [link](https://huggingface.co/xiuyul/Lloco-7b-qmsum/) |
24
+ | Lloco-7b-nqa | 8 | NarrativeQA | [link](https://huggingface.co/xiuyul/Lloco-7b-nqa/) |
25
+ | Lloco-7b-hqa | 8 | HotpotQA | [link](https://huggingface.co/xiuyul/Lloco-7b-hqa/) |
26
+
27
+ ## Citation
28
+ If you find this project useful, please consider citing:
29
+
30
+ ```
31
+ @article{tan2024lloco,
32
+ title={LLoCO: Learning Long Contexts Offline},
33
+ author={Tan, Sijun and Li, Xiuyu and Patil, Shishir and Wu, Ziyang and Zhang, Tianjun and Keutzer, Kurt and Gonzalez, Joseph E and Popa, Raluca Ada},
34
+ journal={arXiv preprint arXiv:2404.07979},
35
+ year={2024}
36
+ }
37
+ ```
38
+
39
+ ## Evaluation
40
+ Check out [LLoCO: Learning Long Contexts Offline](https://arxiv.org/abs/2404.07979) for evaluation results on various long-context tasks such as long document question answering and summarization.