|
--- |
|
license: cc-by-nc-4.0 |
|
--- |
|
This model is a generation model trained via [semiparametric token-sequence co-supervision](https://github.com/kaistAI/Semiparametric_Token-Sequence_Co-Supervision) on top of Llama2-7B. |
|
The embedding model which constructs the nonparametric sequence embedding spaces is in [here](https://huggingface.co/kaist-ai/cosupervision-emb_seq-Llama2_7b). |
|
The models are trained on information-seeking datasets provided by [self-rag](https://selfrag.github.io/) with co-supervision from next token prediction (NTP) and next sequence prediction (NSP). |
|
In the inference step, the model generates a response by retrieving relevant sequences. |
|
See full descriptions in our paper. |
|
|
|
### Usage |
|
Here, we show an easy way to quickly download our model from HuggingFace. |
|
Make sure to install dependencies listed at requirements.txt. |
|
To run our full inference pipeline with embedding model, please use our [code](https://github.com/kaistAI/Semiparametric_Token-Sequence_Co-Supervision). |
|
|
|
``` |
|
from transformers import AutoTokenizer, LlamaForCausalLM |
|
|
|
model = LlamaForCausalLM.from_pretrained( |
|
"kaist-ai/cosupervision-emb_seq-Llama2_7b", |
|
load_in_8bit=True if train_config.quantization else None, |
|
device_map="auto" if train_config.quantization else None, |
|
) |
|
``` |