akariasai commited on
Commit
5b0e658
1 Parent(s): 7371e09

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -2
README.md CHANGED
@@ -6,7 +6,7 @@ This model is a 7B [Self-RAG](https://selfrag.github.io/) model that generates o
6
 
7
  Self-RAG is trained on our instruction-following corpora with interleaving passages and reflection tokens using the standard next-token prediction objective, enabling efficient and stable learning with fine-grained feedback.
8
  At inference, we leverage reflection tokens covering diverse aspects of generations to sample the best output aligning users' preferences.
9
- See full descriptions in [our paper](https://akariasai.github.io/files/adaptive_retrieval_augmented_lm_arxiv.pdf).
10
 
11
  ## Usage
12
  Here, we show an easy way to quickly download our model from HuggingFace and run with `vllm` with pre-given passages. Make sure to install dependencies listed at [self-rag/requirements.txt](https://github.com/AkariAsai/self-rag/requirements.txt).
@@ -49,12 +49,19 @@ or
49
  If you have additional input.
50
  You can insert paragraphs anywhere after `### Response:\n"`, but make sure to mark paragraphs as paragraph tokens (i.e., `<paragraph>{0}</paragraph>`).
51
 
 
 
 
 
 
52
  ## Citation and contact
53
  If you use this model, please cite our work:
54
  ```
55
  @article{asai2023selfrag,
56
  author = {Asai, Akari and Wu, Zeqiu and Wang, Yizhong and Sil, Avirup and Hajishirzi, Hannaneh},
57
- title = {{Self-RAG}: Learning to Retrieve, Generate, and Critique through Self-Reflection},
58
  year = {2023},
 
 
59
  }
60
  ```
 
6
 
7
  Self-RAG is trained on our instruction-following corpora with interleaving passages and reflection tokens using the standard next-token prediction objective, enabling efficient and stable learning with fine-grained feedback.
8
  At inference, we leverage reflection tokens covering diverse aspects of generations to sample the best output aligning users' preferences.
9
+ See full descriptions in [our paper](hhttps://arxiv.org/abs/2310.11511).
10
 
11
  ## Usage
12
  Here, we show an easy way to quickly download our model from HuggingFace and run with `vllm` with pre-given passages. Make sure to install dependencies listed at [self-rag/requirements.txt](https://github.com/AkariAsai/self-rag/requirements.txt).
 
49
  If you have additional input.
50
  You can insert paragraphs anywhere after `### Response:\n"`, but make sure to mark paragraphs as paragraph tokens (i.e., `<paragraph>{0}</paragraph>`).
51
 
52
+ ## Training details
53
+ Our training data is available at the HuggingFace dataset [selfrag_train_data](https://huggingface.co/datasets/selfrag/selfrag_train_data).
54
+ See our official repository for the training details.
55
+ We used 8 A100 40GB for training on the Stability HPC server.
56
+
57
  ## Citation and contact
58
  If you use this model, please cite our work:
59
  ```
60
  @article{asai2023selfrag,
61
  author = {Asai, Akari and Wu, Zeqiu and Wang, Yizhong and Sil, Avirup and Hajishirzi, Hannaneh},
62
+ title = {{Self-RAG}: Learning to Retrieve, Generate, and Critique through Self-Reflection},
63
  year = {2023},
64
+ journal = { arXiv preprint arXiv:2310.11511 },
65
+ URL = {https://arxiv.org/abs/2310.11511}
66
  }
67
  ```