akariasai commited on
Commit
125a51f
1 Parent(s): a2d59cf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -2
README.md CHANGED
@@ -4,7 +4,7 @@ license: mit
4
  This model is a 7B [Self-RAG](https://selfrag.github.io/) model that generates outputs to diverse user queries as well as *reflection tokens* to call the retrieval system adaptively and criticize its own output and retrieved passages.
5
 
6
  Self-RAG is trained on our instruction-following corpora with interleaving passages and reflection tokens using the standard next-token prediction objective, enabling efficient and stable learning with fine-grained feedback.
7
- At inference, we leverage reflection tokens covering diverse aspect of generations to sample the best output aligning users' preferences.
8
  See full descriptions in [our paper](https://akariasai.github.io/files/adaptive_retrieval_augmented_lm_arxiv.pdf).
9
 
10
  ## Usage
@@ -53,12 +53,19 @@ or
53
  If you have additional input.
54
  You can insert paragraphs anywhere after `### Response:\n"`, but make sure to mark paragraphs as paragraph tokens (i.e., `<paragraph>{0}</paragraph>`).
55
 
 
 
 
 
 
56
  ## Citation and contact
57
  If you use this model, please cite our work:
58
  ```
59
  @article{asai2023selfrag,
60
  author = {Asai, Akari and Wu, Zeqiu and Wang, Yizhong and Sil, Avirup and Hajishirzi, Hannaneh},
61
- title = {{Self-RAG}: Learning to Retrieve, Generate, and Critique through Self-Reflection},
62
  year = {2023},
 
 
63
  }
64
  ```
 
4
  This model is a 7B [Self-RAG](https://selfrag.github.io/) model that generates outputs to diverse user queries as well as *reflection tokens* to call the retrieval system adaptively and criticize its own output and retrieved passages.
5
 
6
  Self-RAG is trained on our instruction-following corpora with interleaving passages and reflection tokens using the standard next-token prediction objective, enabling efficient and stable learning with fine-grained feedback.
7
+ At inference, we leverage reflection tokens covering diverse aspects of generations to sample the best output aligning users' preferences.
8
  See full descriptions in [our paper](https://akariasai.github.io/files/adaptive_retrieval_augmented_lm_arxiv.pdf).
9
 
10
  ## Usage
 
53
  If you have additional input.
54
  You can insert paragraphs anywhere after `### Response:\n"`, but make sure to mark paragraphs as paragraph tokens (i.e., `<paragraph>{0}</paragraph>`).
55
 
56
+ ## Training details
57
+ Our training data is available at the HuggingFace dataset [selfrag_train_data](https://huggingface.co/datasets/selfrag/selfrag_train_data).
58
+ See our official repository for the training details.
59
+ We used 8 A100 40GB for training on the Stability HPC server.
60
+
61
  ## Citation and contact
62
  If you use this model, please cite our work:
63
  ```
64
  @article{asai2023selfrag,
65
  author = {Asai, Akari and Wu, Zeqiu and Wang, Yizhong and Sil, Avirup and Hajishirzi, Hannaneh},
66
+ title = {{Self-RAG}: Learning to Retrieve, Generate, and Critique through Self-Reflection},
67
  year = {2023},
68
+ journal = { arXiv preprint arXiv:2310.11511 },
69
+ URL = {https://arxiv.org/abs/2310.11511}
70
  }
71
  ```