Text Generation
Transformers
PyTorch
Safetensors
English
llama
conversational
Inference Endpoints
text-generation-inference
hamishivi commited on
Commit
563c4f6
1 Parent(s): 74f90ca

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -6
README.md CHANGED
@@ -20,6 +20,9 @@ Tulu is a series of language models that are trained to act as helpful assistant
20
  Tulu V2 DPO 70B is a fine-tuned version of Llama 2 that was trained on on a mix of publicly available, synthetic and human datasets using [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290).
21
  This model is a strong alternative to Llama 2 70b Chat.
22
 
 
 
 
23
 
24
  ## Model description
25
 
@@ -137,12 +140,13 @@ The following hyperparameters were used during DPO training:
137
  If you find Tulu 2 is useful in your work, please cite it with:
138
 
139
  ```
140
- @misc{ivison2023changing,
141
- title={Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2},
142
- author={Hamish Ivison and Yizhong Wang and Valentina Pyatkin and Nathan Lambert and Matthew Peters and Pradeep Dasigi and Joel Jang and David Wadden and Noah A. Smith and Iz Beltagy and Hannaneh Hajishirzi},
143
- year={2023},
144
- archivePrefix={arXiv},
145
- primaryClass={cs.CL}
 
146
  }
147
  ```
148
 
 
20
  Tulu V2 DPO 70B is a fine-tuned version of Llama 2 that was trained on on a mix of publicly available, synthetic and human datasets using [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290).
21
  This model is a strong alternative to Llama 2 70b Chat.
22
 
23
+ For more details, read the paper: [Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2
24
+ ](https://arxiv.org/abs/2311.10702).
25
+
26
 
27
  ## Model description
28
 
 
140
  If you find Tulu 2 is useful in your work, please cite it with:
141
 
142
  ```
143
+ @misc{ivison2023camels,
144
+ title={Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2},
145
+ author={Hamish Ivison and Yizhong Wang and Valentina Pyatkin and Nathan Lambert and Matthew Peters and Pradeep Dasigi and Joel Jang and David Wadden and Noah A. Smith and Iz Beltagy and Hannaneh Hajishirzi},
146
+ year={2023},
147
+ eprint={2311.10702},
148
+ archivePrefix={arXiv},
149
+ primaryClass={cs.CL}
150
  }
151
  ```
152