Llama-3.1_OpenScholar-8B-Jaya

This is a fine-tuned version of the OpenScholar/Llama-3.1_OpenScholar-8B model, tailored by Jaya using the LLaMA-Factory framework, incorporating self-awareness data from the identity dataset.

Model Sources

License

This model is licensed under the MIT License.

Citation

If you use this model in your work, please cite it as:

@article{openscholar,
  title={{OpenScholar}: Synthesizing Scientific Literature with Retrieval-Augmented Language Models},
  author={ Asai, Akari and He*, Jacqueline and Shao*, Rulin and Shi, Weijia and Singh, Amanpreet and Chang, Joseph Chee  and Lo,  Kyle and Soldaini, Luca and Feldman, Tian, Sergey and Mike, D’arcy and Wadden, David and Latzke, Matt and Minyang and Ji, Pan and Liu, Shengyan and Tong, Hao and Wu, Bohao and Xiong, Yanyu and Zettlemoyer, Luke and Weld, Dan and Neubig, Graham and Downey, Doug and Yih, Wen-tau and Koh, Pang Wei and Hajishirzi, Hannaneh},
  journal={Arxiv},
  year={2024},
}
Downloads last month
7
Safetensors
Model size
8.03B params
Tensor type
BF16
·
Inference API
Unable to determine this model's library. Check the docs .

Model tree for real-jiakai/OpenScholar-Jaya

Finetuned
(1)
this model