--- language: - en - zh pipeline_tag: text-generation --- ## hongyin/awareness-en-zh-bilingual-2.5b This is a English-Chinese bilingual autoregressive language model based on OPT with a parameter size of 2.5b. The model is trained based on the NTP task on large-scale unstructured text. It is worth noting that the model is suitable for further training as an initialization parameter, and does not have the ability to be used as a chatbot directly unless it is fine-tuned on a multi-round dialogue corpus. This model is a basic model, used as a raw material for alchemy. ## Bibtex entry and citation info Please cite if you find it helpful. ``` @article{zhu2023metaaid, title={MetaAID 2.0: An Extensible Framework for Developing Metaverse Applications via Human-controllable Pre-trained Models}, author={Zhu, Hongyin}, journal={arXiv preprint arXiv:2302.13173}, year={2023} } ``` --- license: other ---