--- license: cc-by-nc-4.0 --- JAX weights converted from Torch checkpoint at `facebook/galactica-6.7b`. ```python (env) ubuntu@vm:~$ JAX_PLATFORM_NAME=cpu python3 >>> import jax >>> print(jax.devices()) [CpuDevice(id=0)] # Ensure that model weights are loaded into CPU RAM, not accelerator memory. >>> from transformers import FlaxOPTForCausalLM >>> model = FlaxOPTForCausalLM.from_pretrained("facebook/galactica-6.7b", from_pt=True) >>> model.push_to_hub(hf_model_repo) ``` ## Citation and Attribution Citation from the original repo is reproduced below as per the cc-by-nc-4.0 licsense. ```bibtex @inproceedings{GALACTICA, title={GALACTICA: A Large Language Model for Science}, author={Ross Taylor and Marcin Kardas and Guillem Cucurull and Thomas Scialom and Anthony Hartshorn and Elvis Saravia and Andrew Poulton and Viktor Kerkez and Robert Stojnic}, year={2022} } ``` > Research supported with Cloud TPUs from Google's TPU Research Cloud (TRC)