Edit model card

This model pairs with the PRETRAINED variant of amharic llama. Available here: https://huggingface.co/iocuydi/llama-2-amharic-3784m/tree/main/pretrained

It will also require Llama2 weights and this clip model: https://huggingface.co/openai/clip-vit-large-patch14-336

More information on running the model here: https://github.com/iocuydi/amharic-llama-llava

Cite:

@misc{andersland2024amharic,
      title={Amharic LLaMA and LLaVA: Multimodal LLMs for Low Resource Languages}, 
      author={Michael Andersland},
      year={2024},
      eprint={2403.06354},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}
Downloads last month
0
Unable to determine this model's library. Check the docs .