Instruction Tuned Models
Collection
Monolingual or multilingual instruction tuning: Which makes a better Alpaca? https://arxiv.org/abs/2309.08958
•
279 items
•
Updated
This HF repository contains base LLMs instruction tuned (SFT) with full-parameter fine-tuning and then used to study whether monolingual or multilingual instruction tuning is more favourable.
The model checkpoint should be loaded using transformers
library.
Please refer to our Github repository HERE for inference and training instructions.
@inproceedings{chen-etal-2024-monolingual,
title="Monolingual or multilingual instruction tuning: Which makes a better {Alpaca}",
author="Pinzhen Chen and Shaoxiong Ji and Nikolay Bogoychev and Andrey Kutuzov and Barry Haddow and Kenneth Heafield",
year="2024",
booktitle = "Findings of the Association for Computational Linguistics: EACL 2024",
}