--- license: openrail --- Experimental Tagalog loras: safe or accurate outputs not guaranteed (not for production use)! # lt2_08162023 * Fine tuned on a small dataset of 14 items, manually edited * 1 epoch (barely any noticable results) * From chat LLaMA-2-7b * Lora of chat-tagalog v0.1 # lt2_08162023a * Fine tuned on a small dataset of 14 items, manually edited * 20 epochs (more observable effects) * From chat LLaMA-2-7b * Lora of chat-tagalog v0.1a # lt2_08162023b * Fine tuned on a small dataset of 14 items, manually edited * 10 epochs * From chat LLaMA-2-7b * Lora of chat-tagalog v0.1b # lt2_08162023c * Fine tuned on a small dataset of 14 items, manually edited * 50 epochs (overfitted) * From chat LLaMA-2-7b * Lora of chat-tagalog v0.1c # lt2_08162023d * Fine tuned on a small dataset of 14 items, manually edited * 30 epochs (v0.1a further trained and cut-off before overfit) * From chat LLaMA-2-7b * Lora of chat-tagalog v0.1d