|
--- |
|
license: openrail |
|
--- |
|
Experimental Tagalog loras: safe or accurate outputs not guaranteed (not for production use)! |
|
|
|
# lt2_08162023 |
|
* Fine tuned on a small dataset of 14 items, manually edited |
|
* 1 epoch (barely any noticable results) |
|
* From chat LLaMA-2-7b |
|
* Lora of chat-tagalog v0.1 |
|
|
|
# lt2_08162023a |
|
* Fine tuned on a small dataset of 14 items, manually edited |
|
* 20 epochs (more observable effects) |
|
* From chat LLaMA-2-7b |
|
* Lora of chat-tagalog v0.1a |
|
|
|
# lt2_08162023b |
|
* Fine tuned on a small dataset of 14 items, manually edited |
|
* 10 epochs |
|
* From chat LLaMA-2-7b |
|
* Lora of chat-tagalog v0.1b |
|
|
|
# lt2_08162023c |
|
* Fine tuned on a small dataset of 14 items, manually edited |
|
* 50 epochs (overfitted) |
|
* From chat LLaMA-2-7b |
|
* Lora of chat-tagalog v0.1c |
|
|
|
# lt2_08162023d |
|
* Fine tuned on a small dataset of 14 items, manually edited |
|
* 30 epochs (v0.1a further trained and cut-off before overfit) |
|
* From chat LLaMA-2-7b |
|
* Lora of chat-tagalog v0.1d |
|
|