--- license: openrail --- Experimental Tagalog loras: safe or accurate outputs not guaranteed (not for production use)! # lt2_08162023 * Fine tuned on a small dataset of 14 items, manually edited * 1 epoch (barely any noticable results) * From chat LLaMA-2-7b * Lora of chat-tagalog v0.1 # lt2_08162023a * Fine tuned on a small dataset of 14 items, manually edited * 20 epochs (more observable effects) * From chat LLaMA-2-7b * Lora of [chat-tagalog v0.1a](https://huggingface.co/922-Narra/llama-2-7b-chat-tagalog-v0.1a) # lt2_08162023b * Fine tuned on a small dataset of 14 items, manually edited * 10 epochs * From chat LLaMA-2-7b * Lora of chat-tagalog v0.1b # lt2_08162023c * Fine tuned on a small dataset of 14 items, manually edited * 50 epochs (overfitted) * From chat LLaMA-2-7b * Lora of chat-tagalog v0.1c # lt2_08162023d * Fine tuned on a small dataset of 14 items, manually edited * 30 epochs (v0.1a further trained and cut-off before overfit) * From chat LLaMA-2-7b * Lora of [chat-tagalog v0.1d](https://huggingface.co/922-Narra/llama-2-7b-chat-tagalog-v0.1d) # llama-2-7b-tagalog-v0.2 loras (08/26/2023) * Fine tuned on dataset of ~10k items (mixed) * 2/2a/2b fine-tuned for 1/2/3 epochs * From chat LLaMA-2-7b * Future attempt planned with cleaner chat/dialogue data # hopia-3b-v0.1 (08/26/2023) * Fine tuned on a small dataset of 14 items, manually edited * 20 epochs * From Open LLaMA 3b