|
--- |
|
license: openrail |
|
--- |
|
Experimental Tagalog loras: safe or accurate outputs not guaranteed (not for production use)! |
|
|
|
Note: better/best results with |
|
* Prompting in Tagalog |
|
* Using format "Human: (prompt)\nAssistant:" |
|
|
|
Example: |
|
"Ito ay isang chat log sa pagitan ng AI Assistant na nagta-Tagalog at isang Pilipino. Magsimula ng chat:\nHuman: Hello po?\nAssistant:" |
|
|
|
# lt2_08162023 |
|
* Fine tuned on a small dataset of 14 items, manually edited |
|
* 1 epoch (barely any noticable results) |
|
* From chat LLaMA-2-7b |
|
* Lora of chat-tagalog v0.1 |
|
|
|
# lt2_08162023a |
|
* Fine tuned on a small dataset of 14 items, manually edited |
|
* 20 epochs (more observable effects) |
|
* From chat LLaMA-2-7b |
|
* Lora of [chat-tagalog v0.1a](https://huggingface.co/922-Narra/llama-2-7b-chat-tagalog-v0.1a) |
|
|
|
# lt2_08162023b |
|
* Fine tuned on a small dataset of 14 items, manually edited |
|
* 10 epochs |
|
* From chat LLaMA-2-7b |
|
* Lora of chat-tagalog v0.1b |
|
|
|
# lt2_08162023c |
|
* Fine tuned on a small dataset of 14 items, manually edited |
|
* 50 epochs (overfitted) |
|
* From chat LLaMA-2-7b |
|
* Lora of chat-tagalog v0.1c |
|
|
|
# lt2_08162023d |
|
* Fine tuned on a small dataset of 14 items, manually edited |
|
* 30 epochs (v0.1a further trained and cut-off before overfit) |
|
* From chat LLaMA-2-7b |
|
* Lora of [chat-tagalog v0.1d](https://huggingface.co/922-Narra/llama-2-7b-chat-tagalog-v0.1d) |
|
|
|
# llama-2-7b-tagalog-v0.2 loras (08/26/2023) |
|
* Fine tuned on dataset of ~10k items (mixed) |
|
* 2/2a/2b fine-tuned for 1/2/3 epochs |
|
* From chat LLaMA-2-7b |
|
* Future attempt planned with cleaner chat/dialogue data |
|
|
|
# hopia-3b-v0.1 (08/26/2023) |
|
* Fine tuned on a small dataset of 14 items, manually edited |
|
* 20 epochs |
|
* From Open LLaMA 3b |
|
|
|
# llama-2-7b-tagalog-v0.3 loras (09/01/2023) |
|
* Fine tuned on a dataset of ~1k items (Tagalog-focused dataset, based off Tagalog sentences augmented by LLaMA-2-13b base to create a 3-turn dialogue dataset between Human and Assistant) |
|
* 3/3a fine-tuned for 1/2 epochs |
|
* From chat LLaMA-2-7b |
|
* Experiment on partially synthetic data (and observing capability of LLaMA-2 base on generating Tagalog): will be further curating dataset |
|
* Loras for [chat-tagalog v0.3)](https://huggingface.co/922-Narra/llama-2-7b-chat-tagalog-v0.3) and [chat-tagalog v0.3](https://huggingface.co/922-Narra/llama-2-7b-chat-tagalog-v0.3a) |
|
|
|
# llama-2-7b-tagalog-v0.3WC2 (09/01/2023) |
|
* Fine tuned on experimental dataset of ~6k items (Tagalog-focused dataset, based off Tagalog sentences and Wiki entries augmented by LLaMA-2-13b to create a dialogue-QnA dataset between Human and Assistant) |
|
* 1 epoch |
|
* From chat LLaMA-2-7b |
|
|
|
# llama-2-13b-tagalog-v0.3 loras (09/01-02/2023) |
|
* Fine tuned on experimental datasets of ~1k items (Tagalog-focused dataset, based off Tagalog sentences augmented by LLaMA-2-13b base to create a 3-turn dialogue dataset between Human and Assistant) |
|
* 3 fine-tuned for 1 epoch, rank = 16, lora alpha = 32 |
|
* 3a with rank = 8 |
|
* 3b for 2 epochs |
|
* 3c for 1 epoch, lr = 1e-4, warmup steps = 0.1 |
|
* 3d with lr = 2e-4, rank = 32, lora alpha = 64 |
|
* 3e for 2 epochs |
|
* From LLaMA-2-13b |
|
* Trying LLaMA-2-13b chat/other base and curated dataset for next attempts |