metadata
license: openrail
Experimental Tagalog loras: safe or accurate outputs not guaranteed (not for production use)!
Note: better/best results with
- Prompting in Tagalog
- Using format "Human: (prompt)\nAssistant:"
Example: "Ito ay isang chat log sa pagitan ng AI Assistant na nagta-Tagalog at isang Pilipino. Magsimula ng chat:\nHuman: Hello po?\nAssistant:"
lt2_08162023
- Fine tuned on a small dataset of 14 items, manually edited
- 1 epoch (barely any noticable results)
- From chat LLaMA-2-7b
- Lora of chat-tagalog v0.1
lt2_08162023a
- Fine tuned on a small dataset of 14 items, manually edited
- 20 epochs (more observable effects)
- From chat LLaMA-2-7b
- Lora of chat-tagalog v0.1a
lt2_08162023b
- Fine tuned on a small dataset of 14 items, manually edited
- 10 epochs
- From chat LLaMA-2-7b
- Lora of chat-tagalog v0.1b
lt2_08162023c
- Fine tuned on a small dataset of 14 items, manually edited
- 50 epochs (overfitted)
- From chat LLaMA-2-7b
- Lora of chat-tagalog v0.1c
lt2_08162023d
- Fine tuned on a small dataset of 14 items, manually edited
- 30 epochs (v0.1a further trained and cut-off before overfit)
- From chat LLaMA-2-7b
- Lora of chat-tagalog v0.1d
llama-2-7b-tagalog-v0.2 loras (08/26/2023)
- Fine tuned on dataset of ~10k items (mixed)
- 2/2a/2b fine-tuned for 1/2/3 epochs
- From chat LLaMA-2-7b
- Future attempt planned with cleaner chat/dialogue data
hopia-3b-v0.1 (08/26/2023)
- Fine tuned on a small dataset of 14 items, manually edited
- 20 epochs
- From Open LLaMA 3b
llama-2-7b-tagalog-v0.3 loras (09/01/2023)
- Fine tuned on a dataset of ~1k items (Tagalog-focused dataset, based off Tagalog sentences augmented by LLaMA-2-13b base to create a 3-turn dialogue dataset between Human and Assistant)
- 3/3a fine-tuned for 1/2 epochs
- From chat LLaMA-2-7b
- Experiment on partially synthetic data (and observing capability of LLaMA-2 base on generating Tagalog): will be further curating dataset
- Loras for chat-tagalog v0.3) and chat-tagalog v0.3
llama-2-7b-tagalog-v0.3WC2 (09/01/2023)
- Fine tuned on experimental dataset of ~6k items (Tagalog-focused dataset, based off Tagalog sentences and Wiki entries augmented by LLaMA-2-13b to create a dialogue-QnA dataset between Human and Assistant)
- 1 epoch
- From chat LLaMA-2-7b
llama-2-13b-tagalog-v0.3 loras (09/01-02/2023)
- Fine tuned on experimental datasets of ~1k items (Tagalog-focused dataset, based off Tagalog sentences augmented by LLaMA-2-13b base to create a 3-turn dialogue dataset between Human and Assistant)
- 3 fine-tuned for 1 epoch, rank = 16, lora alpha = 32
- 3a with rank = 8
- 3b for 2 epochs
- 3c for 1 epoch, lr = 1e-4, warmup steps = 0.1
- 3d with lr = 2e-4, rank = 32, lora alpha = 64
- 3e for 2 epochs
- From LLaMA-2-13b
- Trying LLaMA-2-13b chat/other base and curated dataset for next attempts