Update README.md
Browse files
README.md
CHANGED
@@ -2,10 +2,9 @@
|
|
2 |
license: llama2
|
3 |
---
|
4 |
# Taga-llama-v0.3:
|
5 |
-
* Test Tagalog model
|
6 |
-
* Fine tuned on dataset of ~1k items (
|
7 |
-
*
|
8 |
-
* May still switch to Taglish or English: please see usage
|
9 |
* [QLoras (hf and GGML)](https://huggingface.co/922-Narra/tagalog-lm-lora-tests/tree/main/llama-2-7b-chat-tagalog-0.3)
|
10 |
|
11 |
### USAGE
|
@@ -26,8 +25,10 @@ Use "Human" and "Assistant" and prompt with Tagalog. Example:
|
|
26 |
* grad steps: 4
|
27 |
|
28 |
### WARNINGS AND DISCLAIMERS
|
29 |
-
Note that aside from formatting and other minor edits, dataset used is mostly as is
|
30 |
|
31 |
-
There is a chance that the model may switch back to English (albeit still understand Tagalog inputs) as conversations grow longer, resulting in English-Tagalog conversations: this may be because of the limited 3-turn nature of the dataset.
|
|
|
|
|
32 |
|
33 |
Finally, this model is not guaranteed to output aligned or safe outputs nor is it meant for production use - use at your own risk!
|
|
|
2 |
license: llama2
|
3 |
---
|
4 |
# Taga-llama-v0.3:
|
5 |
+
* Test Tagalog model
|
6 |
+
* Fine tuned on an experimental Tagalog-focused dataset of ~1k items (based off Tagalog sentences augmented by LLaMA-2-13b base to create a mostly 3-turn dialogue dataset between Human and Assistant)
|
7 |
+
* Base: LLaMA-2 7b chat
|
|
|
8 |
* [QLoras (hf and GGML)](https://huggingface.co/922-Narra/tagalog-lm-lora-tests/tree/main/llama-2-7b-chat-tagalog-0.3)
|
9 |
|
10 |
### USAGE
|
|
|
25 |
* grad steps: 4
|
26 |
|
27 |
### WARNINGS AND DISCLAIMERS
|
28 |
+
Note that aside from formatting and other minor edits, dataset used is mostly as is augmented by LM. As such, while this version may be better at coherency or chatting than previous Tagalog ones, conversations may still switch between languages or easily derail.
|
29 |
|
30 |
+
There is a chance that the model may switch back to English (albeit still understand Tagalog inputs) as conversations grow longer, resulting in English-Tagalog conversations: this may be because of the limited 3-turn nature of the dataset. Additionally, Taglish occuring in the dataset or any use of English may sometimes make the model more likely to output Taglish or even English responses.
|
31 |
+
|
32 |
+
Note that we use a partially synthetic dataset due to the lack of readily available Tagalog dialogue datasets, but take this as an opportunity to observe the Tagalog capability of LLaMA-2. However, we plan to further curate the dataset (and fine tune later model versions on this) and release a final cleaned version.
|
33 |
|
34 |
Finally, this model is not guaranteed to output aligned or safe outputs nor is it meant for production use - use at your own risk!
|