huseinzol05 commited on
Commit
eb1b736
·
verified ·
1 Parent(s): 471001d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -2
README.md CHANGED
@@ -13,7 +13,7 @@ Finetuned https://huggingface.co/mesolitica/nanot5-small-malaysian-cased using 2
13
  - This model natively code switching.
14
  - This model maintain `\n`, `\t`, `\r` as it is.
15
 
16
- **Still in training session**, Wandb at https://wandb.ai/huseinzol05/nanot5-small-malaysian-cased-translation-v4?nw=nwuserhuseinzol05
17
 
18
  ## Supported prefix
19
 
@@ -78,4 +78,8 @@ Output,
78
  ' Bayangkan PH dan menang PRU-14. Terdapat pelbagai pintu belakang. Akhirnya, Ismail Sabri naik. Itulah sebabnya saya tidak lagi bercakap tentang politik. Saya bersumpah sudah berputus asa.']
79
  ```
80
 
81
- Input text can be any languages that speak in Malaysia, as long you use proper prefix, it should be able to translate to target language.
 
 
 
 
 
13
  - This model natively code switching.
14
  - This model maintain `\n`, `\t`, `\r` as it is.
15
 
16
+ **Still in training session**, Wandb at https://wandb.ai/huseinzol05/nanot5-small-malaysian-cased-translation-v4-multipack/workspace
17
 
18
  ## Supported prefix
19
 
 
78
  ' Bayangkan PH dan menang PRU-14. Terdapat pelbagai pintu belakang. Akhirnya, Ismail Sabri naik. Itulah sebabnya saya tidak lagi bercakap tentang politik. Saya bersumpah sudah berputus asa.']
79
  ```
80
 
81
+ Input text can be any languages that speak in Malaysia, as long you use proper prefix, it should be able to translate to target language.
82
+
83
+ ## how to finetune your own dataset?
84
+
85
+ We finetuned using T5 SDPA multipacking forked at https://github.com/mesolitica/t5-sdpa-multipack, super undocumented, but scripts from https://github.com/huggingface/transformers/tree/main/examples/pytorch/translation should work also.