Update README.md
Browse files
README.md
CHANGED
@@ -17,11 +17,10 @@ The model architecture and config are the same as [M2M-100](https://huggingface.
|
|
17 |
|
18 |
SMaLL-100 is a seq-to-seq model for the translation task. The input to the model is ```source:[tgt_lang_code] + src_tokens + [EOS]``` and ```target: tgt_tokens + [EOS]```.
|
19 |
|
|
|
|
|
20 |
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
-
# small-100-th inference
|
25 |
```
|
26 |
from transformers import M2M100ForConditionalGeneration
|
27 |
from tokenization_small100 import SMALL100Tokenizer
|
|
|
17 |
|
18 |
SMaLL-100 is a seq-to-seq model for the translation task. The input to the model is ```source:[tgt_lang_code] + src_tokens + [EOS]``` and ```target: tgt_tokens + [EOS]```.
|
19 |
|
20 |
+
# `small-100-th` is the fine-tuned version of SMALL-100 for Thai
|
21 |
+
The dataset can be acquired from [Vistec](https://github.com/vistec-AI/thai2nmt/releases/tag/scb-mt-en-th-2020%2Bmt-opus_v1.0)
|
22 |
|
23 |
+
## small-100-th inference
|
|
|
|
|
|
|
24 |
```
|
25 |
from transformers import M2M100ForConditionalGeneration
|
26 |
from tokenization_small100 import SMALL100Tokenizer
|