File size: 2,481 Bytes
b1a1ff3
 
 
 
 
 
 
 
 
 
a5094ce
bb1c658
 
 
b1d6755
 
5458f45
 
41960f0
b1a1ff3
 
 
 
ef4f5ec
b1a1ff3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5d35a79
 
 
 
 
 
 
 
b1a1ff3
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
---
license: apache-2.0
pipeline_tag: text-generation
---

**Typhoon2-Qwen2.5-7B**: Thai Large Language Model (Instruct)

**Typhoon2-Qwen2.5-7B** is a pretrained only Thai ๐Ÿ‡น๐Ÿ‡ญ large language model with 7 billion parameters, and it is based on Qwen2.5-7B.


For technical-report. please see our [arxiv](https://arxiv.org/abs/2412.13702).

## **Performance**

| Model                 | ThaiExam | ONET   | IC        | A-Level   | TGAT      | TPAT      | M3Exam    | Math       | Science    | Social     | Thai       |
|------------------------|----------|--------|-----------|-----------|-----------|-----------|-----------|------------|------------|------------|------------|
| **Typhoon2 Qwen2.5 7B Base** | **58.86%** | **58.64%** | 65.26%    | **55.11%** | 66.15%    | 49.13%    | **59.90%** | 42.98%     | **59.42%** | **75.62%** | **61.59%** |
| **Qwen2.5 7B**         | 55.74%   | 51.23% | **60.00%** | 41.73%    | **72.30%** | **53.44%** | 55.65%    | **46.15%** | 54.10%     | 66.54%     | 55.82%     |
| **Typhoon1.5 Llama3 8B Base**   | 48.82%   | 41.35% | 41.05%    | 40.94%    | 70.76%    | 50.00%    | 43.88%    | 22.62%     | 43.47%     | 62.81%     | 46.63%     |


## **Model Description**

- **Model type**: A 7B decoder-only model based on Qwen2 architecture.
- **Requirement**: transformers 4.45.0 or newer.
- **Primary Language(s)**: Thai ๐Ÿ‡น๐Ÿ‡ญ and English ๐Ÿ‡ฌ๐Ÿ‡ง
- **License**: Apache-2.0


## **Intended Uses & Limitations**

This model is a pretrained base model. Thus, it may not be able to follow human instructions without using one/few-shot learning or instruction fine-tuning. The model does not have any moderation mechanisms, and may generate harmful or inappropriate responses.

## **Follow us**

**https://twitter.com/opentyphoon**

## **Support**

**https://discord.gg/CqyBscMFpg**

## **Citation**

- If you find Typhoon2 useful for your work, please cite it using:
```
@misc{typhoon2,
      title={Typhoon 2: A Family of Open Text and Multimodal Thai Large Language Models}, 
      author={Kunat Pipatanakul and Potsawee Manakul and Natapong Nitarach and Warit Sirichotedumrong and Surapon Nonesung and Teetouch Jaknamon and Parinthapat Pengpun and Pittawat Taveekitworachai and Adisai Na-Thalang and Sittipong Sripaisarnmongkol and Krisanapong Jirayoot and Kasima Tharnpipitchai},
      year={2024},
      eprint={2412.13702},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2412.13702}, 
}
```