Update README.md
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ license_link: LICENSE
|
|
16 |
|
17 |
## Model Introduction
|
18 |
|
19 |
-
The A13B models released by Tencent Hunyuan this time: [Tencent-Hunyuan-A13B-Pretrain](https://huggingface.co/tencent/Hunyuan-A13B-Pretrain) , [Tencent-Hunyuan-A13B-Instruct](https://huggingface.co/tencent/Hunyuan-A13B-Instruct)
|
20 |
|
21 |
### Introduction to Technical Advantages
|
22 |
|
|
|
16 |
|
17 |
## Model Introduction
|
18 |
|
19 |
+
The A13B models released by Tencent Hunyuan this time: [Tencent-Hunyuan-A13B-Pretrain](https://huggingface.co/tencent/Hunyuan-A13B-Pretrain) , [Tencent-Hunyuan-A13B-Instruct](https://huggingface.co/tencent/Hunyuan-A13B-Instruct) and [Tencent-Hunyuan-A13B-Instruct-FP8](https://huggingface.co/tencent/Tencent-Hunyuan-A13B-Instruct-FP8), use better data allocation and training, have strong performance, and have achieved a good balance between computing and performance. It stands out from many large-scale language models and is currently one of the strongest Chinese Mixture of Experts (MoE) models, featuring a total of 80 billion parameters and 13 billion active parameters.
|
20 |
|
21 |
### Introduction to Technical Advantages
|
22 |
|