Update README.md
Browse files
README.md
CHANGED
|
@@ -16,7 +16,7 @@ license_link: LICENSE
|
|
| 16 |
|
| 17 |
## Model Introduction
|
| 18 |
|
| 19 |
-
The A13B models released by Tencent Hunyuan this time: [Tencent-Hunyuan-A13B-Pretrain](https://huggingface.co/tencent/Hunyuan-A13B-Pretrain) , [Tencent-Hunyuan-A13B-Instruct](https://huggingface.co/tencent/Hunyuan-A13B-Instruct)
|
| 20 |
|
| 21 |
### Introduction to Technical Advantages
|
| 22 |
|
|
|
|
| 16 |
|
| 17 |
## Model Introduction
|
| 18 |
|
| 19 |
+
The A13B models released by Tencent Hunyuan this time: [Tencent-Hunyuan-A13B-Pretrain](https://huggingface.co/tencent/Hunyuan-A13B-Pretrain) , [Tencent-Hunyuan-A13B-Instruct](https://huggingface.co/tencent/Hunyuan-A13B-Instruct) and [Tencent-Hunyuan-A13B-Instruct-FP8](https://huggingface.co/tencent/Tencent-Hunyuan-A13B-Instruct-FP8), use better data allocation and training, have strong performance, and have achieved a good balance between computing and performance. It stands out from many large-scale language models and is currently one of the strongest Chinese Mixture of Experts (MoE) models, featuring a total of 80 billion parameters and 13 billion active parameters.
|
| 20 |
|
| 21 |
### Introduction to Technical Advantages
|
| 22 |
|