renillhuang
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -8,10 +8,11 @@ widget:
|
|
8 |
text: "Hello! How can I help you today?"
|
9 |
pipeline_tag: text-generation
|
10 |
---
|
11 |
-
|
12 |
<!-- markdownlint-disable first-line-h1 -->
|
13 |
<!-- markdownlint-disable html -->
|
14 |
-
|
|
|
|
|
15 |
|
16 |
<div align="center">
|
17 |
<h1>
|
@@ -26,7 +27,7 @@ pipeline_tag: text-generation
|
|
26 |
<p>
|
27 |
<b>πEnglish</b> |
|
28 |
<a href="https://huggingface.co/OrionStarAI/Orion-14B-Base-Int4/blob/main/README_zh.md">π¨π³δΈζ</a><br><br>
|
29 |
-
π€ <a href="https://huggingface.co/OrionStarAI" target="_blank">HuggingFace Mainpage</a> | π€ <a href="https://modelscope.cn/organization/OrionStarAI" target="_blank">ModelScope Mainpage</a><br>π¬ <a href="https://huggingface.co/spaces/OrionStarAI/Orion-14B-App-Demo" target="_blank">HuggingFace Demo</a> | π« <a href="https://modelscope.cn/studios/OrionStarAI/Orion-14B-App-Demo/summary" target="_blank">ModelScope Demo</a>
|
30 |
<p>
|
31 |
</h4>
|
32 |
|
@@ -40,20 +41,34 @@ pipeline_tag: text-generation
|
|
40 |
- [π Model Download](#model-download)
|
41 |
- [π Model Benchmark](#model-benchmark)
|
42 |
- [π Model Inference](#model-inference)
|
43 |
-
- [π₯ Company Introduction](#company-introduction)
|
44 |
- [π Declarations & License](#declarations-license)
|
|
|
45 |
|
46 |
# 1. Model Introduction
|
47 |
|
48 |
-
- Orion-14B-
|
49 |
-
|
50 |
-
|
51 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
52 |
|
53 |
- Orion-14B series models including:
|
54 |
- **Orion-14B-Base:** A multilingual large language foundational model with 14 billion parameters, pretrained on a diverse dataset of 2.5 trillion tokens.
|
55 |
- **Orion-14B-Chat:** A chat-model fine-tuned on a high-quality corpus aims to provide an excellence interactive experience for users in the large model community.
|
56 |
-
- **Orion-14B-LongChat:**
|
57 |
- **Orion-14B-Chat-RAG:** A chat-model fine-tuned on a custom retrieval augmented generation dataset, achieving superior performance in retrieval augmented generation tasks.
|
58 |
- **Orion-14B-Chat-Plugin:** A chat-model specifically tailored for plugin and function calling tasks, ideal for agent-related scenarios where the LLM acts as a plugin and function call system.
|
59 |
- **Orion-14B-Base-Int4:** A quantized base model utilizing 4-bit integer weights. It significantly reduces the model size by 70% and increases the inference speed by 30% while incurring a minimal performance loss of only 1%.
|
@@ -321,7 +336,7 @@ Truly Useful Robots", OrionStar empowers more people through AI technology.
|
|
321 |
|
322 |
**The core strengths of OrionStar lies in possessing end-to-end AI application capabilities,** including big data preprocessing, large model pretraining, fine-tuning, prompt engineering, agent, etc. With comprehensive end-to-end model training capabilities, including systematic data processing workflows and the parallel model training capability of hundreds of GPUs, it has been successfully applied in various industry scenarios such as government affairs, cloud services, international e-commerce, and fast-moving consumer goods.
|
323 |
|
324 |
-
Companies with demands for deploying large-scale model applications are welcome to contact us
|
325 |
**Enquiry Hotline: 400-898-7779**<br>
|
326 |
**E-mail: [email protected]**
|
327 |
|
|
|
8 |
text: "Hello! How can I help you today?"
|
9 |
pipeline_tag: text-generation
|
10 |
---
|
|
|
11 |
<!-- markdownlint-disable first-line-h1 -->
|
12 |
<!-- markdownlint-disable html -->
|
13 |
+
<div align="center">
|
14 |
+
<img src="./assets/imgs/orion_start.PNG" alt="logo" width="50%" />
|
15 |
+
</div>
|
16 |
|
17 |
<div align="center">
|
18 |
<h1>
|
|
|
27 |
<p>
|
28 |
<b>πEnglish</b> |
|
29 |
<a href="https://huggingface.co/OrionStarAI/Orion-14B-Base-Int4/blob/main/README_zh.md">π¨π³δΈζ</a><br><br>
|
30 |
+
π€ <a href="https://huggingface.co/OrionStarAI" target="_blank">HuggingFace Mainpage</a> | π€ <a href="https://modelscope.cn/organization/OrionStarAI" target="_blank">ModelScope Mainpage</a><br>π¬ <a href="https://huggingface.co/spaces/OrionStarAI/Orion-14B-App-Demo" target="_blank">HuggingFace Demo</a> | π« <a href="https://modelscope.cn/studios/OrionStarAI/Orion-14B-App-Demo/summary" target="_blank">ModelScope Demo</a><br>πΊ <a href="https://github.com/OrionStarAI/Orion" target="_blank">GitHub</a><br>π <a href="https://github.com/OrionStarAI/Orion/blob/master/doc/Orion14B_v3.pdf" target="_blank">Tech Report</a>
|
31 |
<p>
|
32 |
</h4>
|
33 |
|
|
|
41 |
- [π Model Download](#model-download)
|
42 |
- [π Model Benchmark](#model-benchmark)
|
43 |
- [π Model Inference](#model-inference)
|
|
|
44 |
- [π Declarations & License](#declarations-license)
|
45 |
+
- [π₯ Company Introduction](#company-introduction)
|
46 |
|
47 |
# 1. Model Introduction
|
48 |
|
49 |
+
- Orion-14B series models are open-source multilingual large language models trained from scratch by OrionStarAI. The base model is trained on 2.5T multilingual corpus, including Chinese, English, Japanese, Korean, etc, and it exhibits superior performance in these languages. For details, please refer to [tech report](https://github.com/OrionStarAI/Orion/blob/master/doc/Orion14B_v3.pdf).
|
50 |
+
|
51 |
+
- The Orion-14B series models exhibit the following features:
|
52 |
+
- Among models with 20B-parameter scale level, Orion-14B-Base model shows outstanding performance in comprehensive evaluations.
|
53 |
+
- Strong multilingual capabilities, significantly outperforming in Japanese and Korean testsets.
|
54 |
+
- The fine-tuned models demonstrate strong adaptability, excelling in human-annotated blind tests.
|
55 |
+
- The long-chat version supports extremely long texts, performing exceptionally well at a token length of 200k and can support up to a maximum of 320k.
|
56 |
+
- The quantized versions reduce model size by 70%, improve inference speed by 30%, with performance loss less than 1%.
|
57 |
+
<table style="border-collapse: collapse; width: 100%;">
|
58 |
+
<tr>
|
59 |
+
<td style="border: none; padding: 10px; box-sizing: border-box;">
|
60 |
+
<img src="./assets/imgs/opencompass_en.png" alt="opencompass" style="width: 100%; height: auto;">
|
61 |
+
</td>
|
62 |
+
<td style="border: none; padding: 10px; box-sizing: border-box;">
|
63 |
+
<img src="./assets/imgs/model_cap_en.png" alt="modelcap" style="width: 100%; height: auto;">
|
64 |
+
</td>
|
65 |
+
</tr>
|
66 |
+
</table>
|
67 |
|
68 |
- Orion-14B series models including:
|
69 |
- **Orion-14B-Base:** A multilingual large language foundational model with 14 billion parameters, pretrained on a diverse dataset of 2.5 trillion tokens.
|
70 |
- **Orion-14B-Chat:** A chat-model fine-tuned on a high-quality corpus aims to provide an excellence interactive experience for users in the large model community.
|
71 |
+
- **Orion-14B-LongChat:** The long-context version excels at handling extremely lengthy texts, performing exceptionally well at a token length of 200k and can support up to a maximum of 320k.
|
72 |
- **Orion-14B-Chat-RAG:** A chat-model fine-tuned on a custom retrieval augmented generation dataset, achieving superior performance in retrieval augmented generation tasks.
|
73 |
- **Orion-14B-Chat-Plugin:** A chat-model specifically tailored for plugin and function calling tasks, ideal for agent-related scenarios where the LLM acts as a plugin and function call system.
|
74 |
- **Orion-14B-Base-Int4:** A quantized base model utilizing 4-bit integer weights. It significantly reduces the model size by 70% and increases the inference speed by 30% while incurring a minimal performance loss of only 1%.
|
|
|
336 |
|
337 |
**The core strengths of OrionStar lies in possessing end-to-end AI application capabilities,** including big data preprocessing, large model pretraining, fine-tuning, prompt engineering, agent, etc. With comprehensive end-to-end model training capabilities, including systematic data processing workflows and the parallel model training capability of hundreds of GPUs, it has been successfully applied in various industry scenarios such as government affairs, cloud services, international e-commerce, and fast-moving consumer goods.
|
338 |
|
339 |
+
Companies with demands for deploying large-scale model applications are welcome to contact us.<br>
|
340 |
**Enquiry Hotline: 400-898-7779**<br>
|
341 |
**E-mail: [email protected]**
|
342 |
|