File size: 3,548 Bytes
9ba07a3
 
 
 
 
 
 
 
 
f069b48
160a6b1
fd81fc7
4081464
945a98b
f069b48
 
 
 
 
 
 
 
 
bcc56bc
4cefa76
f069b48
fd81fc7
e9d9b6c
fd81fc7
 
 
43cfb6e
fd81fc7
43cfb6e
fd81fc7
 
e9d9b6c
fd81fc7
 
 
 
 
 
43cfb6e
fd81fc7
 
e9d9b6c
fd81fc7
 
 
 
 
43cfb6e
fd81fc7
 
 
 
 
4cefa76
 
4e137dc
ef001a4
4e137dc
 
 
 
 
 
 
ef001a4
94a7b93
3e7b7de
160a6b1
 
4cefa76
ef001a4
160a6b1
4e137dc
 
 
 
 
 
 
3e7b7de
4e137dc
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
---
title: README
emoji: 🐠
colorFrom: gray
colorTo: green
sdk: static
pinned: false
---


<div align="center">
  <img src="https://cdn-uploads.huggingface.co/production/uploads/656d9adce8bf55919aca7c3f/xcy1rwFGbrVZ1N68LQcEI.gif" style="width: 65%">
</div>

<p style="margin-top: 0px;" align="center">
  <a rel="nofollow" href="https://discord.gg/hYUwWddeAu">
    <img style="width:20px; vertical-align: middle; display: inline-block; margin-right: 5px; margin-left: 10px; margin-top: 0px; margin-bottom: 0px;" alt="Discord Logo" src="https://cloud.githubusercontent.com/assets/6291467/26705903/96c2d66e-477c-11e7-9f4e-f3c0efe96c9a.png">
    <span class="link-text">Discord</span>
  </a> |
  <a rel="nofollow" href="https://arxiv.org/abs/2403.04652">
    <img style="width:20px; vertical-align: middle; display: inline-block; margin-right: 5px; margin-left: 10px; margin-top: 0px; margin-bottom: 0px;" alt="ArXiv Logo" src="https://github.com/alpayariyak/openchat/blob/master/assets/arxiv-logomark-small-square-border.png?raw=true">
    <span class="link-text">Paper</span>
  </a> 
</p>


<div class="grid lg:grid-cols-3 gap-x-4 gap-y-7">
	<a href="https://www.01.ai/" class="block overflow-hidden group">
		<div
			class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center bg-[#FFFFFF]"
		>
			<img alt="" src="https://cdn-uploads.huggingface.co/production/uploads/656d9adce8bf55919aca7c3f/ADRH_f61dt8uVWBsAehkG.gif" class="w-40" />
		</div>
		<div align="center">Base Models<br/> (Yi-6B/9B/34B)</div>
	</a>
	<a
		href="https://www.01.ai/"
		class="block overflow-hidden"
	>
		<div class="flex items-center h-40 bg-[#FFFFFF] rounded-lg px-4 mb-2">
            <img alt="" src="https://cdn-uploads.huggingface.co/production/uploads/656d9adce8bf55919aca7c3f/JuiI5Zun1XD5BuHCK0L1I.gif" class="w-40" />
		</pre>
		</div>
		<div align="center">Chat Models <br/> (Yi-6B/9B/34B Chat)</div>
	</a>
	<a
		href="https://www.01.ai/"
		class="block overflow-hidden group"
	>
		<div class="flex items-center h-40 bg-[#FFFFFF] rounded-lg px-4 mb-2">
			<img alt="" src="https://cdn-uploads.huggingface.co/production/uploads/656d9adce8bf55919aca7c3f/7D6SjExHLIO1cH0tmQyxh.gif" class="w-40" />
		</div>
		<div align="center" class="underline">Multimodal Models <br/> (Yi-VL-6B/34B)</div>
	</a>
	<div class="lg:col-span-3">
		<p class="mb-4">
		</p>

Welcome to Yi! 😘 

Yi model family is a series of language and multimodal models. It is based on 6B and 34B pretrained language models, then extended to **chat** models, 200K **long context** models, depth-upscaled models (9B), and **vision**-language models.

# ⭐️ Highlights

- **Strong performance**: Yi-1.5-34B is on par with or surpasses GPT-3.5 in commonsense reasoning, college exams, math, coding, reading comprehension, and human preference win-rate on multiple evaluation benchmarks.

- **Cost-effective**: For 6B, 9B, and 34B, you can perform inference on consumer-grade hardware (like the RTX 4090). Additionally, 34B is large enough with complex reasoning and emergent abilities, giving a nice performance-cost balance.

# πŸ“Š Benchmarks

TBD

<div align="center">
  <img src="https://cdn-uploads.huggingface.co/production/uploads/656d9adce8bf55919aca7c3f/kVHWz7yEY3UJlcRD2nwf2.png" style="width: 65%">
</div>


# πŸ“° News

- 2024-03-16: The Yi-9B-200K is open-sourced and available to the public.

- 2024-03-08: Yi Tech Report is published!

- 2024-03-07: The long text capability of the Yi-34B-200K has been enhanced.

For complete news history, see [News](xx.md).