From your work, I find a new way to do model ensemble
1
#14 opened 11 months ago
by
xxx1
Adding Evaluation Results
#12 opened 11 months ago
by
leaderboard-pr-bot
![](https://cdn-avatars.huggingface.co/v1/production/uploads/655506df9dc61e22c5f9c732/IZGvup0FdVlioPPIPnzZv.jpeg)
The function_calling and translation abilities are weaker than Mixtral 8x7b
1
#11 opened about 1 year ago
by
bingw5
Add mixture of experts tag
#10 opened about 1 year ago
by
davanstrien
![](https://cdn-avatars.huggingface.co/v1/production/uploads/1627505688463-60107b385ac3e86b3ea4fc34.jpeg)
how this model goes work,can you share you idea or traning process? thanks
#9 opened about 1 year ago
by
zachzhou
Add merge tag
2
#8 opened about 1 year ago
by
osanseviero
![](https://cdn-avatars.huggingface.co/v1/production/uploads/6032802e1f993496bc14d9e3/w6hr-DEQot4VVkoyRIBiy.png)
Vram
2
#7 opened about 1 year ago
by
DKRacingFan
source code and paper?
8
#6 opened about 1 year ago
by
josephykwang
How does the MoE work?
3
#5 opened about 1 year ago
by
PacmanIncarnate
Quant pls?
6
#4 opened about 1 year ago
by
Yhyu13
What is your config?
1
#3 opened about 1 year ago
by
Weyaxi
![](https://cdn-avatars.huggingface.co/v1/production/uploads/6468ce47e134d050a58aa89c/ApFcPlOzgI6Cjr0SYPpk6.png)
Should not be called mixtral, the models made into the moe are yi based
9
#2 opened about 1 year ago
by
teknium
![](https://cdn-avatars.huggingface.co/v1/production/uploads/6317aade83d8d2fd903192d9/erOwgMXc_CZih3uMoyTAp.jpeg)
Add merge tags
#1 opened about 1 year ago
by
JusticeDike