-
Memory Augmented Language Models through Mixture of Word Experts
Paper • 2311.10768 • Published • 18 -
Mixtral of Experts
Paper • 2401.04088 • Published • 158 -
MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts
Paper • 2401.04081 • Published • 71 -
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
Paper • 2401.06066 • Published • 51
Collections
Discover the best community collections!
Collections trending this week
-
joseluhf11/clinical_case_symptoms_diseases_dataset
Viewer • Updated • 1.19k • 138 • 9 -
BLACKBUN/imaginary_patient_cases
Viewer • Updated • 4.97k • 10 • 1 -
mahfoos/Patient-Doctor-Conversation
Viewer • Updated • 3.33k • 338 • 16 -
medalpaca/medical_meadow_wikidoc_patient_information
Viewer • Updated • 5.94k • 616 • 26