File size: 1,721 Bytes
fa3860b
16abc95
664bf62
 
 
fa3860b
 
 
 
 
 
 
 
 
 
 
664bf62
fa3860b
664bf62
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
---
base_model: IBI-CAAI/MELT-Mistral-3x7B-Instruct-v0.1
license: apache-2.0
language:
- en
library_name: transformers
tags:
- 4-bit
- AWQ
- text-generation
- autotrain_compatible
- endpoints_compatible
pipeline_tag: text-generation
inference: false
quantized_by: Suparious
---
# IBI-CAAI/MELT-Mistral-3x7B-Instruct-v0.1 AWQ

- Model creator: [IBI-CAAI](https://huggingface.co/IBI-CAAI)
- Original model: [MELT-Mistral-3x7B-Instruct-v0.1](https://huggingface.co/IBI-CAAI/MELT-Mistral-3x7B-Instruct-v0.1)

## Model Summary

The MELT-Mistral-3x7B-Instruct-v0.1 Large Language Model (LLM) is a pretrained generative text model pre-trained and fine-tuned on using publically avalable medical data.

MELT-Mistral-3x7B-Instruct-v0.1 demonstrated a average 19.7% improvement over Mistral-3x7B-Instruct-v0.1 (MoE of 3 X Mistral-7B-Instruct-v0.1) across 3 USMLE, Indian AIIMS, and NEET medical examination benchmarks.

This is MoE model, thanks to [Charles Goddard](https://huggingface.co/chargoddard) for code/tools.

The Medical Education Language Transformer (MELT) models have been trained on a wide-range of text, chat, Q/A, and instruction data in the medical domain.   

While the model was evaluated using publically avalable [USMLE](https://www.usmle.org/), Indian AIIMS, and NEET medical examination example questions, its use it intented to be more broadly applicable.   

- **Developed by:** [Center for Applied AI](https://caai.ai.uky.edu/)
- **Funded by:** [Institute or Biomedical Informatics](https://www.research.uky.edu/IBI)
- **Model type:** LLM
- **Language(s) (NLP):** English
- **License:** Apache 2.0
- **Finetuned from model:** A MoE x 3 [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1)