Suparious's picture
Added base_model tag in README.md
16abc95 verified
---
base_model: IBI-CAAI/MELT-Mistral-3x7B-Instruct-v0.1
license: apache-2.0
language:
- en
library_name: transformers
tags:
- 4-bit
- AWQ
- text-generation
- autotrain_compatible
- endpoints_compatible
pipeline_tag: text-generation
inference: false
quantized_by: Suparious
---
# IBI-CAAI/MELT-Mistral-3x7B-Instruct-v0.1 AWQ
- Model creator: [IBI-CAAI](https://huggingface.co/IBI-CAAI)
- Original model: [MELT-Mistral-3x7B-Instruct-v0.1](https://huggingface.co/IBI-CAAI/MELT-Mistral-3x7B-Instruct-v0.1)
## Model Summary
The MELT-Mistral-3x7B-Instruct-v0.1 Large Language Model (LLM) is a pretrained generative text model pre-trained and fine-tuned on using publically avalable medical data.
MELT-Mistral-3x7B-Instruct-v0.1 demonstrated a average 19.7% improvement over Mistral-3x7B-Instruct-v0.1 (MoE of 3 X Mistral-7B-Instruct-v0.1) across 3 USMLE, Indian AIIMS, and NEET medical examination benchmarks.
This is MoE model, thanks to [Charles Goddard](https://huggingface.co/chargoddard) for code/tools.
The Medical Education Language Transformer (MELT) models have been trained on a wide-range of text, chat, Q/A, and instruction data in the medical domain.
While the model was evaluated using publically avalable [USMLE](https://www.usmle.org/), Indian AIIMS, and NEET medical examination example questions, its use it intented to be more broadly applicable.
- **Developed by:** [Center for Applied AI](https://caai.ai.uky.edu/)
- **Funded by:** [Institute or Biomedical Informatics](https://www.research.uky.edu/IBI)
- **Model type:** LLM
- **Language(s) (NLP):** English
- **License:** Apache 2.0
- **Finetuned from model:** A MoE x 3 [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1)