Update README.md
Browse files
README.md
CHANGED
@@ -43,7 +43,10 @@ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
|
43 |
## Model Details
|
44 |
|
45 |
|
46 |
-
The BiMediX model, built on a Mixture of Experts (MoE) architecture, leverages the Mixtral-8x7B base
|
|
|
|
|
|
|
47 |
|
48 |
## Dataset
|
49 |
|
@@ -51,7 +54,20 @@ The BiMediX model, built on a Mixture of Experts (MoE) architecture, leverages t
|
|
51 |
|
52 |
## Benchmarks and Performance
|
53 |
|
54 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
55 |
|
56 |
## Limitations and Ethical Considerations
|
57 |
|
|
|
43 |
## Model Details
|
44 |
|
45 |
|
46 |
+
The BiMediX model, built on a Mixture of Experts (MoE) architecture, leverages the [Mixtral-8x7B](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) base model. It features a sophisticated router network to allocate tasks to the most relevant experts, each being a specialized feedforward blocks within the model.
|
47 |
+
This approach enables the model to scale significantly by utilizing a sparse operation method, where less than 13 billion parameters are active during inference, enhancing efficiency.
|
48 |
+
The training utilized the BiMed1.3M dataset, focusing on bilingual medical interactions in both English and Arabic, with a substantial corpus of over 632 million healthcare-specialized tokens.
|
49 |
+
The model's fine-tuning process includes a low-rank adaptation technique (QLoRA) to efficiently adapt the model to specific tasks while keeping computational demands manageable.
|
50 |
|
51 |
## Dataset
|
52 |
|
|
|
54 |
|
55 |
## Benchmarks and Performance
|
56 |
|
57 |
+
The BiMediX model was evaluated across several benchmarks, demonstrating its effectiveness in medical language understanding and question answering in both English and Arabic.
|
58 |
+
|
59 |
+
1. **Medical Benchmarks Used for Evaluation:**
|
60 |
+
- **PubMedQA**: A dataset for question answering from biomedical research papers, requiring reasoning over biomedical contexts.
|
61 |
+
- **MedMCQA**: Multiple-choice questions from Indian medical entrance exams, covering a wide range of medical subjects.
|
62 |
+
- **MedQA**: Questions from US and other medical board exams, testing specific knowledge and patient case understanding.
|
63 |
+
- **Medical MMLU**: A compilation of questions from various medical subjects, requiring broad medical knowledge.
|
64 |
+
|
65 |
+
2. **Results and Comparisons:**
|
66 |
+
- **Bilingual Evaluation**: BiMediX showed superior performance in bilingual (Arabic-English) evaluations, outperforming both the Mixtral-8x7B base model and Jais-30B, a model designed for Arabic. It demonstrated more than 10 and 15 points higher average accuracy, respectively.
|
67 |
+
- **Arabic Benchmark**: In Arabic-specific evaluations, BiMediX outperformed Jais-30B in all categories, highlighting the effectiveness of the BiMed1.3M dataset and bilingual training.
|
68 |
+
- **English Benchmark**: BiMediX also excelled in English medical benchmarks, surpassing other state-of-the-art models like Med42-70B and Meditron-70B in terms of average performance and efficiency.
|
69 |
+
|
70 |
+
These results underscore BiMediX's advanced capability in handling medical queries and its significant improvement over existing models in both languages, leveraging its unique bilingual dataset and training approach.
|
71 |
|
72 |
## Limitations and Ethical Considerations
|
73 |
|