HuggingSara commited on
Commit
994eae0
·
verified ·
1 Parent(s): e3a0d95

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -2
README.md CHANGED
@@ -43,7 +43,10 @@ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
43
  ## Model Details
44
 
45
 
46
- The BiMediX model, built on a Mixture of Experts (MoE) architecture, leverages the Mixtral-8x7B base network. This approach enables the model to scale significantly by utilizing a sparse operation method, where only a subset of its 47 billion parameters are active during inference, enhancing efficiency. It features a sophisticated router network to allocate tasks to the most relevant experts, each being a specialized feedforward block within the model. The training utilized the BiMed1.3M dataset, focusing on bilingual medical interactions in both English and Arabic, with a substantial corpus of over 632 million healthcare-specialized tokens. The model's fine-tuning process includes a low-rank adaptation technique (QLoRA) to efficiently adapt the model to specific tasks while keeping computational demands manageable.
 
 
 
47
 
48
  ## Dataset
49
 
@@ -51,7 +54,20 @@ The BiMediX model, built on a Mixture of Experts (MoE) architecture, leverages t
51
 
52
  ## Benchmarks and Performance
53
 
54
- (Details about benchmarks and results.)
 
 
 
 
 
 
 
 
 
 
 
 
 
55
 
56
  ## Limitations and Ethical Considerations
57
 
 
43
  ## Model Details
44
 
45
 
46
+ The BiMediX model, built on a Mixture of Experts (MoE) architecture, leverages the [Mixtral-8x7B](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) base model. It features a sophisticated router network to allocate tasks to the most relevant experts, each being a specialized feedforward blocks within the model.
47
+ This approach enables the model to scale significantly by utilizing a sparse operation method, where less than 13 billion parameters are active during inference, enhancing efficiency.
48
+ The training utilized the BiMed1.3M dataset, focusing on bilingual medical interactions in both English and Arabic, with a substantial corpus of over 632 million healthcare-specialized tokens.
49
+ The model's fine-tuning process includes a low-rank adaptation technique (QLoRA) to efficiently adapt the model to specific tasks while keeping computational demands manageable.
50
 
51
  ## Dataset
52
 
 
54
 
55
  ## Benchmarks and Performance
56
 
57
+ The BiMediX model was evaluated across several benchmarks, demonstrating its effectiveness in medical language understanding and question answering in both English and Arabic.
58
+
59
+ 1. **Medical Benchmarks Used for Evaluation:**
60
+ - **PubMedQA**: A dataset for question answering from biomedical research papers, requiring reasoning over biomedical contexts.
61
+ - **MedMCQA**: Multiple-choice questions from Indian medical entrance exams, covering a wide range of medical subjects.
62
+ - **MedQA**: Questions from US and other medical board exams, testing specific knowledge and patient case understanding.
63
+ - **Medical MMLU**: A compilation of questions from various medical subjects, requiring broad medical knowledge.
64
+
65
+ 2. **Results and Comparisons:**
66
+ - **Bilingual Evaluation**: BiMediX showed superior performance in bilingual (Arabic-English) evaluations, outperforming both the Mixtral-8x7B base model and Jais-30B, a model designed for Arabic. It demonstrated more than 10 and 15 points higher average accuracy, respectively.
67
+ - **Arabic Benchmark**: In Arabic-specific evaluations, BiMediX outperformed Jais-30B in all categories, highlighting the effectiveness of the BiMed1.3M dataset and bilingual training.
68
+ - **English Benchmark**: BiMediX also excelled in English medical benchmarks, surpassing other state-of-the-art models like Med42-70B and Meditron-70B in terms of average performance and efficiency.
69
+
70
+ These results underscore BiMediX's advanced capability in handling medical queries and its significant improvement over existing models in both languages, leveraging its unique bilingual dataset and training approach.
71
 
72
  ## Limitations and Ethical Considerations
73