Update README.md
Browse files
README.md
CHANGED
@@ -95,26 +95,50 @@ Use the code below to get started with the model.
|
|
95 |
|
96 |
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
97 |
|
98 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
99 |
|
100 |
### Training Procedure
|
101 |
|
102 |
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
103 |
|
104 |
-
#### Preprocessing
|
105 |
|
106 |
-
|
|
|
|
|
|
|
|
|
107 |
|
108 |
|
109 |
#### Training Hyperparameters
|
110 |
|
111 |
-
- **Training regime:**
|
112 |
|
113 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
114 |
|
115 |
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
116 |
|
117 |
-
|
|
|
|
|
|
|
118 |
|
119 |
## Evaluation
|
120 |
|
@@ -160,10 +184,10 @@ Use the code below to get started with the model.
|
|
160 |
|
161 |
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
162 |
|
163 |
-
- **Hardware Type:**
|
164 |
-
- **Hours used:**
|
165 |
-
- **Cloud Provider:**
|
166 |
-
- **Compute Region:**
|
167 |
- **Carbon Emitted:** [More Information Needed]
|
168 |
|
169 |
## Technical Specifications [optional]
|
|
|
95 |
|
96 |
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
97 |
|
98 |
+
The model has been fine-tuned on the **dmedhi/wiki_medical_terms** dataset. This dataset is designed to improve medical terminology comprehension and consists of:
|
99 |
+
|
100 |
+
- Medical definitions and terminologies
|
101 |
+
- Disease symptoms and conditions
|
102 |
+
- Healthcare and clinical knowledge from Wikipedia's medical section
|
103 |
+
|
104 |
+
This dataset ensures that the fine-tuned model performs well in understanding and responding to medical queries with enhanced accuracy.
|
105 |
|
106 |
### Training Procedure
|
107 |
|
108 |
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
109 |
|
110 |
+
#### Preprocessing
|
111 |
|
112 |
+
- The dataset was cleaned and tokenized using the Llama 3.1 tokenizer, ensuring that medical terms were preserved.
|
113 |
+
|
114 |
+
- Special medical terminologies were handled properly to maintain context.
|
115 |
+
|
116 |
+
- The dataset was formatted into a question-answer style to align with the instruction-based nature of Llama 3.1 3B Instruct.
|
117 |
|
118 |
|
119 |
#### Training Hyperparameters
|
120 |
|
121 |
+
- **Training regime:** <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
122 |
|
123 |
+
- **Training regime:** bf16 mixed precision (to balance efficiency and precision)
|
124 |
+
- **Batch Size:** 1 per device
|
125 |
+
- **Gradient Accumulation Steps:** 4 (to simulate a larger batch size)
|
126 |
+
- **Learning Rate:** 2e-4
|
127 |
+
- **Warmup Steps:** 100
|
128 |
+
- **Epochs:** 3
|
129 |
+
- **Optimizer:** paged_adamw_8bit (efficient low-memory optimizer)
|
130 |
+
- **LoRA Rank (r):** 16
|
131 |
+
- **LoRA Alpha:** 32
|
132 |
+
- **LoRA Dropout:** 0.05
|
133 |
+
|
134 |
+
#### Speeds, Sizes, Times
|
135 |
|
136 |
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
137 |
|
138 |
+
- **Training Hardware:** Single GPU (consumer-grade, VRAM-optimized)
|
139 |
+
- **Model Size after Fine-Tuning:** Approx. 3B parameters with LoRA adapters
|
140 |
+
- **Training Time:** ~3-4 hours per epoch on A100 40GB GPU
|
141 |
+
- **Final Checkpoint Size:** ~2.8GB (with LoRA adapters stored separately)
|
142 |
|
143 |
## Evaluation
|
144 |
|
|
|
184 |
|
185 |
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
186 |
|
187 |
+
- **Hardware Type:** A100 40 GB GPU
|
188 |
+
- **Hours used:** Approximatly 3 to 4 hours
|
189 |
+
- **Cloud Provider:** Google Colabs
|
190 |
+
- **Compute Region:** US-East
|
191 |
- **Carbon Emitted:** [More Information Needed]
|
192 |
|
193 |
## Technical Specifications [optional]
|