Karthik2510 commited on
Commit
4dae541
·
verified ·
1 Parent(s): 011980e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -10
README.md CHANGED
@@ -95,26 +95,50 @@ Use the code below to get started with the model.
95
 
96
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
97
 
98
- [More Information Needed]
 
 
 
 
 
 
99
 
100
  ### Training Procedure
101
 
102
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
103
 
104
- #### Preprocessing [optional]
105
 
106
- [More Information Needed]
 
 
 
 
107
 
108
 
109
  #### Training Hyperparameters
110
 
111
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
112
 
113
- #### Speeds, Sizes, Times [optional]
 
 
 
 
 
 
 
 
 
 
 
114
 
115
  <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
116
 
117
- [More Information Needed]
 
 
 
118
 
119
  ## Evaluation
120
 
@@ -160,10 +184,10 @@ Use the code below to get started with the model.
160
 
161
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
162
 
163
- - **Hardware Type:** [More Information Needed]
164
- - **Hours used:** [More Information Needed]
165
- - **Cloud Provider:** [More Information Needed]
166
- - **Compute Region:** [More Information Needed]
167
  - **Carbon Emitted:** [More Information Needed]
168
 
169
  ## Technical Specifications [optional]
 
95
 
96
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
97
 
98
+ The model has been fine-tuned on the **dmedhi/wiki_medical_terms** dataset. This dataset is designed to improve medical terminology comprehension and consists of:
99
+
100
+ - Medical definitions and terminologies
101
+ - Disease symptoms and conditions
102
+ - Healthcare and clinical knowledge from Wikipedia's medical section
103
+
104
+ This dataset ensures that the fine-tuned model performs well in understanding and responding to medical queries with enhanced accuracy.
105
 
106
  ### Training Procedure
107
 
108
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
109
 
110
+ #### Preprocessing
111
 
112
+ - The dataset was cleaned and tokenized using the Llama 3.1 tokenizer, ensuring that medical terms were preserved.
113
+
114
+ - Special medical terminologies were handled properly to maintain context.
115
+
116
+ - The dataset was formatted into a question-answer style to align with the instruction-based nature of Llama 3.1 3B Instruct.
117
 
118
 
119
  #### Training Hyperparameters
120
 
121
+ - **Training regime:** <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
122
 
123
+ - **Training regime:** bf16 mixed precision (to balance efficiency and precision)
124
+ - **Batch Size:** 1 per device
125
+ - **Gradient Accumulation Steps:** 4 (to simulate a larger batch size)
126
+ - **Learning Rate:** 2e-4
127
+ - **Warmup Steps:** 100
128
+ - **Epochs:** 3
129
+ - **Optimizer:** paged_adamw_8bit (efficient low-memory optimizer)
130
+ - **LoRA Rank (r):** 16
131
+ - **LoRA Alpha:** 32
132
+ - **LoRA Dropout:** 0.05
133
+
134
+ #### Speeds, Sizes, Times
135
 
136
  <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
137
 
138
+ - **Training Hardware:** Single GPU (consumer-grade, VRAM-optimized)
139
+ - **Model Size after Fine-Tuning:** Approx. 3B parameters with LoRA adapters
140
+ - **Training Time:** ~3-4 hours per epoch on A100 40GB GPU
141
+ - **Final Checkpoint Size:** ~2.8GB (with LoRA adapters stored separately)
142
 
143
  ## Evaluation
144
 
 
184
 
185
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
186
 
187
+ - **Hardware Type:** A100 40 GB GPU
188
+ - **Hours used:** Approximatly 3 to 4 hours
189
+ - **Cloud Provider:** Google Colabs
190
+ - **Compute Region:** US-East
191
  - **Carbon Emitted:** [More Information Needed]
192
 
193
  ## Technical Specifications [optional]