G-AshwinKumar commited on
Commit
43c6de4
1 Parent(s): ea50dbb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -79
README.md CHANGED
@@ -3,10 +3,17 @@ license: cc-by-nc-4.0
3
  datasets:
4
  - argilla/dpo-mix-7k
5
  - nvidia/HelpSteer
 
 
 
 
 
 
 
 
 
6
  language:
7
  - en
8
- metrics:
9
- - accuracy
10
  library_name: transformers
11
  tags:
12
  - biology
@@ -26,20 +33,19 @@ Aloe is a new family of healthcare LLMs that is highly competitive with all prev
26
  - **Developed by:** [HPAI](https://hpai.bsc.es/)
27
  - **Model type:** Causal decoder-only transformer language model
28
  - **Language(s) (NLP):** English (mainly)
29
- - **License:** [Meta Llama 3 License](https://llama.meta.com/llama3/license/)
30
- - **Finetuned from model :** [meta-llama/Meta-Llama-3-8B · Hugging Face](https://huggingface.co/meta-llama/Meta-Llama-3-8B)
31
 
32
  ### [](https://huggingface.co/templates/model-card-example#model-sources-optional)Model Sources [optional]
33
 
34
  - **Repository:** Coming Soon
35
- - **Paper [optional]:** Coming soon
36
 
37
  ## Uses
38
 
39
  ### Direct Use
40
 
41
- We encourage the use of Aloe for research purposes, as a
42
- stepping stone to build better foundational models for healthcare.
43
 
44
  ### Out-of-Scope Use
45
 
@@ -47,13 +53,13 @@ These models are not to be used for clinical practice, medical diagnosis, or any
47
 
48
  ## Bias, Risks, and Limitations
49
 
50
- First let us consider Healthcare professional impersonation, a fraudulent behaviour which currently generates billions of dollars in profit https://www.justice.gov/opa/pr/justice-department-charges-dozens-12-billion-health-care-fraud. A model such as Aloe could be used to increase the efficacy of such deceiving activities, making them more widespread. The main preventive actions are public literacy on the unreliability of digitised information and the importance of medical registration, and legislation enforcing AI-generated content disclaimers. The second risk we consider is medical decision-making without professional supervision. While this is already an issue in modern societies (\eg self-medication) a model such as Aloe, capable of producing high-quality conversational data, can facilitate self-delusion, particularly in the presence of sycophancy. By producing tailored responses, it can also be used to generate actionable answers. Public literacy on the dangers of self-diagnosis is one of the main defences, together with the introduction of disclaimers and warnings on the models' outputs. The last risk we consider is the access to information on dangerous substances or procedures. While the literature on sensitive content can already be found on different sources (\eg libraries, internet, dark web), LLMs can centralize such access, making it nearly impossible to control the flow of such information. Model alignment can help in that regard, but so far the effects remain insufficient, as jailbreaking methods still overcome it.
51
 
52
- <img src="https://cdn-uploads.huggingface.co/production/uploads/62972c4979f193515da1d38e/JR7AU-DwJRNAmk8vFPmfT.png" width="90%">
53
 
54
  ### Recommendations
55
 
56
- We avoid the use of all personal data in our training. Model safety cannot be guaranteed, as shown in the red teaming results. Aloe can produce toxic content under the appropriate prompts. For these reasons, minors should not be left alone to interact with Aloe without supervision.
57
 
58
  ## How to Get Started with the Model
59
 
@@ -150,25 +156,14 @@ Supervised fine-tuning on top of Llama 3 8B using medical and general domain dat
150
 
151
  ### Training Data
152
 
153
- - Medical and general domain datasets, including synthetic data generated using Mixtral-8x7B and Genstruct
 
 
 
154
  - argilla/dpo-mix-7k
155
  - nvidia/HelpSteer
156
  - Custom preference data with adversarial prompts generated from Anthropic Harmless, Chen et al., and original prompts
157
 
158
- ### Training Procedure
159
-
160
- #### Preprocessing [optional]
161
-
162
- [More Information Needed]
163
-
164
- #### Training Hyperparameters
165
-
166
- - **Training regime:** [More Information Needed]
167
-
168
- #### Speeds, Sizes, Times [optional]
169
-
170
- [More Information Needed]
171
-
172
  ## Evaluation
173
 
174
  ### Testing Data, Factors & Metrics
@@ -180,10 +175,7 @@ Supervised fine-tuning on top of Llama 3 8B using medical and general domain dat
180
  - [PubMedQA](https://huggingface.co/datasets/bigbio/pubmed_qa)
181
  - [MMLU-Medical](https://huggingface.co/datasets/lukaemon/mmlu)
182
  - [MedQA-4-Option](https://huggingface.co/datasets/GBaker/MedQA-USMLE-4-options)
183
-
184
- #### Factors
185
-
186
- [More Information Needed]
187
 
188
  #### Metrics
189
 
@@ -191,20 +183,15 @@ Supervised fine-tuning on top of Llama 3 8B using medical and general domain dat
191
 
192
  ### Results
193
 
194
- <img src="https://cdn-uploads.huggingface.co/production/uploads/62972c4979f193515da1d38e/5viYXiXgsTnlMLVRNa1NP.png" width="90%">
195
-
196
 
197
  #### Summary
198
 
199
- To compare Aloe with the most competitive open models (both general purpose and healthcare-specific) we use popular healthcare datasets (\eg PubMedQA, MedMCQA, MedQA and MMLU for six medical tasks only), together with the new and highly reliable CareQA. We produce the standard MultiMedQA score for reference, by computing the weighted average accuracy on all scores except CareQA. Additionally, we calculate the arithmetic mean across all datasets. The Medical MMLU is calculated by averaging the six medical subtasks: Anatomy, Clinical knowledge, College Biology, College medicine, Medical genetics, and Professional medicine.
200
-
201
- Benchmark results indicate the training conducted on Aloe has boosted its performance slightly above Llama3-8B-Instruct. Llama3-Aloe-8B-Alpha outperforms larger models like Meditron 70B, and is close to larger base models, like Yi-34} For the former, this gain is consistent even when using SC-CoT, using their best-reported variant. All these results make Llama3-Aloe-8B-Alpha the best healthcare LLM of its size.
202
-
203
- With the help of prompting techniques the performance of Llama3-Aloe-8B-Alpha is significantly improved. Medprompting in particular provides a 7\% increase in reported accuracy, after which \aloealpha only lags behind the ten times bigger Llama-3-70B-Instruct. This improvement is mostly consistent across medical fields. Llama3-Aloe-8B-Alpha with medprompting beats the performance of Meditron 70B with their self reported 20 shot SC-CoT in MMLU med and is slightly worse in the other benchmarks.
204
 
205
- ## Model Examination [optional]
206
 
207
- [More Information Needed]
208
 
209
  ## Environmental Impact
210
 
@@ -214,46 +201,6 @@ With the help of prompting techniques the performance of Llama3-Aloe-8B-Alpha is
214
  - **Compute Region:** Spain
215
  - **Carbon Emitted:** 439.25kg
216
 
217
- ## Technical Specifications [optional]
218
-
219
- ### Model Architecture and Objective
220
-
221
- [More Information Needed]
222
-
223
- ### Compute Infrastructure
224
-
225
- [More Information Needed]
226
-
227
- #### Hardware
228
-
229
- [More Information Needed]
230
-
231
- #### Software
232
-
233
- [More Information Needed]
234
-
235
- ## Citation [optional]
236
-
237
- **BibTeX:**
238
-
239
- [More Information Needed]
240
-
241
- **APA:**
242
-
243
- [More Information Needed]
244
-
245
- ## Glossary [optional]
246
-
247
- [More Information Needed]
248
-
249
- ## More Information [optional]
250
-
251
- [More Information Needed]
252
-
253
- ## Model Card Authors [optional]
254
-
255
- [More Information Needed]
256
-
257
  ## Model Card Contact
258
 
259
 
3
  datasets:
4
  - argilla/dpo-mix-7k
5
  - nvidia/HelpSteer
6
+ - jondurbin/airoboros-3.2
7
+ - hkust-nlp/deita-10k-v0
8
+ - LDJnr/Capybara
9
+ - HPAI-BSC/CareQA
10
+ - GBaker/MedQA-USMLE-4-options
11
+ - lukaemon/mmlu
12
+ - bigbio/pubmed_qa
13
+ - openlifescienceai/medmcqa
14
+ - bigbio/med_qa
15
  language:
16
  - en
 
 
17
  library_name: transformers
18
  tags:
19
  - biology
 
33
  - **Developed by:** [HPAI](https://hpai.bsc.es/)
34
  - **Model type:** Causal decoder-only transformer language model
35
  - **Language(s) (NLP):** English (mainly)
36
+ - **License:** This model is based on Meta Llama 3 8B and is governed by the [Meta Llama 3 License](https://llama.meta.com/llama3/license/). All our modifications are available with a [CC BY-NC 4.0](https://creativecommons.org/licenses/by-nc/4.0/) license.
37
+ - **Finetuned from model :** [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B)
38
 
39
  ### [](https://huggingface.co/templates/model-card-example#model-sources-optional)Model Sources [optional]
40
 
41
  - **Repository:** Coming Soon
42
+ - **Paper:** Coming soon
43
 
44
  ## Uses
45
 
46
  ### Direct Use
47
 
48
+ We encourage the use of Aloe for research purposes, as a stepping stone to build better foundational models for healthcare.
 
49
 
50
  ### Out-of-Scope Use
51
 
 
53
 
54
  ## Bias, Risks, and Limitations
55
 
56
+ First let us consider Healthcare professional impersonation, a fraudulent behaviour which currently generates billions of dollars in [profit](https://www.justice.gov/opa/pr/justice-department-charges-dozens-12-billion-health-care-fraud). A model such as Aloe could be used to increase the efficacy of such deceiving activities, making them more widespread. The main preventive actions are public literacy on the unreliability of digitised information and the importance of medical registration, and legislation enforcing AI-generated content disclaimers. The second risk we consider is medical decision-making without professional supervision. While this is already an issue in modern societies (eg self-medication) a model such as Aloe, capable of producing high-quality conversational data, can facilitate self-delusion, particularly in the presence of sycophancy. By producing tailored responses, it can also be used to generate actionable answers. Public literacy on the dangers of self-diagnosis is one of the main defences, together with the introduction of disclaimers and warnings on the models' outputs. The last risk we consider is the access to information on dangerous substances or procedures. While the literature on sensitive content can already be found on different sources (eg libraries, internet, dark web), LLMs can centralize such access, making it nearly impossible to control the flow of such information. Model alignment can help in that regard, but so far the effects remain insufficient, as jailbreaking methods still overcome it.
57
 
58
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/62972c4979f193515da1d38e/T6Jblpf1kmTkM04K716rM.png" width="90%">
59
 
60
  ### Recommendations
61
 
62
+ We avoid the use of all personal data in our training. Model safety cannot be guaranteed. Aloe can produce toxic content under the appropriate prompts. For these reasons, minors should not be left alone to interact with Aloe without supervision.
63
 
64
  ## How to Get Started with the Model
65
 
 
156
 
157
  ### Training Data
158
 
159
+ - Medical domain datasets, including synthetic data generated using Mixtral-8x7B and Genstruct
160
+ - LDJnr/Capybara
161
+ - hkust-nlp/deita-10k-v0
162
+ - jondurbin/airoboros-3.2
163
  - argilla/dpo-mix-7k
164
  - nvidia/HelpSteer
165
  - Custom preference data with adversarial prompts generated from Anthropic Harmless, Chen et al., and original prompts
166
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
167
  ## Evaluation
168
 
169
  ### Testing Data, Factors & Metrics
 
175
  - [PubMedQA](https://huggingface.co/datasets/bigbio/pubmed_qa)
176
  - [MMLU-Medical](https://huggingface.co/datasets/lukaemon/mmlu)
177
  - [MedQA-4-Option](https://huggingface.co/datasets/GBaker/MedQA-USMLE-4-options)
178
+ - [CareQA](https://huggingface.co/datasets/HPAI-BSC/CareQA)
 
 
 
179
 
180
  #### Metrics
181
 
 
183
 
184
  ### Results
185
 
186
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/62972c4979f193515da1d38e/STlPSggXr9P9JeWAvmAsi.png" width="90%">
 
187
 
188
  #### Summary
189
 
190
+ To compare Aloe with the most competitive open models (both general purpose and healthcare-specific) we use popular healthcare datasets (PubMedQA, MedMCQA, MedQA and MMLU for six medical tasks only), together with the new and highly reliable CareQA. We produce the standard MultiMedQA score for reference, by computing the weighted average accuracy on all scores except CareQA. Additionally, we calculate the arithmetic mean across all datasets. The Medical MMLU is calculated by averaging the six medical subtasks: Anatomy, Clinical knowledge, College Biology, College medicine, Medical genetics, and Professional medicine.
 
 
 
 
191
 
192
+ Benchmark results indicate the training conducted on Aloe has boosted its performance above Llama3-8B-Instruct. Llama3-Aloe-8B-Alpha outperforms larger models like Meditron 70B, and is close to larger base models, like Yi-34. For the former, this gain is consistent even when using SC-CoT, using their best-reported variant. All these results make Llama3-Aloe-8B-Alpha the best healthcare LLM of its size.
193
 
194
+ With the help of prompting techniques the performance of Llama3-Aloe-8B-Alpha is significantly improved. Medprompting in particular provides a 7% increase in reported accuracy, after which Llama3-Aloe-8B-Alpha only lags behind the ten times bigger Llama-3-70B-Instruct. This improvement is mostly consistent across medical fields. Llama3-Aloe-8B-Alpha with medprompting beats the performance of Meditron 70B with their self reported 20 shot SC-CoT in MMLU med and is slightly worse in the other benchmarks.
195
 
196
  ## Environmental Impact
197
 
 
201
  - **Compute Region:** Spain
202
  - **Carbon Emitted:** 439.25kg
203
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
204
  ## Model Card Contact
205
 
206