youngking0727 commited on
Commit
d514d13
1 Parent(s): e3e5895

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -43
README.md CHANGED
@@ -7,70 +7,41 @@ datasets:
7
  ---
8
  # BioMedGPT-LM-7B
9
 
10
- In this repo, we present a medical language model named BioMedGPT-LM which is the first commercial-friendly GPT model in the biomedical domain and has demonstrated
11
- superior performance over existing LLMs of the same parameter size. We are releasing a 7B model **BioMedGPT-LM-7B** which is LLaMA2-7b-chat finetuned on the PMC abstracts and papers from the S2ORC.
12
-
13
-
14
-
15
 
16
  ### Training Details
17
 
18
-
19
-
20
  The model was trained with the following hyperparameters:
21
 
22
- * Epochs: 5
23
- * Batch size: 192
24
- * Cutoff length: 2048
25
  * Learning rate: 2e-5
26
 
27
- Overview BioMedGPT-LM-7B was finetuned on over 26 billion tokens highly pertinent to the field of biomedicine. The fine-tuning data are extracted from 5.5 million biomedical papers in S2ORC data using PubMed Central
28
- (PMC)-ID and PubMed ID as criteria.
29
 
 
30
 
31
- ### Model Developers
32
  PharMolix
33
 
34
  ### How to Use
35
- BioMedGPT-LM-7B is a part of **[BioMedGPT-10B](https://github.com/BioFM/OpenBioMed)**, an open-source version of BioMedGPT. BioMedGPT is a multimodal generative pre-trained transformer (GPT) for biomedicine, which bridges the natural language modality and diverse biomed-
36
- ical data modalities via a single GPT model. BioMedGPT aligns different biological modalities with the text modality via BioMedGPT-LM. The details of BioMedGPT-10B and BioMedGPT-LM-7B can be found in the [technical report]().
37
- ![The architecture of BioMedGPT-10B](BioMedGPT-10B.jpeg)
38
-
39
-
40
-
41
 
42
- **Intended Use Cases**
 
43
 
44
- | **Method** | Parameters (B) | Setting | MedMCQA(\%) | PubMedQA(\%) |
45
- |------------------------|----------------|-----------|-------------|--------------|
46
- | Human (pass)* | - | Manual | - | 60.0 |
47
- | Human (expert)* | - | Manual | 90 | 78.0 |
48
- |------------------------|----------------|-----------|-------------|--------------|
49
- | InstructGPT* | 175 | zero-shot | 44.0 | 73.2 |
50
- | ChatGPT* | - | zero-shot | 44.7 | 63.9 |
51
- | Llama* | 7 | zero-shot | 24.3 | 5.2 |
52
- | Llama2 | 7 | zero-shot | 30.6 | 3.7 |
53
- | Llama2-Chat | 7 | zero-shot | 35.5 | 21.9 |
54
- |------------------------|----------------| --------- |-------------|--------------|
55
- | Llama | 7 |Fine-tuing | 48.2 | 73.4 |
56
- | Llama2-Chat | 7 |Fine-tuing | 48.3 | 75.5 |
57
- | PMC-Llama | 7 |Fine-tuing | 50.5 | 69.5 |
58
- |------------------------|----------------|-----------|-------------|--------------|
59
- | **BioMedGPT-LM-7B** | 7 |Fine-tuing | **51.4** | **76.1** |
60
-
61
- **Out-of-scope Uses**
62
 
 
63
 
64
  ### Technical Report
65
- "BioMedGPT: Open Multimodal Generative Pre-trained Transformer for BioMedicine"
66
 
 
67
 
68
  ### github
69
- [https://github.com/BioFM/OpenBioMed](https://github.com/BioFM/OpenBioMed)
70
 
 
71
 
72
  ### Limitations
73
 
74
- [Highlight any limitations or potential issues of your model.]
75
-
76
-
 
7
  ---
8
  # BioMedGPT-LM-7B
9
 
10
+ **BioMedGPT-LM-7B** is the first large generative language model based on Llama2 in the biomedical domain.
11
+ It was fine-tuned from the Llama2-7B-Chat with millions of biomedical papers from the [S2ORC corpus](https://github.com/allenai/s2orc/blob/master/README.md). Through further fine-tuning, BioMedGPT-LM-7B outperforms or is on par with human and significantly larger general-purpose foundation models on several biomedical QA benchmarks.
 
 
 
12
 
13
  ### Training Details
14
 
 
 
15
  The model was trained with the following hyperparameters:
16
 
17
+ * Epochs: 5
18
+ * Batch size: 192
19
+ * Context length: 2048
20
  * Learning rate: 2e-5
21
 
22
+ BioMedGPT-LM-7B is finetuned on over 26 billion tokens highly pertinent to the field of biomedicine. The fine-tuning data are extracted from 5.5 million biomedical papers in S2ORC data using PubMed Central (PMC)-ID and PubMed ID as criteria.
 
23
 
24
+ ### Model Developers
25
 
 
26
  PharMolix
27
 
28
  ### How to Use
 
 
 
 
 
 
29
 
30
+ BioMedGPT-LM-7B is the generative language model of **[BioMedGPT-10B](https://github.com/BioFM/OpenBioMed)**, an open-source version of BioMedGPT.
31
+ BioMedGPT is an open multimodal generative pre-trained transformer (GPT) for biomedicine, which bridges the natural language modality and diverse biomedical data modalities via large generative language models.
32
 
33
+ More technical details of BioMedGPT-LM-7B, BioMedGPT-10B, and BioMedGPT can be found in the [technical report](https://pan.baidu.com/s/1iAMBkuoZnNAylhopP5OgEg?pwd=7a6b).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
 
35
+ ![The architecture of BioMedGPT-10B](BioMedGPT-10B.jpeg)
36
 
37
  ### Technical Report
 
38
 
39
+ "BioMedGPT: Open Multimodal Generative Pre-trained Transformer for BioMedicine"
40
 
41
  ### github
 
42
 
43
+ [https://github.com/BioFM/OpenBioMed](https://github.com/BioFM/OpenBioMed)
44
 
45
  ### Limitations
46
 
47
+ Large-scale generative language models represent a novel technology, with their generated outputs determined by probabilities, possibly resulting in unforeseen issues, such as generating responses that may contain elements of danger, bias, discrimination, or other harmful content. Up to this point, our experiments focus on data within the field of English biomedical domains, leaving many scenarios unaddressed. Therefore, before using BioMedGPT, developers should conduct safety testing and necessary adjustments. While BioMedGPT has expertise in areas like biomedicine and chemistry, it should not be employed for research that is hazardous or could endanger human life, and it cannot replace medical professionals or provide professional treatment advice. Users need to be cautious and take extra care when using these models.