ANGEL_pretrained / README.md
dmis-lab's picture
Update README.md
6de179b verified
---
license: gpl-3.0
language:
- en
metrics:
- accuracy
base_model: facebook/bart-large
---
# Model Card for ANGEL_pretrained
This model card provides detailed information about the ANGEL_pretrained model, designed for biomedical entity linking.
# Model Details
#### Model Description
- **Developed by:** Chanhwi Kim, Hyunjae Kim, Sihyeon Park, Jiwoo Lee, Mujeen Sung, Jaewoo Kang
- **Model type:** Generative Biomedical Entity Linking Model
- **Language(s):** English
- **License:** GPL-3.0
- **Finetuned from model:** BART-large (Base architecture)
#### Model Sources
- **Repository:** https://github.com/dmis-lab/ANGEL
- **Paper:** https://arxiv.org/pdf/2408.16493
# Direct Use
ANGEL_pretrained is pretrained with UMLS dataset.
We recommand to finetune this model to downstream dataset rather directly use.
If you still want to run the model on a single sample, no preprocessing is required.
Simply execute the run_sample.sh script:
```bash
bash script/inference/run_sample.sh pretrained
```
To modify the sample with your own example, refer to the [Direct Use](https://github.com/dmis-lab/ANGEL?tab=readme-ov-file#direct-use) section in our GitHub repository.
If you're interested in training or evaluating the model, check out the [Fine-tuning](https://github.com/dmis-lab/ANGEL?tab=readme-ov-file#fine-tuning) section and [Evaluation](https://github.com/dmis-lab/ANGEL?tab=readme-ov-file#evaluation) section.
# Training Details
#### Training Data
The model was pretrained on the UMLS-2020-AA dataset.
#### Training Procedure
Positive-only Pre-training: Initial training using only positive examples, following the standard approach.
Negative-aware Training: Subsequent training incorporated negative examples to improve the model's discriminative capabilities.
# Evaluation
#### Testing Data
The model was evaluated using multiple biomedical datasets, including NCBI-disease, BC5CDR, COMETA, AAP, and MedMentions.
The fine-tuned scores have also been included.
#### Metrics
**Accuracy at Top-1 (Acc@1)**: Measures the percentage of times the model's top prediction matches the correct entity.
### Results
<table border="1" cellspacing="0" cellpadding="5" style="width: 100%; text-align: center; border-collapse: collapse; margin-left: 0;">
<thead>
<tr>
<th style="text-align: center;"><b>Model</b></th>
<th style="text-align: center;"><b>NCBI-disease</b></th>
<th style="text-align: center;"><b>BC5CDR</b></th>
<th style="text-align: center;"><b>COMETA</b></th>
<th style="text-align: center;"><b>AAP</b></th>
<th style="text-align: center;"><b>MedMentions<br>ST21pv</b></th>
<th style="text-align: center;"><b>Average</b></th>
</tr>
</thead>
<tbody>
<tr>
<td><b>GenBioEL_pretrained</b></td>
<td>58.2</td>
<td>33.1</td>
<td>42.4</td>
<td>50.6</td>
<td>10.6</td>
<td><b>39.0</b></td>
</tr>
<tr>
<td><b>ANGEL_pretrained (Ours)</b></td>
<td>64.6</td>
<td>49.7</td>
<td>46.8</td>
<td>61.5</td>
<td>18.2</td>
<td><b>48.2</b></td>
</tr>
<tr>
<td><b>GenBioEL_pt_ft</b></td>
<td>91.0</td>
<td>93.1</td>
<td>80.9</td>
<td>89.3</td>
<td>70.7</td>
<td><b>85.0</b></td>
</tr>
<tr>
<td><b>ANGEL_pt_ft (Ours)</b></td>
<td>92.8</td>
<td>94.5</td>
<td>82.8</td>
<td>90.2</td>
<td>73.3</td>
<td><b>86.7</b></td>
</tr>
</tbody>
</table>
- In this table, "pt" refers to pre-training, where the model is trained on a large dataset (UMLS in this case), and "ft" refers to fine-tuning, where the model is further refined on specific datasets.
In the pre-training phase, **ANGEL** was trained using UMLS dataset entities that were similar to a given word based on TF-IDF scores but had different CUIs (Concept Unique Identifiers).
This negative-aware pre-training approach improved its performance across the benchmarks, leading to an average score of 48.2, which is **9.2** points higher than the GenBioEL pre-trained model, which scored 39.0 on average.
The performance improvement continued during the fine-tuning phase. After fine-tuning, ANGEL achieved an average score of 86.7, surpassing the GenBioEL model's average score of 85.0, representing a further **1.7** point improvement. The ANGEL model consistently outperformed GenBioEL across all datasets in this phase.
The results demonstrate that the negative-aware training introduced by ANGEL not only enhances performance during pre-training but also carries over into fine-tuning, helping the model generalize better to unseen data.
# Citation
If you use the ANGEL_ncbi model, please cite:
```bibtex
@article{kim2024learning,
title={Learning from Negative Samples in Generative Biomedical Entity Linking},
author={Kim, Chanhwi and Kim, Hyunjae and Park, Sihyeon and Lee, Jiwoo and Sung, Mujeen and Kang, Jaewoo},
journal={arXiv preprint arXiv:2408.16493},
year={2024}
}
```
# Contact
For questions or issues, please contact chanhwi_[email protected].