Update README.md
Browse files
README.md
CHANGED
@@ -31,11 +31,9 @@ license: mit
|
|
31 |
- **Paper [optional]:** https://arxiv.org/abs/2412.13126
|
32 |
- **Demo [optional]:** [More Information Needed]
|
33 |
|
34 |
-
## Uses
|
35 |
-
|
36 |
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
37 |
|
38 |
-
|
39 |
|
40 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
41 |
|
@@ -69,9 +67,7 @@ text_feature = model.encode_text(token_input)
|
|
69 |
|
70 |
<!-- This section describes the evaluation protocols and provides the results. -->
|
71 |
|
72 |
-
### Testing Data
|
73 |
-
|
74 |
-
#### Testing Data
|
75 |
|
76 |
<!-- This should link to a Dataset Card if possible. -->
|
77 |
|
@@ -111,7 +107,7 @@ We present benchmark results for a range of representative tasks. A complete set
|
|
111 |
| CPTAC-NSCLC | 0.647 | 0.607 | 0.643 | 0.836 | **0.863** |
|
112 |
| EBRAINS | 0.096 | 0.093 | 0.325 | 0.371 | **0.456** |
|
113 |
|
114 |
-
|
115 |
|
116 |
Validated on 18 diverse benchmarks with more than 14,000 whole slide images (WSIs), KEEP achieves state-of-the-art performance in zero-shot cancer diagnostic tasks. Notably, for cancer detection, KEEP demonstrates an average sensitivity of 89.8% at a specificity of 95.0% across 7 cancer types, significantly outperforming vision-only foundation models and highlighting its promising potential for clinical application. For cancer subtyping, KEEP achieves a median balanced accuracy of 0.456 in subtyping 30 rare brain cancers, indicating strong generalizability for diagnosing rare tumors.
|
117 |
|
@@ -120,7 +116,6 @@ Validated on 18 diverse benchmarks with more than 14,000 whole slide images (WSI
|
|
120 |
|
121 |
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
122 |
|
123 |
-
**BibTeX:**
|
124 |
```
|
125 |
@article{zhou2024keep,
|
126 |
title={A Knowledge-enhanced Pathology Vision-language Foundation Model for Cancer Diagnosis},
|
|
|
31 |
- **Paper [optional]:** https://arxiv.org/abs/2412.13126
|
32 |
- **Demo [optional]:** [More Information Needed]
|
33 |
|
|
|
|
|
34 |
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
35 |
|
36 |
+
## Direct Use
|
37 |
|
38 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
39 |
|
|
|
67 |
|
68 |
<!-- This section describes the evaluation protocols and provides the results. -->
|
69 |
|
70 |
+
### Testing Data
|
|
|
|
|
71 |
|
72 |
<!-- This should link to a Dataset Card if possible. -->
|
73 |
|
|
|
107 |
| CPTAC-NSCLC | 0.647 | 0.607 | 0.643 | 0.836 | **0.863** |
|
108 |
| EBRAINS | 0.096 | 0.093 | 0.325 | 0.371 | **0.456** |
|
109 |
|
110 |
+
### Summary
|
111 |
|
112 |
Validated on 18 diverse benchmarks with more than 14,000 whole slide images (WSIs), KEEP achieves state-of-the-art performance in zero-shot cancer diagnostic tasks. Notably, for cancer detection, KEEP demonstrates an average sensitivity of 89.8% at a specificity of 95.0% across 7 cancer types, significantly outperforming vision-only foundation models and highlighting its promising potential for clinical application. For cancer subtyping, KEEP achieves a median balanced accuracy of 0.456 in subtyping 30 rare brain cancers, indicating strong generalizability for diagnosing rare tumors.
|
113 |
|
|
|
116 |
|
117 |
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
118 |
|
|
|
119 |
```
|
120 |
@article{zhou2024keep,
|
121 |
title={A Knowledge-enhanced Pathology Vision-language Foundation Model for Cancer Diagnosis},
|