File size: 1,668 Bytes
ad38abf
 
7958455
24c3399
 
 
fadc4fa
 
20497b9
2367f50
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e783c6c
1025210
 
 
 
 
 
6159219
1025210
 
 
 
 
2231d4f
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
---
license: apache-2.0
metrics:
- mse
- mae
- r_squared
tags:
- medical
model-index:
- name: Vit-COVID-19-severity
  results:
  - task:
      name: Image Classification
      type: image-classification
    metrics:
    - name: MSE
      type: mse
      value: 0.5135470628738403
    - name: MAE
      type: mae
      value: 0.5676153898239136
    - name: R^ 2
      type: r_squared
      value: 0.5378701686859131
pipeline_tag: image-classification
---

## Diagnosing Medical Images for COVID-19 Severity (Regression task)

The COVID-19 pandemic has strained healthcare resources and prompted discussion about how machine learning can alleviate physician burdens and contribute to diagnosis. Chest x-rays (CXRs) are used for diagnosis of COVID-19, but few studies predict the severity of a patient’s condition from CXRs. 
In this study, we produce a large COVID severity dataset by merging three sources and investigate the efficacy of transfer learning vision transformers (ViTs) in severity regression task. 
The ViT had the best regression results, with an MSE of 0.5135. Code developed in this project is available at https://github.com/stwhitfield/covid-severity.

## Model description

We were fine-tuned the Vision Transformer (ViT) from HuggingFaces on our training set for 300 epochs using an AdamW optimizer with a learning rate of 0.00001 and a batch size of 10. 
Early stopping was applied to the model to prevent over-fitting.
For ViT, training stopped after 70 epochs. The only difference between the preprocessed data in the ViT and the other models was that it normalized to a tensor image with mean and standard deviation during data augmentation.