Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,46 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
---
|
4 |
+
|
5 |
+
## Overview
|
6 |
+
|
7 |
+
[SatlasPretrain](https://satlas-pretrain.allen.ai) is a large-scale remote sensing image understanding dataset.
|
8 |
+
The models here are Swin Transformer backbones pre-trained on either the high-resolution images or the Sentinel-2 images in SatlasPretrain.
|
9 |
+
|
10 |
+
- `satlas-model-v1-highres.pth` is applicable for downstream tasks involving 0.5-2.0 m/pixel satellite or aerial imagery.
|
11 |
+
- `satlas-model-v1-lowres.pth` is applicable for downstream tasks involving [Sentinel-2 satellite images](https://sentinel.esa.int/web/sentinel/missions/sentinel-2).
|
12 |
+
|
13 |
+
The pre-trained backbones are expected to improve performance on a wide range of remote sensing and geospatial tasks, such as planetary and environmental monitoring.
|
14 |
+
They have already been deployed to develop robust models for detecting solar farms, wind turbines, offshore platforms, and tree cover in [Satlas](https://satlas.allen.ai), a platform for global geospatial data generated by AI from satellite imagery.
|
15 |
+
|
16 |
+
## Usage and Input Normalization
|
17 |
+
|
18 |
+
The backbones can be loaded for fine-tuning on downstream tasks:
|
19 |
+
|
20 |
+
import torch
|
21 |
+
import torchvision
|
22 |
+
model = torchvision.models.swin_transformer.swin_v2_b()
|
23 |
+
full_state_dict = torch.load('satlas-model-v1-highres.pth')
|
24 |
+
# Extract just the Swin backbone parameters from the full state dict.
|
25 |
+
swin_prefix = 'backbone.backbone.'
|
26 |
+
swin_state_dict = {k[len(swin_prefix):]: v for k, v in full_state_dict.items() if k.startswith(swin_prefix)}
|
27 |
+
model.load_state_dict(swin_state_dict)
|
28 |
+
|
29 |
+
The expected input is as follows:
|
30 |
+
|
31 |
+
- `satlas-model-v1-highres.pth`: inputs 8-bit RGB high-resolution images, with 0-255 RGB values normalized to 0-1 by dividing by 255.
|
32 |
+
- `satlas-model-v1-lowres.pth`: inputs the TCI image from Sentinel-2 L1C scenes, which is an 8-bit image already processed from the B04 (red), B03 (green), and B02 (blue) bands. Normalize the 0-255 RGB values to 0-1 by dividing by 255.
|
33 |
+
|
34 |
+
Please see [the SatlasPretrain github](https://github.com/allenai/satlas/blob/main/SatlasPretrain.md) for more examples and usage options.
|
35 |
+
Models that use nine Sentinel-2 bands are also available there.
|
36 |
+
|
37 |
+
## Code
|
38 |
+
|
39 |
+
The training code and SatlasPretrain dataset are at https://github.com/allenai/satlas/.
|
40 |
+
|
41 |
+
SatlasPretrain is [a paper](https://arxiv.org/abs/2211.15660) appearing at the International Conference on Computer Vision in October 2023.
|
42 |
+
|
43 |
+
## Feedback
|
44 |
+
|
45 |
+
We welcome any feedback about the model or training data.
|
46 |
+
To contact us, please [open an issue on the SatlasPretrain github](https://github.com/allenai/satlas/issues).
|