favyen commited on
Commit
8e4887f
1 Parent(s): 147304d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -0
README.md CHANGED
@@ -1,3 +1,46 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+
5
+ ## Overview
6
+
7
+ [SatlasPretrain](https://satlas-pretrain.allen.ai) is a large-scale remote sensing image understanding dataset.
8
+ The models here are Swin Transformer backbones pre-trained on either the high-resolution images or the Sentinel-2 images in SatlasPretrain.
9
+
10
+ - `satlas-model-v1-highres.pth` is applicable for downstream tasks involving 0.5-2.0 m/pixel satellite or aerial imagery.
11
+ - `satlas-model-v1-lowres.pth` is applicable for downstream tasks involving [Sentinel-2 satellite images](https://sentinel.esa.int/web/sentinel/missions/sentinel-2).
12
+
13
+ The pre-trained backbones are expected to improve performance on a wide range of remote sensing and geospatial tasks, such as planetary and environmental monitoring.
14
+ They have already been deployed to develop robust models for detecting solar farms, wind turbines, offshore platforms, and tree cover in [Satlas](https://satlas.allen.ai), a platform for global geospatial data generated by AI from satellite imagery.
15
+
16
+ ## Usage and Input Normalization
17
+
18
+ The backbones can be loaded for fine-tuning on downstream tasks:
19
+
20
+ import torch
21
+ import torchvision
22
+ model = torchvision.models.swin_transformer.swin_v2_b()
23
+ full_state_dict = torch.load('satlas-model-v1-highres.pth')
24
+ # Extract just the Swin backbone parameters from the full state dict.
25
+ swin_prefix = 'backbone.backbone.'
26
+ swin_state_dict = {k[len(swin_prefix):]: v for k, v in full_state_dict.items() if k.startswith(swin_prefix)}
27
+ model.load_state_dict(swin_state_dict)
28
+
29
+ The expected input is as follows:
30
+
31
+ - `satlas-model-v1-highres.pth`: inputs 8-bit RGB high-resolution images, with 0-255 RGB values normalized to 0-1 by dividing by 255.
32
+ - `satlas-model-v1-lowres.pth`: inputs the TCI image from Sentinel-2 L1C scenes, which is an 8-bit image already processed from the B04 (red), B03 (green), and B02 (blue) bands. Normalize the 0-255 RGB values to 0-1 by dividing by 255.
33
+
34
+ Please see [the SatlasPretrain github](https://github.com/allenai/satlas/blob/main/SatlasPretrain.md) for more examples and usage options.
35
+ Models that use nine Sentinel-2 bands are also available there.
36
+
37
+ ## Code
38
+
39
+ The training code and SatlasPretrain dataset are at https://github.com/allenai/satlas/.
40
+
41
+ SatlasPretrain is [a paper](https://arxiv.org/abs/2211.15660) appearing at the International Conference on Computer Vision in October 2023.
42
+
43
+ ## Feedback
44
+
45
+ We welcome any feedback about the model or training data.
46
+ To contact us, please [open an issue on the SatlasPretrain github](https://github.com/allenai/satlas/issues).