This is the official repository for the ICLR 2024 paper "DiffusionSat: A Generative Foundation Model For Satellite Imagery".
Authors: Samar Khanna 1, Patrick Liu, Linqi (Alex) Zhou, Chenlin Meng, Robin Rombach, Marshall Burke, David B. Lobell, Stefano Ermon.
Please refer to INSTALLATION.md
.
Model checkpoint files can be found at this link.
Note that this is a temporary location, and checkpoint files may be moved to another location in the future (I will update the README in this case).
(While the files are on Google Drive, you can use gdown to download them).
This section covers image-generation using single-image DiffusionSat, without control signal inputs.
The relevant jupyter notebook can be found in notebooks/single-image.ipynb
.
The relevant model checkpoints can be found here:
Resolution | Link |
---|---|
512 x 512 | Download |
256 x 256 | Download |
Coming soon, stay tuned...
These sections describe how to launch training using accelerate
.
To train the (text, metadata) -> single_image
model, use the following command:
./launch_scripts/launch_256_fmow_satlas_spacenet_img_txt_md.sh launch_accelerate_configs/single_gpu_accelerate_config.yaml
Here we provide an example config file to use with accelerate
, but you can also configure your own
file by running accelerate config
and following the steps. This will save the config file in the
cache location (eg: .cache/huggingface/accelerate/default_config.yaml
),
and you can simply copy over the .yaml
file to launch_accelerate_configs/
or remove the
--config_file
argument from accelerate launch
in the bash script.
The datasets we use are in webdataset
format.
You will need to prepare your datasets in this format to be able to train using the given code,
or you can modify the data-loading to use your own custom dataset formats.
We have provided example shardlists in webdataset_shards
. The training code will read the relevant file,
and load data using the data paths in this file. The advantage of using webdataset
is that your data
does not need to only be on disk, and you can stream data from buckets in AWS S3 as well.
Example format for each entry in the fMoW webdataset .tar
file.
__key__: fmow-{cls_name}-{instance_id} # eg: fmow-airport-airport_0
output.cls: label_idx # eg: 32
input.npy: (h,w,c) numpy array
metadata.json: {'img_filename': ..., 'gsd': ..., 'cloud_cover': ..., 'timestamp': ..., 'country_code': ...}
Note that fMoW also requires a metadata .csv
file.
If you find our project helpful, please cite our paper:
@inproceedings{
khanna2024diffusionsat,
title={DiffusionSat: A Generative Foundation Model for Satellite Imagery},
author={Samar Khanna and Patrick Liu and Linqi Zhou and Chenlin Meng and Robin Rombach and Marshall Burke and David B. Lobell and Stefano Ermon},
booktitle={The Twelfth International Conference on Learning Representations},
year={2024},
url={https://openreview.net/forum?id=I5webNFDgQ}
}