Skip to content

Commit

Permalink
Update README.md
Browse files Browse the repository at this point in the history
  • Loading branch information
robmarkcole committed Jun 12, 2023
1 parent 4c9dc8c commit a2c8ab5
Showing 1 changed file with 13 additions and 1 deletion.
14 changes: 13 additions & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -1980,6 +1980,12 @@ Change detection is a vital component of remote sensing analysis, enabling the m

8.105. [SiamCRNN](https://github.com/ChenHongruixuan/SiamCRNN) -> code for 2020 paper: Change Detection in Multisource VHR Images via Deep Siamese Convolutional Multiple-Layers Recurrent Neural Network

8.106. [Graph-based methods for change detection in remote sensing images](https://github.com/jfflorez/Graph-based-methods-for-change-detection-in-remote-sensing-images) -> code for paper: Graph Learning Based on Signal Smoothness Representation for Homogeneous and Heterogeneous Change Detection

8.107. [TransUNetplus2](https://github.com/aj1365/TransUNetplus2) -> code for 2023 paper: TransU-Net++: Rethinking attention gated TransU-Net for deforestation mapping

8.108. [AR-CDNet](https://github.com/guanyuezhen/AR-CDNet) -> code for 2023 paper: Towards Accurate and Reliable Change Detection of Remote Sensing Images via Knowledge Review and Online Uncertainty Estimation


#
## 9. Time series
Expand Down Expand Up @@ -2598,6 +2604,8 @@ Data fusion is a technique for combining information from different sources such

17.36. [deforestation-from-data-fusion](https://github.com/felferrari/deforestation-from-data-fusion) -> Fusing Sentinel-1 and Sentinel-2 images for deforestation detection in the Brazilian Amazon under diverse cloud conditions

17.37. [sct-fusion](https://git.tu-berlin.de/rsim/sct-fusion) -> code for 2023 [paper](https://arxiv.org/abs/2306.01523): Transformer-based Multi-Modal Learning for Multi Label Remote Sensing Image Classification


#
## 18. Generative Adversarial Networks (GANs)
Expand Down Expand Up @@ -2808,7 +2816,9 @@ Image Captioning is the task of automatically generating a textual description o

Visual Question Answering (VQA) is the task of automatically answering a natural language question about an image. In remote sensing, VQA enables users to interact with the images and retrieve information using natural language questions. For example, a user could ask a VQA system questions such as "What is the type of land cover in this area?", "What is the dominant crop in this region?" or "What is the size of the city in this image?". The system would then analyze the image and generate an answer based on its understanding of the image content.

22.1 [VQA-easy2hard](https://gitlab.lrz.de/ai4eo/reasoning/VQA-easy2hard) -> code for 2022 [paper](https://arxiv.org/abs/2205.03147): From Easy to Hard: Learning Language-guided Curriculum for Visual Question Answering on Remote Sensing Data
22.1. [VQA-easy2hard](https://gitlab.lrz.de/ai4eo/reasoning/VQA-easy2hard) -> code for 2022 [paper](https://arxiv.org/abs/2205.03147): From Easy to Hard: Learning Language-guided Curriculum for Visual Question Answering on Remote Sensing Data

22.2. [lit4rsvqa](https://git.tu-berlin.de/rsim/lit4rsvqa) -> code for [paper](https://arxiv.org/abs/2306.00758): LiT-4-RSVQA: Lightweight Transformer-based Visual Question Answering in Remote Sensing

#
## 23. Mixed data learning
Expand Down Expand Up @@ -3035,6 +3045,8 @@ Federated learning is an approach to distributed machine learning where a centra

28.2. [Semantic-Segmentation-UNet-Federated](https://github.com/PratikGarai/Semantic-Segmentation-UNet-Federated) -> code for paper: FedUKD: Federated UNet Model with Knowledge Distillation for Land Use Classification from Satellite and Street Views

28.3. [MM-FL](https://git.tu-berlin.de/rsim/MM-FL) -> code for paper: Learning Across Decentralized Multi-Modal Remote Sensing Archives with Federated Learning

#
## 29. Transformers
Vision transformers are state-of-the-art models for vision tasks such as image classification and object detection. They differ from CNNs as they use self-attention instead of convolution to learn global relations between all pixels in the image. Vision transformers employ a transformer encoder architecture, composed of multi-layer blocks with multi-head self-attention and feed-forward layers, enabling the capture of rich contextual information for more accurate predictions.
Expand Down

0 comments on commit a2c8ab5

Please sign in to comment.