Skip to content
/ SMMCL Public

[WACV 2024] Understanding Dark Scenes by Contrasting Multi-Modal Observations

Notifications You must be signed in to change notification settings

palmdong/SMMCL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

99 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

SMMCL

PyTorch implementation of "Understanding Dark Scenes by Contrasting Multi-Modal Observations"
[WACV paper, supp]

Updates

[2023/10/24] Paper was accepted by WACV2024. We thank anonymous reviewers from ICCV2023 and WACV2024 for their suggestions to our paper. See you in Hawaii.
[2024/04/22] The code for experiments on low-light indoor scenes (the LLRGBD dataset) was uploaded. [2024/08/02] Updated the supplementary material. Previously there were errors with Figures III, V, and VI.

Abstract

Understanding dark scenes based on multi-modal image data is challenging, as both the visible and auxiliary modalities provide limited semantic information for the task. Previous methods focus on fusing the two modalities but neglect the correlations among semantic classes when minimizing losses to align pixels with labels, resulting in inaccurate class predictions. To address these issues, we introduce a supervised multi-modal contrastive learning approach to increase the semantic discriminability of the learned multi-modal feature spaces by jointly performing cross-modal and intra-modal contrast under the supervision of the class correlations. The cross-modal contrast encourages same-class embeddings from across the two modalities to be closer and pushes different-class ones apart. The intra-modal contrast forces same-class or different-class embeddings within each modality to be together or apart. We validate our approach on a variety of tasks that cover diverse light conditions and image modalities. Experiments show that our approach can effectively enhance dark scene understanding based on multi-modal images with limited semantics by shaping semantic-discriminative feature spaces. Comparisons with previous methods demonstrate our state-of-the-art performance.

Experiments

Preparation

  • Python 3.10.6, Torch 1.12.1, CUDA 10.2, requirements.txt
  • Download the datasets and put them in /datasets.

Train

Download the pretrained weight of SegNext-B and put it in /pretrained/segnext.

cd /path/to/SMMCL_LLRGBD or /path/to/SMMCL_MFNet or /path/to/SMMCL_NYU
# modify config.py
python train.py -d 0-3 

Test

Quick Start: Download our pretrained weights and put them in /SMMCL_XXX/log_XXX_mscan_b/checkpoint.

cd /path/to/SMMCL_LLRGBD
python eval.py -d 0-3 -e 500 
cd /path/to/SMMCL_MFNet
python eval.py -d 0-3 -e 300 
cd /path/to/SMMCL_NYU
python eval.py -d 0-3 -e 600

Acknowledgement

Our code was built based on the repositories of CMX and MSCSCL. We thank the authors for their efforts.

Citation

@InProceedings{Dong2024SMMCL,
  author    = {Dong, Xiaoyu and Yokoya, Naoto},
  title     = {Understanding Dark Scenes by Contrasting Multi-Modal Observations},
  booktitle = {WACV},
  year      = {2024}
}

About

[WACV 2024] Understanding Dark Scenes by Contrasting Multi-Modal Observations

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages