Skip to content

[IEEE Transactions on Information Forensics & Security 2023] Imperceptible and Robust Backdoor Attack in 3D Point Cloud

Notifications You must be signed in to change notification settings

KuofengGao/IRBA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Imperceptible and Robust Backdoor Attack in 3D Point Cloud

This repository provides the pytorch implementatin of our TIFS 2023 work: Imperceptible and Robust Backdoor Attack in 3D Point Cloud.

Abstract

With the thriving of deep learning in processing point cloud data, recent works show that backdoor attacks pose a severe security threat to 3D vision applications. The attacker injects the backdoor into the 3D model by poisoning a few training samples with trigger, such that the backdoored model performs well on clean samples but behaves maliciously when the trigger pattern appears. Existing attacks often insert some additional points into the point cloud as the trigger, or utilize a linear transformation (e.g., rotation) to construct the poisoned point cloud. However, the effects of these poisoned samples are likely to be weakened or even eliminated by some commonly used pre-processing techniques for 3D point cloud, e.g., outlier removal or rotation augmentation. In this paper, we propose a novel imperceptible and robust backdoor attack (IRBA) to tackle this challenge. We utilize a nonlinear and local transformation, called weighted local transformation (WLT), to construct poisoned samples with unique transformations. As there are several hyper-parameters and randomness in WLT, it is difficult to produce two similar transformations. Consequently, poisoned samples with unique transformations are likely to be resistant to aforementioned pre-processing techniques. Besides, as the controllability and smoothness of the distortion caused by a fixed WLT, the generated poisoned samples are also imperceptible to human inspection. Extensive experiments on three benchmark datasets and four models show that IRBA achieves 80%+ ASR in most cases even with pre-processing techniques, which is significantly higher than previous state-of-the-art attacks.

Pipeline of IRBA

Installation

This code is tested on our local environment (python=3.7, cuda=11.1), and we recommend you to use anaconda to create a vitural environment:

conda create -n IRBA python=3.7

Then, activate the environment:

conda activate IRBA

Install PyTorch:

pip install torch==1.8.0+cu111 torchvision==0.9.0+cu111 torchaudio==0.8.0 -f https://download.pytorch.org/whl/torch_stable.html

and other requirements:

pip install -r requirements.txt

Data Preparation

Please download ModelNet dataset from its official website to .data/.

Backdoor attack

IRBA backdoor attack in our paper can be run as follows.

python backdoor_attack.py --dataset modelnet10 --num_category 10 --model pointnet_cls --poisoned_rate 0.1 --target_label 8 --num_anchor 16 --R_alpha 5 --S_size 5 --process_data --use_uniform_sample --gpu 0

Calculate the Chamfer Distance

Chamfer Distance between the clean samples and the poisoned samples in our paper is calculated as follows.

cd cuda/chamfer_dist && python setup.py install --user && cd .. && cd ..
python -m tools.calculate_cd --process_data --use_uniform_sample

Citation

@article{gao2023imperceptible,
  title={Imperceptible and robust backdoor attack in 3d point cloud},
  author={Gao, Kuofeng and Bai, Jiawang and Wu, Baoyuan and Ya, Mengxi and Xia, Shu-Tao},
  journal={IEEE Transactions on Information Forensics and Security},
  volume={19},
  pages={1267--1282},
  year={2023},
  publisher={IEEE}
}

Acknowledgements

This respository is mainly based on Pointnet_Pointnet2_pytorch. Thanks for the wonderful work!

About

[IEEE Transactions on Information Forensics & Security 2023] Imperceptible and Robust Backdoor Attack in 3D Point Cloud

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published