Skip to content

zhufengx/SRN_multilabel

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Spatial Regularization Network

This repository contains training code, testing code and trained models for

Feng Zhu, Hongsheng Li, Wanli Ouyang, Nenghai Yu, Xiaogang Wang, "Learning Spatial Regularization with Image-level Supervisions for Multi-label Image Classification", CVPR 2017. pdf.

Directories and Files

  • caffe/: an early version of Yuanjun Xiong's caffe, with OpenMPI-based Multi-GPU support.
  • tools/: demo code for model testing and evaluation.
  • run_test.sh: script for model testing.
  • evaluation.m: matlab code for classification results evaluation.

Prepare data

  • Download train/test split files (Google Drive, BaiduYun) for NUS-WIDE, MS-COCO and WIDR-Attribute, and extract it to datasets/.
  • Download dataset images
    • NUS-WIDE: this dataset contains many untagged images, and some download links are invalid now. By removing invalid and untagged images, we finally get 119,986 images for training and 80,283 images for testing.
    • MS-COCO_2014: 82,783 images in "train2014" for training, and 40,504 images in "val2014" for testing.
    • WIDER-Attribute: original images are provided here, cropped images for each human bounding box can be downloaded here. 28,340 cropped images in "train" and "val" for training, 29,177 cropped images in "test" for testing.
  • Download trained models (Google Drive, BaiduYun), and extract it to models/
    • the released models containing:
      • trained models for NUS-WIDE, MS-COCO and WIDR-Attribute.
      • a ResNet-101 model pretrained on ImageNet.
  • (Optional) Download reference classification results (Google Drive, BaiduYun), and extract it to results/.

Build

See Yuanjun Xiong's github for building this version of caffe.

Run Test

  • Edit run_test.sh
    • uncomment to specify settings of one dataset.
    • modify variable "ROOT": the root directory holding images of each dataset.
    • modify parameters of "--gpus" to specify available gpus for testing.
  • Edit tools/model_test.py
    • add "path to your caffe" to the search path of python at line 4.

Run Training

  • Download training scripts (Google Drive, BaiduYun), and extract it to training_script/.

  • Edit paths in the scripts.

  • using 'run_train_[datasetname].sh' to train models on one specific dataset.

Evaluation

  • Specify one dataset in line 4 of evaluation.m.
  • Default settings will evaluate [reference classification results](https://drive. .com/open?id=0B7lJth6WXHffc0NGSmJidkNjS2M).

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published