Skip to content

Latest commit

 

History

History

English | 简体中文

SCRFD C++ Deployment Example

This directory provides examples that infer.cc fast finishes the deployment of SCRFD on CPU/GPU and GPU accelerated by TensorRT.

Before deployment, two steps require confirmation

Taking the CPU inference on Linux as an example, the compilation test can be completed by executing the following command in this directory. FastDeploy version 0.7.0 or above (x.x.x>=0.7.0) is required to support this model.

mkdir build
cd build
# Download the FastDeploy precompiled library. Users can choose your appropriate version in the `FastDeploy Precompiled Library` mentioned above 
wget https://bj.bcebos.com/fastdeploy/release/cpp/fastdeploy-linux-x64-x.x.x.tgz
tar xvf fastdeploy-linux-x64-x.x.x.tgz
cmake .. -DFASTDEPLOY_INSTALL_DIR=${PWD}/fastdeploy-linux-x64-x.x.x
make -j

# Download the official converted SCRFD model files and test images 
wget https://bj.bcebos.com/paddlehub/fastdeploy/scrfd_500m_bnkps_shape640x640.onnx
wget https://github.com/raw/DefTruth/lite.ai.toolkit/main/examples/lite/resources/test_lite_face_detector_3.jpg


# CPU inference
./infer_demo scrfd_500m_bnkps_shape640x640.onnx test_lite_face_detector_3.jpg 0
# GPU inference
./infer_demo scrfd_500m_bnkps_shape640x640.onnx test_lite_face_detector_3.jpg 1
# TensorRT inference on GPU
./infer_demo scrfd_500m_bnkps_shape640x640.onnx test_lite_face_detector_3.jpg 2

The visualized result after running is as follows

The above command works for Linux or MacOS. For SDK use-pattern in Windows, refer to:

SCRFD C++ Interface

SCRFD Class

fastdeploy::vision::facedet::SCRFD(
        const string& model_file,
        const string& params_file = "",
        const RuntimeOption& runtime_option = RuntimeOption(),
        const ModelFormat& model_format = ModelFormat::ONNX)

SCRFD model loading and initialization, among which model_file is the exported ONNX model format

Parameter

  • model_file(str): Model file path
  • params_file(str): Parameter file path. Only passing an empty string when the model is in ONNX format
  • runtime_option(RuntimeOption): Backend inference configuration. None by default, which is the default configuration
  • model_format(ModelFormat): Model format. ONNX format by default

Predict Function

SCRFD::Predict(cv::Mat* im, FaceDetectionResult* result,
                float conf_threshold = 0.25,
                float nms_iou_threshold = 0.5)

Model prediction interface. Input images and output detection results.

Parameter

  • im: Input images in HWC or BGR format
  • result: Detection results, including detection box and confidence of each box. Refer to Vision Model Prediction Result for FaceDetectionResult
  • conf_threshold: Filtering threshold of detection box confidence
  • nms_iou_threshold: iou threshold during NMS processing

Class Member Variable

Pre-processing Parameter

Users can modify the following pre-processing parameters to their needs, which affects the final inference and deployment results

  • size(vector<int>): This parameter changes the size of the resize used during preprocessing, containing two integer elements for [width, height] with default value [640, 640]
  • padding_value(vector<float>): This parameter is used to change the padding value of images during resize, containing three floating-point elements that represent the value of three channels. Default value [114, 114, 114]
  • is_no_pad(bool): Specify whether to resize the image through padding or not. is_no_pad=ture represents no paddling. Default is_no_pad=false
  • is_mini_pad(bool): This parameter sets the width and height of the image after resize to the value nearest to the size member variable and to the point where the padded pixel size is divisible by the stride member variable. Default is_mini_pad=false
  • stride(int): Used with the stris_mini_pad member variable. Default stride=32
  • downsample_strides(vector<int>): This parameter is used to change the down-sampling multiple of the feature map that generates anchor, containing three integer elements that represent the default down-sampling multiple for generating anchor. Default [8, 16, 32]
  • landmarks_per_face(int): Modify the number of face keypoints if we use an output with face keypoints. Default landmarks_per_face=5
  • use_kps(bool): Whether to use keypoints or not. If the ONNX file has no keypoint output, set use_kps=false and landmarks_per_face=0. Default use_kps=true
  • num_anchors(int): Set the number predicted by each anchor. The parameters of the trained model need modification accordingly. Default num_anchors=2