|
| 1 | +# Code for Face Detection experiments with RNNPool |
| 2 | +## Requirements |
| 3 | +1. Follow instructions to install requirements for EdgeML operators and the EdgeML operators [here](https://github.com/microsoft/EdgeML/blob/master/pytorch/README.md). |
| 4 | +2. Install requirements for face detection model using |
| 5 | +``` pip install -r requirements.txt ``` |
| 6 | +We have tested the installation and the code on Ubuntu 18.04 with Cuda 10.2 and CuDNN 7.6 |
| 7 | + |
| 8 | +## Dataset |
| 9 | +1. Download WIDER face dataset images and annotations from http://shuoyang1213.me/WIDERFACE/ and place them all in a folder with name 'WIDER_FACE'. That is, download WIDER_train.zip, WIDER_test.zip, WIDER_val.zip, wider_face_split.zip and place it in WIDER_FACE folder, and unzip files using: |
| 10 | + |
| 11 | +```shell |
| 12 | +cd WIDER_FACE |
| 13 | +unzip WIDER_train.zip |
| 14 | +unzip WIDER_test.zip |
| 15 | +unzip WIDER_val.zip |
| 16 | +unzip wider_face_split.zip |
| 17 | +cd .. |
| 18 | + |
| 19 | +``` |
| 20 | + |
| 21 | +2. In `data/config.py` , set _C.HOME to the parent directory of the above folder, and set the _C.FACE.WIDER_DIR to the folder path. |
| 22 | +That is, if the WIDER_FACE folder is created in /mnt folder, then _C.HOME='/mnt' |
| 23 | +_C.FACE.WIDER_DIR='/mnt/WIDER_FACE'. |
| 24 | +Similarly, change `data/config_qvga.py` to set _C.HOME and _C.FACE.WIDER_DIR. |
| 25 | +3. Run |
| 26 | +``` python prepare_wider_data.py ``` |
| 27 | + |
| 28 | + |
| 29 | +# Usage |
| 30 | + |
| 31 | +## Training |
| 32 | + |
| 33 | +```shell |
| 34 | + |
| 35 | +IS_QVGA_MONO=0 python train.py --batch_size 32 --model_arch RPool_Face_Quant --cuda True --multigpu True --save_folder weights/ --epochs 300 --save_frequency 5000 |
| 36 | + |
| 37 | +``` |
| 38 | + |
| 39 | +For QVGA: |
| 40 | +```shell |
| 41 | + |
| 42 | +IS_QVGA_MONO=1 python train.py --batch_size 64 --model_arch RPool_Face_QVGA_monochrome --cuda True --multigpu True --save_folder weights/ --epochs 300 --save_frequency 5000 |
| 43 | + |
| 44 | +``` |
| 45 | +This will save checkpoints after every '--save_frequency' number of iterations in a weight file with 'checkpoint.pth' at the end and weights for the best state in a file with 'best_state.pth' at the end. These will be saved in '--save_folder'. For resuming training from a checkpoint, use '--resume <checkpoint_name>.pth' with the above command. For example, |
| 46 | + |
| 47 | + |
| 48 | +```shell |
| 49 | + |
| 50 | +IS_QVGA_MONO=1 python train.py --batch_size 64 --model_arch RPool_Face_QVGA_monochrome --cuda True --multigpu True --save_folder weights/ --epochs 300 --save_frequency 5000 --resume <checkpoint_name>.pth |
| 51 | + |
| 52 | +``` |
| 53 | + |
| 54 | +If IS_QVGA_MONO is 0 then training input images will be 640x640 and RGB. |
| 55 | +If IS_QVGA_MONO is 1 then training input images will be 320x320 and converted to monochrome. |
| 56 | + |
| 57 | +Input images for training models are cropped and reshaped to square to maintain consistency with [S3FD](https://arxiv.org/abs/1708.05237). However testing can be done on any size of images, thus we resize testing input image size to have area equal to VGA (640x480)/QVGA (320x240), so that aspect ratio is not changed. |
| 58 | + |
| 59 | +The architecture RPool_Face_QVGA_monochrome is for QVGA monochrome format while RPool_Face_C and RPool_Face_Quant are for VGA RGB format. |
| 60 | + |
| 61 | + |
| 62 | +## Test |
| 63 | +There are two modes of testing the trained model -- the evaluation mode to generate bounding boxes for a set of sample images, and the test mode to compute statistics like mAP scores. |
| 64 | + |
| 65 | +#### Evaluation Mode |
| 66 | + |
| 67 | +Given a set of images in <your_image_folder>, `eval/py` generates bounding boxes around faces (where the confidence is higher than certain threshold) and write the images in <your_save_folder>. To evaluate the `rpool_face_best_state.pth` model (stored in ./weights), execute the following command: |
| 68 | + |
| 69 | +```shell |
| 70 | +IS_QVGA_MONO=0 python eval.py --model_arch RPool_Face_Quant --model ./weights/RPool_Face_Quant_best_state.pth --image_folder <your_image_folder> --save_dir <your_save_folder> |
| 71 | +``` |
| 72 | + |
| 73 | +For QVGA: |
| 74 | +```shell |
| 75 | +IS_QVGA_MONO=1 python eval.py --model_arch RPool_Face_QVGA_monochrome --model ./weights/RPool_Face_QVGA_monochrome_best_state.pth --image_folder <your_image_folder> --save_dir <your_save_folder> |
| 76 | +``` |
| 77 | + |
| 78 | +This will save images in <your_save_folder> with bounding boxes around faces, where the confidence is high. Here is an example image with a single bounding box. |
| 79 | + |
| 80 | + |
| 81 | + |
| 82 | +If IS_QVGA_MONO=0 the evaluation code accepts an image of any size and resizes it to 640x480x3 while preserving original image aspect ratio. |
| 83 | + |
| 84 | +If IS_QVGA_MONO=1 the evaluation code accepts an image of any size and resizes and converts it to monochrome to make it 320x240x1 while preserving original image aspect ratio. |
| 85 | + |
| 86 | +#### WIDER Set Test |
| 87 | +In this mode, we test the generated model against the provided WIDER_FACE validation and test dataset. |
| 88 | + |
| 89 | +For this, first run the following to generate predictions of the model and store output in the '--save_folder' folder. |
| 90 | + |
| 91 | +```shell |
| 92 | +IS_QVGA_MONO=0 python wider_test.py --model_arch RPool_Face_Quant --model ./weights/RPool_Face_Quant_best_state.pth --save_folder rpool_face_quant_val --subset val |
| 93 | +``` |
| 94 | + |
| 95 | +For QVGA: |
| 96 | +```shell |
| 97 | +IS_QVGA_MONO=1 python wider_test.py --model_arch RPool_Face_QVGA_monochrome --model ./weights/RPool_Face_QVGA_monochrome_best_state.pth --save_folder rpool_face_qvgamono_val --subset val |
| 98 | +``` |
| 99 | + |
| 100 | +The above command generates predictions for each image in the "validation" dataset. For each image, a separate prediction file is provided (image_name.txt file in appropriate folder). The first line of the prediction file contains the total number of boxes identified. |
| 101 | +Then each line in the file corresponds to an identified box. For each box, five numbers are generated: length of the box, height of the box, x-axis offset, y-axis offset, confidence value for presence of a face in the box. |
| 102 | + |
| 103 | +If IS_QVGA_MONO=1 then testing is done by converting images to monochrome and QVGA, else if IS_QVGA_MONO=0 then testing is done on VGA RGB images. |
| 104 | + |
| 105 | +The architecture RPool_Face_QVGA_monochrome is for QVGA monochrome format while RPool_Face_C and RPool_Face_Quant are for VGA RGB format. |
| 106 | + |
| 107 | +###### For calculating MAP scores: |
| 108 | +Now using these boxes, we can compute the standard MAP score that is widely used in this literature (see [here](https://medium.com/@jonathan_hui/map-mean-average-precision-for-object-detection-45c121a31173) for more details) as follows: |
| 109 | + |
| 110 | +1. Download eval_tools.zip from http://shuoyang1213.me/WIDERFACE/support/eval_script/eval_tools.zip and unzip in a folder of same name in this directory. |
| 111 | + |
| 112 | +Example code: |
| 113 | + |
| 114 | +```shell |
| 115 | +wget http://shuoyang1213.me/WIDERFACE/support/eval_script/eval_tools.zip |
| 116 | +unzip eval_tools.zip |
| 117 | +``` |
| 118 | + |
| 119 | +2. Set up scripts to use the Matlab '.mat' data files in eval_tools/ground_truth folder for MAP calculation: The following installs python files that provide the same functionality as the '.m' matlab scripts in eval_tools folder. |
| 120 | +``` |
| 121 | +cd eval_tools |
| 122 | +git clone https://github.com/wondervictor/WiderFace-Evaluation.git |
| 123 | +cd WiderFace-Evaluation |
| 124 | +python3 setup.py build_ext --inplace |
| 125 | +``` |
| 126 | + |
| 127 | +3. Run ```python3 evaluation.py -p <your_save_folder> -g <groud truth dir>``` in WiderFace-Evaluation folder |
| 128 | + |
| 129 | +where `prediction_dir` is the '--save_folder' used for `wider_test.py` above and <groud truth dir> is the subfolder `eval_tools/ground_truth`. That is in, WiderFace-Evaluation directory, run: |
| 130 | + |
| 131 | +```shell |
| 132 | +python3 evaluation.py -p <your_save_folder> -g ../ground_truth |
| 133 | +``` |
| 134 | +This script should output the MAP for the WIDER-easy, WIDER-medium, and WIDER-hard subsets of the dataset. Our best performance using RPool_Face_Quant model is: 0.80 (WIDER-easy), 0.78 (WIDER-medium), 0.53 (WIDER-hard). |
| 135 | + |
| 136 | + |
| 137 | +##### Dump RNNPool Input Output Traces and Weights |
| 138 | + |
| 139 | +To save model weights and/or input output pairs for each patch through RNNPool in numpy format use the command below. Put images which you want to save traces for in <your_image_folder> . Specify output folder for saving model weights in numpy format in <your_save_model_numpy_folder>. Specify output folder for saving input output traces of RNNPool in numpy format in <your_save_traces_numpy_folder>. Note that input traces will be saved in a folder named 'inputs' and output traces in a folder named 'outputs' inside <your_save_traces_numpy_folder>. |
| 140 | + |
| 141 | +```shell |
| 142 | +python3 dump_model.py --model ./weights/RPool_Face_QVGA_monochrome_best_state.pth --model_arch RPool_Face_Quant --image_folder <your_image_folder> --save_model_npy_dir <your_save_model_numpy_folder> --save_traces_npy_dir <your_save_traces_numpy_folder> |
| 143 | +``` |
| 144 | +If you wish to save only model weights, do not specify --save_traces_npy_dir. If you wish to save only traces do not specify --save_model_npy_dir. |
| 145 | + |
| 146 | +Code has been built upon https://github.com/yxlijun/S3FD.pytorch |
0 commit comments