We propose a new face verification method that uses multiple deep convolutional neural networks (DCNNs) and a deep ensemble, that extracts two types of low dimensional but discriminative and high-level abstracted features from each DCNN, then combines them as a descriptor for face verification. Our DCNNs are built from stacked multi-scale convolutional layer blocks to present multi-scale abstraction. To train our DCNNs, we use different resolutions of triplets that consist of reference images, positive images, and negative images, and triplet-based loss function that maximize the ratio of distances between negative pairs and positive pairs and minimize the absolute distances between positive face images. A deep ensemble is generated from features extracted by each DCNN, and used as a descriptor to train the joint Bayesian learning and its transfer learning method. On the LFW, although we use only 198,018 images and only four different types of networks, the proposed method with the joint Bayesian learning and its transfer learning method achieved 98.33% accuracy. In addition to further increase the accuracy, we combine the proposed method and high dimensional LBP based joint Bayesian method, and achieved 99.08% accuracy on the LFW. Therefore, the proposed method helps to improve the accuracy of face verification when training data is insufficient to train DCNNs.
Refer to papers for more details.
Figure 1. Overall procedure of the proposed method. To train deep neural network, we use triplets of faces. With triplets of faces, we train our DCNN with the proposed loss functions to obtain discriminative features. We also train 4 different DCNNs per different resolutions. After training, we extract features from each DCNN model. For test, given face images, these images are passed to multiple DCNNs and then we extract features from each DCNN models. With these extracted features, we classify whether these two face images are same or not using Joint Bayesian Classifier.
Figure 2. Hybridization with high-dimensional LBP based joint Bayesian method in the manner of the score-level fusion.
Table 1. Comparison of the number of DCNNs, the number of images, the dimensionality of feature, and the accuracy of the proposed method with the state-of-the-art on the LFW.
Download the executable binary file: FR_PIMNET_v2.0.zip
This research was supported by the MSIT (Ministry of Science, ICT), Korea, under the SW Starlab support program (IITP-2017-0-00897) supervised by the IITP (Institute for Information & communications Technology Promotion) and also supported by the MSIT, Korea, under the ICT Consilience Creative program (IITP-2017-R0346-16-1007) supervised by the IITP.