9
A Comparative Study of Face Re-identification Systems under Real- World Conditions Brown, G., Martinez del Rincon, J., & Miller, P. (2018). A Comparative Study of Face Re-identification Systems under Real-World Conditions. In Irish Machine Vision and Image Processing Conference Proceedings 2018: Proceedings (pp. 137-144). Irish Pattern Recognition & Classification Society. Published in: Irish Machine Vision and Image Processing Conference Proceedings 2018: Proceedings Document Version: Peer reviewed version Queen's University Belfast - Research Portal: Link to publication record in Queen's University Belfast Research Portal Publisher rights © 2018 The Authors and The Irish Pattern Recognition & Classification Society. This work is made available online in accordance with the publisher’s policies. Please refer to any applicable terms of use of the publisher. General rights Copyright for the publications made accessible via the Queen's University Belfast Research Portal is retained by the author(s) and / or other copyright owners and it is a condition of accessing these publications that users recognise and abide by the legal requirements associated with these rights. Take down policy The Research Portal is Queen's institutional repository that provides access to Queen's research output. Every effort has been made to ensure that content in the Research Portal does not infringe any person's rights, or applicable UK laws. If you discover content in the Research Portal that you believe breaches copyright or violates any law, please contact [email protected]. Download date:05. Apr. 2019

A Comparative Study of Face Re-identification Systems under … · 2019. 4. 5. · A Comparative Study of Face Re-identification Systems under Real-World Conditions. Authors Institution

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: A Comparative Study of Face Re-identification Systems under … · 2019. 4. 5. · A Comparative Study of Face Re-identification Systems under Real-World Conditions. Authors Institution

A Comparative Study of Face Re-identification Systems under Real-World Conditions

Brown, G., Martinez del Rincon, J., & Miller, P. (2018). A Comparative Study of Face Re-identification Systemsunder Real-World Conditions. In Irish Machine Vision and Image Processing Conference Proceedings 2018:Proceedings (pp. 137-144). Irish Pattern Recognition & Classification Society.

Published in:Irish Machine Vision and Image Processing Conference Proceedings 2018: Proceedings

Document Version:Peer reviewed version

Queen's University Belfast - Research Portal:Link to publication record in Queen's University Belfast Research Portal

Publisher rights© 2018 The Authors and The Irish Pattern Recognition & Classification Society.This work is made available online in accordance with the publisher’s policies. Please refer to any applicable terms of use of the publisher.

General rightsCopyright for the publications made accessible via the Queen's University Belfast Research Portal is retained by the author(s) and / or othercopyright owners and it is a condition of accessing these publications that users recognise and abide by the legal requirements associatedwith these rights.

Take down policyThe Research Portal is Queen's institutional repository that provides access to Queen's research output. Every effort has been made toensure that content in the Research Portal does not infringe any person's rights, or applicable UK laws. If you discover content in theResearch Portal that you believe breaches copyright or violates any law, please contact [email protected].

Download date:05. Apr. 2019

Page 2: A Comparative Study of Face Re-identification Systems under … · 2019. 4. 5. · A Comparative Study of Face Re-identification Systems under Real-World Conditions. Authors Institution

A Comparative Study of Face Re-identification Systemsunder Real-World Conditions.

Authors

Institution

May 2018

Abstract

Face re-identification is largely thought of as a solved problem in the research community, with State-of-the-art systems attaining human-level performance on unconstrained image datasets. However, these resultsdo not seem to translate to the real world. In systems matching people in surveillance-like footage to high-quality images, reported performance is much lower than what the literature would suggest. In this work, wecontribute a multi-modal dataset for evaluating real-world performance of facial re-identification systems.We then perform a verification and re-identification evaluation for state-of-art systems on both this datasetand the popular benchmarking dataset Labelled Faces in the Wild (LFW).

Keywords: Deep Learning, Person Re-identification, Face Recognition, Image-to-Video

1 Introduction

Face verification is a well-known problem with interesting applications such as biometrics, access control anduser authentication. Facial re-identification is an evolution of the previous problem in a much more complexsetting, where the identity of a user in the probe (which has never been seen before) should be paired against afull gallery of users.

The emergence of Deep Learning has revolutionised the computer vision field by bringing unprecedentedperformance. Deep Neural Networks (DNNs) may have different network architectures that vary in size, ef-ficiency, and performance, but the weights learned through training are defined by the dataset the networkis trained on. As a requirement, deep neural networks require large and complex datasets to be trained on,which are representative of the real world complexity. Face Re-identification systems are no different, and asubstantial part of the field’s recent success can be attributed to the availability of large-scale datasets to trainon, such as VGGFace [Parkhi et al., 2015], CelebFaces+ [Sun et al., 2014], CASIA-WebFace [Yi et al., 2014],Ms-Celeb-1M [Guo et al., 2016], and others.

Among the available datasets with facial imagery, Labelled Faces in the Wild (LFW) [Huang et al., 2007]is of particular interest due to its highly unconstrained nature, and the large number of images and people itcontains. While this dataset leads in complexity among the available standard facial datasets, it can be argued ifits complexity is enough to develop and evaluate systems truly ready for real-life problems. For instance, due tothe acquiring process, that dataset consists mostly of static images from high-quality sources. Thus, the classeswithin these datasets tend to be unimodal: Every image in a class has roughly similar properties to every otherimage in the same class, making the reidentification easier by not requiring domain adaptation.

This is not always desirable though since it does not always reflect real-world acquisition conditions. For ex-ample: when matching still-images of people facing towards the camera to frames from surveillance video, thetwo sets of images have different intrinsic properties, i.e. belong to different domains, besides the actual subject

Page 3: A Comparative Study of Face Re-identification Systems under … · 2019. 4. 5. · A Comparative Study of Face Re-identification Systems under Real-World Conditions. Authors Institution

in the image. This bias will severely impact the performance of models trained on standard datasets when ap-plied to realistic and complex scenarios. Dataset bias has been noted before in the literature [Klare et al., 2015]leading to researchers performing cross-dataset experiments [Cao et al., 2017, Bansal et al., 2017].

In order to prove the severity of this effect, in this work, we perform a comparative study of state-of-the-artdeep learning models for re-identification under a complex cross-dataset setting. All models are first trained onunconstrained still image datasets and then tested on a still-to-video dataset that we created and released to thecommunity. Doing this, we aim to show that using a one-size-fits-all dataset for evaluating systems may leadto less than expected performance in the real world.

2 State of the Art

Still-to-Still (S2S). Many approaches have successfully been applied for facial verification on unconstrainedimages. In [Schroff et al., 2015], the authors introduced the triplet-loss, a metric learning loss function thataims to embed images into Euclidian space with similar images pulled towards each other and dissimilar imagespushed apart. Similarly, VGG16 architecture [Parkhi et al., 2015] combines the triplet-loss with a CNN and alarge unconstrained dataset. This was further extended [Cao et al., 2017] with an even larger dataset, and anetwork based on Resnet50 [He et al., 2016] with Squeeze-and-Excitation blocks [Hu et al., 2017].

In spite of their good performance in standard sets, authors in [Klare et al., 2015] found that the perfor-mance of re-identification systems (specifically an unnamed Government-Off-The-Self system and OpenBR[Klontz et al., 2013]) was heavily dependant on the subject’s pose. Unconstrained frontal images can be recog-nised with near human-level performance, but images with full pose variation such as profile shots cannotreliably be detected or recognised.

Still to Video (S2V). Recently, research approaches have emerged to tackled Still-to-Video (S2V) re-identification, where still images are enrolled in a system and then video sequences are matched againstthem. These approaches have screening as an intended application, aiming to match mugshots in a watchlistwith CCTV footage of suspects. This task introduces extra domain adaptation challenges between camerasthat increases its complexity. A common approach is to use application-specific models that mitigate themulti-modality of the gallery and probe images. One such S2V model is the HaarNet [Parchami et al., 2017a],which consists of a trunk network which extracts features, with three branch networks based on Haar-likefeatures. Another model is the Canonical Face Representation CNN (CFR-CNN) [Parchami et al., 2017b],which uses a supervised auto-encoder network to generate canonical face representations. The performance ofthe CFR-CNN does not surpass that of the HaarNet, but it is comparable and is much less complex: requiring3 orders of magnitude fewer operations, with only a tenth of the parameters as the HaarNet.

3 MMF Dataset

In this section, we present an independent dataset called Muti-Modal Faces (MMF) that has been recordedfor evaluating applications such as watchlist screening or e-gates in airports. The dataset also aims to providemultiple modalities to better reflect the complexity of the real world.

Our dataset captures 77 subjects, consisting of 2 sets A and B. Set A consists of frontal images whereas SetB consists of images mined from surveillance-like videos. The cameras chosen at the time to record the videoand stills were simple off-the-shelf hardware and were not high quality.

3.1 Data Collection

The original data was collected in many sessions, over several days. Inside a rectangular room, a temporarydivision was put up along the centre. Three checkpoints were set up. Each checkpoint had a face camera

Page 4: A Comparative Study of Face Re-identification Systems under … · 2019. 4. 5. · A Comparative Study of Face Re-identification Systems under Real-World Conditions. Authors Institution

recording still images, and a body camera recording video overlooking it. Figure 1 shows the layout with facecameras labelled A and full body cameras labelled B.

Figure 1: Room Layout - Not to scaleEach subject would enter from the door and proceed to Checkpoint C1. Once there, they were to pause

and look straight into the appropriate face camera. After a few moments, they were to proceed to the nextcheckpoint and repeat the process. Once all three checkpoints were visited, the subject left the room.

3.2 Data Processing

To extract faces from the raw data, we adopted an approach similar to LFW, by using a Viola-Jones detector[Viola and Jones, 2001]. Both the still images and videos had face detection and extraction applied to them.No alignment was done at this stage. The videos were fed into the detector frame-by-frame, with the outputforming set B. The Viola-Jones detector was used to make the resulting dataset more comparable to LFW. Theimages were not resized after extraction. Images were excluded from the sets based on the following criteria:1) If an image does not include a face. 2) If the image was cropped in such a way that the face is not at thecentre of the image. 3) If the image was in set A and the subject was not facing the camera. Sample imagesfrom each camera are shown in Figure 2.

The final dataset contains 77 identities, with 60,924 images: 15,202 in Set A and 45,722 in Set B. Figure 3shows both the number of images in each class per set and the number of images from each camera.

Dataset Variance: MMF differs from other unconstrained datasets in that its intra-class similarity is multi-modal: that is, each class contains images with intrinsically different properties, such that the images formmultiple natural clusters around these properties. These are shown in Figure 4. Figure 4 (left) shows that theaverage intra-class similarity for LFW over all identities is largely unimodal, whereas Figure 4 (centre) showsthat the MMF average intra-class similarity has distinctive peaks and modalities, with clear separation betweenSet A and Set B. Moreover, when observing individual classes distribution, the muti-modalities due to the threecheckpoints can be more easily observed.

4 Methods

Two main state-of-art systems were tested for face re-identification and face verification: a ResNet50[He et al., 2016] based DNN and the VGG16 [Parkhi et al., 2015] DNN.

The ResNet50 based network was pre-trained on the VGGFace2 dataset [Cao et al., 2017] using the soft-max loss function. As the network was pre-trained on a classification problem, we remove the final layer inorder to get image embeddings as output. We chose this network because it is State-of-the-Art with regards tounconstrained facial re-identification.

Page 5: A Comparative Study of Face Re-identification Systems under … · 2019. 4. 5. · A Comparative Study of Face Re-identification Systems under Real-World Conditions. Authors Institution

(a) Camera A1 (b) Camera A2 (c) Camera A3

(d) Camera B1 (e) Camera B2 (f) Camera B3

Figure 2: Sample images captured by each camera for Identity 1

Figure 3: MMF Dataset Summary Statistics

We also compare with the successful VGG16 architecture trained on the original VGGFace dataset[Parkhi et al., 2015] using the triplet-loss function [Schroff et al., 2015]. Like with the ResNet50 networkabove, we remove the final layer to get out feature vectors. We included this network in our comparisonbecause it is a shorter, simpler network while still attaining close to State-of-the-Art performance.

All models are executed in Keras [Chollet et al., 2015] with the TensorFlow backend [Abadi et al., 2015].All pre-trained weights were downloaded from VGG’s website and converted into Keras / Tensorflow models.

Face alignment (FA) has been proved [Bansal et al., 2017] as crucial for a correct identification, particularlyin unconstrained environments. With the aim to validate this in the context of our scenario, all tested method-ologies were evaluated with and without it. It is provided by DLIB [King, 2009] and OpenCV [Bradski, 2000].DLIB is used to detect the positions of facial landmarks, namely the eyes. Knowing their location and the rela-tive angle between them, OpenCV’s warpAffine function was used to place the eyes into standard positions,rotating the image as necessary so the eyes are horizontal relative to each other. Since we are explicitly testingperformance without FA, we are using the original un-aligned LFW dataset.

5 Experiments and Results

Two types of experiments were performed in this evaluation, Face Verification and Face Re-Identification. Weaim for a comparison between LFW and our own dataset, as well as the methods themselves in both scenarios.

Page 6: A Comparative Study of Face Re-identification Systems under … · 2019. 4. 5. · A Comparative Study of Face Re-identification Systems under Real-World Conditions. Authors Institution

Figure 4: Average Intra-Class Modalities for LFW (left) and MMF (centre), and intra-class modality for asingle MMF class (right).

As specified in Section 4, all training was done on either VGGFace or VGGFace2 depending on the network.LFW and MMF are used for evaluation only, being this validation a cross-dataset setting.

5.1 Experimental Setup

5.1.1 Verification

Face verification consists of taking two images, and determining if both are of the same person. Standardpractice for evaluation is to create a set of labelled pairs; pass them into face verification system which producessome distance measurement -Euclidean distance in our case-; which is then thresholded by a value τ to producepositive or negative labels. In our work, τ is found per fold per experiment by performing a grid-search,optimising for verification accuracy. Since τ is obtained from our test datasets, we do not report the best valuefor τ, nor the derived verification accuracy.

LFW. For LFW the standard "Unrestricted with Outside Data" protocol will be used [Learned-Miller, 2014].This consists of 10 sets of 600 test pairs, provided by the authors of LFW. The pairs of images in each fold (afold consisting of 9 of the pair sets, with a different holdout set each time) are then passed through the systemwhich predicts if both images contain the same person. The ROC Curve is then generated for each of the 10folds. The final result is the average ROC curve and averaged accuracy over the folds.

MMF. For MMF, we mirror the protocol used by LFW as much as is practical. In our case, we generate 4620pairs with 50% positive matches and 50% negative matches (same as LFW). Each image in Set A is matcheduniformly to two randomly selected images in Set B: one of the same class, and the other of a different class.The rest of the protocol follows LFW, with the final results averaged over 10 folds.

5.1.2 Re-identification

Re-identification is where one matches a single probe image to a gallery of images where neither the probe norgallery images were in the original training set. We will evaluate performance with the Cumulative MatchingCharacteristic (CMC) Curve. Re-identification naturally gets more difficult the larger the gallery: a systemusing purely random chance has probability 0.5 matching a probe to a gallery of 2, and probability 0.001matching to a gallery of 1000. To mitigate this, we make the gallery size of both LFW and MMF identical.

LFW. To construct galleries and probes for LFW, only the classes with 2 or more images are selected (ofwhich there are 1680). Then, we divide the classes into groups so each group contains 77 classes (with the lastgroup taking the remainder). This results in 22 class groups. We then construct a probe and gallery for eachgroup. Galleries are constructed by randomly selecting an image from each class, with probes then containingthe images not used in the gallery. The CMC scores are then computed for each gallery-probe pair, with thefinal result being the mean of all 22 CMC curves.

Page 7: A Comparative Study of Face Re-identification Systems under … · 2019. 4. 5. · A Comparative Study of Face Re-identification Systems under Real-World Conditions. Authors Institution

MMF. Given the size of the dataset and that our probe and gallery will be made from disjoint images Set Aand Set B, we construct the MMF probe and gallery differently to LFW. Iterating over each class, we randomlyselect 15 images from the class in Set A, and 30 images from the class in Set B. We then construct 15 gallerieswith the images from Set A (every class represented in each gallery), and combine the all of the images fromSet B into a single probe.

5.2 Results

Verification. The resulting ROC curve comparing the difficulty of the verification problem on LFW andMMF can be seen in Figure 5: the Area Under Curve in square brackets in the legend). The best results wereall achieved on LFW, regardless of model or whether FA was used. This shows the more complex conditionsencompassed in our dataset, which is more obvious if no FA is applied. Regarding methods, ResNet50 performsbetter and seems better generalisation properties than VGG16 in this cross-dataset setting.

Figure 5: ROC Curve comparing LFW and MMF Datasets

Re-identification. Figure 6 shows the resulting CMC curves for the re-identification experiments, with Table1 showing the CMC scores for ranks 1, 5, and 20. In this more complex setting, MMF still exhibits morecomplex conditions than LFW, showing even bigger performance drops, in like-for-like examples (same modeland using FA or not). The only exception is for VGG16 without FA. However, since FA is not applied, whichappears to be a requirement for good re-identification, the results are inherently unreliable. Regarding methods,although ResNet50 is clearly better in LFW, a dataset closer to the VGG training sets, VGG16 obtains betterresults on MMF than ResNet50. In spite of VGG16 containing more parameters (138,357,544 vs 25,636,712for ResNet50) due to the bottleneck structure of ResNet50, VGG16 is a shorter network, which means it mayexhibit smaller levels of overfitting for this complex setting.

6 Conclusions

In this work, we contributed a dataset to evaluate facial re-identification systems. By comparing our datasetMMF to a widely used, unconstrained, facial dataset LFW, we show that State-of-the-Art systems are still notfully robust, and give a much poorer performance when used on real-world still-to-video data. We also show

Page 8: A Comparative Study of Face Re-identification Systems under … · 2019. 4. 5. · A Comparative Study of Face Re-identification Systems under Real-World Conditions. Authors Institution

Figure 6: CMC Curve comparing LFW and MFW DatasetsExperiment Rank 1 Rank 5 Rank 20ResN50 LFW 0.572 0.790 0.933ResN50 MMF 0.068 0.287 0.628ResN50 LFW+FA 0.946 0.969 0.978ResN50 MMF+FA 0.573 0.786 0.878VGG16 LFW 0.669 0.856 0.955VGG16 MMF 0.732 0.909 0.996VGG16 LFW+FA 0.862 0.935 0.964VGG16 MMF+FA 0.596 0.839 0.986

Table 1: Resulting CMC Scores for each experiment ar Rank 1, 5, and 20.

that face alignment gives a large boost to current systems. Furthermore, MMF complexity and the use ofcross-dataset experimentation may unravel relevant conclusions such as overfitting of current architectures.

Regarding possible future work, for real-world performance of facial re-identification systems to improve,application specific datasets need to be created which are large enough to facilitate training: either entirely, orwith fine-tuning. Until such datasets are available, via either harvesting or data augmentation, the field willstruggle to implement robust solutions that reliably work in real-world applications.

References

[Abadi et al., 2015] Abadi, M., Agarwal, A., Barham, P., Brevdo, E., Chen, Z., Citro, C., Corrado, G. S., Davis,A., Dean, J., Devin, M., Ghemawat, S., Goodfellow, I., Harp, A., Irving, G., Isard, M., Jia, Y., Jozefowicz,R., Kaiser, L., Kudlur, M., Levenberg, J., Mané, D., Monga, R., Moore, S., Murray, D., Olah, C., Schuster,M., Shlens, J., Steiner, B., Sutskever, I., Talwar, K., Tucker, P., Vanhoucke, V., Vasudevan, V., Viégas, F.,Vinyals, O., Warden, P., Wattenberg, M., Wicke, M., Yu, Y., and Zheng, X. (2015). TensorFlow: Large-scalemachine learning on heterogeneous systems. Software available from tensorflow.org.

[Bansal et al., 2017] Bansal, A., Castillo, C., Ranjan, R., and Chellappa, R. (2017). The do’s and don’tsfor cnn-based face verification. In 2017 IEEE International Conference on Computer Vision Workshops(ICCVW), pages 2545–2554.

Page 9: A Comparative Study of Face Re-identification Systems under … · 2019. 4. 5. · A Comparative Study of Face Re-identification Systems under Real-World Conditions. Authors Institution

[Bradski, 2000] Bradski, G. (2000). The OpenCV Library. Dr. Dobb’s Journal of Software Tools.

[Cao et al., 2017] Cao, Q., Shen, L., Xie, W., Parkhi, O. M., and Zisserman, A. (2017). Vggface2: A datasetfor recognising faces across pose and age. CoRR, abs/1710.08092.

[Chollet et al., 2015] Chollet, F. et al. (2015). Keras. https://keras.io.

[Guo et al., 2016] Guo, Y., Zhang, L., Hu, Y., He, X., and Gao, J. (2016). Ms-celeb-1m: A dataset andbenchmark for large-scale face recognition. CoRR, abs/1607.08221.

[He et al., 2016] He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep residual learning for image recognition.In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 770–778.

[Hu et al., 2017] Hu, J., Shen, L., and Sun, G. (2017). Squeeze-and-excitation networks. CoRR,abs/1709.01507.

[Huang et al., 2007] Huang, G. B., Ramesh, M., Berg, T., and Learned-Miller, E. (2007). Labeled faces inthe wild: A database for studying face recognition in unconstrained environments. Technical Report 07-49,University of Massachusetts, Amherst.

[King, 2009] King, D. E. (2009). Dlib-ml: A machine learning toolkit. Journal of Machine Learning Research,10:1755–1758.

[Klare et al., 2015] Klare, B. F., Klein, B., Taborsky, E., Blanton, A., Cheney, J., Allen, K., Grother, P., Mah,A., Burge, M., and Jain, A. K. (2015). Pushing the frontiers of unconstrained face detection and recognition:Iarpa janus benchmark a. In 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),pages 1931–1939.

[Klontz et al., 2013] Klontz, J. C., Klare, B. F., Klum, S., Jain, A. K., and Burge, M. J. (2013). Open sourcebiometric recognition. In 2013 IEEE Sixth International Conference on Biometrics: Theory, Applicationsand Systems (BTAS), pages 1–8.

[Learned-Miller, 2014] Learned-Miller, G. B. H. E. (2014). Labeled faces in the wild: Updates and newreporting procedures. Technical Report UM-CS-2014-003, University of Massachusetts, Amherst.

[Parchami et al., 2017a] Parchami, M., Bashbaghi, S., and Granger, E. (2017a). Video-based face recognitionusing ensemble of haar-like deep convolutional neural networks. In 2017 International Joint Conference onNeural Networks (IJCNN), pages 4625–4632.

[Parchami et al., 2017b] Parchami, M., Bashbaghi, S., Granger, E., and Sayed, S. (2017b). Using deep au-toencoders to learn robust domain-invariant representations for still-to-video face recognition. In 2017 14thIEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pages 1–6.

[Parkhi et al., 2015] Parkhi, O. M., Vedaldi, A., and Zisserman, A. (2015). Deep face recognition. In BritishMachine Vision Conference.

[Schroff et al., 2015] Schroff, F., Kalenichenko, D., and Philbin, J. (2015). Facenet: A unified embedding forface recognition and clustering. In 2015 IEEE Conference on Computer Vision and Pattern Recognition(CVPR), pages 815–823.

[Sun et al., 2014] Sun, Y., Wang, X., and Tang, X. (2014). Deep learning face representation from predicting10,000 classes. In 2014 IEEE Conference on Computer Vision and Pattern Recognition, pages 1891–1898.

[Viola and Jones, 2001] Viola, P. and Jones, M. (2001). Robust real-time face detection. In Proceedings EighthIEEE International Conference on Computer Vision. ICCV 2001, volume 2, pages 747–747.

[Yi et al., 2014] Yi, D., Lei, Z., Liao, S., and Li, S. Z. (2014). Learning face representation from scratch.CoRR, abs/1411.7923.