Heterogeneous Face Recognition by Margin-Based Cross-Modality Metric Learning

Abstract
Heterogeneous face recognition deals with matching face images from different modalities or sources. The main challenge lies in cross-modal differences and variations and the goal is to make cross-modality separation among subjects. A margin-based cross-modality metric learning ((MCML)-L-2) method is proposed to address the problem. A cross-modality metric is defined in a common subspace where samples of two different modalities are mapped and measured. The objective is to learn such metrics that satisfy the following two constraints. The first minimizes pairwise, intrapersonal cross-modality distances. The second forces a margin between subject specific intrapersonal and interpersonal cross-modality distances. This is achieved by defining a hinge loss on triplet-based distance constraints for efficient optimization. It allows the proposed method to focus more on optimizing distances of those subjects whose intrapersonal and interpersonal distances are hard to separate. The proposed method is further extended to a kernelized (MCML)-L-2 ((KMCML)-L-2). Both methods have been evaluated on an ID card face dataset and two other cross-modality benchmark datasets. Various feature extraction methods have also been incorporated in the study, including recent deep learned features. In extensive experiments and comparisons with the state-of-the-art methods, the (MCML)-L-2 and (KMCML)-L-2 methods achieved marked improvements in most cases.
Funding Information
  • National Science Foundation of China (61432008, 61673203, 61603193)
  • Young Elite Scientists Sponsorship Program by CAST (YESS 20160035)
  • Collaborative Innovation Center of Novel Software Technology and Industrialization
  • China Scholarship Council through the University of Manchester

This publication has 37 references indexed in Scilit: