Ecludian Distance based Partial Face Recognition

DOI : 10.17577/IJERTV6IS070108

Download Full-Text PDF Cite this Publication

Text Only Version

Ecludian Distance based Partial Face Recognition

Aher Pravin D. ME(VLSI and Embedded Systems) Pravara Rural Engg. College, loni

Ahmednagar, India

Galande S.G.

Associate Professor (E&TC Department) Pravara Rural Engg. College, loni Ahmednagar, India

Abstract Various approaches have been proposed to recognize complete faces. However, few have dealt with the problem of recognizing an arbitrary partial face. In real-world scenarios, human faces might easily be occluded by other objects. This made traditional face recognition algorithms, which heavily rely on face alignment and face normalization, infeasible. In this paper, we propose an approach for partial face recognition which will be based on matching feature set, and this approach will be able to match partial face patches to gallery data set faces automatically and is robust to occlusions as well as illumination changes.Then, we propose arobust point set matching method to discriminatively match thesetwo extracted local feature sets, where both the textural informationand geometrical information of local features are explicitlyused for matching simultaneously. Finally, the similarity of two faces is been extracted and distance is been calculated based on Euclidian distance formula.

Keywords Face recognition, partial face recognition,feature set matching, feature alignment, image matching,biometrics.

  1. INTRODUCTION

    Face recognition is a technique of identifying an individual by comparing live image or digital image data with the stored record for that person. A face recognition system is a computer application capable of identifying or verifying a person from a digital image or a video frame from a video source. One of the ways to do this is by comparing selected facial features from the image and a face database. It is typically used in security systems and is similar to other techniques such as fingerprint or eye iris recognition systems. Recently, it has also become popular as a commercial identification and marketing tool.

    FACE recognition (FR) is the problem of verifying or identifying a face from its image. It has received substantial attention over the last three decades due to its value both in understanding how FR process works in humans as well as in addressing many challenging real-world applications, including reduplication of identity documents (e.g. passport, driver license), access control and video surveillance. The performance of automatic FR systems has advanced significantly. While face recognition in controlled conditions (frontal face of cooperative users and controlled indoor illumination) has already achieved impressive performance over large-scale galleries, as indicated in a recent IEEE T- PAMI special issue on real-word face recognition [1], there still exist many challenges for face recognition in uncontrolled environments, such as partial occlusions, large pose variations, and extreme ambient illumination.

    Typical applications of face recognition in uncontrolled environments include recognition of individuals in video surveillance frames and images captured by handheld devices (e.g. mobile phones), where a face may be captured in arbitrary pose without user cooperation and knowledge. In such scenarios, it is quite likely that the captured image contains only a partial face. Table 1 lists a categorization of partial face images and some further illustrations are given in We call the resulting problem a Partial Face Recognition (PFR) problem, so as to differentiate it from the holistic face recognition problem. Commercial off-the-shelf (COTS) face recognition systems are not able to handle the general PFR problem since they need to align faces by facial landmarks that may be occluded. For example, Face VACS [3] requires localization of the two eyes, and PittPatt [4] detects several predefined landmarks for facealignment. Therefore, research in PFR is important to advance the state of the art in face recognition and enlarge the application domain.

    This paper is an extended version of our previous workpresented at IEEE ICCV 2013 [15]. There are several newcontributions in this work compared to its conference version:

    • We have developed a new feature set matching approachfor partial face matching. In our previous conferenceversion [15], the matching algorithm was built based onChuis work [16], where no constraint was enforced onthe affine transformation matrix, so that unrealistic imagewarping can be generated if the difference between theprobe patch and gallery image is large. In this work,we explicitly constrain the affine matrix to address thislimitation. Experimental results show that our new featureset matching method achieves better performance.

    • We have conducted more partial face recognition experimentsto further evaluate the performance of ourapproach. The newly extensions include:

      1. More resultson additional datasets,

      2. More face verification evaluations,and

      3. More detailed parameter analysis of theproposed approach.

        Feature set matching [7] has been a hot topic in pattern recognition. [24] was the first work that used graph matching for face recognition. However, their work relies heavily on manual landmarks labeling. Chui and Rangarajan [6] presented Robust Point set Matching (RPM) to align two feature sets according to their geometry distribution by learning a non-affine transformation function through iterative updates. However, it neglects textural information offeature points. Liao et al. [15] utilized SRC to reconstruct probe local feature set with gallery feature sets, and they used the reconstruction error as distance metric. The main

        drawback of their method is that they neglected the geometry information of feature sets and their approach is computationally intensive. To address the partial face recognition problem, we propose a new partial face recognition approach by using feature set matching, and devise a Metric Learned Extended Robust Point set Matching (MLERPM) approach to register the extracted local features according to their geometric distribution and textural information. Based on the matching result, a point set distance metric is proposed to describe the similarity of two faces.

  2. PROBLEM DEFINITION

    Most existing approaches exploit handcrafted blur features that are optimized for a certain uniform blur across the image, which is unrealistic in a real blind de-convolution setting, where the blur type is often unknown. To deal with this issue, we aim at identifying the blur type for each input image patch, and then estimating the kernel parameter in this paper. A learning-based method using a pre-trained deep neural network (DNN) and a general regression neural network (GRNN) is proposed to first classify the blur type and then estimate its parameters, taking advantages of both the classification ability of DNN and the regression ability of GRNN.

  3. PROPOSED APPROACH

    We propose to use local features instead of holistic featuresfor partial face representation. Specifically, we applythe Scale-Invariant Feature Transform (SIFT) [17] featuredetector to detect local feature keypoints, which are then concatenated with the Speeded Up Robust Features (SURF)[2]. Before matching, keypoints selection is performed to filter out obvious outliers. These selected keypoints of probe and gallery images are then matched by our MLERPMbased on their geometric distribution and textural information, through which we obtain a one-to-one point set correspondencematrix to indicate the genuine matching pairs, as well as non-affine transformation function to registergeometric distributions of these matched keypoints. With matched keypoint pairs at hand, we design a point set distancemetric to describe the difference between two facesbased on MLERPM, where the lowest matching distanceachieved would be reckoned as positive match. The facematching process is illustrated in Figure1. Throughout therest of the paper, matrix transposition is denoted by.

    Gig .3.1 Flow Chart of proposed system

    Following is the process flow for designing the proposed approach:

      1. Input Image:

        System can give input image by selecting from system/dataset or system can also take image input from real time camera. Any real time partial image can be taken as input from camera or from dataset.

      2. Image Preprocessing:

        In image preprocessing system will preprocess the input image, extract features of image and convert the image in grayscale image using grayscale algorithm.

        All grayscale algorithms utilize the same basic three-step process:

        1. Get the red, green, and blue values of a pixel

        2. Use fancy math to turn those numbers into a single gray value.

        3. Replace the original red, green, and blue values with the new gray value.

        When describing grayscale algorithms, Im going to focus on step 2 using math to turn color values into a grayscale value. So, when you see a formula like this:

        Gray = (Red + Green + Blue) / 3

      3. HAAR Features Detection:

        HAAR-like features are digital image features used in

        Learned Robust Point Matching After feature extraction and keypoints selection, for the probe partial face image, its

        object recognition. They owe their name to their intuitive

        geometry feature set is {gP

        , gP , …gP

        }, with its

        1 2 NP

        similarity with HAAR wavelets and were used in the first

        correspondent texture feature set as {tP , tP , …t P

        }, where

        1 2 NP

        real-time face detector. Historically, working with only image intensities (i.e., the RGB pixel values at each and every pixel of image) made the task of feature calculation computationally expensive. A publication by Papageorgiou et al.[2] discussed working with an alternate feature set based on HAAR wavelets instead of the usual image intensities. Viola and Jones[1] adapted the idea of using HAAR wavelets and developed the so-called HAAR -like features. A HAAR – like feature considers adjacent rectangular regions at a specific location in a detection window, sums up the pixel intensities in each region and calculates the difference between these sums. This difference is then used to categorize subsections of an image. For example, let us say we have an image database with human faces. It is a common observation that among all faces the region of the eyes is darker than the region of the cheeks. Therefore a common HAAR feature for face detection is a set of two adjacent rectangles that lie above the eye and the cheek region. The position of these rectangles is defined relative to a detection window that acts like a bounding box to the target object (the face in this case).

      4. Feature Extraction

        Since there exist rotation, translation, scaling and evenocclusions between probe image and gallery images ofsame identity, it is very difficult to normalize them to eyepositions. Without proper face alignment, holistic featureswould fail to work. Hence, we proposed to use local features. Firstly, we detect keypoints with SIFT feature detector.Normally for a typical 128 × 128 face image, SIFT featuredetector could output hundreds of feature points. Thegeometric feature of each keypoint, denoted as g, recordsits relative position in the image frame.To describe the texture features of these detected keypoints, we combined the strength of SIFT and SURF keypoint descriptor by simple concatenation. SURF keypoint

        descriptor was introduced as a complement to SIFT forits greater robustness against illumination variations [14].Hence, this augmented texture feature, denoted as t, is robustagainst in-plane rotation, scale as well as illumination change.

      5. Keypoint Selection

        As we have indicated previously, the number of keypoints of facial image could be up to hundreds. Matchingpoint sets at this scale is computationally intensive. Moreover, irrelevant keypoints might hamper point set matchingprocess, such as misleading the matching process to a localminimum, especially when genuine matching pairs are fewamong all matching features. Hence, its beneficial to filterout obvious outliers before point matching.We applied the idea of Lowes matching scheme [17] for keypoint selection, which is to compare the ratio of distance of the closest neighbor to the one of the second-closest neighbor to a predefined threshold. The threshold was setas 0.5 in our experiments. These coarsely matched keypoint. Pairs are then selected for our MLERPM for finer matching. 2.3. Metric

        NP is the number of keypoints in probe feature set. Similarly, for the gallery image, we have {gG1 , gG2 , …gGNG} and {tG1 , tG2 , …tGNG} correspondingly. To align a probe partial face image to a gallery image automatically, we need match their correspondent geometric features and textural features respectively, which should have three characteristics:

        • Subset matching: since the probe image and gallery images are not identical, some keypoints in the probe image couldnt find their correspondences in the gallery image. Likewise, not all keypoints in gallery images are ensured to be matched. Hence, this point set matching is a subset point matching problem.

        • One-to-one point correspondence: this trait is obvious as keypoints of different positions in the probe image shouldnt be matched to a single keypoint in the gallery image.

        • Non-affine transformation: the appearance of face changes when the perspective or facial expression changes. Such changes, when projected into the 2D image, are non-affine.

      6. Euclidian Distance:

    In mathematics, the Euclidean distance or Euclidean metric is the "ordinary" straight-line distance between two points in Euclidean space. With this distance, Euclidean space becomes a metric space. The associated norm is called the Euclidean norm. Older literature refers to the metric as Pythagorean metric. A generalized term for the Euclidean norm is the L2 norm or L2 distance.In the context of Euclidean geometry, a metric is established in one dimension by fixing two points on a line, and choosing one to be the origin. The length of the line segment between these points defines the unit of distance and the direction from the origin to the second point is defined as the positive direction. This line segment may be translated along the line to build longer segments whose lengths correspond to multiples of the unit distance. In this manner real numbers can be associated to points on the line (as the distance from the origin to the point) and these are the Cartesian coordinates of the points on what may now be called the real line. As an alternate way to establish the metric, instead of choosing two points on the line, choose one point to be the origin, a unit of length and a direction along the line to call positive. The second point is then uniquely determined as the point on the line that is at a distance of one positive unit from the origin.

    Fig: 1 Our proposed partial face recognition framework. (a) Feature extraction: keypoints detected by SIFT keypoint detector are marked out as greendots on both images. The left image is the probe partial face image, and the right one is the gallery face image. (b) Keypoint selection by Lowes matching scheme: roughly matched keypoints of these two images are connected by green lines, while two pairs of imposter matches are linked by red lines. (c) RPSMprocedure: point set of probe image marked out as blue diamonds are iteratively aligned to the red-marked point et of gallery image.

    During the trust region shrinkage process, there might occur a scenario where all elements in a row (column) of M are close to 0, this happens when an outlier has been detected. Note that the keypoint selection part doesnt guarantee the selected keypoint pairs are genuine matching pairs, some of which are imposter pairs. The geometric distribution of these imposter pairs lies inharmoniously against the rest. Detecting and removing them from the matching process not only accelerates the matching process, but also improve matching accuracy.

    The Scale-invariant Feature Transform This section reviews the basics of the SIFT algorithm, which according to [4] consists of four computational stages: (i) scale-space extrema detection, (ii) removal of unreliable keypoints, (iii) orientation assignment, and (iv)keypoint descriptor calculation. 2.1 Scale-space extrema detection. In the first stage, interest points called keypoints, are identified in the scale space by looking for image locations that represent maxima or minima of the difference-of-Gaussian function.

  4. CONCLUSION

We have proposed a partial face recognition method by using robust feature set matching and Euclidian distance method. The proposed RPSM method is able to align the probe partial face to gallery facial images robustly even with the presence of occlusion, random partial crop, and exaggerated facial expressions. After face alignment, partial face recognition is achieved by measuring face similarity based on the proposed point set distance,which can be readily acquired with the face alignment result.The hallmark of the RPSM is its robust matching scheme,which considers both the geometric distribution consistency and the textural similarity.

REFERENCES

  1. M. S. Bartlett, J. R. Movellan, and T. J. Sejnowski. Face recognition by independent component analysis. Neural Networks, IEEE Transactions on, 13(6):14501464, 2002.

  2. H. Bay, T. Tuytelaars, and L. Van Gool. Surf: Speeded up robust features. ECCV, pages 404417, 2006.

  3. P. N. Belhumeur, J. P. Hespanha, and D. J. Kriegman. Eigenfaces vs. fisherfaces: Recognition using class specific linear projection. PAMI, 19(7):711720, 1997.

  4. F. L. Bookstein. Principal warps: Thin-plate splines and the decomposition of deformations. PAMI, 11(6):567585, 1989.

  5. W. Chen and Y. Gao. Recognizing partially occluded faces from a single sample per class using string-based matching. In ECCV, pages 496509. 2010.

  6. H. Chui and A. Rangarajan.A new point matching algorithm for non-rigid registration. Computer Vision and Image Understanding, 89(2):114141, 2003. 1, 3, 7

  7. D. Conte, P. Foggia, C. Sansone, and M. Vento. Thirty years of graph matching in pattern recognition. PAMI, 18(03):265298, 2004. 1

  8. T. F. Cootes, G. J. Edwards, and C. J. Taylor.Active appearance models. PAMI, 23(6):681685, 2001. 1

  9. T. F. Cootes, C. J. Taylor, D. H. Cooper, J. Graham, et al. Active shape models-their training and application. Computer vision and image understanding, 61(1):3859, 1995.

  10. P. Dreuw, P. Steingrube, H. Hanselmann, H. Ney, and G. Aachen,SURF-face: Face recognition under viewpoint consistency constraints,in Proc. BMVC, 2009, pp. 111.

  11. G. B. Huang, M. Ramesh, T. Berg, and E. Learned-Miller, Labeledfaces in the wild: A database for studying face recognition inunconstrained environments, Dept. Comput. Sci., Univ. MassachusettsAmherst, Amherst, MA, USA, Tech. Rep. 07-49, 2007.

  12. X. Xiong and F. de la Torre, Supervised descent method andits applications to face alignment, in Proc. CVPR, Jun. 2013,pp. 532539.

  13. X. Cao, Y. Wei, F. Wen, and J. Sun, Face alignment by explicit shaperegression, in Proc. CVPR, Jun. 2012, pp. 2887 2894.

  14. X. Zhu and D. Ramanan, Face detection, pose estimation, and landmarklocalization in the wild, in Proc. CVPR, Jun. 2012, pp. 28792886.

  15. R. Weng, J. Lu, J. Hu, G. Yang, and Y.-P. Tan, Robust feature setmatching for partial face recognition, in Proc. ICCV, Dec. 2013,pp. 601608.

  16. H. Chui and A. Rangarajan, A new point matching algorithm for nonrigidregistration, Comput. Vis. Image Understand., vol. 89, nos. 23,pp. 114141, Feb. 2003.

  17. J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma, Robust facerecognition via sparse representation, IEEE Trans. Pattern Anal. Mach.Intell., vol. 31, no. 2, pp. 210227, Feb. 2009.

  18. L. Zhang, M. Yang, and X. Feng, Sparse representation or collaborativerepresentation: Which helps face recognition? in Proc. ICCV,Nov. 2011, pp. 471478.

  19. E. Elhamifar and R. Vidal, Robust classification using structured sparserepresentation, in Proc. CVPR, Jun. 2011, pp. 18731879.

  20. Y. Liu, F. Wu, Z. Zhang, Y. Zhuang, and S. Yan, Sparse representationusing nonnegative curds and whey, in Proc. CVPR, Jun. 2010,pp. 35783585.

  21. K. Jia, T.-H.Chan, and Y. Ma, Robust and practical face recognitionvia structured sparsity, in Proc. ECCV, 2012, pp. 331344.

  22. W. Chen and Y. Gao, Recognizing partially occluded faces from asingle sample per class using string-based matching, in Proc. ECCV,2010, pp. 496509.

  23. R. Min, A. Hadid, and J. Dugelay, Improving the recognition of facesoccluded by facial accessories, in Proc. IEEE Int. Conf. FG, Mar. 2011,pp. 442447.

  24. K. Pan, S. Liao, Z. Zhang, S. Z. Li, and P. Zhang, Part-based facerecognition using near infrared images, in Proc. CVPR, Jun. 2007,pp. 16.

  25. Y. Rubner, C. Tomasi, and L. J. Guibas, The earth movers distanceas a metric for image retrieval, Int. J. Comput. Vis., vol. 40, no. 2,pp. 99121, Nov. 2000.

  26. Y. Tang, R. Salakhutdinov, and G. Hinton, Robust Boltzmannmachines for recognition and denoising, in Proc. CVPR, Jun. 2012,pp. 22642271.

Leave a Reply