Regular Articles

Weighted joint sparse representation-based classification method for robust alignment-free face recognition

[+] Author Affiliations
Bo Sun

Beijing Normal University, College of Information Science and Technology, Xinjiekouwai Street No. 19, Beijing 100875, China

Feng Xu

Beijing Normal University, College of Information Science and Technology, Xinjiekouwai Street No. 19, Beijing 100875, China

Guoyan Zhou

Beijing Normal University, College of Information Science and Technology, Xinjiekouwai Street No. 19, Beijing 100875, China

Jun He

Beijing Normal University, College of Information Science and Technology, Xinjiekouwai Street No. 19, Beijing 100875, China

Fengxiang Ge

Beijing Normal University, College of Information Science and Technology, Xinjiekouwai Street No. 19, Beijing 100875, China

J. Electron. Imaging. 24(1), 013018 (Jan 26, 2015). doi:10.1117/1.JEI.24.1.013018
History: Received September 11, 2014; Accepted December 29, 2014
Text Size: A A A

Open Access Open Access

Abstract.  This work proposes a weighted joint sparse representation (WJSR)-based classification method for robust alignment-free face recognition, in which an image is represented by a set of scale-invariant feature transform descriptors. The proposed method considers the correlation and the reliability of the query descriptors. The reliability is measured by the similarity information between the query descriptors and the atoms in the dictionary, which is incorporated into the l0l2-norm minimization to seek the optimal WJSR. Compared with the related state-of-art methods, the performance is advanced, as verified by the experiments on the benchmark face databases.

Figures in this Article

Face recognition has received substantial attention for a long time. Many typical methods have been proposed to perform face recognition.14 Since Wright et al. presented the sparse representation-based classification (SRC) method,5 it has been widely studied in many pattern recognition applications due to its promising results, such as face recognition,6,7 along with gender,8 digit,9,10 biology data,11,12 and medical image13,14 classification. Although many improved SRC-based methods have been proposed for robust face recognition,1519 most of them require rigid image alignment, where all images of an object or objects of interest are aligned to a fixed canonical template. Until now, much work has been performed to address the alignment problem.20,21 However, such alignment is still difficult to achieve in real scenarios, such as partial face, scale, or pose variation face recognition. To address the alignment problem in SRC, some methods22,23,24 introduced the scale-invariant feature transform (SIFT)22 or the speeded-up robust features25 descriptor to the recognition method. However, most of these methods pay little attention to the correlation among the query descriptors, which is found to be useful for classification. Thus, it is necessary to study a method exploiting the correlation of the query descriptors for robust alignment-free face recognition, which is the focus of this paper.

Supposing that one image is represented with a set of SIFT descriptors22 which are robust to handle scale variations and rotation, the SIFT-based method can solve the problem of alignment. The simple matching method22 obtains the identification for each query descriptor separately according to the best matching. By voting with the separate results, the final identification is determined. Rather than matching, Liao and Jain presented a multikeypoint descriptors-based SRC (MKD-SRC) method.23 The method solves the sparse representation (SR) problem for each query descriptor separately and determines the image identification using all of the reconstruction residuals. By exploring the discrimination of the atoms in a SIFT dictionary, Sun et al. proposed a clustering-weighted SIFT-based classification method via SR24 and obtained better robustness for alignment-free face recognition with sufficient samples.

After analyzing these methods, we find that the above-described methods treat each query descriptor independently and equally. For each query descriptor, there may be some similar atoms distributing in different classes in the dictionary, which will influence the classification performance.24 Therefore, if we solve the SR problem for each descriptor extracted from a query image separately, some false identities may be obtained, resulting in errors in the final image classification. As a result, it is beneficial to solve the SR problem simultaneously for all query descriptors by their correlation. To handle this problem, the concept of the joint sparse representation (JSR) is introduced.

In this paper, we propose a weighted joint sparse representation-based classification (WJSRC) method. There are three contributions in our work. (1) For exploration of the correlation among query descriptors, the concept of JSR is introduced. (2) Considering the reliability of the query descriptors, a modified JSR model with a weighted sparsity constraint is introduced. (3) A WJSRC algorithm is proposed to solve the modified model. Because the proposed method studies the correlation among the query descriptors and their own reliability, the performance of the alignment-free recognition is improved.

The remainder of this paper is organized as follows. In Sec. 2, we review the JSR algorithm. Section 3 proposes the WJSRC method. The experimental results using the proposed method on the Yale database,26 the Olivetti Research Laboratory database,27 and the AR database (The AR database is a public face database created by Aleix Martneza and Robert Benavente.)28 are described in Sec. 4. The conclusions are presented in Sec. 5.

The original SRC method5 solves the SR problem for query descriptors separately. To explore the correlation among the query descriptors, the JSR is introduced.

As far as we know, there are two types of JSR methods. (1) The first group of JSR methods utilizes multiple types of keypoint features and dictionaries.2931 If we extract the shape, color, and texture features from a face image, which are different from each other, it is necessary to construct a single dictionary for each type of feature. Thus, three dictionaries are obtained. For a test image, multiple types of query descriptors should be extracted, each of which can be just sparsely represented by its corresponding dictionary. However, the SR for all query descriptors should be performed under the JSR constraint. The workflow is shown in Fig. 1(a). (2) The second group of JSR methods utilizes multiple keypoint features and a single dictionary,3234 whose workflow is shown in Fig. 1(b). This method supposes that multiview sample images for an object are obtained, and the queries of the object are also multiview images. With the same type of features, a single dictionary is constructed. Because the query images are similar, just one atom is selected from the dictionary to represent them at each iteration step of the atom selection process. After adequate iteration, one set of features from the dictionary can be selected to represent all query images simultaneously under the joint representation constraint.34 Thus, the sparse coefficients share the same sparsity pattern at the atom level,34 but the coefficient value is different, which is illustrated in Fig. 2(b). This method is different from the original SRC,5 which solves the SR problem for each view image separately. The sparse coefficients’ sparsity and value in the SRC method are different from each other, which are depicted in Fig. 2(a). In real scenarios, the multiview images of an object may not be well represented by the same features. In Refs. 32 and 33, Zhang et al. optimized it by proposing a joint dynamic sparse representation (JDSR) method, which chooses different features from the same class to represent each view image at each iteration step of the atom selection process. The sparse coefficients share the same sparsity pattern at the class level, but not at the atom level, which is depicted in Fig. 2(c).

Graphic Jump LocationF1 :

The workflows of the two types of joint sparse representation methods. (a) The workflow of the multiple types of features and dictionaries-based JSR method. (b) The workflow of the multiple features and single dictionary-based JSR method.

Graphic Jump LocationF2 :

The sparsity pattern of multiple task sparse representation,23 joint sparse representation34 and joint dynamic sparse representation.33 Each column vector denotes a coefficient vector and each block denotes a coefficient value. The white block denotes a zero value and others denote different nonzero values. (a) Multiple task sparse representation. It solves the SR problem for each query feature separately. The coefficient sparsity and value of each query feature may be different. (b) Joint sparse representation. Sparse coefficients share the same sparsity pattern at atom level, i.e., selecting the same atoms for all query vectors simultaneously, but with different coefficient values. (c) Joint dynamic sparse representation. The atoms on the same arrow line represent one set of features selected at each iteration step of the atom selection process. From one iteration to the next iteration, the algorithm keeps the existing atoms in the set and tries to find the next best atoms to add to the set. Sparse coefficients share the same sparsity pattern at class-level.

Although the problem of face recognition with SIFT descriptors belongs to the second type, the existing methods cannot solve it perfectly for three reasons. (1) The query descriptors are quite different from each other, for example, the descriptors extracted from an eye are different from those of a mouth, which is obviously different from the characteristic of the query features in Refs. 32 to 34. (2) The number of query descriptors is often large, typically in the hundreds; as a result, the query descriptors are challenging to join. (3) Not all of the query descriptors contain correct identity information in practice. For example, the descriptors extracted from the occluded module cannot be treated equally to those from the clear module. The unreliable descriptors will mislead the JSR, which has been verified by experiments in 32. As a result, a more robust method is required to solve our problem.

Given samples collected from c classes, the SIFT descriptors extracted from the samples of the k’th class constructed the k’th sub-dictionary Dk=[d1,d2,,dNk], where Nk denotes the quantity of the descriptors of the k’th class, and dRN×1(N=128) is the SIFT descriptor. All of the sub-dictionaries are pooled together to construct the dictionary D=[D1,D2,,Dc]=[d1,d2,,dN], where N=k=1cNk denotes the quantity of descriptors in all classes. Given a query image Y, it can be represented by a set of SIFT descriptors, i.e., Y=[y1,y2,,ym], where yi is called a query descriptor.

WJSR Model

If a query image belongs to one of the given classes, then the query descriptors extracted from it can be well represented by the ones of the corresponding class. Because the SIFT descriptors are discriminative, for example, the descriptors of an eye are different from those of a mouth, they should be represented by different atoms of the same class, i.e., the sparse coefficients share the same sparsity pattern at the class level32,33 but not at the atom level.34 As mentioned above, the query descriptors should not be treated equally. Thus, we proposed a model of WJSR, whose mathematical model is Display Formula

{x^i}1m=argmin{xi}i=1mwiyiDxi22s.t.Xg0K,(1)
where ·0 and ·2 denote l0-norm and l2-norm, respectively, K is the sparsity, and wi is the weight for the classification reliability of the i’th query descriptor. Supposing the coefficients matrix Display Formula
X=[x11x21xm1x12x22xm2x1Nx2NxmN]
and the nonzero coefficients of the selected atoms matrix Display Formula
X=[x11x21xm1x12x22xm2],
xgk=[x1k,x2k,,xmk] represents the k’th selected set, the atoms of which belong to the same class, and Xg=[xg12,xg22,]T is a matrix constraint term enabling the sparse coefficients to be in line with the weighted joint sparsity pattern. In Eq. (1), wi is not only a measurement of the reliability for the i’th query descriptor but also a balance factor for the residuals. To obtain the whole minimal residuals, we must make the residuals of the larger weight query descriptors smaller, i.e., the representation of the more reliable descriptors must be a more accurate approximation because they contain the correct classification information. Thus, in Algorithm 1, the reliable query descriptors lead the atoms’ selection.

Table Grahic Jump Location
Algorithm 1The weighted joint sparse representation based classification (WJSRC).
WJSR-Based Classification Algorithm
Calculating the weight for each query descriptor

For classification, not all query descriptors contribute equally. In this paper, we measure the importance of each query descriptor by the similarity ci between the query descriptor yi and the dictionary D, i.e., ci=max(yiT·D), and then the weight of yi can be defined as Display Formula

wi={1cic¯cic0c¯c0c¯>cic00c0>ci,(2)
where c¯=(i=1mci)/m, and c0=min{ci}i=1,,m is the least similarity value. Then we can construct the weight vector as W=[w1,w2,,wm].

Solving the WJSR problem

Solving Eq. (1) is an NP-hard problem due to the l0l2 mixed-norm minimization with a weighted joint sparsity constraint. In this paper, we propose a greedy algorithm, i.e., the WJSRC algorithm, to solve this problem, which is described in detail in Algorithm 1. The algorithm is similar to the orthogonal matching pursuit algorithm,35 with a major difference in the atom selection criteria. In the WJSRC algorithm, the most relevant set of atoms belonging to the same class is selected at each iteration step in the atom selection process. To minimize the whole residuals, we propose the weighted atoms selection criteria, which is automatically led by the larger weight descriptors.

Determining the identity of the query image

The identity of the image Y is determined by combining the residuals of all of the query descriptors. Display Formula

mincRc(Y)=1mi=1myiDcδc(xi)2,(3)
where δc(·) is a function to select the coefficients belonging to the c’th class.

Summary

The proposed WJSRC method is summarized as follows:

  • Extract the SIFT descriptors from the sample and query images to construct the dictionary D and the multi query-descriptor matrix Y, respectively.
  • Calculate the weight for each query descriptor using equation (2) to form the weight vector W.
  • Solve the WJSR for all of the query descriptors using Algorithm 1 to obtain the sparse coefficient matrix X.

Determine the identity of the query image using Eq. (3).

In this section, we present the performance of the “proposed method” on three public databases: (1) the Yale database,26 (2) the Olivetti Research Laboratory (ORL) database,27 and (3) the AR database.28 We focus on three scenarios of alignment-free face recognition: (1) arbitrary patches of holistic faces; (2) faces with arbitrary pose and expression variations; and (3) faces with occlusions. A performance comparison among the related methods, namely the SIFT matching,22 MKD-SRC,23 CWS-SRC,24 JDSRC,33 and the original SRC methods,5 is conducted. The three experiments are performed on gray images. The SIFT descriptors extracted from images are of dimension 128.

Determination of the Parameters

In the experiments, one parameter should be set manually, i.e., the sparsity K, which is the number of iterations in Algorithm 1. At each iteration step in the atom selection process, one set of descriptors is selected to represent the query descriptors. Therefore, with the increase in K, the representation for most of the query descriptors becomes more approximate. To ascertain the relationship between the recognition performance and the sparsity K, we examined different values of K on the ORL database and evaluated the resulting performance in terms of the accuracy. The curve is depicted in Fig. 3, which shows that when K is greater than 7, the recognition accuracy is stable, i.e., the approximation is adequate. Therefore, K is set as 7, which has been proven to also be suitable for other databases.

Graphic Jump LocationF3 :

The relationship between the recognition accuracy and the sparsity K.

Partial Face Recognition with an Arbitrary Patch

This experiment is conducted on the Yale database, which consists of 165 frontal face images of 15 subjects with an image size of 170×230. Two, three and four images (per subject) are randomly selected as samples. From each of the remaining images in the three settings, one patch of random size h×w at a random position is cropped as a query, where h and w are randomly selected from (100,160) and (80,110). The queries are all partial faces. Some examples of the sample and query images are shown in Fig. 4.

Graphic Jump LocationF4 :

Some examples of the sample and partial query images. (a) The examples of partial query images. (b) The examples of the sample images.

For each experimental setting, we use 10 random splits of the data in the experiment. The average results are presented in Table 1. Because the original SRC method is not suitable in this scenario, the other five algorithms are compared. The descriptors extracted from a query partial face are relatively insufficient, and the classification information is limited. Thus, it is necessary to join all of the query descriptors by their correlation. The results in Table 1 show that the WJSRC method achieves the highest recognition rate in the three settings, which indicates the validity and advantage of the proposed method in the scenario of incomplete classification information.

Table Grahic Jump Location
Table 1The average recognition performance of the partial faces.
Face Recognition with Pose and Expression Variation

This experiment is conducted on the ORL database, which contains 400 images of 40 subjects with different expressions, frontal poses, and slight scale variations. We randomly selected two, three, four, and five images from each subject as samples and the remaining as queries. Some examples of the sample and query images are shown in Fig. 5. For each experimental setting, we use 10 random splits of the data in the experiment. The average results are presented in Table 2.

Graphic Jump LocationF5 :

Some examples of the sample and query images. (a) The examples of the sample images. (b) The examples of the query images.

Table Grahic Jump Location
Table 2The average recognition performance on the ORL database.

In this experiment, the recognition rate of the proposed WJSRC method is found to be outstanding. The original SRC method does not work ideally due to the alignment problem. As the database exhibits great changes in pose and expressions and the dictionary does not have sufficient samples, there are many unreliable query descriptors. Because the proposed method considers the query descriptors holistically and joints all of the reliable ones, it achieves a better performance than the others.

Holistic Face Recognition with Occlusion

This experiment is conducted on the AR database. The AR database contains 120 subjects, including 65 males and 55 females. For each subject, 26 images were taken in two sessions, of which 14 images are nonoccluded and the remainder are occluded by various objects, such as scarves and sunglasses. Experiments are performed on the images of two separate sessions. We selected nonoccluded face images in one session as the samples for each subject. The remaining occluded face images in that session are selected as the queries. Therefore, there are 840 samples and 720 queries in each experimental setting. All images were cropped to 128×170pixels. No alignment was performed between the queries and the samples. Some examples of the samples and the queries are shown in Fig. 6.

Graphic Jump LocationF6 :

Some examples of the sample and query images. (a) The examples of the sample images. (b) The examples of the query images.

The recognition performance is presented in Table 3. The performance of the proposed method is found to be outstanding. The MKD-SRC and CWS-SRC methods also work well. As is known, most of the SIFT descriptors extracted from the occluded module in a face image are not reliable. Compared to the results of the JDSRC method, we can see that consideration of the reliability of the query descriptors is practical. Thus, the calculation of the weight of the query descriptors is important. Analyzing the misrecognition images by WJSRC, we find that our method works poorly for the face images containing too many descriptors, especially in the case where most of these descriptors are unreliable. Our future work will focus on this issue of face images with too many descriptors.

Table Grahic Jump Location
Table 3The recognition performance on the AR database.

In this work, a novel framework for robust alignment-free face recognition was proposed. The approach studies the reliability of the query descriptors holistically and utilizes the correlation among them. We demonstrated promising experimental results on images of partial faces, occluded faces, and faces with variations due to different poses and expressions. Comparison of the proposed algorithm with the related algorithms indicated that the proposed method is more robust for alignment-free scenarios. Meanwhile, some methods may be used to improve the robustness, such as optimizing the approach of weight calculation, which will be studied in the future.

This work was supported by the Fundamental Research Funds for the Central Universities (2014KJJCA15), the State Key Laboratory of Acoustics, Chinese Academy of Sciences (SKLA201304), the National Natural Science Foundation of China (61431004), and the Fundamental Research Funds for the Central Universities (2013NT55). We thank Prof. Xiaoming Zhu for helping us revise the organizational structure and grammar issues of the paper.

Jiang  X. D., “Asymmetric principal component and discriminant analyses for pattern classification,” IEEE Trans. Patt. Anal. Mach. Intell.. 31, (5 ), 931 –937 (2009). 0162-8828 CrossRef
Comon  P., “Independent component analysis, a new concept?,” Signal Process.. 36, , 287 –314 (1994). 0165-1684 CrossRef
Heisele  B., Ho  P., Poggio  T., “Face recognition with support vector machines: Global versus component-based approach,” in  Eighth IEEE International Conference on Computer Vision, 2001. ICCV 2001. Proceedings , Vol. 2, pp. 688 –694 (2001).CrossRef
Naseem  I., Togneri  R., Bennamoun  M., “Linear regression for face recognition,” IEEE Trans. Pattern Anal. Mach. Intell.. 32, (11 ), 2106 –2112 (2010). 0162-8828 CrossRef
Wright  J. et al., “Robust face recognition via sparse representation,” IEEE Trans. Pattern Anal. Mach. Intell.. 31, (2 ), 210 –227 (2009). 0162-8828 CrossRef
He  R. et al., “Two-stage nonnegative sparse representation for large-scale face recognition,” IEEE Trans. Neural Netw. Learn. Syst.. 24, (1 ), 35 –46 (2013). 1045-9227 CrossRef
Huang  J., Huang  X., Metaxas  D., “Simultaneous image transformation and sparse representation recovery,” in  IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2008 , pp. 1 –8 (2008).CrossRef
Khorsandi  R., Abdel-Mottaleb  M., “Gender classification using 2-D ear images and sparse representation,” in  2013 IEEE Workshop on Applications of Computer Vision (WACV) , pp. 461 –466 (2013).CrossRef
Ramirez  I., Sprechmann  P., Sapiro  G., “Classification and clustering via dictionary learning with structured incoherence and shared features,” in  2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) , pp. 3501 –3508 (2010).CrossRef
Yang  J., Yu  K., Huang  T., “Supervised translation-invariant sparse coding,” in  2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) , pp. 3517 –3524 (2010).CrossRef
Cao  H. et al., “Classification of multicolor fluorescence in situ hybridization (M-FISH) images with sparse representation,” IEEE Trans. Nanobiosci.. 11, (2 ), 111 –118 (2012). 1536-1241 CrossRef
Li  Y., Ngom  A., “Fast sparse representation approaches for the classification of high-dimensional biological data,” in  IEEE Int. Conf. Bioinformatics and Biomedicine , pp. 1 –6 (2012).CrossRef
Julazadeh  A., Alirezaie  J., Babyn  P., “A novel automated approach for segmenting lateral ventricle in MR images of the brain using sparse representation classification and dictionary learning,” in  11th Int. Conf. Information Science, Signal Processing and their Applications , pp. 888 –893 (2012).CrossRef
Xu  M. et al., “Tumor classification via sparse representation based on metasample,” in  2th Int. Symposium on Knowledge Acquisition and Modeling , pp. 31 –34 (2009).CrossRef
Lai  J., Jiang  X., “Modular weighted global sparse representation for robust face recognition,” IEEE Signal Process. Lett.. 19, (9 ), 571 –574 (2012). 1070-9908 CrossRef
Estabridis  K., “Automatic target recognition via sparse representation,” Proc. SPIE. 7696, , 76960O  (2010). 0277-786X CrossRef
Chen  Y., Do  T., Tran  T., “Robust face recognition using locally adaptive sparse representation,” in  17th IEEE Int. Conf. Image Processing (ICIP) , pp. 1657 –1660 (2010).CrossRef
Wagner  J. et al., “Toward a practical face recognition system: Robust registration and illumination by sparse representation,” in  Conference on IEEE Computer Vision and Pattern Recognition, 2009. CVPR 2009 , pp. 597 –604 (2009).CrossRef
Lu  C.-Y. et al., “Face recognition via weighted sparse representation,” J. Vis. Commun. Image R. 24, , 111 –116 (2013). 1047-3203 CrossRef
Cox  M. et al., “Least squares congealing for unsupervised alignment of images,” in  IEEE Conference on Computer Vision and Pattern Recognition, 2008. CVPR 2008 , pp. 1 –8 (2008).CrossRef
Learned-Miller  E., “Data driven image models through continuous joint alignment,” Pattern Anal. Mach. Intell.. 28, (2 ), 236 –250 (2006). 0162-8828 CrossRef
Lowe  G., “Distinctive image features from scale-invariant keypoints,” Int. J. Comput. Vision. 60, , 91 –110 (2004). 0920-5691 CrossRef
Liao  S., Jain  A. K., “Partial face recognition: an alignment free approach,” in  2011 International Joint Conference on Biometrics (IJCB) , pp. 1 –8 (2011).CrossRef
Sun  B., Xu  F., He  J., “Clustering-weighted SIFT-based classification method via sparse representation,” J. Electron. Imaging. 23, (4 ), 043007  (2014). 1017-9909 CrossRef
Bay  H. et al., “SURF: speeded up robust features,” in  9th European Conference on Computer Vision , Vol. 3951, pp. 404 –417 (2006).CrossRef
Belhumeur  P. N., Hespanha  J. P., Kriegman  D. J., “Eigenfaces vs. Fisherfaces: recognition using class specific linear projection,” IEEE Trans. Pattern Anal. Mach. Intell.. 19, (7 ), 711 –720 (1997). 0162-8828 CrossRef
Samaria  F. S., Harter  A. C., “Parameterisation of a stochastic model for human face identification,” in  Proceedings of the Second IEEE Workshop on Applications of Computer Vision, 1994 , pp. 138 –142 (1994).CrossRef
Martinez  A., Benavente  R., “The AR face database, Technical report,” CVC Technical Report (1998).
Yuan  X.-T., Liu  X., Yan  S., “Visual classification with multitask joint sparse representation,” IEEE Trans. Image Process.. 21, (10 ), 4349 –4360 (2012). 1057-7149 CrossRef
Nam  H. et al., “Robust multi-sensor classification via joint sparse representation,” in  2011 Proceedings of the 14th International Conference on Information Fusion (FUSION) , pp. 1 –8 (2011).
Shekhar  S. et al., “Joint sparse representation for robust multimodal biometrics recognition,” IEEE Trans. Pattern Anal. Mach Intell.. 36, (1 ), 113 –126 (2014). 0162-8828 CrossRef
Zhang  H. et al., “Multi-observation visual recognition via joint dynamic sparse representation,” in  2011 IEEE International Conference on Computer Vision (ICCV) , pp. 595 –602 (2011).CrossRef
Zhang  H. et al., “Multi-view face recognition via joint dynamic sparse representation,” in  2011 18th IEEE International Conference on Image Processing (ICIP) , pp. 3025 –3028 (2011).CrossRef
Zhang  H. et al., “Multi-view automatic target recognition using joint sparse representation,” IEEE Trans. Aerosp. Electron. Syst.. 48, (3 ), 2481 –2497 (2012). 0018-9251 CrossRef
Pati  Y. C., Rezaiifar  R., Krishnaprasad  P. S., “Orthogonal matching pursuits: Recursive function approximation with applications to wavelet decomposition,” in  1993 Conference Record of the Twenty-Seventh Asilomar Conference on Signals, Systems and Computers , Vol. 1, pp. 40 –44(1993).CrossRef

Bo Sun received a BSc in computer science from Beihang University, China, and MSc and PhD degrees from Beijing Normal University, China. He is currently a professor in the Department of Computer Science and Technology at Beijing Normal University. His research interests include pattern recognition, natural language processing, and information systems. He is a member of ACM and a senior member of the China Society of Image and Graphics.

Feng Xu received a BSc in electronic science and technology from Beijing Normal University in 2009. He is currently working toward the MSc degree in computer application technology at Beijing Normal University. His research interests include pattern recognition and signal processing.

Guoyan Zhou received a BSc in computer science and technology from Beijing Normal University, 2009. She is currently working toward the MSc degree in Computer Application Technology at Beijing Normal University. Her research interest includes signal processing.

Jun He received a BSc in optical engineering and a PhD in physical electronics from Beijing Institute of Technology, China, in 1998 and 2003, respectively. Since 2003, she has been with the College of Information Science and Technology of Beijing Normal University, China. She was elected as a lecturer and an assistant professor in 2003 and 2010, respectively. Her research interests include image processing application and pattern recognition.

Fengxiang Ge received the PhD degree in communication and information systems from Tsinghua University in 2003. From 2003 to 2005, he was a postdoctoral research associate at the University of Hong Kong. In 2005, he joined Intel Corporation as a senior researcher and an architect. In November 2011, he joined the College of Information Science and Technology, Beijing Normal University, China. His research interests include signal processing and its applications.

© The Authors. Published by SPIE under a Creative Commons Attribution 3.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.

Citation

Bo Sun ; Feng Xu ; Guoyan Zhou ; Jun He and Fengxiang Ge
"Weighted joint sparse representation-based classification method for robust alignment-free face recognition", J. Electron. Imaging. 24(1), 013018 (Jan 26, 2015). ; http://dx.doi.org/10.1117/1.JEI.24.1.013018


Figures

Graphic Jump LocationF3 :

The relationship between the recognition accuracy and the sparsity K.

Graphic Jump LocationF4 :

Some examples of the sample and partial query images. (a) The examples of partial query images. (b) The examples of the sample images.

Graphic Jump LocationF5 :

Some examples of the sample and query images. (a) The examples of the sample images. (b) The examples of the query images.

Graphic Jump LocationF2 :

The sparsity pattern of multiple task sparse representation,23 joint sparse representation34 and joint dynamic sparse representation.33 Each column vector denotes a coefficient vector and each block denotes a coefficient value. The white block denotes a zero value and others denote different nonzero values. (a) Multiple task sparse representation. It solves the SR problem for each query feature separately. The coefficient sparsity and value of each query feature may be different. (b) Joint sparse representation. Sparse coefficients share the same sparsity pattern at atom level, i.e., selecting the same atoms for all query vectors simultaneously, but with different coefficient values. (c) Joint dynamic sparse representation. The atoms on the same arrow line represent one set of features selected at each iteration step of the atom selection process. From one iteration to the next iteration, the algorithm keeps the existing atoms in the set and tries to find the next best atoms to add to the set. Sparse coefficients share the same sparsity pattern at class-level.

Graphic Jump LocationF1 :

The workflows of the two types of joint sparse representation methods. (a) The workflow of the multiple types of features and dictionaries-based JSR method. (b) The workflow of the multiple features and single dictionary-based JSR method.

Graphic Jump LocationF6 :

Some examples of the sample and query images. (a) The examples of the sample images. (b) The examples of the query images.

Tables

Table Grahic Jump Location
Table 1The average recognition performance of the partial faces.
Table Grahic Jump Location
Table 2The average recognition performance on the ORL database.
Table Grahic Jump Location
Algorithm 1The weighted joint sparse representation based classification (WJSRC).
Table Grahic Jump Location
Table 3The recognition performance on the AR database.

References

Jiang  X. D., “Asymmetric principal component and discriminant analyses for pattern classification,” IEEE Trans. Patt. Anal. Mach. Intell.. 31, (5 ), 931 –937 (2009). 0162-8828 CrossRef
Comon  P., “Independent component analysis, a new concept?,” Signal Process.. 36, , 287 –314 (1994). 0165-1684 CrossRef
Heisele  B., Ho  P., Poggio  T., “Face recognition with support vector machines: Global versus component-based approach,” in  Eighth IEEE International Conference on Computer Vision, 2001. ICCV 2001. Proceedings , Vol. 2, pp. 688 –694 (2001).CrossRef
Naseem  I., Togneri  R., Bennamoun  M., “Linear regression for face recognition,” IEEE Trans. Pattern Anal. Mach. Intell.. 32, (11 ), 2106 –2112 (2010). 0162-8828 CrossRef
Wright  J. et al., “Robust face recognition via sparse representation,” IEEE Trans. Pattern Anal. Mach. Intell.. 31, (2 ), 210 –227 (2009). 0162-8828 CrossRef
He  R. et al., “Two-stage nonnegative sparse representation for large-scale face recognition,” IEEE Trans. Neural Netw. Learn. Syst.. 24, (1 ), 35 –46 (2013). 1045-9227 CrossRef
Huang  J., Huang  X., Metaxas  D., “Simultaneous image transformation and sparse representation recovery,” in  IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2008 , pp. 1 –8 (2008).CrossRef
Khorsandi  R., Abdel-Mottaleb  M., “Gender classification using 2-D ear images and sparse representation,” in  2013 IEEE Workshop on Applications of Computer Vision (WACV) , pp. 461 –466 (2013).CrossRef
Ramirez  I., Sprechmann  P., Sapiro  G., “Classification and clustering via dictionary learning with structured incoherence and shared features,” in  2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) , pp. 3501 –3508 (2010).CrossRef
Yang  J., Yu  K., Huang  T., “Supervised translation-invariant sparse coding,” in  2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) , pp. 3517 –3524 (2010).CrossRef
Cao  H. et al., “Classification of multicolor fluorescence in situ hybridization (M-FISH) images with sparse representation,” IEEE Trans. Nanobiosci.. 11, (2 ), 111 –118 (2012). 1536-1241 CrossRef
Li  Y., Ngom  A., “Fast sparse representation approaches for the classification of high-dimensional biological data,” in  IEEE Int. Conf. Bioinformatics and Biomedicine , pp. 1 –6 (2012).CrossRef
Julazadeh  A., Alirezaie  J., Babyn  P., “A novel automated approach for segmenting lateral ventricle in MR images of the brain using sparse representation classification and dictionary learning,” in  11th Int. Conf. Information Science, Signal Processing and their Applications , pp. 888 –893 (2012).CrossRef
Xu  M. et al., “Tumor classification via sparse representation based on metasample,” in  2th Int. Symposium on Knowledge Acquisition and Modeling , pp. 31 –34 (2009).CrossRef
Lai  J., Jiang  X., “Modular weighted global sparse representation for robust face recognition,” IEEE Signal Process. Lett.. 19, (9 ), 571 –574 (2012). 1070-9908 CrossRef
Estabridis  K., “Automatic target recognition via sparse representation,” Proc. SPIE. 7696, , 76960O  (2010). 0277-786X CrossRef
Chen  Y., Do  T., Tran  T., “Robust face recognition using locally adaptive sparse representation,” in  17th IEEE Int. Conf. Image Processing (ICIP) , pp. 1657 –1660 (2010).CrossRef
Wagner  J. et al., “Toward a practical face recognition system: Robust registration and illumination by sparse representation,” in  Conference on IEEE Computer Vision and Pattern Recognition, 2009. CVPR 2009 , pp. 597 –604 (2009).CrossRef
Lu  C.-Y. et al., “Face recognition via weighted sparse representation,” J. Vis. Commun. Image R. 24, , 111 –116 (2013). 1047-3203 CrossRef
Cox  M. et al., “Least squares congealing for unsupervised alignment of images,” in  IEEE Conference on Computer Vision and Pattern Recognition, 2008. CVPR 2008 , pp. 1 –8 (2008).CrossRef
Learned-Miller  E., “Data driven image models through continuous joint alignment,” Pattern Anal. Mach. Intell.. 28, (2 ), 236 –250 (2006). 0162-8828 CrossRef
Lowe  G., “Distinctive image features from scale-invariant keypoints,” Int. J. Comput. Vision. 60, , 91 –110 (2004). 0920-5691 CrossRef
Liao  S., Jain  A. K., “Partial face recognition: an alignment free approach,” in  2011 International Joint Conference on Biometrics (IJCB) , pp. 1 –8 (2011).CrossRef
Sun  B., Xu  F., He  J., “Clustering-weighted SIFT-based classification method via sparse representation,” J. Electron. Imaging. 23, (4 ), 043007  (2014). 1017-9909 CrossRef
Bay  H. et al., “SURF: speeded up robust features,” in  9th European Conference on Computer Vision , Vol. 3951, pp. 404 –417 (2006).CrossRef
Belhumeur  P. N., Hespanha  J. P., Kriegman  D. J., “Eigenfaces vs. Fisherfaces: recognition using class specific linear projection,” IEEE Trans. Pattern Anal. Mach. Intell.. 19, (7 ), 711 –720 (1997). 0162-8828 CrossRef
Samaria  F. S., Harter  A. C., “Parameterisation of a stochastic model for human face identification,” in  Proceedings of the Second IEEE Workshop on Applications of Computer Vision, 1994 , pp. 138 –142 (1994).CrossRef
Martinez  A., Benavente  R., “The AR face database, Technical report,” CVC Technical Report (1998).
Yuan  X.-T., Liu  X., Yan  S., “Visual classification with multitask joint sparse representation,” IEEE Trans. Image Process.. 21, (10 ), 4349 –4360 (2012). 1057-7149 CrossRef
Nam  H. et al., “Robust multi-sensor classification via joint sparse representation,” in  2011 Proceedings of the 14th International Conference on Information Fusion (FUSION) , pp. 1 –8 (2011).
Shekhar  S. et al., “Joint sparse representation for robust multimodal biometrics recognition,” IEEE Trans. Pattern Anal. Mach Intell.. 36, (1 ), 113 –126 (2014). 0162-8828 CrossRef
Zhang  H. et al., “Multi-observation visual recognition via joint dynamic sparse representation,” in  2011 IEEE International Conference on Computer Vision (ICCV) , pp. 595 –602 (2011).CrossRef
Zhang  H. et al., “Multi-view face recognition via joint dynamic sparse representation,” in  2011 18th IEEE International Conference on Image Processing (ICIP) , pp. 3025 –3028 (2011).CrossRef
Zhang  H. et al., “Multi-view automatic target recognition using joint sparse representation,” IEEE Trans. Aerosp. Electron. Syst.. 48, (3 ), 2481 –2497 (2012). 0018-9251 CrossRef
Pati  Y. C., Rezaiifar  R., Krishnaprasad  P. S., “Orthogonal matching pursuits: Recursive function approximation with applications to wavelet decomposition,” in  1993 Conference Record of the Twenty-Seventh Asilomar Conference on Signals, Systems and Computers , Vol. 1, pp. 40 –44(1993).CrossRef

Some tools below are only available to our subscribers or users with an online account.

Related Content

Customize your page view by dragging & repositioning the boxes below.

Related Book Chapters

Topic Collections

Advertisement
  • Don't have an account?
  • Subscribe to the SPIE Digital Library
  • Create a FREE account to sign up for Digital Library content alerts and gain access to institutional subscriptions remotely.
Access This Article
Sign in or Create a personal account to Buy this article ($20 for members, $25 for non-members).
Access This Proceeding
Sign in or Create a personal account to Buy this article ($15 for members, $18 for non-members).
Access This Chapter

Access to SPIE eBooks is limited to subscribing institutions and is not available as part of a personal subscription. Print or electronic versions of individual SPIE books may be purchased via SPIE.org.