Zhenyong Fu
Pairwise constraint propagation via low-rank matrix recovery
Zhenyong Fu1
As a kind ofweaker supervisory information,pairwise constraints can be exploited to guide the data analysis process,such as data clustering.This paper formulates pairwise constraint propagation,which aims to predict the large quantity of unknown constraints from scarce known constraints,as a low-rank matrix recovery(LMR)problem.Although recent advances in transductive learning based on matrix completion can be directly adopted to solve this problem,our work intends to develop a more general low-rank matrix recovery solution for pairwise constraint propagation,which not only completes the unknown entries in the constraint matrix but also removes the noise from the data matrix.The problem can be e ff ectively solved using an augmented Lagrange multiplier method.Experimental results on constrained clustering tasks based on the propagated pairwise constraints have shown that our method can obtain more stable results than state-of-the-art algorithms, and outperform them.
semi-supervised learning; pairwise constraint propagation;low-rank matrix recovery(LMR);constrained clustering; matrix completion
Pairwise constraints provide prior knowledge as to whether two data points belong to the same class or not,known as must-link constraints and cannot-link constraints,respectively.Generally,we cannot infer instance labels from only pairwise constraints,especially for multi-class data.Thismeans that pairwise constraints are weaker and thus more general than explicit labels on data.Pairwise constrains have been widely used in the context of clustering with side information,also called semi-supervised clustering or constrained clustering, where it has been shown that the presence of appropriate pairwise constraints can often improve the performance[1-5].While it is possible to infer pairwise constraints from domain knowledge or user feedback,in practice,the availability of such constraints is scarce.Pairwise constraint propagation aimsto producea largequantity ofpairwise constraints from scarce known constraints.
One way to utilize pairwise constraintsfor constrained clustering is to trivially set the similarities between the constrained data to 1 and 0 for must-link and cannot-link constraints, respectively[3].This approach only adjusts the similarities between constrained data, without propagating the constraint information to other data.In contrast,much research shows that it is more e ff ective to propagate known constraint information to other unconstrained data[6-8].In Ref.[9],the pairwise constraints are propagated to unconstrained data using a Gaussian process.However,as noted in Ref.[9],this method makes certain assumptions about constraint propagation,especially with respect to two-class problems.
Recently,some research has tried to solve the problem of pairwise constraint propagation under the semi-supervised learning framework that is usually based on the manifold assumption(objects on the same manifold structure tend to have similar pairwise constraint settings)[10-15].For instance, in Ref.[7],Li et al.applied the manifold assumption inversely.They proposed to learn a mapping with known pairwise constraints so that two must-link objects are mapped to the same manifold whiletwo cannot-link objects are mapped to di ff erent manifolds.Propagation of the constraint relationship is achieved by making the mapping smooth on the manifolds.Li et al.’s method can be formulated as a semide fi nite programming(SDP)problem.Although this method can tackle multi-class problems,current SDP solvers are severely limited in the size of problems they can solve.Other work also based on the manifold assumption is the exhaustive and efficient pairwise constraint propagation(E2CP) method proposed by Lu and Ip[8].It decomposes the constraint propagation problem into a series of two-class label propagation subproblems,each of which can be solved through semi-supervised learning based onk-nearest neighbor graphs.The basic underpinning of Lu and Ip’s method is still the manifold assumption,i.e.,that similar objects should have similar propagation results.
However,the manifold assumption only concerns local structure.In other words,it does not concern those objects that are not close each other or do not share the same manifold structure.In this paper, we make attempt to capture global structure in the constraint propagation problem.We observe that the constraint matrix,whose entries are the propagated pairwise constraints,should be globally low-rank,which has not been noted in previous work on constraint propagation.Therefore,we address constraint propagation via the low-rank assumption, i.e.,among all of the solutions for constraint propagation,we aim to choose the lowest-rank one.
With the low-rank assumption,we formulate pairwise constraint propagation as a low-rank matrix recovery(LMR)problem.In our work,we first present a matrix completion formulation,which is a special and simpler case of the low-rank matrix recovery problem,for constraint propagation.Our experimentsshow thatthismatrix completion method works well in practice as long as there are enough known pairwise constraints.However, as the available pairwise constraints decrease,the performance of such a simple matrix completion method degrades very quickly.This indicates that when lacking sufficientconstraints,to improve the propagation performance, we need some other information.Recent progress in transductive learning addressesthe semi-supervised learning problem using a matrix completion framework[16, 17].The main novelty of these methods is to make the data matrix and the label matrix jointly lowrank.In the following,we will refer to such methods as joint-MC.Although joint-MC methods can be directly applied to solve the constraint propagation problem,these methods neglect the in fl uence of noise in the data.In this paper,we adopt a joint lowrank approach for constraint propagation,but use a more general low-rank matrix recovery method which explicitly deals with the data noise.More concretely,joint-MC methods assume that the noisy data matrix and the constraint matrix should be jointly low-rank,while our method assumes that the clean data matrix and the constraint matrix should be jointly low-rank.Our experimental results show that the latter assumption is more reasonable.The proposed method is formulated as a low-rank matrix recovery problem,which can be e ff ectively solved with an augmented Lagrange multiplier algorithm.
For comparison with the existing propagation approaches,we use the propagated constraints obtained by our method to adjust the similarity between data points,as do previous works,so that they can be incorporated into the data clustering process.To evaluate the e ff ectiveness of the proposed method,we select two real-life image data sets for constrained clustering tasks.Our experimental results show that the proposed method can achieve more stable and better performance than the stateof-the-art.
The main contributions of this paper can be summarized as follows:
·We address the constraint propagation problem on the basis of the low-rank nature of this problem,which has not been noted by previous work.
·Our work shows that when there are enough known pairwise constraints,the simplest matrix completion formulation is very e ff ective for constraint propagation.
·Wefurtherproposeamoregenerallowrank matrix recovery solution for constraint propagation that works well even with insufficient pairwise constraints.
The remainder of this paper is organized as follows. Section 2 presents our low-rank matrix recovery algorithm for constraint propagation.Section 3 explains how to apply the proposed method toconstrained clustering tasks.In Section 4,our methods are evaluated on two real-life image data sets.Finally,Section 5 gives the conclusions drawn from the experimental results.
2.1 Low-rank structure in constraint propagation
Given a data set ofnobjectsX={x1,···,xn}and two sets of pairwise constraints,denoted respectively byM={(xi,xj)}wherexiandxjshould be in the same class andC={(xi,xj)}wherexiandxjshould be in di ff erent classes,the goal of pairwise constraint propagation is to propagate these pairwise constraints across the entire data set.
To propagate the initial pairwise constraints onX,we first represent the two types of pairwise constraints,i.e.,MandC,using a single constraint matrixY={Yij}n×n:
GivenXandY,the problem of pairwise constraint propagation is to construct a matrixF={Fij}n×nso that for any pair instances(xi,xj)∈X×X, the elementFijcan be used to predict the pairwise constraint relationship betweenxiandxj,i.e.,Fij>0 means thatxiandxjshould be must-link whileFij<0 means thatxiandxjshould be cannot-link.
As mentioned above,one e ff ective method to solve this problem is to view the constraint propagation as a two-class classi fi cation problem,where the“positive class”is the must-link relationship and the“negative class”is the cannot-link relationship[8, 14].However,in this paper,we view constraint propagation as a low-rank matrix recovery problem.
Fig.1 Red and blue respectively represent+1(a must-link constraint)and?1(a cannot-link constraint)in the constraint matrix.Constraint propagation can be viewed as a process thatrecoverstheunderlyinglow-rankmatrixfrom partial observation.
As shown in Fig.1,in the problem of pairwise constraint propagation,we have a partially observed constraint matrixYand wish to obtain the whole constraint matrix,that is,we have to complete the missing entries in the constraint matrix.This process satis fi es the requirement of matrix completion.More importantly,we notice that the constraint matrix has low-rank structure.In fact,each row(and column)of the constraint matrix can be viewed as the constraint vector of an object in the data set.It can be observed that in a desired constraint matrix,objects from the same class should have the same constraint vector.Thus,the matrix rows(and columns)from the objects that belong to a given special class form a rank-1 subspace and the total rank of the desired constraint matrix should equal the number of classes in the data set.In summary,we can view constraint propagation as a low-rank matrix completion problem,which can be formulated as the following optimization problem:
The above matrix completion problem is difficult to solve due to the discrete nature of the rank function.As suggested by Refs.[18,19],Problem (2)can be replaced by solving the following convex surrogate:
Here,‖·‖?denotes the nuclear norm[20]of a matrix, i.e.,the sum of the singular values of the matrix.
Fig.2 Clustering performance comparison between NCuts, E2CP,and simple-MC on the Scene data set.
Problem(3)is the simplest matrix completion formulation for constraint propagation.In matrix completion research [19,21],Problem (3)has been recently carefully studied,and can be solved by singular value thresholding(SVT)[22], fi xed pointcontinuation (FPC)[23],oraugmented Lagrange multiplier(ALM)[24].Figure 2 shows a clustering performance comparison between simple-MC,E2CP,and NCuts on the Scene data set(the experimental details will be described in Sections 3 and 4).It can be seen that when the known pairwise constraints are fewer than 0.5%,simple-MC even does not perform better than NCuts, which is e ff ectively a spectral clustering algorithm but without considering pairwise constraints.On the other hand,when the known constraints are fewer than 1%,the performance of simple-MC is still worse than that of E2CP,which is the baseline constraint propagation approach.Such results verify the conclusion that we can exactly recover a low-rank matrix only when we have enough observed matrix entries[19],i.e.,known initial pairwise constraints in the context of constraint propagation.
2.2 Low-rank matrix recovery formulation for constraint propagation
Since the known initial pairwise constraints are usually sparse,to further improve the performance of the low-rank matrix recovery based propagation method,wehaveto resortto someauxiliary information.Joint-MC methods from the area of transductive learning assume that the data matrix and the label matrix should be jointly low-rank[16, 17].Although these matrix completion based transduction learning algorithms can be directly used to solve constraint propagation,as discussed earlier,joint-MC methods may be unreliable in the presence of unclean data due to the lack of robustness to noise.In this section,we attempt to develop a general low-rank matrix recovery solution for constraint propagation.
LetDdenote the data matrix.Joint-MC methods consider that the matrix(D,F)should be jointly low-rank.In this paper,we assume that the data matrixDis the sum of a clean data matrixAand an additive error matrixE,i.e.,D=A+E.Unlike the joint-MC methods,we intend to make the matrix (A,F),which concatenates the clean data matrixAand the propagated constraint matrixF,jointly lowrank.This idea can be formulated as the following optimization problem:
There are two issues with Formulation(4).As in Problem(2),rank()is a non-convex function and difficult to optimize.We can relax the rank function to the convex nuclear norm‖·‖?.In addition,the discrepancy betweenAandD,i.e.,the noise matrixE,should be minimized.These two issues lead to the following convex optimization problem for constraint propagation:
whereλis a positive weighting parameter. In Formulation(5),we use the?1norm(the sum of the absolute values of the matrix entries)to measure the additive error matrixE,because the?1norm can handle arbitrarily large entries inE.The?1norm works well in our experiments.In addition,the extension of the proposed method from the?1norm to other matrix norms,such as the Frobenius norm or?2,1norm[25]is straightforward.
It should be noted that there are three considerations in Formulation(5).On one hand,we wish to remove the noise from a submatrixDin the joint matrix(D,F).On the other hand,we wish to complete the other submatrixFin the joint matrix(D,F)from the partially observed matrixY.Finally,we wish the new joint matrix(A,F)to be low-rank.This di ff ers from the formulations of robust principal component analysis(RPCA)[26]or matrix completion[19].In fact,Formulation(5)can be viewed as a combination of RPCA and matrix completion,specially designed for the problem of pairwise constraint propagation.
2.3 Solving the optimization problem
In this section,we apply the augmented Lagrange multiplier(ALM)method to solve Formulation(5).We first convert Problem(5)to the following equivalent problem:
where?=M∪Candπ?:?n×n→?n×nis a linear operator that keeps the entries in?unchanged and sets those outside?(i.e.,inˉ?)to zero.Following recent work in low-rank matrix recovery[24],in Problem(6),we use the matrixRto compensate for the unknown entries ofY.LetB=(A,F),M=(E,R),andN=(D,Y).Problem(6)can be reformulated as follows:
Problem(7)can be solved by minimizing the following augmented Lagrangian function:
whereμ>0 is a penalty parameter andH= (He,Hr)is the Lagrangian multiplier,in which the submatricesHeandHrare the Lagrange multipliers that correspond to the constraintsA+E=DandF+R=Yrespectively.The above problem can be solved either by exact or inexact ALM algorithms,as suggested in Ref.[24].For efficiency,we choose the inexact ALM,in which in each iteration we need to alternatively update the matricesB,E,R,andH.
The following soft-thresholding (shrinkage) operator is used in the process of solving Problem (7):
The update to matrixBoptimizes the following subproblem:
and the parameterμis updated viaμk+1=ρμkwithρ>1(see Ref.[28]for further details).The convergence properties of the inexact ALM have been proved in Ref.[24].
Once we have the converged matrixBk=(Ak,Fk), we can extract the submatrixFkfromBkand setF=Fkas the constraint propagation result.We summarise the proposed method in Algorithm 1, which we call the low-rank matrix recovery based pairwise constraint propagation(LMRPCP).
Algorithm 1: Low-rankmatrixrecoverybased pairwise constraint propagation Input:Data matrixD,observed constraint matrixY. Initialize:B=0,E=0,R=0,H=0,μ=1e?7,λ=1e?3,ρ=1.2,∈=1e?8. while not converged dok[S]VT. 3)Ek+1=Sλμ?11) (U,S,V)=svd(N?Mk+μ?1kHk), 2)Bk+1=USμ?14)Rk+1=πˉ?(Y?Fk+1+μ?1kHr,k) 5)Hk+1=Hk+μk(N?Bk+1?Mk+1) 6)μk+1=ρμk7)k=k+1 8) Check the convergence condition:‖N?Bk?Mk‖∞<∈end while Output:F=Fkk(D?Ak+1+μ?1kHe,k)
In previous work,the propagation resultFis further used to improve the clustering algorithm.For fair comparison,we use spectral clustering as the basic clustering algorithm.As in Ref.[8],we incorporateFinto the spectral clustering algorithm by adjusting the similarities between the data points according toFusing the following adjustment method:
In this section,we apply the propagated pairwise constraints to constrained clustering tasks and test the performance of the proposed method on two reallife image data sets.For comparison,the results of three well-known,most closely related algorithms are also reported:Lu and Carreira-Perpin′an’s affinity propagation[9],Lu and Ip’s exhaustive and efficient constraint propagation[8],and Goldberg et al.’s joint-MC[17].We use normalized cuts(NCuts)[30], which is e ff ectively a spectral clustering algorithm but without considering pairwise constraints,as the baseline.
We first describe the experimental setup,including similarity matrix computation,the performance measure,and parameter selection.Then we compare the proposed algorithm to the other four approaches on two image data sets.
4.1 Experimental setup
We tested the proposed algorithm on two di ff erent image data sets.The first one contained eight scene categories from MIT[31],including four man-made scenes and four natural scenes.The total number of images was 2688.The size of each image in this Scene data set was 256×256 pixels.The second data set contained images from a Corel collection.We selected 15 categories including bus,sunrise/sunset, plane,foxes,horses,coins,gardens,eagles,models, sailing,stream trains,racing car,pumpkins,rockies, and fields.Each of these categories contained 100 images.Therefore,this selected Corel data set had a total of 1500 images.The size of each image in this data set was 384×256 or 256×384 pixels.Some image examples from these two data sets are given in Fig.3.
For these two image data sets,we chose two di ff erent feature sets as used in Refs.[32]and[33], respectively.Following Ref.[32],SIFT descriptors were used for the Scene data set,while,following Ref.[33],joint color and Gabor features were used for the Corel data set.These features were chosen to ensure a fair comparison to state-of-theart techniques.More concretely,for the Scene data set,we extracted SIFT descriptors for 16×16 pixel blocks computed over a regular grid with spacing of 8 pixels.For the Corel data set,we divided each image into blocks of 16×16 pixels and then extracted a joint color/texture feature vector from each block.Here,the texture features are represented as the means and standard deviations of the coefficients of a bank of Gabor fi lters(with 3 scales and 4 orientations),and the color features are the mean values of HSV color components.Finally, for each image data set,we performedk-means clustering on the extracted feature vectors to form a vocabulary of 400 visual keywords.Since all of the constraint propagation algorithms finally apply spectral clustering to form the clusters,and the spectral clustering is a graph-based algorithm,based on this visual vocabulary,we then de fi ned a spatial Markov kernel[33]as the similarity matrixWfor graph construction.Moreover,to model local neighborhood relationships between the data points, we constructed ak-nearest neighbor graph,in whichwij=0 ifxjwas not among thek-nearest neighbors (k-NN)ofxi.Thek-NN graph construction was necessary for the spectral clustering[29].In addition, we fi xedk=20 in all experiments.
In order to evaluate these algorithms, we compared the clustering results with the available ground-truth data labels,and employed normalized mutualinformation(NMI)astheperformance measure[34].For two random variablesXandY, NMI is de fi ned as
Fig.3 Sample images from 8 categories of the Scene data set and 15 categories of the Corel data set.
whereI(X,Y)is the mutual information betweenXandY,andH(X)andH(Y)are the entropies ofXandYrespectively.Note that 0≤ NMI≤ 1, and the larger NMI is,the better is the result.To evaluate the algorithms under di ff erent settings of pairwise constraints,we applied the ground-truth data labels to generate a varying number of pairwise constraints randomly for each data set.We randomly chose a pair of data points from each data set.If they had the same class labels,we generated a mustlink constraint,otherwise a cannot-link constraint.In the experiments,we ran these algorithms 20 times with random initialization,and we report the average NMI(avg),the minimal NMI(min),and the standard variation(std).
4.2 Results on Scene data set
We firstchosetheSceneimagedata setto test the proposed algorithm.In the experiments, we compared the clustering performance of the fi ve algorithms with a varied number of pairwise constraints,from 0.5%to 4%.The clustering results are shown in Table 1,which shows that the proposed LMRPCP algorithm performs better than the other four algorithms in most cases.Whether there are fewer or more pairwise constraints,our LMRPCP generally performs well.The e ff ectiveness of our propagation method is veri fied by the fact that our LMRPCP consistently obtains better results in the constrained clustering task.In contrast,joint-MC performs unsatisfactorily,and in the case of few pairwise constraints,its performance is even worse than that of NCuts.This may be because when there are fewer pairwise constraints,the noise in the data matrix will heavily in fl uence the matrix completion process and degrade its performance.On the other hand,as the number of constraints grows,the performance of our LMRPCP algorithm improves more rapidly than that of AP and E2CP: our method can propagate the pairwise constraints more e ff ectively than AP and E2CP.
Moreover,another important observation is that our LMRPCP performs very stably under di ff erent settings of pairwise constraints.However,the other three constraint propagation methods are less stable (measured by the standard variation).Note that as the pairwise constraints increase,the standard variationofE2CP grows,becauseE2CP only considers the local structure in the propagation process which therefore inevitably leads to the over-propagation phenomenon.With fewer pairwise constraints,joint-MC also has big variations due to data noise.
4.3 Results on the Corel data set
We also tested our proposed method on the Corel image data set.Clustering results are shown in Table 2,from which we may draw the same conclusions as for the Scene data set.Generally,LMRPCP performs better and more stably than the other three constraint propagation methods.This is because LMRPCP can not only consider the global low-rank structure but also remove the in fl uence of the data noise.
To make it clearer how our LMRPCP exploits the pairwise constraints for spectral clustering,in Fig.4, we show the distance matrices of the low-dimensional data representations obtained by NCuts,AP,E2CP, joint-MC,and LMRPCP when there are 1%pairwise constraints.We can see that the block structure of the distance matrices of the data representations obtained by LMRPCP on the two image data sets is signi fi cantly more obvious compared to those of the data representations obtained by NCuts,AP,E2CP,and joint-MC:after incorporating the pairwise constraints produced by our method,each cluster associated with the new data representation becomes more compact and di ff erent clusters become more separate.We can conclude that LMRPCP is e ff ective as the pairwise constraints produced by LMRPCP can correctly re fl ect the category relationships between the data points.
Table 1 Clustering results(NMI)on Scene image data set under di ff erent settings of pairwise constraints,from 0.5%to 4%.The NCuts result is avg=0.54,min=0.53,and std=0.01.
Table 2 Clustering results(NMI)on the Corel image data set under di ff erent settings of pairwise constraints,from 0.5%to 4%.The NCuts result is avg=0.51,min=0.50,and std=0.01.
Fig.4 Distance matrices of the low-dimensional representations for the two image data sets obtained by NCuts,AP,E2CP, joint-MC,and LMRPCP.The darker a pixel,the smaller the distance.
In this paper,we have proposed a novel constraint propagation approach,which isbased on the low-rank nature of the propagation problem.We propose to make the clean data matrix and the constraint matrix jointly low-rank to remove data noise.Ourapproach isformulated asa lowrank matrix recovery(LMR)problem,which can be e ff ectively solved with a modi fied augmented Lagrange multiplier(ALM)method.Experimental results on two real-life image data sets demonstrate the superiority of our proposed algorithm over the state-of-the-art.
The work described in this paper was supported by the National Natural Science Foundation of China (No.61300164).
Open Access This article is distributed under the terms of the Creative Commons Attribution License which permits any use,distribution,and reproduction in any medium,provided the original author(s)and the source are credited.
[1]Basu,S.;Bilenko,M.;Mooney,R.J.A probabilistic framework for semi-supervised clustering. In: Proceedings of the tenth ACM SIGKDD international conference on Knowledge discovery and data mining, 59-68,2004.
[2]Basu,S.;Davidson,I.;Wagsta ff,K.Constrained Clustering:Advances in Algorithms,Theory,and Applications.Chapman&Hall/CRC,2008.
[3]Kamvar,S.D.;Klein,D.;Manning,C.D.Spectral learning.In:Proceedings of the 18th international joint conference on Arti fi cial intelligence,561-566,2003.
[4]Kulis,B.;Basu,S.;Dhillon,I.;Mooney,R.Semisupervised graph clustering:A kernel approach.In: Proceedings of the 22nd international conference on Machine learning,457-464,2005.
[5]Xing,E.P.;Jordan,M.I.;Russell,S.J.;Ng, A.Y.Distancemetriclearningwith application to clustering with side-information.In: Advances in Neural Information Processing Systems 15, 2002.Available at http://papers.nips.cc/paper/2164-distance-metric-learning-with-application-toclustering-with-side-information.pdf.
[6]Li,Z.;Liu,J.Constrained clustering by spectral kernel learning.In:IEEE 12th International Conference on Computer Vision,421-427,2009.
[7]Li, Z.; Liu, J.; Tang, X.Pairwise constraint propagation by semide fi nite programming for semisupervised classi fi cation.In:Proceedings of the 25th international conference on Machine learning,576-583, 2008.
[8]Lu,Z.;Ip,H.H.S.Constrained spectral clustering via exhaustive and efficient constraint propagation.Lecture Notes in Computer ScienceVol.6316,1-14, 2010.
[9]Lu, Z.; Carreira-Perpin′an, M.A.Constrained spectral clustering through affinity propagation.In: IEEE Conference on Computer Vision and Pattern Recognition,1-8,2008.
[10]Fu,Z.;Ip,H.H.S.;Lu,H.;Lu,Z.Multimodal constraint propagation for heterogeneous image clustering.In: Proceedingsofthe 19th ACM international conference on Multimedia,143-152,2011.
[11]Fu,Z.;Lu,H.;Ip,H.H.S.;Lu,Z.Modalities consensus for multi-modal constraint propagation.In: Proceedings of the 20th ACM international conference on Multimedia,773-776,2012.
[12]Fu, Z.; Lu, H.; Li, W.Incrementalvisual objects clustering with the growing vocabulary tree.Multimedia Tools and ApplicationsVol.56,No.3, 535-552,2012.
[13]Fu,Z.;Lu,Z.;Ip,H.H.S.;Lu,H.;Wang, Y.Local similarity learning for pairwise constraint propagation.MultimediaToolsandApplicationsVol.74,No.11,3739-3758,2015.
[14]Fu, Z.; Lu, Z.; Ip, H.H.S.; Peng, Y.; Lu, H.Symmetric graph regularized constraint propagation.In: Proceedings of the Twenty-Fifth AAAI Conference on Arti fi cial Intelligence,350-355, 2011.
[15]Zhu,X.;Goldberg,A.B.Introduction to semisupervised learning.Synthesis Lectures on Arti fi cial Intelligence and Machine LearningVol.3,No.1,1-130,2009.
[16]Cabral,R.S.;de la Torre,F.;Costeira,J.P.; Bernardino, A. Matrix completion for multilabel image classi fi cation.In: Advances in Neural Information Processing Systems 24,2011.Available at http://papers.nips.cc/paper/4419-matrix-completionfor-multi-label-image-classi fi cation.pdf.
[17]Goldberg,A.;Recht,B.;Xu,J.;Nowak,R.;Zhu, X.Transductionwith matrixcompletion: Three birdswith onestone.In: Advancesin Neural Information Processing Systems 23,2010.Available at http://papers.nips.cc/paper/3932-transduction-withmatrix-completion-three-birds-with-one-stone.pdf.
[18]Cand`es,E.J.;Li,X.;Ma,Y.;Wright,J.Robust principal component analysis?Journal of the ACMVol.58,No.3,Article No.11,2009.
[19]Cand`es,E.J.;Recht,B.Exact matrix completion via convex optimization.Foundations of Computational MathematicsVol.9,No.6,717-772,2009.
[20]Fazel, M. Matrix rank minimization with applications. Ph.D. Thesis. Stanford University, 2002.
[21]Cand`es,E.J.;Tao,T.The power of convex relaxation: Near-optimal matrix completion.IEEE Transactions on Information TheoryVol.56,No.5,2053-2080,2010.
[22]Cai,J.-F.;Cand`es,E.J.;Shen,Z.A singular value thresholding algorithm for matrix completion.SIAM Journal on OptimizationVol.20,No.4,1956-1982, 2010.
[23]Ma,S.; Goldfarb,D.; Chen,L.Fixed point andBregmaniterativemethodsformatrixrank minimization.Mathematical ProgrammingVol.128, Nos.1-2,321-353,2011.
[24]Lin,Z.;Chen,M.;Ma,Y.The augmented Lagrange multiplier method for exact recovery of corrupted lowrank matrices.arXiv:1009.5055,2010.
[25]Liu, G.; Lin, Z.; Yu, Y.Robust subspace segmentation by low-rank representation. In: Proceedingsof the 27th InternationalConference on Machine Learning, 2010. Available at http://www.icml2010.org/papers/521.pdf.
[26]Wright,J.; Ganesh,A.; Rao, S.; Peng, Y.; Ma, Y. Robust principal component analysis: Exact recovery of corrupted low-rank matrices via convex optimization.In:Advances in Neural Information Processing Systems 22,2009.Available at http://papers.nips.cc/paper/3704-robust-principalcomponent-analysis-exact-recovery-of-corrupted-lowrank-matrices-via-convex-optimization.pdf.
[27]Hale,E.T.;Yin,W.;Zhang,Y.Fixed-point continuation forl1-minimization:Methodology and convergence.SIAM Journal on OptimizationVol.19, No.3,1107-1130,2008.
[28]Nocedal,J.;Wright,S.Numerical Optimization.New York,NY,USA:Springer,1999.
[29]Von Luxburg, U. A tutorial on spectral clustering.Statistics and ComputingVol.17,No.4, 395-416,2007.
[30]Shi,J.; Malik,J.Normalized cutsand image segmentation.IEEE Transactions on Pattern Analysis and Machine IntelligenceVol.22,No.8,888-905,2000.
[31]Oliva,A.;Torralba,A.Modelingtheshapeof the scene: A holistic representation of the spatial envelope.International Journal of Computer VisionVol.42,No.3,145-175,2001.
[32]Bosch, A.; Zisserman, A.; Mu?noz, X.Scene classi fi cation via pLSA.Lecture Notes in Computer ScienceVol.3954,517-530,2006.
[33]Lu,Z.;Ip,H.H.S.Image categorization by learning with context and consistency.In:IEEE Conference on Computer Vision and Pattern Recognition,2719-2726, 2009.
[34]Strehl,A.;Ghosh,J.Cluster ensembles—a knowledge reuse framework for combining multiple partitions.The Journal of Machine Learning ResearchVol.3,583-617, 2003.
Zhenyong Fu received thePh.D. degree in the Department of Computer Science and Engineering,Shanghai Jiao Tong University,China.He is currently an assistant professor in the College of Computer,Nanjing University of Posts and Telecommunications,China. His research interests include machine learning,computer vision,and multimedia processing.
Other papers from this open access journal are available at no cost from http://www.springer.com/journal/41095. To submit a manuscript,please go to https://www. editorialmanager.com/cvmj.
1 College of Computer,Nanjing University of Posts and Telecommunications,Nanjing 210003,China.E-mail: fuzy@njupt.edu.cn.
Manuscript received:2014-12-05;accepted:2015-03-18
Computational Visual Media2015年3期