
This Article  
 
Share  
Bibliographic References  
Add to:  
Digg Furl Spurl Blink Simpy Del.icio.us Y!MyWeb  
Search  
 
ASCII Text  x  
Charanpal Dhanjal, Steve R. Gunn, John ShaweTaylor, "Efficient Sparse Kernel Feature Extraction Based on Partial Least Squares," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 31, no. 8, pp. 13471361, August, 2009.  
BibTex  x  
@article{ 10.1109/TPAMI.2008.171, author = {Charanpal Dhanjal and Steve R. Gunn and John ShaweTaylor}, title = {Efficient Sparse Kernel Feature Extraction Based on Partial Least Squares}, journal ={IEEE Transactions on Pattern Analysis and Machine Intelligence}, volume = {31}, number = {8}, issn = {01628828}, year = {2009}, pages = {13471361}, doi = {http://doi.ieeecomputersociety.org/10.1109/TPAMI.2008.171}, publisher = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, }  
RefWorks Procite/RefMan/Endnote  x  
TY  JOUR JO  IEEE Transactions on Pattern Analysis and Machine Intelligence TI  Efficient Sparse Kernel Feature Extraction Based on Partial Least Squares IS  8 SN  01628828 SP1347 EP1361 EPD  13471361 A1  Charanpal Dhanjal, A1  Steve R. Gunn, A1  John ShaweTaylor, PY  2009 KW  Machine learning KW  kernel methods KW  feature extraction KW  partial least squares (PLS). VL  31 JA  IEEE Transactions on Pattern Analysis and Machine Intelligence ER   
[1] B.E. Boser, I.M. Guyon, and V.N. Vapnik, “A Training Algorithm for Optimal Margin Classifiers,” Proc. Fifth Ann. Conf. Computational Learning Theory, pp. 144152, 1992.
[2] J. Weston, S. Mukherjee, O. Chapelle, M. Pontil, T. Poggio, and V. Vapnik, “Feature Selection for SVMs,” Advances in Neural Information Processing Systems 13, T.K. Leen, T.G. Dietterich, and V. Tresp, eds., pp. 668674, MIT Press, 2000.
[3] J. Bi, K.P. Bennett, M.J. Embrechts, C.M. Breneman, and M. Song, “Dimensionality Reduction via Sparse Support Vector Machines,” J. Machine Learning Research, vol. 3, pp. 12291243, 2003.
[4] J.H. Friedman and J.W. Tukey, “A Projection Pursuit Algorithm for Exploratory Data Analysis,” IEEE Trans. Computers, vol. 23, no. 9, pp. 881889, Sept. 1974.
[5] P.J. Huber, “Projection Pursuit,” The Annals of Statistics, vol. 13, no. 2, pp. 435475, 1985.
[6] A.J. Smola, O.L. Mangasarian, and B. Scholkopf, “Sparse Kernel Feature Analysis,” technical report, Data Mining Inst., Univ. of WisconsinMadison, 1999.
[7] H. Hotelling, “Analysis of a Complex of Statistical Variables into Principle Components,” J. Educational Psychology, vol. 24, pp. 417441 and 498520, 1933.
[8] H. Wold, “Estimation of Principal Components and Related Models by Iterative Least Squares,” Multivariate Analysis, pp. 391420, 1966.
[9] M. Momma and K.P. Bennett, “Constructing Orthogonal Latent Features for Arbitrary Loss,” Feature Extraction, Foundations and Applications, I.M. Guyon, S.R. Gunn, M. Nikravesh, and L. Zadeh, eds., pp. 551583, 2005.
[10] N. Cristianini, J. ShaweTaylor, A. Elisseeff, and J.S. Kandola, “On KernelTarget Alignment,” Advances in Neural Information Processing Systems 14, T.G. Dietterich, S. Becker, and Z. Ghahramani, eds., pp. 367373, 2001.
[11] C. Dhanjal, S.R. Gunn, and J. ShaweTaylor, “Sparse Feature Extraction Using Generalised Partial Least Squares,” Proc. IEEE Int'l Workshop Machine Learning for Signal Processing, pp. 2732, 2006.
[12] W. Massy, “Principal Components Regression in Exploratory Statistical Research,” J. Am. Statistical Assoc., vol. 60, no. 309, pp.234256, 1965.
[13] B. Schölkopf, A.J. Smola, and K.R. Müller, “Nonlinear Component Analysis as a Kernel Eigenvalue Problem,” Neural Computation, vol. 10, no. 5, pp. 12991319, 1998.
[14] A. d'Aspremont, L. El Ghaoui, M.I. Jordan, and G.R.G. Lanckriet, “A Direct Formulation for Sparse PCA Using Semidefinite Programming,” Technical Report UCB/CSD041330, Electrical Eng. and Computer Science Dept., Univ. of California, Berkeley, June 2004.
[15] B. Moghaddam, Y. Weiss, and S. Avidan, “Spectral Bounds for Sparse PCA: Exact and Greedy Algorithms,” Advances in Neural Information Processing Systems 18, Y. Weiss, B. Schölkopf, and J.Platt, eds., pp. 915922, 2006.
[16] B. Moghaddam, Y. Weiss, and S. Avidan, “Generalized Spectral Bounds for Sparse LDA,” Proc. 23rd Int'l Conf. Machine Learning, W.W.Cohen and A. Moore, eds., pp. 641648, 2006.
[17] M. Barker and W. Rayens, “Partial Least Squares for Discrimination,” J. Chemometrics, vol. 17, no. 3, pp. 166173, 2003.
[18] J. ShaweTaylor and N. Cristianini, Kernel Methods for Pattern Analysis. Cambridge Univ. Press, 2004.
[19] R. Rosipal and N. Kramer, “Overview and Recent Advances in Partial Least Squares,” Subspace, Latent Structure and Feature Selection Techniques, pp. 3451, 2006.
[20] R. Manne, “Analysis of Two PartialLeastSquares Algorithms for Multivariate Calibration,” Chemometrics and Intelligent Laboratory Systems, vol. 2, no. 1, pp. 187197, 1987.
[21] M. Stone and R.J. Brooks, “Continuum Regression: CrossValidated Sequentially Constructed Prediction Embracing Ordinary Least Squares, Partial Least Squares and Principal Components Regression,” J. Royal Statistical Soc. Series B (Methodological), vol. 52, no. 2, pp. 237269, 1990.
[22] R. Rosipal and L.J. Trejo, “Kernel Partial Least Squares Regression in Reproducing Kernel Hilbert Space,” J. Machine Learning Research, vol. 2, pp. 97123, 2001.
[23] R. Rosipal, L.J. Trejo, and B. Matthews, “Kernel PLSSVC for Linear and Nonlinear Classification,” Proc. 12th Int'l Conf. Machine Learning, A. Prieditis and S.J. Russell, eds., pp. 640647, 2003.
[24] Y. Freund and R. Schapire, “A Short Introduction to Boosting,” J.Japanese Soc. for Artificial Intelligence, vol. 14, no. 5, pp. 771780, Sept. 1999.
[25] K. Crammer, J. Keshet, and Y. Singer, “Kernel Design Using Boosting,” Advances in Neural Information Processing Systems 15, S.Becker, S. Thrun, and K. Obermayer, eds., pp. 537544, 2002.
[26] M. Momma and K.P. Bennett, “Sparse Kernel Partial Least Squares Regression,” Proc. 16th Ann. Conf. Computational Learning Theory, B. Schölkopf and M.K. Warmuth, eds., pp. 216230, 2003.
[27] H. Drucker, C.J.C. Burges, L. Kaufman, A. Smola, and V. Vapnik, “Support Vector Regression Machines,” Advances in Neural Information Processing Systems 9, M. Mozer, M.I. Jordan, and T.Petsche, eds., pp. 155161, 1997.
[28] M. Momma, “Efficient Computations via Scalable Sparse Kernel Partial Least Squares and Boosted Latent Features,” Proc. 11th ACM Int'l Conf. Knowledge Discovery in Data Mining, pp. 654659, 2005.
[29] J. ArenasGarcía, K.B. Petersen, and L.K. Hansen, “Sparse Kernel Orthonormalized PLS for Feature Extraction in Large Data Sets,” Advances in Neural Information Processing Systems 19, B. Schölkopf, J.C. Platt, and T. Hoffman, eds., pp. 3340, 2006.
[30] K.J. Worsley, J.B. Poline, K.J. Friston, and A.C. Evans, “Characterizing the Response of PET and fMRI Data Using Multivariate Linear Models,” Neuroimage, vol. 6, no. 4, pp. 305319, 1997.
[31] G. Strang, Introduction to Linear Algebra, third ed. Wellesley Cambridge Press, 2003.
[32] J. ShaweTaylor, C.K.I. Williams, N. Cristianini, and J.S. Kandola, “On the Eigenspectrum of the Gram Matrix and the Generalization Error of Kernel PCA,” IEEE Trans. Information Theory, vol. 51, no. 7, pp. 25102522, 2005.
[33] M. Ledoux and M. Talagrand, Probability in Banach Spaces: Isoperimetry and Processes. Springer, May 1991.
[34] P.L. Bartlett and S. Mendelson, “Rademacher and Gaussian Complexities: Risk Bounds and Structural Results,” J. Machine Learning Research, vol. 3, pp. 463482, 2003.
[35] H.A. Guvenir and I. Uysal, Bilkent University Function Approximation Repository, http:/funapp.cs.bilkent.edu.tr, 2000.
[36] D.J. Newman, S. Hettich, C.L. Blake, and C.J. Merz, UCI Repository of Machine Learning Databases, http://www.ics.uci.edu/mlearnMLRepository.html , 1998.
[37] T. Rose, M. Stevenson, and M. Whitehead, “The Reuters Corpus Volume 1—From Yesterday's News to Tomorrow's Language Resources,” Proc. Third Int'l Conf. Language Resources and Evaluation, pp. 827832, 2002.
[38] C.C. Chang and C.J. Lin, LIBSVM: A Library for Support Vector Machines, software available at http://www.csie.ntu.edu.tw/ ~cjlinlibsvm , 2001.
[39] V.N. Vapnik, Statistical Learning Theory. WileyInterscience, Sept. 1998.
[40] T. Joachims, “Text Categorization with Support Vector Machines: Learning with Many Relevant Features,” Proc. 10th European Conf. Machine Learning, C. Nédellec and C. Rouveirol, eds., pp. 137142, 1998.
[41] T. Joachims, “Training Linear SVMs in Linear Time,” Proc. 12th ACM SIGKDD Int'l Conf. Knowledge Discovery and Data Mining, T.EliassiRad, L.H. Ungar, M. Craven, and D. Gunopulos, eds., pp. 217226, 2006.
[42] W.Y. Zhao, R. Chellappa, P.J. Phillips, and A. Rosenfeld, “Face Recognition: A Literature Survey,” ACM Computing Surveys, vol. 35, no. 4, pp. 399458, 2003.
[43] M.A. Turk and A.P. Pentland, “Eigenfaces for Recognition,” J.Cognitive Neuroscience, vol. 3, no. 1, pp. 7186, 1991.
[44] MIT CBCL Face Database 1. Center for Biological and Computational Learning, MIT, http://www.ai.mit.edu/projectscbcl, 1996.
[45] H.A. Rowley, S. Baluja, and T. Kanade, “Neural NetworkBased Face Detection,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 20, no. 1, pp. 2338, Jan. 1998.
[46] P. Viola and M.J. Jones, “Robust RealTime Face Detection,” Int'l J. Computer Vision, vol. 57, no. 2, pp. 137154, May 2004.
[47] B. Heisele, T. Poggio, and M. Pontil, “Face Detection in Still Gray Images,” A.I. Memo 1687, Center for Biological and Computational Learning, Massachusetts Inst. of Tech nology, 2000.