nips nips2008 nips2008-143 nips2008-143-reference knowledge-graph by maker-knowledge-mining
Source: pdf
Author: Shuiwang Ji, Liang Sun, Rong Jin, Jieping Ye
Abstract: We present a multi-label multiple kernel learning (MKL) formulation in which the data are embedded into a low-dimensional space directed by the instancelabel correlations encoded into a hypergraph. We formulate the problem in the kernel-induced feature space and propose to learn the kernel matrix as a linear combination of a given collection of kernel matrices in the MKL framework. The proposed learning formulation leads to a non-smooth min-max problem, which can be cast into a semi-infinite linear program (SILP). We further propose an approximate formulation with a guaranteed error bound which involves an unconstrained convex optimization problem. In addition, we show that the objective function of the approximate formulation is differentiable with Lipschitz continuous gradient, and hence existing methods can be employed to compute the optimal solution efficiently. We apply the proposed formulation to the automated annotation of Drosophila gene expression pattern images, and promising results have been reported in comparison with representative algorithms.
[1] S. Agarwal, K. Branson, and S. Belongie. Higher order learning with graphs. In ICML, pages 17–24, 2006.
[2] D. Zhou, J. Huang, and B. Sch¨ lkopf. Learning with hypergraphs: Clustering, classification, and embedo ding. In NIPS, pages 1601–1608. 2007.
[3] Z. H. Zhou and M. L. Zhang. Multi-instance multi-label learning with application to scene classification. In NIPS, pages 1609–1616. 2007.
[4] D. R. Hardoon, S. R. Szedmak, and J. R. Shawe-taylor. Canonical correlation analysis: An overview with application to learning methods. Neural Computation, 16(12):2639–2664, 2004.
[5] Y. Nesterov. Introductory Lectures on Convex Optimization: A Basic Course. Springer, 2003.
[6] Y. Nesterov. Smooth minimization of non-smooth functions. Mathematical Programming, 103(1):127– 152, 2005.
[7] P. Tomancak and et al. Systematic determination of patterns of gene expression during Drosophila embryogenesis. Genome Biology, 3(12), 2002.
[8] S. Ji, L. Sun, R. Jin, S. Kumar, and J. Ye. Automated annotation of Drosophila gene expression patterns using a controlled vocabulary. Bioinformatics, 24(17):1881–1888, 2008.
[9] K. Grauman and T. Darrell. Approximate correspondences in high dimensions. In NIPS, pages 505–512. 2006.
[10] F. R. K. Chung. Spectral Graph Theory. American Mathematical Society, 1997.
[11] S. Sch¨ lkopf and A. Smola. Learning with Kernels: Support Vector Machines,Regularization, Optimizao tion and Beyond. MIT Press, 2002.
[12] G. R. G. Lanckriet, N. Cristianini, P. Bartlett, L. E. Ghaoui, and M. I. Jordan. Learning the kernel matrix with semidefinite programming. Journal of Machine Learning Research, 5:27–72, 2004.
[13] R. Hettich and K. O. Kortanek. Semi-infinite programming: Theory, methods, and applications. SIAM Review, 35(3):380–429, 1993.
[14] S. Sonnenburg, G. R¨ tsch, C. Sch¨ fer, and B. Sch¨ lkopf. Large scale multiple kernel learning. Journal of a a o Machine Learning Research, 7:1531–1565, July 2006.
[15] K. Mikolajczyk and C. Schmid. A performance evaluation of local descriptors. IEEE Transactions on Pattern Analysis and Machine Intelligence, 27(10):1615–1630, 2005.