nips nips2004 nips2004-30 nips2004-30-reference knowledge-graph by maker-knowledge-mining

30 nips-2004-Binet-Cauchy Kernels


Source: pdf

Author: Alex J. Smola, S.v.n. Vishwanathan

Abstract: We propose a family of kernels based on the Binet-Cauchy theorem and its extension to Fredholm operators. This includes as special cases all currently known kernels derived from the behavioral framework, diffusion processes, marginalized kernels, kernels on graphs, and the kernels on sets arising from the subspace angle approach. Many of these kernels can be seen as the extrema of a new continuum of kernel functions, which leads to numerous new special cases. As an application, we apply the new class of kernels to the problem of clustering of video sequences with encouraging results. 1


reference text

[1] C. J. C. Burges and V. Vapnik. A new method for constructing artificial neural networks. Interim technical report, ONR contract N00014 - 94-c-0186, AT&T; Bell Laboratories, 1995.

[2] K. De Cock and B. De Moor. Subspace angles between ARMA models. Systems and Control Letter, 46:265 – 270, 2002.

[3] C. Cortes, P. Haffner, and M. Mohri. Rational kernels. In Proceedings of Neural Information Processing Systems 2002, 2002. in press.

[4] G. Doretto, A. Chiuso, Y.N. Wu, and S. Soatto. Dynamic textures. International Journal of Computer Vision, 51(2):91 – 109, 2003.

[5] J. D. Gardiner, A. L. Laub, J. J. Amato, and C. B. Moler. Solution of the Sylvester matrix equation AXB ⊤ + CXD⊤ = E. ACM Transactions on Mathematical Software, 18(2):223 – 231, 1992.

[6] T. G¨ rtner, P.A. Flach, and S. Wrobel. On graph kernels: Hardness results and effia cient alternatives. In B. Sch¨ lkopf and M. Warmuth, editors, Sixteenth Annual Cono ference on Computational Learning Theory and Seventh Kernel Workshop, COLT. Springer, 2003.

[7] W. Gr¨ bner. Matrizenrechnung. BI Hochschultaschenb¨ cher, 1965. o u

[8] H. Kashima, K. Tsuda, and A. Inokuchi. Marginalized kernels between labeled graphs. In Proceedings of the 20th International Conference on Machine Learning (ICML), Washington, DC, United States, 2003.

[9] I.R. Kondor and J. D. Lafferty. Diffusion kernels on graphs and other discrete structures. In Proceedings of the ICML, 2002.

[10] R.J. Martin. A metric for ARMA processes. IEEE Transactions on Signal Processing, 48(4):1164 – 1170, 2000.

[11] A. Pinkus. Spectral properties of totally positive kernels and matrices. In M. Gasca and C. A. Miccheli, editors, Total Positivity and its Applications, volume 359 of Mathematics and its Applications, pages 1–35. Kluwer, March 1996.

[12] S. Roweis and L. K. Saul. Nonlinear dimensionality reduction by locally linear embedding. Science, 290:2323 – 2326, December 2000.

[13] A.J. Smola and I.R. Kondor. Kernels and regularization on graphs. In B. Sch¨ lkopf o and M. K. Warmuth, editors, Proceedings of the Annual Conference on Computational Learning Theory, Lecture Notes in Computer Science. Springer, 2003.

[14] A.J. Smola, R. Vidal, and S.V.N. Vishwanathan. Kernels and dynamical systems. Automatica, 2004. submitted.

[15] S.V.N. Vishwanathan and A.J. Smola. Fast kernels on strings and trees. In Proceedings of Neural Information Processing Systems 2002, 2002.

[16] J. C. Willems. From time series to linear system. I. Finite-dimensional linear time invariant systems. Automatica J. IFAC, 22(5):561 – 580, 1986.

[17] L. Wolf and A. Shashua. Learning over sets using kernel principal angles. Jounal of Machine Learning Research, 4:913 – 931, 2003.