Multiple Kernel Learning: A Unifying Probabilistic Viewpoint

Multiple Kernel Learning: A Unifying Probabilistic Viewpoint
Notice: This research summary and analysis were automatically generated using AI technology. For absolute accuracy, please refer to the [Original Paper Viewer] below or the Original ArXiv Source.

We present a probabilistic viewpoint to multiple kernel learning unifying well-known regularised risk approaches and recent advances in approximate Bayesian inference relaxations. The framework proposes a general objective function suitable for regression, robust regression and classification that is lower bound of the marginal likelihood and contains many regularised risk approaches as special cases. Furthermore, we derive an efficient and provably convergent optimisation algorithm.


šŸ’” Research Summary

This paper presents a unified probabilistic framework that bridges the two dominant paradigms in kernel learning: regularised risk minimisation (the ā€œclassicalā€ SVM/ridge‑regression view) and Bayesian Gaussian‑process (GP) modelling. The authors start by formalising the kernel as a linear combination of base kernels, K(Īø)=āˆ‘ā‚˜Īøā‚˜Kā‚˜, with non‑negative coefficients Īø. In the regularised‑risk setting the objective is
ā€ƒā€ƒuįµ€K(Īø)^{-1}u + Cāˆ‘ā‚āæā„“(yįµ¢,uįµ¢) ,
where ā„“ is a loss function and C a trade‑off parameter. By the representer theorem the solution can be expressed as u=K(Īø)α, turning the problem into a finite‑dimensional optimisation over α.

From a Bayesian perspective the same setting is described by a GP prior P(u|Īø)=š’©(u|0,K(Īø)) and a likelihood P(y|u). The joint MAP estimate then minimises
ā€ƒā€ƒuįµ€K(Īø)^{-1}u – 2āˆ‘ā‚āæ log P(yįµ¢|uįµ¢) + log|K(Īø)| .
The log‑determinant term log|K(Īø)| is the normalising constant of the prior; it grows without bound when any Īøā‚˜ā†’āˆž, thereby automatically penalising overly complex kernels (an Occam’s‑razor effect).

Multiple kernel learning (MKL) traditionally adds a convex regulariser Ī»ā€–Īøā€–ā‚šįµ– (or an ā„“ā‚šā€‘norm ball constraint) to the risk objective, yielding
ā€ƒā€ƒĻ†_MKL(Īø)=min_u


Comments & Academic Discussion

Loading comments...

Leave a Comment