Statistics – Machine Learning
Scientific paper
2012-03-02
Statistics
Machine Learning
57 pages. arXiv admin note: substantial text overlap with arXiv:1103.0431
Scientific paper
We investigate the learning rate of multiple kernel leaning (MKL) with $\ell_1$ and elastic-net regularizations. The elastic-net regularization is a composition of an $\ell_1$-regularizer for inducing the sparsity and an $\ell_2$-regularizer for controlling the smoothness. We focus on a sparse setting where the total number of kernels is large but the number of non-zero components of the ground truth is relatively small, and show sharper convergence rates than the learning rates ever shown for both $\ell_1$ and elastic-net regularizations. Our analysis shows there appears a trade-off between the sparsity and the smoothness when it comes to selecting which of $\ell_1$ and elastic-net regularizations to use; if the ground truth is smooth, the elastic-net regularization is preferred, otherwise the $\ell_1$ regularization is preferred.
Sugiyama Masashi
Suzuki Taiji
No associations
LandOfFree
Fast Learning Rate of Multiple Kernel Learning: Trade-Off between Sparsity and Smoothness does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Fast Learning Rate of Multiple Kernel Learning: Trade-Off between Sparsity and Smoothness, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Fast Learning Rate of Multiple Kernel Learning: Trade-Off between Sparsity and Smoothness will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-344473