Learning Rates for -Regularized Kernel Classifiers

We consider a family of classification algorithms generated from a regularization kernel scheme associated with -regularizer and convex loss function. Our main purpose is to provide an explicit convergence rate for the excess misclassification error of the produced classifiers. The error decompositi...

Full description

Saved in:
Bibliographic Details
Main Authors: Hongzhi Tong, Di-Rong Chen, Fenghong Yang
Format: Article
Language:English
Published: Wiley 2013-01-01
Series:Journal of Applied Mathematics
Online Access:http://dx.doi.org/10.1155/2013/496282
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:We consider a family of classification algorithms generated from a regularization kernel scheme associated with -regularizer and convex loss function. Our main purpose is to provide an explicit convergence rate for the excess misclassification error of the produced classifiers. The error decomposition includes approximation error, hypothesis error, and sample error. We apply some novel techniques to estimate the hypothesis error and sample error. Learning rates are eventually derived under some assumptions on the kernel, the input space, the marginal distribution, and the approximation error.
ISSN:1110-757X
1687-0042