Please use this identifier to cite or link to this item: http://hdl.handle.net/20.500.11861/7564
Title: Generalized regularized least-squares learning with predefined features in a Hilbert space
Authors: Li, Wenye 
Lee, Kin-Hong 
Prof. LEUNG Kwong Sak 
Issue Date: 2007
Source: Advances in Neural Information Processing Systems, 2007, pp. 881 - 888
Journal: Advances in Neural Information Processing Systems 
Abstract: Kernel-based regularized learning seeks a model in a hypothesis space by minimizing the empirical error and the model's complexity. Based on the representer theorem, the solution consists of a linear combination of translates of a kernel. This paper investigates a generalized form of representer theorem for kernel-based learning. After mapping predefined features and translates of a kernel simultaneously onto a hypothesis space by a specific way of constructing kernels, we proposed a new algorithm by utilizing a generalized regularizer which leaves part of the space unregularized. Using a squared-loss function in calculating the empirical error, a simple convex solution is obtained which combines predefined features with translates of the kernel. Empirical evaluations have confirmed the effectiveness of the algorithm for supervised learning tasks.
Type: Conference Paper
URI: http://hdl.handle.net/20.500.11861/7564
ISBN: 978-026219568-3
ISSN: 10495258
Appears in Collections:Applied Data Science - Publication

Show full item record

Page view(s)

30
Last Week
1
Last month
checked on Nov 24, 2024

Google ScholarTM

Impact Indices

PlumX

Metrics


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.