Please use this identifier to cite or link to this item:
http://hdl.handle.net/20.500.11861/7564
Title: | Generalized regularized least-squares learning with predefined features in a Hilbert space |
Authors: | Li, Wenye Lee, Kin-Hong Prof. LEUNG Kwong Sak |
Issue Date: | 2007 |
Source: | Advances in Neural Information Processing Systems, 2007, pp. 881 - 888 |
Journal: | Advances in Neural Information Processing Systems |
Abstract: | Kernel-based regularized learning seeks a model in a hypothesis space by minimizing the empirical error and the model's complexity. Based on the representer theorem, the solution consists of a linear combination of translates of a kernel. This paper investigates a generalized form of representer theorem for kernel-based learning. After mapping predefined features and translates of a kernel simultaneously onto a hypothesis space by a specific way of constructing kernels, we proposed a new algorithm by utilizing a generalized regularizer which leaves part of the space unregularized. Using a squared-loss function in calculating the empirical error, a simple convex solution is obtained which combines predefined features with translates of the kernel. Empirical evaluations have confirmed the effectiveness of the algorithm for supervised learning tasks. |
Type: | Conference Paper |
URI: | http://hdl.handle.net/20.500.11861/7564 |
ISBN: | 978-026219568-3 |
ISSN: | 10495258 |
Appears in Collections: | Applied Data Science - Publication |
Find@HKSYU Show full item record
Page view(s)
30
Last Week
1
1
Last month
checked on Nov 24, 2024
Google ScholarTM
Impact Indices
PlumX
Metrics
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.