Back to Search
Start Over
ESTIMATING THE APPROXIMATION ERROR IN LEARNING THEORY
- Source :
- Analysis and Applications. :17-41
- Publication Year :
- 2003
- Publisher :
- World Scientific Pub Co Pte Lt, 2003.
-
Abstract
- Let B be a Banach space and (ℋ,‖·‖ℋ) be a dense, imbedded subspace. For a ∈ B, its distance to the ball of ℋ with radius R (denoted as I(a, R)) tends to zero when R tends to infinity. We are interested in the rate of this convergence. This approximation problem arose from the study of learning theory, where B is the L2 space and ℋ is a reproducing kernel Hilbert space. The class of elements having I(a, R) = O(R-r) with r > 0 is an interpolation space of the couple (B, ℋ). The rate of convergence can often be realized by linear operators. In particular, this is the case when ℋ is the range of a compact, symmetric, and strictly positive definite linear operator on a separable Hilbert space B. For the kernel approximation studied in Learning Theory, the rate depends on the regularity of the kernel function. This yields error estimates for the approximation by reproducing kernel Hilbert spaces. When the kernel is smooth, the convergence is slow and a logarithmic convergence rate is presented for analytic kernels in this paper. The purpose of our results is to provide some theoretical estimates, including the constants, for the approximation error required for the learning theory.
Details
- ISSN :
- 17936861 and 02195305
- Database :
- OpenAIRE
- Journal :
- Analysis and Applications
- Accession number :
- edsair.doi...........0cdd4f510fb3c080a2fe7b11380356d5
- Full Text :
- https://doi.org/10.1142/s0219530503000089