Book contents
- Frontmatter
- Contents
- Preface
- 1 Introduction
- Part one Pattern Classification with Binary-Output Neural Networks
- Part two Pattern Classification with Real-Output Networks
- Part three Learning Real-Valued Functions
- 16 Learning Classes of Real Functions
- 17 Uniform Convergence Results for Real Function Classes
- 18 Bounding Covering Numbers
- 19 Sample Complexity of Learning Real Function Classes
- 20 Convex Classes
- 21 Other Learning Problems
- Part four Algorithmics
- Appendix 1 Useful Results
- Bibliography
- Author index
- Subject index
16 - Learning Classes of Real Functions
Published online by Cambridge University Press: 26 February 2010
- Frontmatter
- Contents
- Preface
- 1 Introduction
- Part one Pattern Classification with Binary-Output Neural Networks
- Part two Pattern Classification with Real-Output Networks
- Part three Learning Real-Valued Functions
- 16 Learning Classes of Real Functions
- 17 Uniform Convergence Results for Real Function Classes
- 18 Bounding Covering Numbers
- 19 Sample Complexity of Learning Real Function Classes
- 20 Convex Classes
- 21 Other Learning Problems
- Part four Algorithmics
- Appendix 1 Useful Results
- Bibliography
- Author index
- Subject index
Summary
Introduction
This part of the book examines supervised learning problems in which we require a learning system to model the relationship between a pattern and a real-valued quantity. For example, in using a neural network to predict the future price of shares on the stock exchange, or to estimate the probability that a particular patient will experience problems during a surgical procedure, the predictions are represented by the real-valued output of the network.
In the pattern classification problems studied in Parts 1 and 2, the (x, y) pairs are generated by a probability distribution on the product space X × {0, 1}. In a similar way, we assume in this part of the book that the data is generated by a probability distribution P on X × ℝ. This is a generalization of the pattern classification model, and includes a number of other data-generating processes as special cases. For example, it can model a deterministic relationship between patterns and their labels, where each (x, y) pair satisfies y = f(x) for some function f. It can model a deterministic relationship with additive independent observation noise, where yi = f(xi) + ηi, and the ηi are independent and identically distributed random variables. It can also model a noisy relationship in which the observation noise variables ηi are mutually independent, but the distribution of ηi depends on the pattern xi.
- Type
- Chapter
- Information
- Neural Network LearningTheoretical Foundations, pp. 231 - 240Publisher: Cambridge University PressPrint publication year: 1999