Skip Nav Destination
Close Modal
Update search
NARROW
Format
Journal
Date
Availability
1-1 of 1
Naotake Fujita
Close
Follow your search
Access your saved searches in your account
Would you like to receive an alert when new items match your search?
Sort by
Journal Articles
Publisher: Journals Gateway
Neural Computation (1992) 4 (4): 605–618.
Published: 01 July 1992
Abstract
View article
PDF
If machines are learning to make decisions given a number of examples, the generalization error ε( t ) is defined as the average probability that an incorrect decision is made for a new example by a machine when trained with t examples. The generalization error decreases as t increases, and the curve ε( t ) is called a learning curve. The present paper uses the Bayesian approach to show that given the annealed approximation, learning curves can be classified into four asymptotic types. If the machine is deterministic with noiseless teacher signals, then (1) ε ∼ at -1 when the correct machine parameter is unique, and (2) ε ∼ at -2 when the set of the correct parameters has a finite measure. If the teacher signals are noisy, then (3) ε ∼ at -1/2 for a deterministic machine, and (4) ε ∼ c + at -1 for a stochastic machine.