What Size Net Gives Valid Generalization?

نویسندگان

  • Eric B. Baum
  • David Haussler
چکیده

We address the question of when a network can be expected to generalize from m random training examples chosen from some arbitrary probability distribution, assuming that future test examples are drawn from the same distribution. Among our results are the following bounds on appropriate sample vs, network size. Assume 0 < E 5 1/8. We show that if m 2 0($209!) random examples can be loaded on a feedforward network of linear threshold functions with N nodes and W weights, so that at least a fraction 1 5 of the examples are correctly classified, then one has confidence approaching certainty that the network will correctly classify a fraction 1 E of future test examples drawn from the same distribution. Conversely, for fully-connected feedforward nets with one hidden layer, any learning algorithm using fewer than O ( F ) random training examples will, for some distributions of examples consistent with an appropriate weight choice, fail at least some fixed fraction of the time to find a weight choice that will correctly classify more than a 1 E fraction of the future test examples.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Training Recurrent Networks Using the Extended Kalman Filter

The extended Kalman lter (EKF) can be used as an on-line algorithm to determine the weights in a recurrent network given target outputs as it runs. This paper notes some relationships between the EKF as applied to recurrent net learning and some simpler techniques that are more widely used. In particular, making certain simpliications to the EKF gives rise to an algorithm essentially identical ...

متن کامل

What Size Neural Network Gives Optimal Generalization? Convergence Properties of Backpropagation

One of the most important aspects of any machine learning paradigm is how it scales according to problem size and complexity. Using a task with known optimal training error, and a pre-specified maximum number of training updates, we investigate the convergence of the backpropagation algorithm with respect to a) the complexity of the required function approximation, b) the size of the network in...

متن کامل

Generalization of a CFD Model to Predict the Net Power in PEM Fuel Cells

Qualitatively, it is known that the reactants content within the catalyst layer (CL) is the driving moments for the kinetics of reaction within the CL. This paper aimed to quantitatively express the level of enhancement in electrical power due to enrichment in the oxygen content. For a given MEA, a flow field (FF) designer is always willing to design a FF to maximize the content of oxygen in al...

متن کامل

On sumfree subsets of hypercubes

We consider the possible sizes of large sumfree sets contained in the discrete hypercube {1, . . . , n}k, and we determine upper and lower bounds for the maximal size as n becomes large. We also discuss a continuous analogue in which our lower bound remains valid and our upper bound can be strengthened, and we consider the generalization of both problems to l-fold-sumfree sets.

متن کامل

Adaptive Elastic Net GMM Estima- tion with Many Invalid Moment Conditions: Simultaneous Model and Moment Selection CENTER FOR POLICY RESEARCH –Spring 2015

This paper develops the adaptive elastic net GMM estimator in large dimensional models with many possibly invalid moment conditions, where both the number of structural parameters and the number of moment conditions may increase with the sample size. The basic idea is to conduct the standard GMM estimation combined with two penalty terms: the quadratic regularization and the adaptively weighted...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • Neural Computation

دوره 1  شماره 

صفحات  -

تاریخ انتشار 1988