Note on Weight Noise Injection During Training a MLP

نویسندگان

  • Kevin Ho
  • John Sum
چکیده

Although many analytical works have been done to investigate the change of prediction error of a trained NN if its weights are injected by noise, seldom of them has truly investigated on the dynamical properties (such as objective functions and convergence behavior) of injecting weight noise during training. In this paper, four different online weight noise injection training algorithms for multilayer perceptron (MLP) are analyzed and their objective functions are derived. Most importance, the objective function of injecting multiplicative weight noise during training is shown to be different from the prediction error of a trained MLP if its weights are injected by the same multiplicative weight noise. It provides a firm response to a question being posed for 14 years [8]: Can deterministic penalty terms model the effects of synaptic weight noise on network fault-tolerance?. Besides, we show that the objective function of injecting additive weight noise during training is equivalent to adding a regularizer penalizing the magnitude of the gradient vector of the MLP output with respect to its weight vector. Finally, the issue on their convergence proofs will be discussed.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Convergence analysis of on-line weight noise injection training algorithms for MLP networks

Injecting weight noise during training has been proposed for almost two decades as a simple technique to improve fault tolerance and generalization of a multilayer perceptron (MLP). However, little has been done regarding their convergence behaviors. Therefore, we presents in this paper the convergence proofs of two of these algorithms for MLPs. One is based on combining injecting multiplicativ...

متن کامل

On Weight-Noise-Injection Training

While injecting weight noise during training has been proposed for more than a decade to improve the convergence, generalization and fault tolerance of a neural network, not much theoretical work has been done to its convergence proof and the objective function that it is minimizing. By applying the Gladyshev Theorem, it is shown that the convergence of injecting weight noise during training an...

متن کامل

Enhanced MLP performance and fault tolerance resulting from synaptic weight noise during training

We analyze the effects of analog noise on the synaptic arithmetic during multilayer perceptron training, by expanding the cost function to include noise-mediated terms. Predictions are made in the light of these calculations that suggest that fault tolerance, training quality and training trajectory should be improved by such noise-injection. Extensive simulation experiments on two distinct cla...

متن کامل

SNIWD: Simultaneous Weight Noise Injection with Weight Decay for MLP Training

Despite noise injecting during training has been demonstrated with success in enhancing the fault tolerance of neural network, theoretical analysis on the dynamic of this noise injection-based online learning algorithm has far from complete. In particular, the convergence proofs for those algorithms have not been shown. In this regards, this paper presents an empirical study on the non-converge...

متن کامل

Synaptic Weight Noise during Mlp Learning Enhances Fault-tolerance, Generalisation and Learning Tra Jectory

We analyse the eeects of analog noise on the synaptic arithmetic during MultiLayer Perceptron training, by expanding the cost function to include noise-mediated penalty terms. Predictions are made in the light of these calculations which suggest that fault tolerance, generalisation ability and learning trajectory should be improved by such noise-injection. Extensive simulation experiments on tw...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2009