The Curious Case of Convex Neural Networks

نویسندگان

چکیده

This paper investigates a constrained formulation of neural networks where the output is convex function input. We show that convexity constraints can be enforced on both fully connected and convolutional layers, making them applicable to most architectures. The include restricting weights (for all but first layer) non-negative using non-decreasing activation function. Albeit simple, these have profound implications generalization abilities network. draw three valuable insights: (a) Input Output Convex Neural Networks (IOC-NNs) self regularize significantly reduce problem overfitting; (b) Although heavily constrained, they outperform base multi layer perceptrons achieve similar performance as compared architectures (c) IOC-NNs robustness noise in train labels. demonstrate efficacy proposed idea thorough experiments ablation studies six commonly used image classification datasets with different network appendix codes for this are available at: https://github.com/sarathsp1729/Convex-Networks.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

pricing unemployment insurance : the case of iran

employees always concern about losing their job , or in other word , losing their income resources. for this purpose, every government requires strong system for covering these concerns. the unemployment insurance (ui) program’s can be used for achieving this goal. in this thesis, we price ui based on the insurance history of employee and the duration of being unemployed. we use the weibull dis...

Neural Tube Closure: The Curious Case of Shrinking Junctions

Your brain and spinal cord began as a flat sheet, which narrowed, elongated, and rolled up to form a tube. A new study identifies a key molecular link underlying vertebrate neural tube formation, connecting planar cell polarity patterning to contraction of specific cell-cell junctions.

متن کامل

Input Convex Neural Networks

This paper presents the input convex neural network architecture. These are scalar-valued (potentially deep) neural networks with constraints on the network parameters such that the output of the network is a convex function of (some of) the inputs. The networks allow for efficient inference via optimization over some inputs to the network given others, and can be applied to settings including ...

متن کامل

Convex Neural Networks

Convexity has recently received a lot of attention in the machine learning community, and the lack of convexity has been seen as a major disadvantage of many learning algorithms, such as multi-layer artificial neural networks. We show that training multi-layer neural networks in which the number of hidden units is learned can be viewed as a convex optimization problem. This problem involves an ...

متن کامل

When Sparse Traditional Models Outperform Dense Neural Networks: the Curious Case of Discriminating between Similar Languages

We present the results of our participation in the VarDial 4 shared task on discriminating closely related languages. Our submission includes simple traditional models using linear support vector machines (SVMs) and a neural network (NN). The main idea was to leverage language group information. We did so with a two-layer approach in the traditional model and a multi-task objective in the neura...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Lecture Notes in Computer Science

سال: 2021

ISSN: ['1611-3349', '0302-9743']

DOI: https://doi.org/10.1007/978-3-030-86486-6_45