نتایج جستجو برای: modified subgradient method
تعداد نتایج: 1831354 فیلتر نتایج به سال:
In this paper, we propose a strongly convergent variant of Robinson’s subgradient algorithm for solving a system of vector convex inequalities in Hilbert spaces. The advantage of the proposed method is that it converges strongly, when the problem has solutions, under mild assumptions. The proposed algorithm also has the following desirable property: the sequence converges to the solution of the...
We study finite convergence of the modified cyclic subgradient projections (MCSP) algorithm for the convex feasibility problem (CFP) in the Euclidian space. Expanding control sequences allow the indices of the sets of the CFP to re-appear and be used again by the algorithm ∗Currently with the Department of Radiation Oncology, Massachusetts General Hospital and Harvard Medical School, Boston, Ma...
We consider minimization of nonsmooth functions which can be represented as the composition of a positively homogeneous convex function and a smooth mapping. This is a sufficiently rich class that includes max-functions, largest eigenvalue functions, and norm-1 regularized functions. The bundle method uses an oracle that is able to compute separately the function and subgradient information for...
We consider constrained optimization problems with a nonsmooth objective function in the form of mathematical expectation. The Sample Average Approximation (SAA) is used to estimate and variable sample size strategy employed. proposed algorithm combines an SAA subgradient spectral coefficient order provide suitable direction which improves performance first method as shown by numerical results....
Proximity control is a well-known mechanism in bundle method for nonsmooth optimization. Here we show that it can be used to optimize a large class of nonconvex and nonsmooth functions with additional structure. This includes for instance nonconvex maximum eigenvalue functions, and also infinite suprema of such functions.
We relate the minimax game of generative adversarial networks (GANs) to finding the saddle points of the Lagrangian function for a convex optimization problem, where the discriminator outputs and the distribution of generator outputs play the roles of primal variables and dual variables, respectively. This formulation shows the connection between the standard GAN training process and the primal...
نمودار تعداد نتایج جستجو در هر سال
با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید