A Free Line Search Steepest Descent Method for Solving Unconstrained Optimization Problems

نویسنده

  • Narges Bidabadi Assistant Professor, Department of Mathematical Sciences, Yazd University, Yazd, Iran
چکیده مقاله:

In this paper, we solve unconstrained optimization problem using a free line search steepest descent method. First, we propose a double parameter scaled quasi Newton formula for calculating an approximation of the Hessian matrix. The approximation obtained from this formula is a positive definite matrix that is satisfied in the standard secant relation. We also show that the largest eigen value of this matrix is not greater than the number of variables of the problem. Then, using this double parameter scaled quasi Newton formula, an explicit formula for calculating the step length in the steepest descent method is presented and therefore, this method does not require the use of approximate methods for calculating step length. The numerical results obtained from the implementation of the algorithm in MATLAB software environment are presented for some optimization problems. These results show the efficiency of the proposed method in comparison with other existing methods.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Modified Algorithm of Steepest Descent Method for Solving Unconstrained Nonlinear Optimization Problems

The steepest descent method (SDM), which can be traced back to Cauchy (1847), is the simplest gradient method for unconstrained optimization problem. The SDM is effective for well-posed and low-dimensional nonlinear optimization problems without constraints; however, for a large-dimensional system, it converges very slowly. Therefore, a modified steepest decent method (MSDM) is developed to dea...

متن کامل

SVM-Optimization and Steepest-Descent Line Search

We consider (a subclass of) convex quadratic optimization problems and analyze decomposition algorithms that perform, at least approximately, steepest-descent exact line search. We show that these algorithms, when implemented properly, are within ǫ of optimality after O(log 1/ǫ) iterations for strictly convex cost functions, and after O(1/ǫ) iterations in the general case. Our analysis is gener...

متن کامل

Steepest descent method for solving zero-one nonlinear programming problems

In this paper we use steepest descent method for solving zero-one nonlinear programming problem. Using penalty function we transform this problem to an unconstrained optimization problem and then by steepest descent method we obtain the original problem optimal solution. 2007 Elsevier Inc. All rights reserved.

متن کامل

A New Steepest Descent Differential Inclusion-Based Method for Solving General Nonsmooth Convex Optimization Problems

In this paper, we investigate a steepest descent neural network for solving general nonsmooth convex optimization problems. The convergence to optimal solution set is analytically proved. We apply the method to some numerical tests which confirm the effectiveness of the theoretical results and the performance of the proposed neural network.

متن کامل

A derivative–free nonmonotone line search technique for unconstrained optimization

A tolerant derivative–free nonmonotone line search technique is proposed and analyzed. Several consecutive increases in the objective function and also non descent directions are admitted for unconstrained minimization. To exemplify the power of this new line search we describe a direct search algorithm in which the directions are chosen randomly. The convergence properties of this random metho...

متن کامل

On the Complexity of Steepest Descent, Newton's and Regularized Newton's Methods for Nonconvex Unconstrained Optimization Problems

It is shown that the steepest descent and Newton’s method for unconstrained nonconvex optimization under standard assumptions may be both require a number of iterations and function evaluations arbitrarily close to O(ǫ) to drive the norm of the gradient below ǫ. This shows that the upper bound of O(ǫ) evaluations known for the steepest descent is tight, and that Newton’s method may be as slow a...

متن کامل

منابع من

با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ذخیره در منابع من قبلا به منابع من ذحیره شده

{@ msg_add @}


عنوان ژورنال

دوره 6  شماره 26

صفحات  159- 166

تاریخ انتشار 2020-10-22

با دنبال کردن یک ژورنال هنگامی که شماره جدید این ژورنال منتشر می شود به شما از طریق ایمیل اطلاع داده می شود.

میزبانی شده توسط پلتفرم ابری doprax.com

copyright © 2015-2023