A Derivative-Free Algorithm for Least-Squares Minimization
نویسندگان
چکیده
We develop a framework for a class of derivative-free algorithms for the least-squares minimization problem. These algorithms are designed to take advantage of the problem structure by building polynomial interpolation models for each function in the least-squares minimization. Under suitable conditions, global convergence of the algorithm is established within a trust region framework. Promising numerical results indicate the algorithm is both efficient and robust. Numerical comparisons are made with standard derivative-free software packages that do not exploit the special structure of the least-squares problem or that use finite differences to approximate the gradients.
منابع مشابه
On the local convergence of a derivative-free algorithm for least-squares minimization
In Zhang et al. (accepted by SIAM J. Optim., 2010), we developed a class of derivative-free algorithms, called DFLS, for least-squares minimization. Global convergence of the algorithm as well as its excellent numerical performance within a limited computational budget was established and discussed in the same paper. Here we would like to establish the local quadratic convergence of the algorit...
متن کاملImproving adaptive resolution of analog to digital converters using least squares mean method
This paper presents an adaptive digital resolution improvement method for extrapolating and recursive analog-to-digital converters (ADCs). The presented adaptively enhanced ADC (AE-ADC) digitally estimates the digital equivalent of the input signal by utilizing an adaptive digital filter (ADF). The least mean squares (LMS) algorithm also determines the coefficients of the ADF block. In this sch...
متن کاملA Derivative-free Algorithm for Finding Least Squares Solutions of Quasi-linear and Linear Systems
A novel derivative-free algorithm for solving quasilinear systems is presented. It resembles “classical” optimization approach but greatly simplifies computation, resulting in fast execution and numerical stability. Though the global convergence cannot be guaranteed, it turns out that the presented algorithm finds a solution as successfully as other commonly accepted methods. The algorithm is c...
متن کاملMinimization and Parameter Estimation for Seminorm Regularization Models with I-Divergence Constraints
In this papers we analyze the minimization of seminorms ‖L · ‖ on R under the constraint of a bounded I-divergence D(b,H ·) for rather general linear operators H and L. The I-divergence is also known as Kullback-Leibler divergence and appears in many models in imaging science, in particular when dealing with Poisson data. Often H represents, e.g., a linear blur operator and L is some discrete d...
متن کاملImproved Iteratively Reweighted Least Squares for Unconstrained
In this paper, we first study q minimization and its associated iterative reweighted algorithm for recovering sparse vectors. Unlike most existing work, we focus on unconstrained q minimization, for which we show a few advantages on noisy measurements and/or approximately sparse vectors. Inspired by the results in [Daubechies et al., Comm. Pure Appl. Math., 63 (2010), pp. 1–38] for constrained ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- SIAM Journal on Optimization
دوره 20 شماره
صفحات -
تاریخ انتشار 2010