Globally convergent algorithms for solving unconstrained optimization problems
- Authors: Taheri, Sona , Mammadov, Musa , Seifollahi, Sattar
- Date: 2013
- Type: Text , Journal article
- Relation: Optimization Vol. , no. (2013), p. 1-15
- Full Text:
- Reviewed:
- Description: New algorithms for solving unconstrained optimization problems are presented based on the idea of combining two types of descent directions: the direction of anti-gradient and either the Newton or quasi-Newton directions. The use of latter directions allows one to improve the convergence rate. Global and superlinear convergence properties of these algorithms are established. Numerical experiments using some unconstrained test problems are reported. Also, the proposed algorithms are compared with some existing similar methods using results of experiments. This comparison demonstrates the efficiency of the proposed combined methods.
A new method for solving linear ill-posed problems
- Authors: Zhang, Jianjun , Mammadov, Musa
- Date: 2012
- Type: Text , Journal article
- Relation: Applied Mathematics and Computation Vol. 218, no. 20 (2012), p.10180-10187
- Full Text:
- Reviewed:
- Description: In this paper, we propose a new method for solving large-scale ill-posed problems. This method is based on the Karush-Kuhn-Tucker conditions, Fisher-Burmeister function and the discrepancy principle. The main difference from the majority of existing methods for solving ill-posed problems is that, we do not need to choose a regularization parameter in advance. Experimental results show that the proposed method is effective and promising for many practical problems. © 2012.
Solving systems of nonlinear equations using a globally convergent optimization algorithm
- Authors: Taheri, Sona , Mammadov, Musa
- Date: 2012
- Type: Text , Journal article
- Relation: Global Journal of Technology & Optimization Vol. 3, no. (2012), p. 132-138
- Full Text:
- Reviewed:
- Description: Solving systems of nonlinear equations is a relatively complicated problem for which a number of different approaches have been presented. In this paper, a new algorithm is proposed for the solutions of systems of nonlinear equations. This algorithm uses a combination of the gradient and the Newton’s methods. A novel dynamic combinatory is developed to determine the contribution of the methods in the combination. Also, by using some parameters in the proposed algorithm, this contribution is adjusted. We use the gradient method due to its global convergence property, and the Newton’s method to speed up the convergence rate. We consider two different combinations. In the first one, a step length is determined only along the gradient direction. The second one is finding a step length along both the gradient and the Newton’s directions. The performance of the proposed algorithm in comparison to the Newton’s method, the gradient method and an existing combination method is explored on several well known test problems in solving systems of nonlinear equations. The numerical results provide evidence that the proposed combination algorithm is generally more robust and efficient than other mentioned methods on someimportant and difficult problems.
Solving a system of nonlinear integral equations by an RBF network
- Authors: Golbabai, A. , Mammadov, Musa , Seifollahi, Sattar
- Date: 2009
- Type: Text , Journal article
- Relation: Computers & Mathematics with Applications Vol. 57, no. 10 (2009), p. 1651-1658
- Full Text: false
- Reviewed:
- Description: In this paper, a novel learning strategy for radial basis function networks (RBFN) is proposed. By adjusting the parameters of the hidden layer, including the RBF centers and widths, the weights of the output layer are adapted by local optimization methods. A new local optimization algorithm based on a combination of the gradient and Newton methods is introduced. The efficiency of some local optimization methods to Update the weights of RBFN is Studied in solving systems of nonlinear integral equations. (C) 2009 Elsevier Ltd. All rights reserved.