Author: C. T. Kelley
Release Date: 1999-01-01
Genre: Iterative methods (Mathematics)
This book presents a carefully selected group of methods for unconstrained and bound constrained optimization problems and analyzes them in depth both theoretically and algorithmically. It focuses on clarity in algorithmic description and analysis rather than generality, and while it provides pointers to the literature for the most general theoretical results and robust software, the author thinks it is more important that readers have a complete understanding of special cases that convey essential ideas. A companion to Kelley's book, Iterative Methods for Linear and Nonlinear Equations (SIAM, 1995), this book contains many exercises and examples and can be used as a text, a tutorial for self-study, or a reference. Iterative Methods for Optimization does more than cover traditional gradient-based optimization: it is the first book to treat sampling methods, including the Hooke-Jeeves, implicit filtering, MDS, and Nelder-Mead schemes in a unified way, and also the first book to make connections between sampling methods and the traditional gradient-methods. Each of the main algorithms in the text is described in pseudocode, and a collection of MATLAB codes is available. Thus, readers can experiment with the algorithms in an easy way as well as implement them in other languages.
Author: Lap Chi Lau
Publisher: Cambridge University Press
Release Date: 2011-04-18
With the advent of approximation algorithms for NP-hard combinatorial optimization problems, several techniques from exact optimization such as the primal-dual method have proven their staying power and versatility. This book describes a simple and powerful method that is iterative in essence and similarly useful in a variety of settings for exact and approximate optimization. The authors highlight the commonality and uses of this method to prove a variety of classical polyhedral results on matchings, trees, matroids and flows. The presentation style is elementary enough to be accessible to anyone with exposure to basic linear algebra and graph theory, making the book suitable for introductory courses in combinatorial optimization at the upper undergraduate and beginning graduate levels. Discussions of advanced applications illustrate their potential for future application in research in approximation algorithms.
Author: C. T. Kelley
Release Date: 1995
Genre: Iterative methods (Mathematics)
Linear and nonlinear systems of equations are the basis for many, if not most, of the models of phenomena in science and engineering, and their efficient numerical solution is critical to progress in these areas. This is the first book to be published on nonlinear equations since the mid-1980s. Although it stresses recent developments in this area, such as Newton-Krylov methods, considerable material on linear equations has been incorporated. This book focuses on a small number of methods and treats them in depth. The author provides a complete analysis of the conjugate gradient and generalized minimum residual iterations as well as recent advances including Newton-Krylov methods, incorporation of inexactness and noise into the analysis, new proofs and implementations of Broyden's method, and globalization of inexact Newton methods. Examples, methods, and algorithmic choices are based on applications to infinite dimensional problems such as partial differential equations and integral equations. The analysis and proof techniques are constructed with the infinite dimensional setting in mind and the computational examples and exercises are based on the MATLAB environment.
The book is designed for researchers, students and practitioners interested in using fast and efficient iterative methods to approximate solutions of nonlinear equations. The following four major problems are addressed. Problem 1: Show that the iterates are well defined. Problem 2: concerns the convergence of the sequences generated by a process and the question of whether the limit points are, in fact solutions of the equation. Problem 3: concerns the economy of the entire operations. Problem 4: concerns with how to best choose a method, algorithm or software program to solve a specific type of problem and its description of when a given algorithm succeeds or fails. The book contains applications in several areas of applied sciences including mathematical programming and mathematical economics. There is also a huge number of exercises complementing the theory. - Latest convergence results for the iterative methods - Iterative methods with the least computational cost - Iterative methods with the weakest convergence conditions - Open problems on iterative methods
The purpose of this book is to provide the mathematical foundations of numerical methods, to analyze their basic theoretical properties and to demonstrate their performances on examples and counterexamples. Within any specific class of problems, the most appropriate scientific computing algorithms are reviewed, their theoretical analyses are carried out and the expected results are verified using the MATLAB software environment. Each chapter contains examples, exercises and applications of the theory discussed to the solution of real-life problems. While addressed to senior undergraduates and graduates in engineering, mathematics, physics and computer sciences, this text is also valuable for researchers and users of scientific computing in a large variety of professional fields.
In this book, which focuses on the use of iterative methods for solving large sparse systems of linear equations, templates are introduced to meet the needs of both the traditional user and the high-performance specialist. Templates, a description of a general algorithm rather than the executable object or source code more commonly found in a conventional software library, offer whatever degree of customization the user may desire. Templates offer three distinct advantages: they are general and reusable; they are not language specific; and they exploit the expertise of both the numerical analyst, who creates a template reflecting in-depth knowledge of a specific numerical technique, and the computational scientist, who then provides "value-added" capability to the general template description, customizing it for specific needs. For each template that is presented, the authors provide: a mathematical description of the flow of algorithm; discussion of convergence and stopping criteria to use in the iteration; suggestions for applying a method to special matrix types; advice for tuning the template; tips on parallel implementations; and hints as to when and why a method is useful.
Author: Tomas B. Co
Publisher: Cambridge University Press
Release Date: 2013-06-28
Genre: Technology & Engineering
Based on course notes from over twenty years of teaching engineering and physical sciences at Michigan Technological University, Tomas Co's engineering mathematics textbook is rich with examples, applications and exercises. Professor Co uses analytical approaches to solve smaller problems to provide mathematical insight and understanding, and numerical methods for large and complex problems. The book emphasises applying matrices with strong attention to matrix structure and computational issues such as sparsity and efficiency. Chapters on vector calculus and integral theorems are used to build coordinate-free physical models with special emphasis on orthogonal co-ordinates. Chapters on ODEs and PDEs cover both analytical and numerical approaches. Topics on analytical solutions include similarity transform methods, direct formulas for series solutions, bifurcation analysis, Lagrange–Charpit formulas, shocks/rarefaction and others. Topics on numerical methods include stability analysis, DAEs, high-order finite-difference formulas, Delaunay meshes, and others. MATLAB® implementations of the methods and concepts are fully integrated.
Matrix algorithms are at the core of scientific computing and are indispensable tools in most applications in engineering. This book offers a comprehensive and up-to-date treatment of modern methods in matrix computation. It uses a unified approach to direct and iterative methods for linear systems, least squares and eigenvalue problems. A thorough analysis of the stability, accuracy, and complexity of the treated methods is given. Numerical Methods in Matrix Computations is suitable for use in courses on scientific computing and applied technical areas at advanced undergraduate and graduate level. A large bibliography is provided, which includes both historical and review papers as well as recent research papers. This makes the book useful also as a reference and guide to further study and research work.
This revised edition discusses numerical methods for computing eigenvalues and eigenvectors of large sparse matrices. It provides an in-depth view of the numerical methods that are applicable for solving matrix eigenvalue problems that arise in various engineering and scientific applications. Each chapter was updated by shortening or deleting outdated topics, adding topics of more recent interest, and adapting the Notes and References section. Significant changes have been made to Chapters 6 through 8, which describe algorithms and their implementations and now include topics such as the implicit restart techniques, the Jacobi-Davidson method, and automatic multilevel substructuring.
Author: Andreas Hohmann
Publisher: Springer Science & Business Media
Release Date: 2012-12-06
This book introduces the main topics of modern numerical analysis: sequence of linear equations, error analysis, least squares, nonlinear systems, symmetric eigenvalue problems, three-term recursions, interpolation and approximation, large systems and numerical integrations. The presentation draws on geometrical intuition wherever appropriate and is supported by a large number of illustrations, exercises, and examples.
Author: David M. Young
Release Date: 2014-05-10
Iterative Solution of Large Linear Systems describes the systematic development of a substantial portion of the theory of iterative methods for solving large linear systems, with emphasis on practical techniques. The focal point of the book is an analysis of the convergence properties of the successive overrelaxation (SOR) method as applied to a linear system where the matrix is "consistently ordered". Comprised of 18 chapters, this volume begins by showing how the solution of a certain partial differential equation by finite difference methods leads to a large linear system with a sparse matrix. The next chapter reviews matrix theory and the properties of matrices, as well as several theorems of matrix theory without proof. A number of iterative methods, including the SOR method, are then considered. Convergence theorems are also given for various iterative methods under certain assumptions on the matrix A of the system. Subsequent chapters deal with the eigenvalues of the SOR method for consistently ordered matrices; the optimum relaxation factor; nonstationary linear iterative methods; and semi-iterative methods. This book will be of interest to students and practitioners in the fields of computer science and applied mathematics.