You may have to Search all our reviewed books and magazines, click the sign up button below to create a free account.
The method of least squares was discovered by Gauss in 1795. It has since become the principal tool to reduce the influence of errors when fitting models to given observations. Today, applications of least squares arise in a great number of scientific areas, such as statistics, geodetics, signal processing, and control. In the last 20 years there has been a great increase in the capacity for automatic data capturing and computing. Least squares problems of large size are now routinely solved. Tremendous progress has been made in numerical methods for least squares problems, in particular for generalized and modified least squares problems and direct and iterative methods for sparse problems. Until now there has not been a monograph that covers the full spectrum of relevant problems and methods in least squares. This volume gives an in-depth treatment of topics such as methods for sparse least squares problems, iterative methods, modified least squares, weighted problems, and constrained and regularized problems. The more than 800 references provide a comprehensive survey of the available literature on the subject.
"Substantial, detailed and rigorous . . . readers for whom the book is intended are admirably served." — MathSciNet (Mathematical Reviews on the Web), American Mathematical Society. Practical text strikes fine balance between students' requirements for theoretical treatment and needs of practitioners, with best methods for large- and small-scale computing. Prerequisites are minimal (calculus, linear algebra, and preferably some acquaintance with computer programming). Text includes many worked examples, problems, and an extensive bibliography.
Matrix algorithms are at the core of scientific computing and are indispensable tools in most applications in engineering. This book offers a comprehensive and up-to-date treatment of modern methods in matrix computation. It uses a unified approach to direct and iterative methods for linear systems, least squares and eigenvalue problems. A thorough analysis of the stability, accuracy, and complexity of the treated methods is given. Numerical Methods in Matrix Computations is suitable for use in courses on scientific computing and applied technical areas at advanced undergraduate and graduate level. A large bibliography is provided, which includes both historical and review papers as well as recent research papers. This makes the book useful also as a reference and guide to further study and research work.
This new book from the authors of the classic book Numerical methods addresses the increasingly important role of numerical methods in science and engineering. More cohesive and comprehensive than any other modern textbook in the field, it combines traditional and well-developed topics with other material that is rarely found in numerical analysis texts, such as interval arithmetic, elementary functions, operator series, convergence acceleration, and continued fractions. Although this volume is self-contained, more comprehensive treatments of matrix computations will be given in a forthcoming volume. A supplementary Website contains three appendices: an introduction to matrix computations; a description of Mulprec, a MATLAB multiple precision package; and a guide to literature, algorithms, and software in numerical analysis. Review questions, problems, and computer exercises are also included. For use in an introductory graduate course in numerical analysis and for researchers who use numerical methods in science and engineering.
The method of least squares, discovered by Gauss in 1795, is a principal tool for reducing the influence of errors when fitting a mathematical model to given observations. Applications arise in many areas of science and engineering. The increased use of automatic data capturing frequently leads to large-scale least squares problems. Such problems can be solved by using recent developments in preconditioned iterative methods and in sparse QR factorization. The first edition of Numerical Methods for Least Squares Problems was the leading reference on the topic for many years. The updated second edition stands out compared to other books on this subject because it provides an in-depth and up-to...
This work addresses the increasingly important role of numerical methods in science and engineering. It combines traditional and well-developed topics with other material such as interval arithmetic, elementary functions, operator series, convergence acceleration, and continued fractions.
The method of least squares: the principal tool for reducing the influence of errors when fitting models to given observations.
Accuracy and Stability of Numerical Algorithms gives a thorough, up-to-date treatment of the behavior of numerical algorithms in finite precision arithmetic. It combines algorithmic derivations, perturbation theory, and rounding error analysis, all enlivened by historical perspective and informative quotations. This second edition expands and updates the coverage of the first edition (1996) and includes numerous improvements to the original material. Two new chapters treat symmetric indefinite systems and skew-symmetric systems, and nonlinear systems and Newton's method. Twelve new sections include coverage of additional error bounds for Gaussian elimination, rank revealing LU factorizations, weighted and constrained least squares problems, and the fused multiply-add operation found on some modern computer architectures.