Computer Solution of Linear Algebraic Systems
Author : George Elmer Forsythe
Publisher :
Page : 172 pages
File Size : 44,69 MB
Release : 1967
Category : Electronic data processing
ISBN :
Author : George Elmer Forsythe
Publisher :
Page : 172 pages
File Size : 44,69 MB
Release : 1967
Category : Electronic data processing
ISBN :
Author : Gerard Meurant
Publisher : Elsevier
Page : 777 pages
File Size : 49,64 MB
Release : 1999-06-16
Category : Mathematics
ISBN : 0080529518
This book deals with numerical methods for solving large sparse linear systems of equations, particularly those arising from the discretization of partial differential equations. It covers both direct and iterative methods. Direct methods which are considered are variants of Gaussian elimination and fast solvers for separable partial differential equations in rectangular domains. The book reviews the classical iterative methods like Jacobi, Gauss-Seidel and alternating directions algorithms. A particular emphasis is put on the conjugate gradient as well as conjugate gradient -like methods for non symmetric problems. Most efficient preconditioners used to speed up convergence are studied. A chapter is devoted to the multigrid method and the book ends with domain decomposition algorithms that are well suited for solving linear systems on parallel computers.
Author : Bruno Codenotti
Publisher : World Scientific
Page : 232 pages
File Size : 38,22 MB
Release : 1991-04-30
Category : Computers
ISBN : 9814506478
This book presents the most important parallel algorithms for the solution of linear systems. Despite the evolution and significance of the field of parallel solution of linear systems, no book is completely dedicated to the subject. People interested in the themes covered by this book belong to two different groups: numerical linear algebra and theoretical computer science, and this is the first effort to produce a useful tool for both. The book is organized as follows: after introducing the general features of parallel algorithms and the most important models of parallel computation, the authors analyze the complexity of solving linear systems in the circuit, PRAM, distributed, and VLSI models. The approach covers both the general case (i.e. dense linear systems without structure) and many important special cases (i.e. banded, sparse, Toeplitz, circulant linear systems).
Author : J. J. Dongarra
Publisher : Society for Industrial and Applied Mathematics (SIAM)
Page : 274 pages
File Size : 17,32 MB
Release : 1991
Category : Computers
ISBN :
Mathematics of Computing -- Parallelism.
Author : Alan George
Publisher : Prentice Hall
Page : 346 pages
File Size : 13,24 MB
Release : 1981
Category : Mathematics
ISBN :
Author : Charles George Broyden
Publisher : Elsevier
Page : 343 pages
File Size : 14,61 MB
Release : 2004-09-08
Category : Mathematics
ISBN : 0080478875
The first four chapters of this book give a comprehensive and unified theory of the Krylov methods. Many of these are shown to be particular examples ofthe block conjugate-gradient algorithm and it is this observation thatpermits the unification of the theory. The two major sub-classes of thosemethods, the Lanczos and the Hestenes-Stiefel, are developed in parallel asnatural generalisations of the Orthodir (GCR) and Orthomin algorithms. Theseare themselves based on Arnoldi's algorithm and a generalised Gram-Schmidtalgorithm and their properties, in particular their stability properties,are determined by the two matrices that define the block conjugate-gradientalgorithm. These are the matrix of coefficients and the preconditioningmatrix.In Chapter 5 the"transpose-free" algorithms based on the conjugate-gradient squared algorithm are presented while Chapter 6 examines the various ways in which the QMR technique has been exploited. Look-ahead methods and general block methods are dealt with in Chapters 7 and 8 while Chapter 9 is devoted to error analysis of two basic algorithms.In Chapter 10 the results of numerical testing of the more important algorithms in their basic forms (i.e. without look-ahead or preconditioning) are presented and these are related to the structure of the algorithms and the general theory. Graphs illustrating the performances of various algorithm/problem combinations are given via a CD-ROM.Chapter 11, by far the longest, gives a survey of preconditioning techniques. These range from the old idea of polynomial preconditioning via SOR and ILU preconditioning to methods like SpAI, AInv and the multigrid methods that were developed specifically for use with parallel computers. Chapter 12 is devoted to dual algorithms like Orthores and the reverse algorithms of Hegedus. Finally certain ancillary matters like reduction to Hessenberg form, Chebychev polynomials and the companion matrix are described in a series of appendices.·comprehensive and unified approach·up-to-date chapter on preconditioners·complete theory of stability·includes dual and reverse methods·comparison of algorithms on CD-ROM·objective assessment of algorithms
Author : Stephen Boyd
Publisher : Cambridge University Press
Page : 477 pages
File Size : 11,34 MB
Release : 2018-06-07
Category : Business & Economics
ISBN : 1316518965
A groundbreaking introduction to vectors, matrices, and least squares for engineering applications, offering a wealth of practical examples.
Author : Timothy A. Davis
Publisher : SIAM
Page : 228 pages
File Size : 20,71 MB
Release : 2006-09-01
Category : Computers
ISBN : 0898716136
The sparse backslash book. Everything you wanted to know but never dared to ask about modern direct linear solvers. Chen Greif, Assistant Professor, Department of Computer Science, University of British Columbia.Overall, the book is magnificent. It fills a long-felt need for an accessible textbook on modern sparse direct methods. Its choice of scope is excellent John Gilbert, Professor, Department of Computer Science, University of California, Santa Barbara.Computational scientists often encounter problems requiring the solution of sparse systems of linear equations. Attacking these problems efficiently requires an in-depth knowledge of the underlying theory, algorithms, and data structures found in sparse matrix software libraries. Here, Davis presents the fundamentals of sparse matrix algorithms to provide the requisite background. The book includes CSparse, a concise downloadable sparse matrix package that illustrates the algorithms and theorems presented in the book and equips readers with the tools necessary to understand larger and more complex software packages.With a strong emphasis on MATLAB and the C programming language, Direct Methods for Sparse Linear Systems equips readers with the working knowledge required to use sparse solver packages and write code to interface applications to those packages. The book also explains how MATLAB performs its sparse matrix computations.Audience This invaluable book is essential to computational scientists and software developers who want to understand the theory and algorithms behind modern techniques used to solve large sparse linear systems. The book also serves as an excellent practical resource for students with an interest in combinatorial scientific computing.Preface; Chapter 1: Introduction; Chapter 2: Basic algorithms; Chapter 3: Solving triangular systems; Chapter 4: Cholesky factorization; Chapter 5: Orthogonal methods; Chapter 6: LU factorization; Chapter 7: Fill-reducing orderings; Chapter 8: Solving sparse linear systems; Chapter 9: CSparse; Chapter 10: Sparse matrices in MATLAB; Appendix: Basics of the C programming language; Bibliography; Index.
Author : Yousef Saad
Publisher : SIAM
Page : 537 pages
File Size : 10,33 MB
Release : 2003-04-01
Category : Mathematics
ISBN : 0898715342
Mathematics of Computing -- General.
Author : James M. Ortega
Publisher : Springer Science & Business Media
Page : 330 pages
File Size : 17,18 MB
Release : 1988-04-30
Category : Computers
ISBN : 9780306428623
Although the origins of parallel computing go back to the last century, it was only in the 1970s that parallel and vector computers became available to the scientific community. The first of these machines-the 64 processor llliac IV and the vector computers built by Texas Instruments, Control Data Corporation, and then CRA Y Research Corporation-had a somewhat limited impact. They were few in number and available mostly to workers in a few government laboratories. By now, however, the trickle has become a flood. There are over 200 large-scale vector computers now installed, not only in government laboratories but also in universities and in an increasing diversity of industries. Moreover, the National Science Foundation's Super computing Centers have made large vector computers widely available to the academic community. In addition, smaller, very cost-effective vector computers are being manufactured by a number of companies. Parallelism in computers has also progressed rapidly. The largest super computers now consist of several vector processors working in parallel. Although the number of processors in such machines is still relatively small (up to 8), it is expected that an increasing number of processors will be added in the near future (to a total of 16 or 32). Moreover, there are a myriad of research projects to build machines with hundreds, thousands, or even more processors. Indeed, several companies are now selling parallel machines, some with as many as hundreds, or even tens of thousands, of processors.