Eigenvalue Algorithms for Symmetric Hierarchical Matrices


Book Description

This thesis is on the numerical computation of eigenvalues of symmetric hierarchical matrices. The numerical algorithms used for this computation are derivations of the LR Cholesky algorithm, the preconditioned inverse iteration, and a bisection method based on LDL factorizations. The investigation of QR decompositions for H-matrices leads to a new QR decomposition. It has some properties that are superior to the existing ones, which is shown by experiments using the HQR decompositions to build a QR (eigenvalue) algorithm for H-matrices does not progress to a more efficient algorithm than the LR Cholesky algorithm. The implementation of the LR Cholesky algorithm for hierarchical matrices together with deflation and shift strategies yields an algorithm that require O(n) iterations to find all eigenvalues. Unfortunately, the local ranks of the iterates show a strong growth in the first steps. These H-fill-ins makes the computation expensive, so that O(n³) flops and O(n²) storage are required. Theorem 4.3.1 explains this behavior and shows that the LR Cholesky algorithm is efficient for the simple structured Hl-matrices. There is an exact LDLT factorization for Hl-matrices and an approximate LDLT factorization for H-matrices in linear-polylogarithmic complexity. This factorizations can be used to compute the inertia of an H-matrix. With the knowledge of the inertia for arbitrary shifts, one can compute an eigenvalue by bisectioning. The slicing the spectrum algorithm can compute all eigenvalues of an Hl-matrix in linear-polylogarithmic complexity. A single eigenvalue can be computed in O(k²n log^4 n). Since the LDLT factorization for general H-matrices is only approximative, the accuracy of the LDLT slicing algorithm is limited. The local ranks of the LDLT factorization for indefinite matrices are generally unknown, so that there is no statement on the complexity of the algorithm besides the numerical results in Table 5.7. The preconditioned inverse iteration computes the smallest eigenvalue and the corresponding eigenvector. This method is efficient, since the number of iterations is independent of the matrix dimension. If other eigenvalues than the smallest are searched, then preconditioned inverse iteration can not be simply applied to the shifted matrix, since positive definiteness is necessary. The squared and shifted matrix (M-mu I)² is positive definite. Inner eigenvalues can be computed by the combination of folded spectrum method and PINVIT. Numerical experiments show that the approximate inversion of (M-mu I)² is more expensive than the approximate inversion of M, so that the computation of the inner eigenvalues is more expensive. We compare the different eigenvalue algorithms. The preconditioned inverse iteration for hierarchical matrices is better than the LDLT slicing algorithm for the computation of the smallest eigenvalues, especially if the inverse is already available. The computation of inner eigenvalues with the folded spectrum method and preconditioned inverse iteration is more expensive. The LDLT slicing algorithm is competitive to H-PINVIT for the computation of inner eigenvalues. In the case of large, sparse matrices, specially tailored algorithms for sparse matrices, like the MATLAB function eigs, are more efficient. If one wants to compute all eigenvalues, then the LDLT slicing algorithm seems to be better than the LR Cholesky algorithm. If the matrix is small enough to be handled in dense arithmetic (and is not an Hl(1)-matrix), then dense eigensolvers, like the LAPACK function dsyev, are superior. The H-PINVIT and the LDLT slicing algorithm require only an almost linear amount of storage. They can handle larger matrices than eigenvalue algorithms for dense matrices. For Hl-matrices of local rank 1, the LDLT slicing algorithm and the LR Cholesky algorithm need almost the same time for the computation of all eigenvalues. For large matrices, both algorithms are faster than the dense LAPACK function dsyev.




Hierarchical Matrices: Algorithms and Analysis


Book Description

This self-contained monograph presents matrix algorithms and their analysis. The new technique enables not only the solution of linear systems but also the approximation of matrix functions, e.g., the matrix exponential. Other applications include the solution of matrix equations, e.g., the Lyapunov or Riccati equation. The required mathematical background can be found in the appendix. The numerical treatment of fully populated large-scale matrices is usually rather costly. However, the technique of hierarchical matrices makes it possible to store matrices and to perform matrix operations approximately with almost linear cost and a controllable degree of approximation error. For important classes of matrices, the computational cost increases only logarithmically with the approximation error. The operations provided include the matrix inversion and LU decomposition. Since large-scale linear algebra problems are standard in scientific computing, the subject of hierarchical matrices is of interest to scientists in computational mathematics, physics, chemistry and engineering.




Numerical Methods for General and Structured Eigenvalue Problems


Book Description

This book is about computing eigenvalues, eigenvectors, and invariant subspaces of matrices. Treatment includes generalized and structured eigenvalue problems and all vital aspects of eigenvalue computations. A unique feature is the detailed treatment of structured eigenvalue problems, providing insight on accuracy and efficiency gains to be expected from algorithms that take the structure of a matrix into account.




Supercomputing Frontiers


Book Description

It constitutes the refereed proceedings of the 4th Asian Supercomputing Conference, SCFA 2018, held in Singapore in March 2018. Supercomputing Frontiers will be rebranded as Supercomputing Frontiers Asia (SCFA), which serves as the technical programme for SCA18. The technical programme for SCA18 consists of four tracks: Application, Algorithms & Libraries Programming System Software Architecture, Network/Communications & Management Data, Storage & Visualisation The 20 papers presented in this volume were carefully reviewed nd selected from 60 submissions.




Advanced Parallel Processing Technologies


Book Description

This book constitutes the refereed proceedings of the 6th International Workshop on Advanced Parallel Processing Technologies, APPT 2005, held in Hong Kong, China in September 2005. The 55 revised full papers presented were carefully reviewed and selected from over 220 submissions. All current aspects in parallel and distributed computing are addressed ranging from hardware and software issues to algorithmic aspects and advanced applications. The papers are organized in topical sections on architecture, algorithm and theory, system and software, grid computing, networking, and applied technologies.




Matrix Computations


Book Description

This revised edition provides the mathematical background and algorithmic skills required for the production of numerical software. It includes rewritten and clarified proofs and derivations, as well as new topics such as Arnoldi iteration, and domain decomposition methods.




Domain Decomposition Methods in Science and Engineering XXVI


Book Description

These are the proceedings of the 26th International Conference on Domain Decomposition Methods in Science and Engineering, which was hosted by the Chinese University of Hong Kong and held online in December 2020. Domain decomposition methods are iterative methods for solving the often very large systems of equations that arise when engineering problems are discretized, frequently using finite elements or other modern techniques. These methods are specifically designed to make effective use of massively parallel, high-performance computing systems. The book presents both theoretical and computational advances in this domain, reflecting the state of art in 2020.




Numerical Algebra, Matrix Theory, Differential-Algebraic Equations and Control Theory


Book Description

This edited volume highlights the scientific contributions of Volker Mehrmann, a leading expert in the area of numerical (linear) algebra, matrix theory, differential-algebraic equations and control theory. These mathematical research areas are strongly related and often occur in the same real-world applications. The main areas where such applications emerge are computational engineering and sciences, but increasingly also social sciences and economics. This book also reflects some of Volker Mehrmann's major career stages. Starting out working in the areas of numerical linear algebra (his first full professorship at TU Chemnitz was in "Numerical Algebra," hence the title of the book) and matrix theory, Volker Mehrmann has made significant contributions to these areas ever since. The highlights of these are discussed in Parts I and II of the present book. Often the development of new algorithms in numerical linear algebra is motivated by problems in system and control theory. These and his later major work on differential-algebraic equations, to which he together with Peter Kunkel made many groundbreaking contributions, are the topic of the chapters in Part III. Besides providing a scientific discussion of Volker Mehrmann's work and its impact on the development of several areas of applied mathematics, the individual chapters stand on their own as reference works for selected topics in the fields of numerical (linear) algebra, matrix theory, differential-algebraic equations and control theory.