LSRN : A PARALLEL ITERATIVE SOLVER FOR STRONGLY OVER- OR UNDERDETERMINED SYSTEMS
We describe a parallel iterative least squares solver named LSRN that is based on random normal projection. LSRN computes the min-length solution to min x∈ℝ n ‖Ax - b‖2, where A ∈ ℝ m × n with m ≫ n or m ≪ n, and where A may be rank-deficient. Tikhonov regularization may also be included. Since A is...
Veröffentlicht in: | SIAM journal on scientific computing : a publication of the Society for Industrial and Applied Mathematics. - 1999. - 36(2014), 2 vom: 01., Seite C95-C118 |
---|---|
1. Verfasser: | |
Weitere Verfasser: | , |
Format: | Aufsatz |
Sprache: | English |
Veröffentlicht: |
2014
|
Zugriff auf das übergeordnete Werk: | SIAM journal on scientific computing : a publication of the Society for Industrial and Applied Mathematics |
Schlagworte: | Journal Article Chebyshev semi-iterative method LAPACK LSQR Tikhonov regularization iterative method linear least squares minimum-length solution over determined system parallel computing mehr... |
Zusammenfassung: | We describe a parallel iterative least squares solver named LSRN that is based on random normal projection. LSRN computes the min-length solution to min x∈ℝ n ‖Ax - b‖2, where A ∈ ℝ m × n with m ≫ n or m ≪ n, and where A may be rank-deficient. Tikhonov regularization may also be included. Since A is involved only in matrix-matrix and matrix-vector multiplications, it can be a dense or sparse matrix or a linear operator, and LSRN automatically speeds up when A is sparse or a fast linear operator. The preconditioning phase consists of a random normal projection, which is embarrassingly parallel, and a singular value decomposition of size ⌈γ min(m, n)⌉ × min(m, n), where γ is moderately larger than 1, e.g., γ = 2. We prove that the preconditioned system is well-conditioned, with a strong concentration result on the extreme singular values, and hence that the number of iterations is fully predictable when we apply LSQR or the Chebyshev semi-iterative method. As we demonstrate, the Chebyshev method is particularly efficient for solving large problems on clusters with high communication cost. Numerical results show that on a shared-memory machine, LSRN is very competitive with LAPACK's DGELSD and a fast randomized least squares solver called Blendenpik on large dense problems, and it outperforms the least squares solver from SuiteSparseQR on sparse problems without sparsity patterns that can be exploited to reduce fill-in. Further experiments show that LSRN scales well on an Amazon Elastic Compute Cloud cluster |
---|---|
Beschreibung: | Date Revised 01.10.2020 published: Print Citation Status PubMed-not-MEDLINE |
ISSN: | 1064-8275 |