
doi: 10.1145/2898348
To face the advent of multicore processors and the ever increasing complexity of hardware architectures, programming models based on DAG parallelism regained popularity in the high performance, scientific computing community. Modern runtime systems offer a programming interface that complies with this paradigm and powerful engines for scheduling the tasks into which the application is decomposed. These tools have already proved their effectiveness on a number of dense linear algebra applications. This article evaluates the usability and effectiveness of runtime systems based on the Sequential Task Flow model for complex applications, namely, sparse matrix multifrontal factorizations that feature extremely irregular workloads, with tasks of different granularities and characteristics and with a variable memory consumption. Most importantly, it shows how this parallel programming model eases the development of complex features that benefit the performance of sparse, direct solvers as well as their memory consumption. We illustrate our discussion with the multifrontal QR factorization running on top of the StarPU runtime system.
[INFO.INFO-DC]Computer Science [cs]/Distributed, G13 [Numerical Analysis]: Numerical Linear Algebra, memory-aware, multicores, Parallel numerical computation, runtime systems, communication-avoiding, Parallel, 004, Computational methods for sparse matrices, Categories and Subject Descriptors: D13 [Programming Techniques]: Concurrent Programming, and Cluster Computing [cs.DC], Numerical algorithms for specific classes of architectures, G4 [Mathematical Software] General Terms: Algorithms, [INFO.INFO-DC] Computer Science [cs]/Distributed, Parallel, and Cluster Computing [cs.DC], Performance Additional Key Words and Phrases: Sparse direct solvers, sparse direct solvers
[INFO.INFO-DC]Computer Science [cs]/Distributed, G13 [Numerical Analysis]: Numerical Linear Algebra, memory-aware, multicores, Parallel numerical computation, runtime systems, communication-avoiding, Parallel, 004, Computational methods for sparse matrices, Categories and Subject Descriptors: D13 [Programming Techniques]: Concurrent Programming, and Cluster Computing [cs.DC], Numerical algorithms for specific classes of architectures, G4 [Mathematical Software] General Terms: Algorithms, [INFO.INFO-DC] Computer Science [cs]/Distributed, Parallel, and Cluster Computing [cs.DC], Performance Additional Key Words and Phrases: Sparse direct solvers, sparse direct solvers
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 33 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Top 10% | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Top 10% | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Top 10% |
