ECCOMAS 2024

MS146 - Advanced Parallel Algorithms for Extreme-Scale Simulations

Organized by: X. Álvarez-Farré (SURF, Netherlands), À. Alsalti-Baldellou (Technical University of Catalonia, Spain), M. Torres Rodrigues (SURF, Netherlands), V. Codreanu (SURF, Netherlands), T. Aoki (Tokyo Institute of Technology, Japan) and F. Trias (Technical University of Catalonia, Spain)
Keywords: Distributed Memory, Extreme-Scale Simulation, Heterogeneous Systems, Linear Solvers, Parallel Computing, Sparse Algebra
The evolution in hardware technologies enables scientific computing to reach further aims. Nowadays, the use of massively parallel supercomputers is rather common in the solution of both industrial and academic scale problems. Unfortunately, the literature reveals that the performance of numerous scientific computing applications is often limited to tiny fractions of the system’s peak performance. While the peak performance has been growing 9 and 4.5 times faster than network and main memory bandwidth per decade over the past 30 years, many algorithms employed in scientific computing have very low operational intensity. Numerical simulation codes are increasingly memory-bounded, making processors suffer from serious data starvation. Moreover, distributed memory systems pose an additional bottleneck: network bandwidth and latency. The relative weight of data exchanges grows as parallel applications are distributed among a greater number of processors, making it advantageous to increase the problem size per node. However, hardware accelerators with high-bandwidth memory, which are major contributors to the increased computational power of individual nodes, have smaller memory capacities than traditional processors. Application bottlenecks becoming more severe exacerbates the already difficult job of architects, programmers, and researchers. In this context, this mini-symposium aims to bring together people working on advanced, cutting-edge parallel methods for solving extreme-scale simulations with a special focus on efficiency, portability, and sustainability. Works on domain-specific languages, parallel linear solvers, sparse algebra kernels, eigenvalue problems, parallel-in-time methods, load balancing, adaptive mesh refinement, and memory footprint reduction techniques, among others, are welcomed.