TY - GEN
T1 - Combining performance aspects of irregular Gauss-Seidel via Sparse tiling
AU - Strout, Michelle Mills
AU - Carter, Larry
AU - Ferrante, Jeanne
AU - Freeman, Jonathan
AU - Kreaseck, Barbara
PY - 2005
Y1 - 2005
N2 - Finite Element problems are often solved using multigrid techniques. The most time consuming part of multigrid is the iterative smoother, such as Gauss-Seidel. To improve performance, iterative smoothers can exploit parallelism, intra-iteration data reuse, and inter-iteration data reuse. Current methods for parallelizing Gauss-Seidel on irregular grids, such as multi-coloring and owner-computes based techniques, exploit parallelism and possibly intra-iteration data reuse but not inter-iteration data reuse. Sparse tiling techniques were developed to improve intra-iteration and inter-iteration data locality in iterative smoothers. This paper describes how sparse tiling can additionally provide parallelism. Our results show the effectiveness of Gauss-Seidel parallelized with sparse tiling techniques on shared memory machines, specifically compared to owner-computes based Gauss-Seidel methods. The latter employ only parallelism and intra-iteration locality. Our results support the premise that better performance occurs when all three performance aspects (parallelism, intra-iteration, and inter-iteration data locality) are combined.
AB - Finite Element problems are often solved using multigrid techniques. The most time consuming part of multigrid is the iterative smoother, such as Gauss-Seidel. To improve performance, iterative smoothers can exploit parallelism, intra-iteration data reuse, and inter-iteration data reuse. Current methods for parallelizing Gauss-Seidel on irregular grids, such as multi-coloring and owner-computes based techniques, exploit parallelism and possibly intra-iteration data reuse but not inter-iteration data reuse. Sparse tiling techniques were developed to improve intra-iteration and inter-iteration data locality in iterative smoothers. This paper describes how sparse tiling can additionally provide parallelism. Our results show the effectiveness of Gauss-Seidel parallelized with sparse tiling techniques on shared memory machines, specifically compared to owner-computes based Gauss-Seidel methods. The latter employ only parallelism and intra-iteration locality. Our results support the premise that better performance occurs when all three performance aspects (parallelism, intra-iteration, and inter-iteration data locality) are combined.
UR - http://www.scopus.com/inward/record.url?scp=33745140812&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=33745140812&partnerID=8YFLogxK
U2 - 10.1007/11596110_7
DO - 10.1007/11596110_7
M3 - Conference contribution
AN - SCOPUS:33745140812
SN - 3540307818
SN - 9783540307815
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 90
EP - 110
BT - Languages and Compilers for Parallel Computing - 15th Workshop, LCPC 2002, Revised Papers
T2 - 15th Workshop on Languages and Compilers for Parallel Computing, LCPC 2002
Y2 - 25 July 2002 through 27 July 2002
ER -