Weak scaling of parallel FMM vs. FFT up to 4096 processes Lorena A. Barba Rio Yokota 10.6084/m9.figshare.92425.v3 https://figshare.com/articles/dataset/Weak_scaling_of_parallel_FMM_vs_FFT_up_to_4096_processes/92425 <p>This figure shows the weak scaling of a parallel FMM-based fluid solver on GPUs, from 1 to 4096 processes. The FMM (fast multipole method) is used as the numerical engine in a vortex method fluid solver, simulating decaying isotropic turbulence. The reference method for this application is the pseudo-spectral method, which uses FFT as the numerical engine. Given the communication pattern of FFT, only 14% parallel efficiency is obtained with the spectral method on 4096 processes (no GPU acceleration). The parallel efficiency of the FMM-based solver is 74% going from one to 4096 processes (one GPU per MPI process, 3 GPUs per node).</p> <p>It is important to note that the results correspond to the full-application codes, not just the FMM and FFT algorithms. The spectral method calculations were done using the 'hit3d' code (see link below). The size of the largest problem corresponds to a 4096^3 mesh, i.e., almost 69 billion points (about 17 million points per process).</p> <p>These calculations were run on the TSUBAME 2.0 system of the Tokyo Institute of Technology, thanks to guest access, during October 2011. </p> <p>We include here the dataset (ASCII files in 1003.zip), the Matlab plotting script (readTimes1003.m) and the figure file (weakGPU.pdf) that is being included in the following paper:</p> <p>“Petascale turbulence simulation using a highly parallel fast multipole method”, Rio Yokota, L A Barba, Tetsu Narumi, Kenji Yasuoka. Comput. Phys. Comm. (online 13 Sept. 2012) doi:10.1016/j.cpc.2012.09.011<br>Preprint arXiv:1106.5273 [cs.NA]</p> <p> </p> <p>The figure is here shared under CC-BY 3.0. Please use the handle and doi above for citation if you use it.</p> 2013-02-06 18:51:09 fmm treecode turbulence vortex method spectral method Mechanical Engineering Computational Physics Applied Computer Science