[MITgcm-support] MPI rules to save compute time?

Menemenlis, Dimitris (3248) Dimitris.Menemenlis at jpl.nasa.gov
Mon Nov 14 11:10:08 EST 2011


Consider number of grid points per process, "including" overlap regions.
The narrower "run1" has larger overlap regions, hence more computations
per process, possibly impacting processor cache too.

Dimitris Menemenlis

On Nov 14, 2011, at 7:47 AM, Chun-Yan Zhou wrote:

Dear all,
 I use verification/exp4 to test the MPI.
 The the Nx=80, Ny=42 and Nr=8.
 I use differnt nPx and Npy to run the same iteration number and the data files are same.
 &PARM04
 usingCartesianGrid=.TRUE.,
 delX=80*5.e3,
 delY=42*5.e3,
 delR= 8*562.5,
 &
           sNx   sNy  nPx nPy processes    time cost
run0    80      42   1       1          1             20mins
run1    80      21   1       2          2             20mins
run2     40     42   2       1          2             12mins
run3     40      21  2        2          4            10mins

comparing run0 and run1,  nPy=2 didn't speed up the computation. But nPx=2 case is much faster than NPx=1(run0 VS. run2). Since the grid spaces are the same in the X and Y direction, what causes the compute time difference?  Any idea about this? Any rules to assign the nPx and nPy values in order to save time?

Thanks in advance!
chunyan


The University of Dundee is a registered Scottish charity, No: SC015096

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mitgcm.org/pipermail/mitgcm-support/attachments/20111114/f00585ea/attachment.htm>


More information about the MITgcm-support mailing list