[MITgcm-support] Errors with eedata and eeset_parms

Jonathan Lauderdale jml1 at mit.edu
Fri Jul 31 14:04:00 EDT 2020


Hi Katie,

When you ran genmake, did you also use the “-mpi” flag? Then when you go to execute the model, as Kaveh mentioned, you’ll need:

> mpirun -np 12 ./mitgcmuv


i.e. requesting the same number of cores as tasks.

Cheers,

Jonathan 
___________________________________________________________________________
Dr. Jonathan M. Lauderdale
Research Scientist
Department of Earth, Atmosphere and Planetary Sciences 
Massachusetts Institute of Technology 
77 Massachusetts Avenue (54-1518)
Cambridge, MA 02139, USA 
Office: +1 617 324 1568 
Cell   : +1 617 304 5661
Email: mailto:jml1 at mit.edu <mailto:jml1 at mit.edu>
Web: http://paocweb.mit.edu/people/jml1 <http://paocweb.mit.edu/people/jml1>
Twitter: https://twitter.com/jon_lauderdale
 <https://twitter.com/jon_lauderdale>Blog: https://seamanticscience.wordpress.com/ <https://seamanticscience.wordpress.com/>
Git: https://github.com/seamanticscience <https://github.com/seamanticscience>
___________________________________________________________________________

> On Jul 31, 2020, at 1:56 PM, Katie Hughes <kestagl at ucdavis.edu> wrote:
> 
> Hello,
> 
> I'm new to MITgcm, so apologies if I leave out any pertinent information.
> 
> I'm trying to run a 2D under-ice simulation with 12 tasks per node (12 horizontal partitions of the grid for parallelization). I'm receiving errors for 'eedata'. 
> 
> I'm running the model with the command mpirun -np 4 ./mitgcmuv but was previously just using ./mitgcmuv and getting the same errors.
> 
> Here is an example of my eedata and the coinciding errors:
> If I use this eedata:
> &EEPARMS
> 
> /
> 
> I get this error:
> 
> (PID.TID 0000.0001) *** ERROR *** INI_PROCS: needs MPI for multi-procs (nPx*nPy=    12) setup
> 
> (PID.TID 0000.0001) *** ERROR *** INI_PROCS:  but presently usingMPI = False (in "eedata")
> 
> (PID.TID 0000.0001) *** ERROR *** EEDIE: earlier error in multi-proc/thread setting
> 
> 
> (PID.TID 0000.0001) *** ERROR *** PROGRAM MAIN: ends with fatal Error
> 
> 
> 
> If I use this eedata:
> 
> &EEPARMS
>  usingMPI=TRUE,
> /
> I get this error:
> (PID.TID 0000.0001) *** ERROR *** INI_PROCS: needs MPI for multi-procs (nPx*nPy=    12) setup
> 
> (PID.TID 0000.0001) *** ERROR *** INI_PROCS:  but presently usingMPI = False (in "eedata")
> 
> (PID.TID 0000.0001) *** ERROR *** EEDIE: earlier error in multi-proc/thread setting
> 
> (PID.TID 0000.0001) *** ERROR *** PROGRAM MAIN: ends with fatal Error
> 
> 
> Any suggestions?
> 
> Thank you in advance!
> Katie Stagl Hughes
> 
> Ph.D. Candidate | Dept. of Civil & Environmental Engineering
> University of California, Davis
> kestagl at ucdavis.edu <mailto:kestagl at ucdavis.edu>_______________________________________________
> MITgcm-support mailing list
> MITgcm-support at mitgcm.org
> http://mailman.mitgcm.org/mailman/listinfo/mitgcm-support

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.mitgcm.org/pipermail/mitgcm-support/attachments/20200731/d2539e50/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: smime.p7s
Type: application/pkcs7-signature
Size: 1864 bytes
Desc: not available
URL: <http://mailman.mitgcm.org/pipermail/mitgcm-support/attachments/20200731/d2539e50/attachment-0001.p7s>


More information about the MITgcm-support mailing list