[MITgcm-support] Difficult with cpl_aim+oce

Chris Horvat horvat at fas.harvard.edu
Fri Sep 20 18:13:39 EDT 2013


Hi Jean-Michel,

mpi runs on my machine, and I can get other MITGCM runs to run in parallel,
so at least it isn't that so far (I just re-read a conversation you and
Taimaz had a while back).

I've updated a submission script to basically handle what is being send to
mpirun (I'm using mvapich2), essentially sending the command

mpirun.lsf -np 1 ./build_cpl/mitgcmuv : -np 1 ./build_ocn/mitgcmuv : -np 1
./build_atm/mitgcmuv

which runs some part of the MITGCM code, but gives the appended output into
the lsf.err file. So something is executing the MITgcm in some way, just
not the way I want? There is no STD.* output in either oce/atm directories

 MITCPLR_init1:            0  Coupler Rank/Size =            0  /
3
 MITCPLR_init1:            1  Coupler Rank/Size =            1  /
3
 MITCPLR_init1:            0  Coupler MPI_Bcast from            0
 59
 MITCPLR_init1:            1  Coupler MPI_Bcast from            0
 59
  ierr=           0
 MITCPLR_init1:            1  Coupler Bcast cbuf=
 Coupler                         x
 MITCPLR_init1:            1  Coupler MPI_Bcast from            1
 59
  ierr=           0
 MITCPLR_init1:            1  Coupler Bcast cbuf=
 Coupler                         x
  ierr=           0
 MITCPLR_init1:            0  Coupler Bcast cbuf=
 Coupler                         x
 MITCPLR_init1:            0  Coupler MPI_Bcast from            1
 59
  ierr=           0
 MITCPLR_init1:            0  Coupler Bcast cbuf=
 Coupler                         x
 MITCPLR_init1:            2  Coupler Rank/Size =            2  /
3
 MITCPLR_init1:            2  Coupler MPI_Bcast from            0
 59
  ierr=           0
 MITCPLR_init1:            2  Coupler Bcast cbuf=
 Coupler                         x
 MITCPLR_init1:            2  Coupler MPI_Bcast from            1
 59
  ierr=           0
 MITCPLR_init1:            2  Coupler Bcast cbuf=
 Coupler                         x
 MITCPLR_init1:            1  Coupler MPI_Bcast from            2
 59
 MITCPLR_init1:            2  Coupler MPI_Bcast from            2
 59
  ierr=           0
 MITCPLR_init1:            2  Coupler Bcast cbuf=
 Coupler                         x
MITCPLR_init1: I can only handle one coupler process
 MITCPLR_init1:            0  Coupler MPI_Bcast from            2
 59
  ierr=           0
 MITCPLR_init1:            0  Coupler Bcast cbuf=
 Coupler                         x
MITCPLR_init1: I can only handle one coupler process
  ierr=           0
 MITCPLR_init1:            1  Coupler Bcast cbuf=
 Coupler                         x
MITCPLR_init1: I can only handle one coupler process





---
Christopher Horvat -- www.chrv.at <http://www.chrv.at/blog.html>



On Fri, Sep 20, 2013 at 12:30 PM, Jean-Michel Campin <jmc at ocean.mit.edu>wrote:

> Hi Chris,
>
> You may need to adapt the script "run_cpl_test" for your MPI version.
> In run_cpl_test, there are 3 cases coded (lines 234 to 247),
> mpich2 or openmpi, mpich-mx and default old mpich (version 1).
> But the way the script select one of these 3 is not
> portable, so may be if you reset (just after line 233) the 3 flags:
> mpichmx, mpich2 and openpi, it might just work.
>
> Cheers,
> Jean-Michel
>
> On Thu, Sep 19, 2013 at 02:18:56PM -0400, Chris Horvat wrote:
> > Hi all,
> >
> > I know there have been a few threads covering this from other directions,
> > but I'm looking for a little help running the coupled ocean/aim
> experiment.
> > Currently the run_cpl_test 3 command returns the following:
> >
> > --------------------------------------
> > execute 'mpirun -p4pg pr_group -v ./build_cpl/mitgcmuv' :
> >       -umask <umask>               # umask for remote process
> >       -path <dirname>              # place to look for executables
> >       -host <hostname>             # host to start on
> >       -soft <spec>                 # modifier of -n value
> >       -arch <arch>                 # arch type to start on (not
> implemented)
> >       -envall                      # pass all env vars in current
> > environment
> >       -envnone                     # pass no env vars
> >       -envlist <list of env var names> # pass current values of these
> vars
> >       -env <name> <value>          # pass this value of this env var
> > mpiexec [global args] [local args] executable args : [local args]
> > executable...
> > mpiexec -gdba jobid                # gdb-attach to existing jobid
> > mpiexec -configfile filename       # filename contains cmd line segs as
> > lines
> >   (See User Guide for more details)
> >
> > Examples:
> >    mpiexec -l -n 10 cpi 100
> >    mpiexec -genv QPL_LICENSE 4705 -n 3 a.out
> >
> >    mpiexec -n 1 -host foo master : -n 4 -host mysmp slave
> > ----------------------------------------------------
> >
> > which I take to mean it isn't liking one of the optional (most likely
> > mpirun -pgp4). Is anyone familiar with this error?
> >
> > If I run the simple command
> >
> > mpd
> > mpirun ./build_cpl/mitgcmuv > xx
> >
> > in a batch script, the job will send and appear to be running, but won't
> > produce any STD* output or error output.
> >
> > Thanks for your help!!!!
> >
> > Chris
> >
> >
> > ---
> > Christopher Horvat -- www.chrv.at <http://www.chrv.at/blog.html>
>
> > _______________________________________________
> > MITgcm-support mailing list
> > MITgcm-support at mitgcm.org
> > http://mitgcm.org/mailman/listinfo/mitgcm-support
>
>
> _______________________________________________
> MITgcm-support mailing list
> MITgcm-support at mitgcm.org
> http://mitgcm.org/mailman/listinfo/mitgcm-support
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mitgcm.org/pipermail/mitgcm-support/attachments/20130920/e780074d/attachment.htm>


More information about the MITgcm-support mailing list