[MITgcm-devel] heff_max...more sea ice issues

Martin Losch Martin.Losch at awi.de
Fri Dec 22 04:09:46 EST 2006


Hi Matt,

I am sorry, but I will now make life more difficult for you:

1. LSR is an implicit scheme, BUT is does not treat the Coriolis term  
implicitly (so it is different from the alogoritm described in  
Jinlun's paper Zhang+Hibler (1998), or rather it uses only the first  
two parts of that algorithm). From my experience with explicit  
Coriolis terms, I would guess that you are limited by approximately  
1h (at least that's true for the ocean) for SEAICE_deltaTdyn (I  
usually set this to deltaTmom for asynchronos time stepping). I am  
not sure that I can take 12h time steps, but see next point:

2. I don't see why changing the time step should save you any time  
anyway: The seaice model (including the lsr or evp solver) is called  
at every time step and changing the time step can only affect the  
stability (if you do an asynchonous time stepping scheme). We have  
not (yet) implemented a scheme that calls the seaice model or parts  
of it only every so often (depending on the time step). This is  
commonly done in other models and seems to be good way of reducing  
the computational cost (if the ice model is slow, why is the ice  
model slow? It's a 2D model? I don't get it!).

3. the EVP solver requires a very short time step. For example, for  
my 2deg model I use SEAICE_deltaTevp=60s which is a 20th of the  
deltaTmom=SEAICE_deltaTdyn. According to the manual of CICE (Hunke's  
ice model) one should use something like a 30th to a 20th of the  
physical time step. Now EVP steps forward the stress evolution  
equations n-times (in my case 20 times) within on SEAICE_deltaTdyn  
time step. For a large domain (and high resolution), this is usually  
still faster than LSR, because in LSR call the routine lsr (or  
seaice_lsr) twice and in each call you do two iterations, one for  
Uice and one for Vice, so you have 4 iterations and if one is only 5  
steps, then you still have 20 steps, which are more complicated than  
the seaice_evp dynamics.

4. metric terms: the dynamics include only the metric terms for the  
spherical grid. I know that this is not exact for a general  
orthogonal coordinate system, but as far as I know Matt is using a  
spherical grid anyway. Including all metric terms (or finding a  
"vector invariant" formulation of the stress) is still a "to do".

Without knowing the details of your configuration I would recommend  
these experiments:
Use SEAICE_deltaTdyn = deltaT(mom) and SEAICE_deltaTtherm = deltaT 
(tracer) in any case (your deltaT is probably on the order 600-900sec  
isn't it?).
1. try the B-grid code (I expect that is will be even worse, but if  
not that's a lot of help for me in terms of debugging. I am praying  
that this experiment will not help you (o:)
2. Try LSR with higher precision maybe even better than 1e-4 (will  
make you model even slower, but maybe stable)
3. Try EVP with SEAICE_deltaTevp = deltaT(mom)/20
4. with C-grid: turn off the ice-ocean stress, for that you'll have  
to edit seaice_ocean_stress: comment out the lines that say fu=(1- 
areaw)*fu + areaw*fuice etc. (But note: Dimitris has found in his  
high res cubed sphere experiments that using ice-ocean stress in the  
C-grid is stable, while on the B-grid it's not, so I personally don't  
think that turning off the stress will help).
5. with C-grid: try the Hibler and Bryan stress, that's could even  
more stable (and is more physical too, but I have in the back of my  
mind that Dimitris had problems with this in his high res runs).

Martin

On 22 Dec 2006, at 01:30, Jinlun Zhang wrote:

> Matt,
> This is a paper that gives a clue about EVP time step.
> http://www.mrcc.uqam.ca/V_f/Publications/articles/ 
> Saucier2004_ClimDyn.pdf.
> Jinlun
>
> Matthew Mazloff wrote:
>
>> Dimitris and Jinlun,
>>
>> Thank you very much for the info.  I will try some runs (in early   
>> January) and let you know how how things work out.
>>
>> -Matt
>>
>> On Dec 21, 2006, at 6:47 PM, Jinlun Zhang wrote:
>>
>>>
>>>
>>> Matthew Mazloff wrote:
>>>
>>>> Thanks for the help...but I am a bit confused.  Two things
>>>>
>>>> 1) Re model efficiency and time stepping...I see there are 3    
>>>> parameters.  I am guessing  SEAICE_deltaTtherm should be the   
>>>> ocean  dynamics time-step as the forcing comes from this.  The   
>>>> other time  stepping parameters are  SEAICE_deltaTdyn  and   
>>>> SEAICE_deltaTevp which  I assume are the timesteps for each   
>>>> dynamic solver (LSR and EVP)  respectively.  And as I  
>>>> understand  it LSR can use the "large"  timestep, but the EVP  
>>>> should use the  "small" timestep...is this  correct?  And I am  
>>>> not using both at  the same time obviously, but you  are saying  
>>>> I should try both  independently because it is not obvious   
>>>> which is faster.
>>>
>>>
>>> Correct.
>>>
>>>>
>>>> 2)More important than efficiency (right now anyway) is   
>>>> stability.   Jinlun, your first email seemed to suggest I try  
>>>> LSR  with a half day  time step and LSR_ERROR=1e-4, or try EVP  
>>>> with  "small" timestep.  Are  either of these methods likely to  
>>>> be more  stable?
>>>
>>>
>>> Although we may use half day time step for LSR, but it is better  
>>> to  use the same ocean dynamics time step for LSR for  
>>> consistency, and  particularly when the code blows up. And using  
>>> 1e-4. I would think,  from the heff_max figure, that the problem  
>>> is most likely due to  the surface ocean stress calculation that  
>>> causes instability.  However, you might also want to try EVP. I  
>>> don't have much  experience with EVP, but people have been  
>>> telling me that very  small time steps should be used for  
>>> stability and for getting rid  of unphysical elastic waves. I  
>>> read one paper about high-res.  (~10km) Hudson Bay simulation,  
>>> the time step is as small as a few  seconds.
>>>
>>> Jinlun
>>>
>>> _______________________________________________
>>> MITgcm-devel mailing list
>>> MITgcm-devel at mitgcm.org
>>> http://mitgcm.org/mailman/listinfo/mitgcm-devel
>>
>>
>> _______________________________________________
>> MITgcm-devel mailing list
>> MITgcm-devel at mitgcm.org
>> http://mitgcm.org/mailman/listinfo/mitgcm-devel
>
>
> -- 
>
> Jinlun Zhang
> Polar Science Center, Applied Physics Laboratory
> University of Washington, 1013 NE 40th St, Seattle, WA 98105-6698
>
> Phone: (206)-543-5569;  Fax: (206)-616-3142
> zhang at apl.washington.edu
> http://psc.apl.washington.edu/pscweb2002/Staff/zhang/zhang.html
>
>
>
>
> _______________________________________________
> MITgcm-devel mailing list
> MITgcm-devel at mitgcm.org
> http://mitgcm.org/mailman/listinfo/mitgcm-devel




More information about the MITgcm-devel mailing list