-
Notifications
You must be signed in to change notification settings - Fork 244
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
1/2 degree coupled model crashes when backscatter is on and a mask_table is used #262
Comments
Hi Niki, I do not know the method to check if the mask table is bad. Are these Zhi On Mon, Mar 28, 2016 at 3:44 PM, Niki Zadeh notifications@github.com
|
@Zhi-Liang you have a good point. I will run the test with DEBUG= True to see where things start to diverge. |
When I set DEBUG=True the issue went away!!!
|
Hi Niki, From my previous experience, the possible reason Is that some variable Zhi On Mon, Mar 28, 2016 at 6:21 PM, Niki Zadeh notifications@github.com
|
Yes, but I am not talking about the "debug" mode. I used the same "prod" executable. DEBUG=True is a MOM6 parameter that prints out lots of checksums. To check my sanity I repeated the experiment a few times and every time it crashed, then I added @adcroft could #override DEBUG = True have any effect on uninitialized variables? |
@adcroft modified MOM_spatial_means.F90 a little for debugging :
This resulted in a crash as before at line 61 above confirming that something goes wrong with the input array var. I'll rerun in "debug" mode for further clues. |
That tells us the bad data is in var, so yes we need to look at it in a debugger. |
The crash does not happen in "debug" mode! It happens in both "prod" and "repro" modes. |
This issue is related to the combination of MOM6 backscatter AND using a mask_table.
So, backscatter code is not compatible with using a mask_table. I edit the title to reflect that. |
- Closes issue mom-ocean#734 and hopefully issue mom-ocean#262 - MEKE%Ku array is allocated on data domain and has a mpp_domain_update following the loop so it needs to be calculated only on compute domain - The problem with larger extents of the loops is Lmixscale array is initialized/calculate on compute domain and is NaN beyond compute domain extents, causing NaN's to lurk into MEKE%Ku and the model.
We do not think that this is an issue with any actively used models. Further work that completes the backscatter capability would have to address this, if it is still an issue there. |
The base 1/2 degree coupled model ( CM4_c96L32_am4g9_2000_OMp5 ) crashes with some specific layouts with mask_tables (both c1 and c3) with
It runs fine for the same layout without the mask_table.
It also runs fine with some other mask_tables (for other layouts).
Here are the work dirs for the test using mask_table and not with the same layout:
@Zhi-Liang is there a way to test if a mask_table is "bad"?
I made all mask_tables using the same check_mask tool.
Moreover, if I change the model (MOM6) parameters it runs fine with the same mask_table that crashed for the base experiment above, e.g.,
This tells me there might be something wrong with some of the mask_tables that I have generated for the 1/2 degree ocean grid:
the ones that worked
/lustre/f1/unswept/Niki.Zadeh/archive/input/verona/John.Dunne_OM4_05_20151028/mosaic_ocean.720x576/mask_table.320.36x45
/lustre/f1/unswept/Niki.Zadeh/archive/input/verona/John.Dunne_OM4_05_20151028/mosaic_ocean.720x576/mask_table.417.45x45
the ones that crashed
/lustre/f1/unswept/Niki.Zadeh/archive/input/verona/John.Dunne_OM4_05_20151028/mosaic_ocean.720x576/mask_table.239.36x36
/lustre/f1/unswept/Niki.Zadeh/archive/input/verona/John.Dunne_OM4_05_20151028/mosaic_ocean.720x576/mask_table.201.32x36
/lustre/f1/unswept/Niki.Zadeh/archive/input/verona/John.Dunne_OM4_05_20151028/mosaic_ocean.720x576/mask_table.359.36x50
/lustre/f1/unswept/Niki.Zadeh/archive/input/verona/John.Dunne_OM4_05_20151028/mosaic_ocean.720x576/mask_table.548.36x72
The text was updated successfully, but these errors were encountered: