Error in executing CESM 1.2.0 T85_gx1v6 compset B_2000

yli_ouc@163_com

New Member
Hi everyoneI am trying to port CESM 1.2.0 to my local cluster. The test with f45_g37 compset X was successful. However, when I tried to execute a new case with T85_gx1v6 and B_2000, it showed that "UrbanInputldomain and input file do not match dims" in the lnd.log.XXXX-XXX file.I'm not sure it is a bug of CESM 1.2.0 or something wrong with my data.PS: all the inputdata was downloaded automatically by the build shellThank you
 

erik

Erik Kluzek
CSEG and Liaisons
Staff member
The problem is that the CLM is trying to run with a domain file at a different resolution than the surface dataset. So there must be a bug in scripts for the datasets.
 

erik

Erik Kluzek
CSEG and Liaisons
Staff member
I've posted the bug-report and fix for this to bug 1999. The fix for cesm1_2_0 (which also fixes the issue for T341 and T85_tx01) is as follows (although not I haven't tested the change)... Index: config_grid.xml===================================================================--- config_grid.xml(revision 161)+++ config_grid.xml(working copy)@@ -986,12 +986,12 @@     domain.lnd.T85_gx1v4.060403.nc-  domain.lnd.fv0.9x1.25_gx1v6.090309.nc+  domain.lnd.T85_gx1v4.060403.nc      domain.lnd.T85_gx1v4.060403.nc-  domain.lnd.fv0.9x1.25_tx0.1v2_070928.nc+  domain.lnd.T85_gx1v4.060403.nc    @@ -1030,7 +1030,7 @@      domain.lnd.T341_gx1v6.111226.nc         -  domain.lnd.fv0.23x0.31_tx0.1v2_070929.nc+  domain.lnd.T341_gx1v6.111226.nc            
 
Thanks a lot.I change the config_grid.xml like you said.    domain.lnd.T85_gx1v4.060403.nc-  domain.lnd.fv0.9x1.25_gx1v6.090309.nc+  domain.lnd.T85_gx1v4.060403.nc the program crashed,return code 1.
Warning: no access to tty (Bad file descriptor).Thus no job control in this shell.WARNING: Unable to read mpd.hosts or list of hosts isn't provided. MPI job will be run on the current machine only.running mpdallexit on cn051LAUNCHED mpd on cn051  viaRUNNING: mpd on cn051(seq_comm_setcomm)  initialize ID (  1 GLOBAL          ) pelist   =     0    11     1 ( npes =    12) ( nthreads =  1)rank 11 in job 1  cn051_60174   caused collective abort of all ranks  exit status of rank 11: return code 1rank 6 in job 1  cn051_60174   caused collective abort of all ranks  exit status of rank 6: killed by signal 9rank 5 in job 1  cn051_60174   caused collective abort of all ranks  exit status of rank 5: return code 1rank 3 in job 1  cn051_60174   caused collective abort of all ranks  exit status of rank 3: return code 1rank 2 in job 1  cn051_60174   caused collective abort of all ranks  exit status of rank 2: killed by signal 9rank 0 in job 1  cn051_60174   caused collective abort of all ranks   exit status of rank 0: return code 1
 

erik

Erik Kluzek
CSEG and Liaisons
Staff member
Hi Hmmm. OK, I can't really tell what went wrong with the message above. Some questions to access are:- What does the new namelist look like?- Did it get further than before?- Can you tell which component it died in?- Did it run out of memory or some other resource?- Does it need more processors? there is a trouble shooting chapter in the CESM User's Guide, you might find that helpful. 
 
Back
Top