Scheduled Downtime
On Tuesday 24 October 2023 @ 5pm MT the forums will be in read only mode in preparation for the downtime. On Wednesday 25 October 2023 @ 5am MT, this website will be down for maintenance and expected to return online later in the morning.
Normal Operations
The forums are back online with normal operations. If you notice any issues or errors related to the forums, please reach out to help@ucar.edu

CESM1.0.3 run error

Weiyi

New Member
Hi, I'm new to using CESM1.0.3 and when I sumit the $CASE.$MACH.run script, I'm getting this error from ccsm.log :(seq_io_init) pio init parameters: before nml read(seq_io_init)    pio_stride   =    -99(seq_io_init)    pio_root     =    -99(seq_io_init)    pio_typename = nothing(seq_io_init)    pio_numtasks =    -99(seq_io_init)    pio_debug_level =      0    pio_async_interface =  F(seq_io_init) pio init parameters: after nml read(seq_io_init)    pio_stride   =     -1(seq_io_init)    pio_root     =      1(seq_io_init)    pio_typename = netcdf(seq_io_init)    pio_numtasks =     -1(seq_io_init) pio init parameters:(seq_io_init)    pio_stride   =      4(seq_io_init)    pio_root     =      1(seq_io_init)    pio_typename = NETCDF(seq_io_init)    pio_numtasks =     27(seq_io_init)    pio_debug_level =      0    pio_async_interface =  F(seq_comm_setcomm)  initialize ID (  7 GLOBAL ) pelist   =     0   107     1 ( npes =   108) ( nthreads =  1)(seq_comm_setcomm)  initialize ID (  2   ATM  ) pelist   =     0   107     1 ( npes =   108) ( nthreads =  1)(seq_comm_setcomm)  initialize ID (  1   LND  ) pelist   =     0   107     1 ( npes =   108) ( nthreads =  1)(seq_comm_setcomm)  initialize ID (  4   ICE  ) pelist   =     0    99     1 ( npes =   100) ( nthreads =  1)(seq_comm_setcomm)  initialize ID (  5   GLC  ) pelist   =     0   107     1 ( npes =   108) ( nthreads =  1)(seq_comm_setcomm)  initialize ID (  3   OCN  ) pelist   =     0    99     1 ( npes =   100) ( nthreads =  1)(seq_comm_setcomm)  initialize ID (  6   CPL  ) pelist   =     0   107     1 ( npes =   108) ( nthreads =  1)(seq_comm_joincomm) initialize ID (  8 CPLATM ) join IDs =     6     2       ( npes =   108) ( nthreads =  1)(seq_comm_joincomm) initialize ID (  9 CPLLND ) join IDs =     6     1       ( npes =   108) ( nthreads =  1)(seq_comm_joincomm) initialize ID ( 10 CPLICE ) join IDs =     6     4       ( npes =   108) ( nthreads =  1)(seq_comm_joincomm) initialize ID ( 11 CPLOCN ) join IDs =     6     3       ( npes =   108) ( nthreads =  1)(seq_comm_joincomm) initialize ID ( 12 CPLGLC ) join IDs =     6     5       ( npes =   108) ( nthreads =  1)[c06n04:31992] *** An error occurred in MPI_Gather[c06n04:31992] *** on communicator MPI COMMUNICATOR 5 CREATE FROM 0[c06n04:31992] *** MPI_ERR_TYPE: invalid datatype[c06n04:31992] *** MPI_ERRORS_ARE_FATAL (your MPI job will now abort)--------------------------------------------------------------------------mpirun has exited due to process rank 74 with PID 31963 onnode c06n07 exiting without calling "finalize". This mayhave caused other processes in the application to beterminated by signals sent by mpirun (as reported here).--------------------------------------------------------------------------[c06n05:04048] 107 more processes have sent help message help-mpi-errors.txt / mpi_errors_are_fatal[c06n05:04048] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages
The MPI is openmpi1.4.3_ifort11.1 and the netcdf is NetCDF4.1.3.ifort11.1.  
Leo 
 

jedwards

CSEG and Liaisons
Staff member
Hi Leo,Do you need to use cesm1.0.3, perhaps 1.2.2 will work?   I believe that the problem is an incompatibility of that pio version with openmpi.   Even if you update pio and solve that problem I think that you may run into others in other parts of the model.  If you must use cesm1.0.3 you should see if you can use an mpich rather than an openmpi mpi librarry.   
 

Weiyi

New Member
Thank you, jedwards.We have finished some experiments on cesm1.0.3 before, and I wanted to do a single forcing experiment on ctrl run. So I have to use cesm1.0.3. We can use openmpi before. When I porting cesm to a new machine, it makes that problem. I have checked the config_machines.xml, env_machopts, Macros, mkbatch and environment settings on .bashrc.  It confused me. Leo
 

jedwards

CSEG and Liaisons
Staff member
You'll need to find the MPI_GATHER call that is failing and examine the data types in the call.   It looks like its failing on the first or nearly the first mpi call in the program so you should look for problems in your mpi install. 
 
Top