Hello there!
Recently, I'm running the CESM2.1.1 on Cheyenne with a compset of the piControl. The simulation has successfully run for 55 years, and gets stuck in MPI error during last weekend. Below shows the MPI error in the log file:
1: Opened file b.e21.B1850.f09_g17.CMIP6-deforest-trop.002.cam.h0.1925-09.nc
1: to write 46
577:MPI error (MPI_File_write_at_all) : I/O error
MPT: Received signal 15
I don't think it's the nc file issue because I've tried to resubmit this run starting from 1925-01, the resubmit has a similar error:
613:MPI error (MPI_File_write_at_all) : I/O error
649:MPI error (MPI_File_write_at_all) : I/O error
Does anyone have ideas on why this happens suddenly, given that the simulations are going well for the first 55 years?
I'm attaching the cesm.log of first time crash and my second try of the resubmit. My run directory is under: /glade/scratch/liyue1/b.e21.B1850.f09_g17.CMIP6-deforest-trop.002/run/
Thanks in advance!
Recently, I'm running the CESM2.1.1 on Cheyenne with a compset of the piControl. The simulation has successfully run for 55 years, and gets stuck in MPI error during last weekend. Below shows the MPI error in the log file:
1: Opened file b.e21.B1850.f09_g17.CMIP6-deforest-trop.002.cam.h0.1925-09.nc
1: to write 46
577:MPI error (MPI_File_write_at_all) : I/O error
MPT: Received signal 15
I don't think it's the nc file issue because I've tried to resubmit this run starting from 1925-01, the resubmit has a similar error:
613:MPI error (MPI_File_write_at_all) : I/O error
649:MPI error (MPI_File_write_at_all) : I/O error
Does anyone have ideas on why this happens suddenly, given that the simulations are going well for the first 55 years?
I'm attaching the cesm.log of first time crash and my second try of the resubmit. My run directory is under: /glade/scratch/liyue1/b.e21.B1850.f09_g17.CMIP6-deforest-trop.002/run/
Thanks in advance!