Scheduled Downtime
On Tuesday 24 October 2023 @ 5pm MT the forums will be in read only mode in preparation for the downtime. On Wednesday 25 October 2023 @ 5am MT, this website will be down for maintenance and expected to return online later in the morning.
Normal Operations
The forums are back online with normal operations. If you notice any issues or errors related to the forums, please reach out to help@ucar.edu

Memusage messages within my output file

I noticed the following messages in my output file while running CAM3 at T85 resolution on Bluevista:

print_memusage: size, rss, share, text, datastack= -1 217700 -1 -1 -1
print_memusage iam 4 End stepon. -1 in the next line means unavailable
print_memusage: size, rss, share, text, datastack= -1 216600 -1 -1 -1
print_memusage iam 19 End stepon. -1 in the next line means unavailable
print_memusage: size, rss, share, text, datastack= -1 212936 -1 -1 -1


Is this something I should be concerned about? Here is bsub script I am using:

## Setting LSF options for batch queue submission.
## 70 8-way nodes available on Bluevista
#BSUB -J cam_eul_som_T85_1pdf
#BSUB -n 32 # total tasks and threads (processors) needed
#BSUB -R "span[ptile=4]" # max number of tasks (MPI) per node
#BSUB -P 36271015 # Project 36271015
#BSUB -B # sends mail at dispatch and intiation times
#BSUB -o /ptmp/scapps/cam_eul_som_T85_1pdf/out.%J # output filename
#BSUB -e /ptmp/scapps/cam_eul_som_T85_1pdf/out.%J # error filename
#BSUB -q regular # queue
#BSUB -W 3:00 # 6 hour wallclock limit (required)
#BSUB -N
##BSUB -x # exclusive use of node (not_shared)
##BSUB -u scapps@uci.edu # email notifications

limit stacksize unlimited
limit datasize unlimited

setenv OMP_NUM_THREADS 4
setenv MP_SHARED_MEMORY yes
setenv XLSMPOPTS "stack=256000000"
setenv AIXTHREAD_SCOPE S
setenv MALLOCMULTIHEAP true
setenv OMP_DYNAMIC false
setenv MP_STDINMODE 0

## POE Environment. Set these for interactive jobs. They're ignored by LoadLeveler
## MP_NODES is the number of nodes. The number chosen should be a power of 2, up to a max of 16 for T42.
setenv MP_NODES 2
setenv MP_TASKS_PER_NODE 2
setenv MP_EUILIB us
setenv MP_RMPOOL 1
# TH: bug fix suggested by Brian Eaton 1/24/03
unsetenv MP_PROCS
 
Top