Users and Sjors:
After looking at a post on the issue in March of 2015, it seemed to be a memory issue.
For the realign_movie_frames step even though specifying --j 12 threads and 6 nodes
for a total of 72 threads, the job ran on 6 nodes using only 1 core per node, based on the top for each node,
so the amount of memory available per node was 32 Gb (2.7 Gb/core).
However, since no intermediate files are produced, it is dealing with a large amount of data since the input
movie star file has a huge number of entries/lines:
grep @Particle cl_1-3_movie.star | wc -l
7578396
this seems very inefficient.
I have 12 frames per movie at about 2.2 e/A2/frame so I am averaging over 3 frames as the ptcls are fairly dense and big.
I don't see how to get any more memory for this job.
Under these conditions is the job able to use all of the 32Gb per node since only one core on each node is active at 100% ?
C Akey
|