Dear Prem, You probably made an error in combining STAR files from the different classes, as the rlnRandomSubset is not always 1 or 2. Be aware that column numbering may be different in different STAR files: the header defines the order of the columns! HTH, Sjors > Dear all, > > I am refining a data set after successful 2D and 3D Classification in > relion 1.4. my input script is > > mpirun -pernode -n 30 `which relion_refine_mpi` --o Ref-class-1-2-3/run1 > --auto_refine --split_random_halves --i Class-1-2-3-merg.star > --particle_diameter 430 --angpix 1.25 --ref run3_6.0A_class001.mrc > --ini_high 50 --ctf --ctf_corrected_ref --flatten_solvent --zero_mask > --oversampling 1 --healpix_order 2 --auto_local_healpix_order 4 > --offset_range 5 --offset_step 2 --sym C1 --low_resol_join_halves 40 > --norm --scale --j 8 --memory_per_thread 4 > -------------- > it is giving following error > ------------------- > Warning: no access to tty (Bad file descriptor). > Thus no job control in this shell. > /usr15/data/psk01/DATA1 > === RELION MPI setup === > + Number of MPI processes = 30 > + Number of threads per MPI process = 8 > + Total number of threads therefore = 240 > + Master (0) runs on host = node54 > + Slave 1 runs on host = node53 > + Slave 2 runs on host = node52 > + Slave 3 runs on host = node51 > + Slave 4 runs on host = node50 > + Slave 5 runs on host = node49 > + Slave 6 runs on host = node48 > + Slave 7 runs on host = node47 > + Slave 8 runs on host = node46 > + Slave 9 runs on host = node45 > + Slave 10 runs on host = node44 > + Slave 11 runs on host = node43 > + Slave 12 runs on host = node42 > + Slave 13 runs on host = node41 > + Slave 14 runs on host = node40 > + Slave 15 runs on host = node39 > + Slave 16 runs on host = node38 > + Slave 17 runs on host = node37 > + Slave 18 runs on host = node36 > + Slave 19 runs on host = node35 > + Slave 20 runs on host = node34 > + Slave 21 runs on host = node33 > + Slave 22 runs on host = node32 > + Slave 23 runs on host = node31 > + Slave 24 runs on host = node30 > + Slave 25 runs on host = node29 > + Slave 26 runs on host = node28 > + Slave 27 runs on host = node27 > + Slave 28 runs on host = node26 > + Slave 29 runs on host = node25 > ================= > Running in double precision. > ERROR Experiment::divideParticlesInRandomHalves: invalid number for random > subset (i.e. not 1 or 2): 7 > File: src/exp_model.cpp line: 211 > -------------------------------------------------------------------------- > MPI_ABORT was invoked on rank 5 in communicator MPI_COMM_WORLD > with errorcode 1. > > NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. > You may or may not see output from other processes, depending on > exactly when Open MPI kills them. > -------------------------------------------------------------------------- > ERROR Experiment::divideParticlesInRandomHalves: invalid number for random > subset (i.e. not 1 or 2): 7 > File: src/exp_model.cpp line: 211 > ERROR Experiment::divideParticlesInRandomHalves: invalid number for random > subset (i.e. not 1 or 2): 7 > File: src/exp_model.cpp line: 211 > [node54:19017] 1 more process has sent help message help-mpi-api.txt / > mpi-abort > [node54:19017] Set MCA parameter "orte_base_help_aggregate" to 0 to see > all help / error messages > --------------------- > Your advice and suggestion would be greatly appreciated > > with regards > -Prem > > -- Sjors Scheres MRC Laboratory of Molecular Biology Francis Crick Avenue, Cambridge Biomedical Campus Cambridge CB2 0QH, U.K. tel: +44 (0)1223 267061 http://www2.mrc-lmb.cam.ac.uk/groups/scheres