Hi Stijn, The attached files, if put in /etc/profile.d/ on CE and all WNs, will hopefully take care of this problem :) Best regards, Antun ----- Antun Balaz Research Assistant E-mail: [log in to unmask] Web: http://scl.phy.bg.ac.yu/ Phone: +381 11 3160260, Ext. 152 Fax: +381 11 3162190 Scientific Computing Laboratory Institute of Physics, Belgrade, Serbia ----- ---------- Original Message ----------- From: Stijn De Smet <[log in to unmask]> To: [log in to unmask] Sent: Mon, 23 Apr 2007 14:01:29 +0200 Subject: Re: [LCG-ROLLOUT] MPI and EDG_WL_SCRATCH > Hello, > > I installed the scripts, but they don't seem to help(I submit the > jobs using the MPICH jobtype). I added some lines to the pbs.pm > script, and the directory looks > OK(/home/betest001//gram_scratch_VThCRN5Ajx). I saw that the pbs > jobtype used for this kind of jobs is 'single', so it starts a > single executable with a number of CPU's reserved for it. If this is > correct, then the Jobwrapper created at the resource broker is run > on the first node. This jobwrapper has some lines like this > included(almost at the start of the script after the function > definitions): > > if [ ! -z "$EDG_WL_SCRATCH" ]; then > cd $EDG_WL_SCRATCH > cleanupDir > fi > > newdir="https_3a_2f_2fgridrb.atlantis.ugent.be_3a9000_2frpFW4r8NnYsI4VjjIG41NQ" > mkdir -p ".mpi/"${newdir} > > So the directory my MPI job ends up in, is not ~/.... but /scratch/.mpi/... > > Regards, > Stijn > > Antun Balaz wrote: > > Hi Stijn, > > > > You need to replace /opt/globus/lib/perl/Globus/GRAM/JobManager/pbs.pm with > > the attached pbs.pm (of course, do the diff!). In order to avoid changing back > > to the old version of pbs.pm after CE reconfiguration, you may also want to > > replace /opt/globus/setup/globus/pbs.in with the attached. This pbs.pm will > > put all jobs that do not use more than 1 CPU to $EDG_WL_SCRATCH, so be sure to > > define it on all WNs! MPI jobs will stay in homes of pool accounts. > > > > I would also suggest that you replace > > /opt/globus/lib/perl/Globus/GRAM/JobManager.pm with the attached > > JobManager.pm, so that you avoid possible problems with proxy renewal (if > > /home is shared over NFS, sometimes it happens that the renewed proxy, > > although availble, is not seen on WNs). > > > > Hope this helps, > > Antun > > > > > > ----- > > Antun Balaz > > Research Assistant > > E-mail: [log in to unmask] > > Web: http://scl.phy.bg.ac.yu/ > > > > Phone: +381 11 3160260, Ext. 152 > > Fax: +381 11 3162190 > > > > Scientific Computing Laboratory > > Institute of Physics, Belgrade, Serbia > > ----- > > > > ---------- Original Message ----------- > > From: Stijn De Smet <[log in to unmask]> > > To: [log in to unmask] > > Sent: Fri, 20 Apr 2007 08:45:58 +0200 > > Subject: [LCG-ROLLOUT] MPI and EDG_WL_SCRATCH > > > >> Hello, > >> > >> I recently configured MPI support on my nodes, but when I try to use > >> it, it always fails because even MPI jobs get started in the > >> EDG_WL_SCRATCH directory, which isn't shared, while my homedirs are. > >> My nodes are configured using YAIM, but for the moment, I don't use > >> the yaim mpi configuration. Is there an easy solution for disabling > >> the SCRATCH directory for MPI jobs, or do I just have to disable > >> scratch space completely? > >> > >> Regards, > >> Stijn > > ------- End of Original Message ------- ------- End of Original Message -------