Hi Stijn,
The attached files, if put in /etc/profile.d/ on CE and all WNs, will
hopefully take care of this problem :)
Best regards, Antun
-----
Antun Balaz
Research Assistant
E-mail: [log in to unmask]
Web: http://scl.phy.bg.ac.yu/
Phone: +381 11 3160260, Ext. 152
Fax: +381 11 3162190
Scientific Computing Laboratory
Institute of Physics, Belgrade, Serbia
-----
---------- Original Message -----------
From: Stijn De Smet <[log in to unmask]>
To: [log in to unmask]
Sent: Mon, 23 Apr 2007 14:01:29 +0200
Subject: Re: [LCG-ROLLOUT] MPI and EDG_WL_SCRATCH
> Hello,
>
> I installed the scripts, but they don't seem to help(I submit the
> jobs using the MPICH jobtype). I added some lines to the pbs.pm
> script, and the directory looks
> OK(/home/betest001//gram_scratch_VThCRN5Ajx). I saw that the pbs
> jobtype used for this kind of jobs is 'single', so it starts a
> single executable with a number of CPU's reserved for it. If this is
> correct, then the Jobwrapper created at the resource broker is run
> on the first node. This jobwrapper has some lines like this
> included(almost at the start of the script after the function
> definitions):
>
> if [ ! -z "$EDG_WL_SCRATCH" ]; then
> cd $EDG_WL_SCRATCH
> cleanupDir
> fi
>
> newdir="https_3a_2f_2fgridrb.atlantis.ugent.be_3a9000_2frpFW4r8NnYsI4VjjIG41NQ"
> mkdir -p ".mpi/"${newdir}
>
> So the directory my MPI job ends up in, is not ~/.... but /scratch/.mpi/...
>
> Regards,
> Stijn
>
> Antun Balaz wrote:
> > Hi Stijn,
> >
> > You need to replace /opt/globus/lib/perl/Globus/GRAM/JobManager/pbs.pm with
> > the attached pbs.pm (of course, do the diff!). In order to avoid changing back
> > to the old version of pbs.pm after CE reconfiguration, you may also want to
> > replace /opt/globus/setup/globus/pbs.in with the attached. This pbs.pm will
> > put all jobs that do not use more than 1 CPU to $EDG_WL_SCRATCH, so be sure to
> > define it on all WNs! MPI jobs will stay in homes of pool accounts.
> >
> > I would also suggest that you replace
> > /opt/globus/lib/perl/Globus/GRAM/JobManager.pm with the attached
> > JobManager.pm, so that you avoid possible problems with proxy renewal (if
> > /home is shared over NFS, sometimes it happens that the renewed proxy,
> > although availble, is not seen on WNs).
> >
> > Hope this helps,
> > Antun
> >
> >
> > -----
> > Antun Balaz
> > Research Assistant
> > E-mail: [log in to unmask]
> > Web: http://scl.phy.bg.ac.yu/
> >
> > Phone: +381 11 3160260, Ext. 152
> > Fax: +381 11 3162190
> >
> > Scientific Computing Laboratory
> > Institute of Physics, Belgrade, Serbia
> > -----
> >
> > ---------- Original Message -----------
> > From: Stijn De Smet <[log in to unmask]>
> > To: [log in to unmask]
> > Sent: Fri, 20 Apr 2007 08:45:58 +0200
> > Subject: [LCG-ROLLOUT] MPI and EDG_WL_SCRATCH
> >
> >> Hello,
> >>
> >> I recently configured MPI support on my nodes, but when I try to use
> >> it, it always fails because even MPI jobs get started in the
> >> EDG_WL_SCRATCH directory, which isn't shared, while my homedirs are.
> >> My nodes are configured using YAIM, but for the moment, I don't use
> >> the yaim mpi configuration. Is there an easy solution for disabling
> >> the SCRATCH directory for MPI jobs, or do I just have to disable
> >> scratch space completely?
> >>
> >> Regards,
> >> Stijn
> > ------- End of Original Message -------
------- End of Original Message -------
|