Print

Print


Summary: you need to modify /opt/lcg/etc/lcg-info-dynamic- 
scheduler.conf to map your new pool groups for ATLAS sgm and prd  
users onto the correct VOView of your batch system.

1. Make a copy of lcg-info-dynamic-scheduler.conf to work on (e.g.,  
lcg-info-dynamic-scheduler-new.conf).

2. Edit lcg-info-dynamic-scheduler-new.conf, modifying the current  
mapping to a VOMS Role based VO view to a simple VO name view, e.g.,

atlas:/VO=atlas/GROUP=/atlas/ROLE=production -> atlasprd:atlas

On the LHS you have the primary group of the pool account (warning  
this may be different at your site - use _yours_), on the RHS you  
have the name of the VO in the VO view.

It's probably worth doing this for all VOs, however it's in fact the  
ATLAS production role that it's absolutely essential that you do.

3. Test the new configuration, running the plugin by hand, e.g.,

/opt/lcg/libexec/lcg-info-dynamic-scheduler -c /opt/lcg/etc/lcg-info- 
dynamic-scheduler-new.conf

Check that the GlueVOViewLocalID=VONAME figures are now correct.

ATLAS should now look like:

dn: GlueVOViewLocalID=atlas,GlueCEUniqueID=svr016.gla.scotgrid.ac.uk: 
2119/jobmanager-lcgpbs-atlas,mds-vo-name=local,o=grid
GlueVOViewLocalID: atlas
GlueCEAccessControlBaseRule: VO:atlas
GlueCEStateRunningJobs: 215
GlueCEStateWaitingJobs: 0
GlueCEStateTotalJobs: 215
GlueCEStateFreeJobSlots: 109
GlueCEStateEstimatedResponseTime: 0
GlueCEStateWorstResponseTime: 0

4. When you get it correrct (should be pretty easy!) then move then  
backup the old configuration and copy in the new one to /opt/lcg/etc/ 
lcg-info-dynamic-scheduler.conf.


Some other notes:

1. If you have a DNS VO name, then give the full name in the RHS:

snemo:/VO=supernemo.vo.eu-egee.org/GROUP=/supernemo.vo.eu-egee.org/ 
ROLE=lcgadmin -> snemosgm:supernemo.vo.eu-egee.org

2. You cannot run the current YAIM config_gip_scheduler_plugin or it  
will put the file back the way it was. Wise idea to backup the new  
file, in case it gets trashed by mistake (there are a few other  
obvious tricks: change /opt/lcg/var/gip//plugin/lcg-info-dynamic- 
scheduler-wrapper; override the YAIM function).

3. If you have no jobs running for a VO, so cannot tell if the  
changes have been effective, then su to an atlas production account  
and qsub in a sleep job to the batch system and check you get a 1 for  
the VOView.

4. If you haven't moved to pool accounts for prd and sgm accounts  
then I'm pretty sure you don't need to do this - but please check the  
VOView is right.

5. These instructions are for torque/pbs. For SGE sites I have no  
idea what plugin you're using to provide this information!


OK, any questions or problems, please ask me. It's really important  
for ATLAS that this change gets made correctly - ATLAS have a lot of  
production jobs right now and fixing this will help the right numbers  
of jobs to your site.

Sites we know need to do this are:

Edinburgh
IC-HEP (hep-ce.cx1.hpc.ic.ac.uk and ce00.hep.ph.ic.ac.uk)
Birmingham
Lancaster
Liverpool
RAL PP
Sheffield
Bristol
Cambridge
Glasgow (DONE)
Oxford (2ce02.physics.ox.ac.uk and t2ce03.physics.ox.ac.uk)


Thanks

Graeme

--
Dr Graeme Stewart - http://wiki.gridpp.ac.uk/wiki/User:Graeme_stewart
ScotGrid - http://www.scotgrid.ac.uk/ http://scotgrid.blogspot.com/