Disclaimer
~~~~~~~~~~
This is not a production release.
Please note this is a testing release and was made primarily to
facilitate further improvements and bug fixes and secondarily to show
you where I have got to on in the process of configuring D-Cache 1.6.6
into LCG.
Instructions for testing
~~~~~~~~~~~~~~~~~~~~~~~~
follow the doc book taking this email as an override for all instructions
that are similar
http://storage.esc.rl.ac.uk/documentation/html/D-Cache-Howto/ch02.html
Please add the following line /etc/apt/sources.list.d/gpp_storage.list
rpm http://storage.esc.rl.ac.uk/ apt/datastore/sl3.0.4 testing
then set the following environment options
export RESET_DCACHE_CONFIGURATION=yes
export RESET_DCACHE_PNFS=yes
export RESET_DCACHE_RDBMS=yes
Otherwise follow the How to but instead use the commands
/opt/lcg/yaim/scripts/install_node ./site-info.def SE_dcache
and followed with
/opt/lcg/yaim/scripts/configure_node ./site-info.def SE_dcache
And you should have a head node with SRM and working pool on the same
computer, (After about a 7-10min wait after installation is completed)
PnfsManager is off line, but that can be changed easily and maybe put
on line next release I shall chat further with D-Cache about default
settings.
I was thinking of splitting door and pool nodes in the default setup
making the two independent yaim variables what do you guys think?
Known bugs
~~~~~~~~~~
SRMClientV1 : put, protocols[1]="dcap"
SRMClientV1 : put, protocols[2]="gsiftp"
SRMClientV1 : put, contacting service httpg://dev01.gridpp.rl.ac.uk:8443/srm/ma
nagerv1
doneAddingJobs is false
copy_jobs is empty
Thu Dec 01 16:04:45 GMT 2005: srm returned requestId = -2147483647
Thu Dec 01 16:04:45 GMT 2005: sleeping 1 seconds ...
Thu Dec 01 16:04:48 GMT 2005: sleeping 4 seconds ...
Then loops for ever, which is a pain I shall fix this first!
Not tested for pool nodes and admin nodes separate or for that matter
doors
Some errors and information is displayed that should not
Configuring config_gip
produces a lot of file not found errors but that's what happens if you
branch CVS.
Starting rgma-gin: [FAILED]
Stopping globus-gridftp: [FAILED]
Starting globus-gridftp:execvp: No such file or directory
[FAILED]
These are made public as I don't see why I should hide such information
psql:/opt/d-cache/etc/psql_install_companion.sql:6: NOTICE: CREATE
TABLE / UNIQUE will create implicit index "cacheinfo_pnfsid_key" for
table "cacheinfo"
psql:/opt/d-cache/etc/psql_install_replicas.sql:99: NOTICE: ALTER TABLE
/ ADD PRIMARY KEY will create implicit index "poolname" for table
"pools"
psql:/opt/d-cache/etc/psql_install_replicas.sql:108: NOTICE: ALTER
TABLE / ADD PRIMARY KEY will create implicit index "replica" for table
"replicas"
psql:/opt/d-cache/etc/psql_install_replicas.sql:117: NOTICE: ALTER
TABLE / ADD PRIMARY KEY will create implicit index "hbprocess" for table
"heartbeat"
cat: /opt/pnfs/etc/pnfs_config: No such file or directory
Known odd reports
~~~~~~~~~~~~~~~~~
s|^\(java=\).*|\1/usr/java/j2sdk1.4.2_08/bin/java| /^java_options=/{ :
join /\\$/{ N b join } s|\(["
\n]\)-Dorg.globus.tcp.port.range=[0-9]*,[0-9]*|\1| s|\(["
\n]\)-XX:MaxDirectMemorySize=[0-9]*m|\1| s|\([" \n]\)-Xmx[0-9]*m|\1|
s|\([" \n]\)-server|\1|
s|[^=]*="|&-Dorg.globus.tcp.port.range=50000,52000 \\\ |
s|[^=]*="|&-XX:MaxDirectMemorySize=256m \\\ | s|[^=]*="|&-Xmx256m \\\ |
s|[^=]*="|&-server \\\ | : loop s|\\\n[ ]*"[ ]*$|"| t loop }
s|^\(serviceLocatorHost=\).*|\1dev01.gridpp.rl.ac.uk|
s|^\(defaultPnfsServer=\).*|\1dev01.gridpp.rl.ac.uk|
s|^\(srmDbHost=\).*|\1dev01.gridpp.rl.ac.uk| s|^\(pnfsSrmPath=\).*|\1/|
s|^\(maxActiveGet=\).*|\11000| s|^\(maxActivePut=\).*|\11000|
s|^\(maxActiveCopy=\).*|\11000|
s|^\(RecursiveDirectoryCreation=\).*|\1true|
s|^\(AdvisoryDelete=\).*|\1true|
s|^\(clientDataPortRange=\).*|\150000,52000| s|^\(ADMIN_NODE[
]*\).*|\1dev01.gridpp.rl.ac.uk| s|^\(GSIDCAP[ ]*\).*|\1no| s|^\(GRIDFTP[
]*\).*|\1yes| s|^\(SRM[ ]*\).*|\1no| s|^\(#srmPort=\).*|srmPort=8443|
and other such rubbish that gets printed is debug information which will
be removed before a production release
My siteinfo.def is below
~~~~~~~~~~~~~~~~~~~~~~~~
# YAIM example site configuration file - adapt it to your site!
MY_DOMAIN=gridpp.rl.ac.uk
CE_HOST=my-ce.$MY_DOMAIN
SE_HOST=dev01.$MY_DOMAIN
RB_HOST=lxn1188.cern.ch
PX_HOST=my-px.$MY_DOMAIN
BDII_HOST=lxn1189.cern.ch
MON_HOST=my-mon.$MY_DOMAIN
REG_HOST=lcgic01.gridpp.rl.ac.uk # there is only 1 central registry for
now
# Set this if you are building a LFC server
# not if you're just using clients
LFC_HOST=my-lfc.$MY_DOMAIN
WN_LIST=/opt/lcg/yaim/examples/wn-list.conf
USERS_CONF=/opt/lcg/yaim/examples/users.conf
FUNCTIONS_DIR=/opt/lcg/yaim/functions
# Pick an apt-get repository appropriate to your OS
#LCG_REPOSITORY="rpm
#http://grid-deployment.web.cern.ch/grid-deployment/gis
#apt/LCG-2_4_0/rh73/en/i386 lcg_rh73 lcg_rh73.updates"
#LCG_REPOSITORY="rpm
#http://grid-deployment.web.cern.ch/grid-deployment/gis
#apt/LCG-2_4_0/sl3/en/i386 lcg_sl3 lcg_sl3.updates"
LCG_REPOSITORY="rpm
http://grid-deployment.web.cern.ch/grid-deployment/gis
apt/LCG-2_6_0/sl3/en/i386 lcg_sl3 lcg_sl3.updates"
CA_REPOSITORY="rpm
http://grid-deployment.web.cern.ch/grid-deployment/gis
apt/LCG_CA/en/i386 lcg"
CA_WGET="http://grid-deployment.web.cern.ch/grid-deployment/download/Rp
mDir/security/index.html"
# For the relocatable (tarball) distribution, ensure
# that INSTALL_ROOT is set correctly
INSTALL_ROOT=/opt
# You'll probably want to change these too for the relocatable dist
OUTPUT_STORAGE=/tmp/jobOutput
JAVA_LOCATION="/usr/java/j2sdk1.4.2_08"
# Set this to '/dev/null' or some other dir if you want
# to turn off yaim's installation of cron jobs
CRON_DIR=/etc/cron.d
GLOBUS_TCP_PORT_RANGE="20000 25000"
MYSQL_PASSWORD=password
GRID_TRUSTED_BROKERS=" "
#GRID_TRUSTED_BROKERS="'broker one' 'broker two'"
GRIDMAP_AUTH="ldap://lcg-registrar.cern.ch/ou=users,o=registrar,dc=lcg,
dc=org"
#GRIDMAP_AUTH="ldap://lcg-registrar.cern.ch/ou=users,o=registrar,dc=lcg
#,dc=org ldap://xxx"
GRIDICE_SERVER_HOST=$SE_HOST
SITE_EMAIL=root@localhost
SITE_NAME=my-site-name
SITE_VERSION=LCG-2_6_0
SE_TYPE=disk
JOB_MANAGER=lcgpbs
CE_BATCH_SYS=torque
CE_CPU_MODEL=PIII
CE_CPU_VENDOR=intel
CE_CPU_SPEED=1001
CE_OS=Redhat
CE_OS_RELEASE=7.3
CE_MINPHYSMEM=513
CE_MINVIRTMEM=1025
CE_SMPSIZE=2
CE_SI00=381
CE_SF00=0
CE_OUTBOUNDIP=TRUE
CE_INBOUNDIP=FALSE
CE_RUNTIMEENV="LCG-2 LCG-2_1_0 LCG-2_1_1 LCG-2_2_0 LCG-2_3_0 LCG-2_3_1
LCG-2_4_0 R-GMA"
CE_CLOSE_SE="SE1 SE2"
CE_CLOSE_SE1_HOST=$SE_HOST
CE_CLOSE_SE1_ACCESS_POINT=/storage
CE_CLOSE_SE2_HOST=another-se.$MY_DOMAIN
CE_CLOSE_SE2_ACCESS_POINT=/somewhere
# dCache-specific settings
DCACHE_ADMIN="dev01.gridpp.rl.ac.uk"
DCACHE_POOLS="dev01.gridpp.rl.ac.uk:/pool"
# Optional
# DCACHE_PORT_RANGE="20000,25000"
# SE_dpm-specific settings
DPM_POOLS="lxb1727:/dpmpool2"
# Optional
# DPM_PORT_RANGE="20000,25000" ??
DPMDATA=$CE_CLOSE_SE1_ACCESS_POINT
DPMDB_PWD=dpmu_Bar
DPMUSER_PWD=dpmu_Bar
DPMCONFIG=/home/dpmuser/DPMCONFIG
DPMLOGS=/var/tmp/DPMLogs
DPMFSIZE=200M
DPM_HOST=$SE_HOST
## Temp
DPMPOOL=dpmpool2
BDII_HTTP_URL="http://grid-deployment.web.cern.ch/grid-deployment/gis/l
cg2-bdii/dteam/lcg2-all-sites.conf"
BDII_REGIONS="CE SE RB PX" # list of the services provided by the site
BDII_CE_URL="ldap://$CE_HOST:2135/mds-vo-name=local,o=grid"
BDII_SE_URL="ldap://$SE_HOST:2135/mds-vo-name=local,o=grid"
BDII_RB_URL="ldap://$RB_HOST:2135/mds-vo-name=local,o=grid"
BDII_PX_URL="ldap://$PX_HOST:2135/mds-vo-name=local,o=grid"
VOS="atlas alice lhcb cms dteam sixt"
QUEUES=${VOS}
VO_SW_DIR=/opt/exp_soft
VO_ATLAS_SW_DIR=$VO_SW_DIR/atlas
VO_ATLAS_DEFAULT_SE=$SE_HOST
VO_ATLAS_SGM=ldap://grid-vo.nikhef.nl/ou=lcgadmin,o=atlas,dc=eu-datagri
d,dc=org
VO_ATLAS_USERS=ldap://grid-vo.nikhef.nl/ou=lcg1,o=atlas,dc=eu-datagrid,
dc=org
VO_ATLAS_STORAGE_DIR=$CE_CLOSE_SE1_ACCESS_POINT/atlas
VO_ATLAS_QUEUES="atlas"
VO_ALICE_SW_DIR=$VO_SW_DIR/alice
VO_ALICE_DEFAULT_SE=$SE_HOST
VO_ALICE_SGM=ldap://grid-vo.nikhef.nl/ou=lcgadmin,o=alice,dc=eu-datagri
d,dc=org
VO_ALICE_USERS=ldap://grid-vo.nikhef.nl/ou=lcg1,o=alice,dc=eu-datagrid,
dc=org
VO_ALICE_STORAGE_DIR=$CE_CLOSE_SE1_ACCESS_POINT/alice
VO_ALICE_QUEUES="alice"
VO_CMS_SW_DIR=$VO_SW_DIR/cms
VO_CMS_DEFAULT_SE=$SE_HOST
VO_CMS_SGM=ldap://grid-vo.nikhef.nl/ou=lcgadmin,o=cms,dc=eu-datagrid,dc
=org
VO_CMS_USERS=ldap://grid-vo.nikhef.nl/ou=lcg1,o=cms,dc=eu-datagrid,dc=o
rg
VO_CMS_STORAGE_DIR=$CE_CLOSE_SE1_ACCESS_POINT/cms
VO_CMS_QUEUES="cms"
VO_LHCB_SW_DIR=$VO_SW_DIR/lhcb
VO_LHCB_DEFAULT_SE=$SE_HOST
VO_LHCB_SGM=ldap://grid-vo.nikhef.nl/ou=lcgadmin,o=lhcb,dc=eu-datagrid,
dc=org
VO_LHCB_USERS=ldap://grid-vo.nikhef.nl/ou=lcg1,o=lhcb,dc=eu-datagrid,dc
=org
VO_LHCB_STORAGE_DIR=$CE_CLOSE_SE1_ACCESS_POINT/lhcb
VO_LHCB_QUEUES="lhcb"
VO_DTEAM_SW_DIR=$VO_SW_DIR/dteam
VO_DTEAM_DEFAULT_SE=$SE_HOST
VO_DTEAM_SGM=ldap://lcg-vo.cern.ch/ou=lcgadmin,o=dteam,dc=lcg,dc=org
VO_DTEAM_USERS=ldap://lcg-vo.cern.ch/ou=lcg1,o=dteam,dc=lcg,dc=org
VO_DTEAM_STORAGE_DIR=$CE_CLOSE_SE1_ACCESS_POINT/dteam
VO_DTEAM_QUEUES="dteam"
VO_SIXT_SW_DIR=$VO_SW_DIR/sixt
VO_SIXT_DEFAULT_SE=$SE_HOST
VO_SIXT_USERS=ldap://lcg-vo.cern.ch/ou=lcg1,o=sixt,dc=lcg,dc=org
VO_SIXT_STORAGE_DIR=$CE_CLOSE_SE1_ACCESS_POINT/sixt
VO_SIXT_QUEUES="sixt"
Summary
~~~~~~~
srmcp currently fails I am sure this is a configuration issue.
This is not a production release so if you do want to see this product
please don't expect production quality but do still send in bugs. I shall
be happy to hear any feedback from anyone.
I am now going to set about working on pool nodes.
Regards
Owen
|