Hi Patrick, You can also check the reverse lookup from you DNS. Your CE and WN should be able to reverse lookup each other, it is a known problem with torque. Regards Sajjad Asghar -----Original Message----- From: LHC Computer Grid - Rollout [mailto:[log in to unmask]] On Behalf Of Patrick Guio Sent: Wednesday, December 07, 2005 6:53 PM To: [log in to unmask] Subject: [LCG-ROLLOUT] maui/torque trouble Dear Support, I am in the process of installing a LCG2.6.0 CE_torque and WN_torque on a cluster machine running Rocks 3.3.0 (Makalu). The native torque is 1.0.1p6-1 and is packed in a different way (single rpm containing everything: pbs and mom server, and mom and other clients (cli and gui) and was working fine. I first installed the LCG2.6.0 "manually" (using diverses solution pure rpm, yum, apt). Run the yaim configure (for dteam VO). Queue system was created properly and I could submit globus jobs to the lcg queue (dteam). I added manually with qmgr a "non-default" default queue which worked fine. Now I wanted to use the yaim install and the meta rpms for nodes that contain nothing else than dependencies. I had to remove the native torque and get installed the 1.0.1p6-11.SL30X.st version which is packed in several rpms. On CE_torque: torque-1.0.1p6-11.SL30X.st torque-resmom-1.0.1p6-11.SL30X.st torque-server-1.0.1p6-11.SL30X.st torque-clients-1.0.1p6-11.SL30X.st torque-devel-1.0.1p6-11.SL30X.st lcg-CE_torque-2.6.0-sl3 on WN_torque: torque-clients-1.0.1p6-11.SL30X.st torque-1.0.1p6-11.SL30X.st lcg-WN_torque-2.6.0-sl3 torque-resmom-1.0.1p6-11.SL30X.st I noticed that configuration files are not in the same place (/opt/torque/ for native torque vs /var/spool/pbs/ for sl30x) and rerun yaim configuration for both CE and WN and restarted maui, pbs_server, pbs_mom on CE and pbs_mom on WN. That's where I get into trouble! It seems that jobs get queued but for some reason never execute. They stay in the 'Q' state forever. % qsub -q default test-pbs.sh 1037.grid.bccs.uib.no % qstat -a grid.bccs.uib.no: Req'd Req'd Elap Job ID Username Queue Jobname SessID NDS TSK Memory Time S Time --------------- -------- -------- ---------- ------ --- --- ------ ----- - ----- 1037.grid.bccs. patrickg default test-pbs.s -- -- -- -- 48:00 Q -- The same happen if I submit a globus job % globus-job-run grid.bccs.uib.no:2119/jobmanager-lcgpbs -queue dteam /bin/hostname % qstat -a grid.bccs.uib.no: Req'd Req'd Elap Job ID Username Queue Jobname SessID NDS TSK Memory Time S Time --------------- -------- -------- ---------- ------ --- --- ------ ----- - ----- 1038.grid.bccs. dteam001 dteam STDIN -- 1 -- -- 48:00 Q -- Running pbsnodes, I get such output compute-0-0.local state = state-unknown,down np = 2 properties = lcgpro ntype = cluster Looking at the pbs_mom log (in /var/spool/pbs/mom_logs), it seems it is first reading the config file from the WM (/var/spool/pbs/mom_priv/config) $clienthost grid.bccs.uib.no $clienthost grid.local $clienthost localhost $clienthost localhost.localdomain $restricted grid.bccs.uib.no $restricted grid.local $logevent 511 $ideal_load 1.6 $max_load 2.1 $usecp grid.bccs.uib.no:/home /home pbs_mom;Svr;Log;Log opened pbs_mom;Svr;restricted;grid.bccs.uib.no [snip] pbs_mom;Svr;usecp;grid.bccs.uib.no:/home /home pbs_mom;n/a;initialize;independent pbs_mom;Svr;pbs_mom;Is up and then many lines with the same message: pbs_mom;Svr;pbs_mom;im_eof, End of File from addr 129.177.120.153:15001 In the pbs_server log (in /var/spool/pbs/server_logs), there are lines: Type disconnect request received from [log in to unmask], sock=9 Type statusqueue request received from [log in to unmask], sock=9 Type statusjob request received from [log in to unmask], sock=9 On the working node pbs_mom log there are many lines (about every three minutes) Premature end of message from addr 129.177.120.153:15001 The maui run is 3.2.6p11-2_SL30X % rpm -qa|grep maui maui-server-3.2.6p11-2_SL30X maui-3.2.6p11-2_SL30X maui-client-3.2.6p11-2_SL30X and I can see in /var/log/maui.log on CE there are lines like INFO: PBS node compute-0-0.local set to state Down (state-unknown,down) INFO: 0 PBS resources detected on RM base WARNING: no resources detected There seems to a problem of communication between CE and WN. Does the torque expect communication through something different than ssh/scp? The Rocks torque works with ssh and rshd is disabled. Could it be the problem? Also there seems to be some problem with packaging % rpm -qa | grep ^torque | xargs rpm -qV S.5....T c /etc/sysconfig/pbs ..?..... /usr/sbin/pbs_mom ..?..... /usr/sbin/pbs_server missing /var/spool/pbs/server_priv/accounting missing /var/spool/pbs/server_priv/acl_groups missing /var/spool/pbs/server_priv/acl_hosts missing /var/spool/pbs/server_priv/acl_svr missing /var/spool/pbs/server_priv/acl_users missing /var/spool/pbs/server_priv/jobs missing /var/spool/pbs/server_priv/queues S.5....T c /var/spool/pbs/server_name even though % rpm -qf /var/spool/pbs/server_priv/acl_svr torque-server-1.0.1p6-11.SL30X.st Do you any idea what is going wrong? I found at Steve Staylen's site (http://hepunx.rl.ac.uk/~traylens) newer versions of torque rpms for sl3 (2.0.0p1-2.sl3.st) than the one provided by CERN but still with same packaging. Are these test rpms? Can they be used for LCG production? Do you think it will solve my problem? Any help is appreciated Sincerely, Patrick