On Mon, 29 Nov 2010, Catalin Condurache wrote:
> Any tips for ICE jobs in a similar state on WMS?
What would you like to do exactly ?
Cheers, Massimo
>
> Catalin
>
>
>
>> -----Original Message-----
>> From: LHC Computer Grid - Rollout [mailto:[log in to unmask]]
>> On Behalf Of Maarten Litmaath
>> Sent: 24 November 2010 17:40
>> To: [log in to unmask]
>> Subject: Re: [LCG-ROLLOUT] condor jobs in WMS
>>
>> Hola Arnau,
>>
>>> We have many jobs in H status in our WMS. Most of them are really
>> old:
>>>
>>> 184308.0 glite 4/27 21:26 0+00:00:00 H 0 9.8
>> JobWrapper.https_3
>>> 197405.0 glite 5/11 22:18 1+17:25:59 H 0 9.8
>> JobWrapper.https_3
>>> 198019.0 glite 5/12 15:37 0+20:36:19 H 0 9.8
>> JobWrapper.https_3
>>> 198058.0 glite 5/12 15:42 1+03:51:48 H 0 9.8
>> JobWrapper.https_3
>>> 198190.0 glite 5/12 15:55 1+10:48:08 H 0 9.8
>> JobWrapper.https_3
>>> 198536.0 glite 5/12 16:33 0+23:34:50 H 0 9.8
>> JobWrapper.https_3
>>> 198598.0 glite 5/12 16:36 0+18:43:36 H 0 9.8
>> JobWrapper.https_3
>>> 198770.0 glite 5/12 16:55 0+21:51:36 H 0 9.8
>> JobWrapper.https_3
>>> 200168.0 glite 5/13 08:44 0+00:00:00 H 0 9.8
>> JobWrapper.https_3
>>> 214206.0 glite 5/27 00:59 0+00:00:00 H 0 9.8
>> JobWrapper.https_3
>>> 214507.0 glite 5/27 10:48 0+00:00:00 H 0 9.8
>> JobWrapper.https_3
>>> [...]
>>>
>>>
>>> Shouldn't be deleted by any wms process? if not, may I delete them
>> by
>>> hand?
>>> # condor_q|grep -c H
>>> 997
>>
>> That is nothing; some of our WMS had >15k before we did some cleanup.
>> A solution is provided in this bug report:
>>
>> https://savannah.cern.ch/bugs/index.php?70401
>
\|||/
-----------0oo----( o o )----oo0-------------------
(_)
INFN Sezione di Padova
Via Marzolo, 8
35131 Padova - Italy E-mail: massimo.sgaravatto [at] pd.infn.it
Tel: ++39 0498275908 Skype: massimo.sgaravatto
Fax: ++39 0498275952
|