Print

Print


Hi Brian,

yes please could you delete:

The directories and contents

Eagle
dp012
dp023
dp030
dp033

DO NOT delete dp010!!!!
and NOT
dp010/dc-mejn1 and jch

my list previously contained dp010 - but that was a HUGE mistake.

Lydia






>>>
>>> They have now been tar'ed archived.
>>>
>>> But do not delete:
>>> dp010/dc-mejn1
>>> and
>>> jch
>>>
>>> dp010/dc-mejn1 contains tars in directories and jch contains tars
>>> and very large files and I would like to keep those.
>>>
>>>
>>>
>>> then delete Eagle
>>> dp010
>>> dp012
>>> dp023

On Thu, 14 Jan 2016, Brian Davies wrote:

> So I am just about to start to delete 7238028 files from castor form the subdirectories I/Lydia listed before.
> I can confirm none of the have "tar" in their filename.
> Brian
>
> -----Original Message-----
> From: DiRAC Users [mailto:[log in to unmask]] On Behalf Of Jensen, Jens (STFC,RAL,SC)
> Sent: 14 January 2016 10:17
> To: [log in to unmask]
> Subject: Re: Backup proposal -option 3
>
> Great stuff!  Now we need to (a) clean up the old stuff (-> Brian), (b) update Document if necessary (Lydia, Jens), (c) get Jon started ASAP (Jon).
>
> Looking at the toplevel directories (ie just under .../cosma5/data) I see about 140TB. Is that what we expect?
>
> Thanks
> -j
>
> On 13/01/2016 17:26, Lydia Heck wrote:
>>
>> Hi Jens,
>>
>> the last of the tars for the 4.2M file job is being done as I a
>> writing this.
>>
>> no more issues with respect to white space in files.
>>
>> I will have to re-do two of the archived directories because of the
>> white space issue, but then I have completed one who project. There
>> are more to go and I would like to start those. But this is the first
>> "clean" transfer.
>>
>> Lydia
>>
>>
>>
>>
>>  On Tue, 12 Jan 2016, Jensen, Jens (STFC,RAL,SC) wrote:
>>
>>> Hi Lydia
>>>
>>> OK, thanks. I will ask Brian to delete the ones we can delete.
>>>
>>> Cheers
>>> --jens
>>>
>>> On 12/01/2016 14:12, Lydia Heck wrote:
>>>>
>>>> Hi Jens
>>>>
>>>> I had this hangup because of previous error messages. I have now
>>>> used 100,000. I think that is enough per tar. It just takes too much
>>>> time otherwise to create or extract.
>>>>
>>>> We could delete all of the other directory structures:
>>>>
>>>> dp010/dc-gord1 dp010/dc-guer1 dp010/dc-hood1 dp010/dc-port2
>>>> dp010/dc-kond1 dp010/dc-maso1
>>>>
>>>> They have now been tar'ed archived.
>>>>
>>>> But do not delete:
>>>> dp010/dc-mejn1
>>>> and
>>>> jch
>>>>
>>>> dp010/dc-mejn1 contains tars in directories and jch contains tars
>>>> and very large files and I would like to keep those.
>>>>
>>>>
>>>>
>>>> then delete Eagle
>>>> dp010
>>>> dp012
>>>> dp023
>>>> dp030
>>>> dp033
>>>>
>>>> Once these have been deleted I will start a parallel stream to
>>>> attack those again.
>>>>
>>>> I will also have to rerun some of the dp010 tars where I was caught
>>>> out file and directory names with white space.
>>>>
>>>> Lydia
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On Tue, 12 Jan 2016, Jensen, Jens (STFC,RAL,SC) wrote:
>>>>
>>>>> Hi Lydia,
>>>>>
>>>>> You should not need an argument list with tar if you use the -T
>>>>> switch (for a file list of filenames) and there is no limit to the
>>>>> number of files you can ask tar to, er, tar. Also files can have
>>>>> spaces in their names (but not newlines; for newlines you will need
>>>>> to use NUL
>>>>> separator)
>>>>>
>>>>> (Otherwise the conventional approach is to use xargs but then you
>>>>> have to call it with tar -u)
>>>>>
>>>>> Anyway, happy that it's going well. Keep it busy. We should talk
>>>>> about deleting the obsolete files.
>>>>>
>>>>> Cheers
>>>>> --jens
>>>>>
>>>>>
>>>>> On 12/01/2016 13:40, Lydia Heck wrote:
>>>>>>
>>>>>> Hi Jens and Brian,
>>>>>>
>>>>>> I have resumed the arriving this morning and the transfers are
>>>>>> going well.
>>>>>>
>>>>>> As I said before I am dealing with a user who has 4.5M files and I
>>>>>> am going through sets of 10,000 each time. I tried with 100,000
>>>>>> but I hit a boundary in that the argument list for tar was too
>>>>>> long. So rather than trial and error I restricted to a maximum of
>>>>>> 10,000.
>>>>>>
>>>>>> at the ~4.1368M file I will hit the issue with name containing
>>>>>> white space.
>>>>>> So this will be with 24 hours or so. Let us see if my script can
>>>>>> handle that.
>>>>>>
>>>>>> Lydia
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>  On Tue, 12 Jan 2016, Jensen, Jens (STFC,RAL,SC) wrote:
>>>>>>
>>>>>>> On 12/01/2016 10:34, Lydia Heck wrote:
>>>>>>>>
>>>>>>>> Hi Jens,
>>>>>>>>
>>>>>>>> my problems were towards the end of the week (Saturday) and in
>>>>>>>> the end I canceled those jobs as they did not lead to anywhere.
>>>>>>>> So they will turn up as "cancelled"
>>>>>>>>
>>>>>>> OK, so most of the transfers seem to have been ticking along
>>>>>>> happily when they were killed but I found one which had clearly stalled:
>>>>>>>
>>>>>>> https://lcgfts06.gridpp.rl.ac.uk:8449/var/log/fts3/2016-01-09/dat
>>>>>>> a.cosma.dur.ac.uk__srm-dirac.gridpp.rl.ac.uk/2016-01-09-1215__dat
>>>>>>> a.cosma.dur.ac.uk__srm-dirac.gridpp.rl.ac.uk__501464354__95f10488
>>>>>>> -f5ae-47af-b4ec-c0e419467880
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> This is the /only one/ within the past week, all the others were
>>>>>>> moving data at about 45MB/s.  Here are the relevant parts of the
>>>>>>> log:
>>>>>>>
>>>>>>> Sat Jan  9 12:34:31 2016 INFO     bytes: 6936682496, avg
>>>>>>> KB/sec:5916,
>>>>>>> inst KB/sec:256, elapsed:1145
>>>>>>> Sat Jan  9 12:34:33 2016 INFO     bytes: 6944808960, avg
>>>>>>> KB/sec:5913,
>>>>>>> inst KB/sec:4176, elapsed:1147
>>>>>>> Sat Jan  9 13:47:37 2016 WARNING  Received signal 15 (Terminated)
>>>>>>> Sat Jan  9 13:47:37 2016 WARNING  TRANSFER
>>>>>>> 95f10488-f5ae-47af-b4ec-c0e419467880 canceled by the user
>>>>>>>
>>>>>>> Curiously the disk server (gdss621) says at that time:
>>>>>>> [14456] Sat Jan  9 12:34:33 2016 :: Finished transferring
>>>>>>> "/284707b3-8e49-4895-e053-01b6f6828eaa".
>>>>>>> [14456] Sat Jan  9 12:34:33 2016 :: Transfer stats:
>>>>>>> DATE=20160109123433.322795 HOST=gdss621.gridpp.rl.ac.uk
>>>>>>> PROG=globus-gridftp-server NL.EVNT=FTP_INFO
>>>>>>> START=20160109121526.124260
>>>>>>> USER=:globus-mapping: FILE=/284707b3-8e49-4895-e053-01b6f6828eaa
>>>>>>> BUFFER=131072 BLOCK=262144 NBYTES=6944808960 VOLUME=/ STREAMS=4
>>>>>>> STRIPES=1 DEST=[129.234.196.65] TYPE=STOR CODE=226
>>>>>>>
>>>>>>> That's about 7GB in ~20 mins, or 35MB/s which does not seem
>>>>>>> unreasonable.
>>>>>>>
>>>>>>> Cheers
>>>>>>> --jens
>>>>>>>
>>>>>
>>>
>