[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: fastest/least bandwidth selective archive removal methods?



Thank you James & Graham for the responses, but lets now put things into perspective, *why* I’m asking what I’m asking:

> On 04 Aug 2020, at 07:21 , Graham Percival <gperciva@tarsnap.com> wrote:
> 
> On Mon, Aug 03, 2020 at 06:14:06PM -0700, james young wrote:
>> 
>> On Aug 3, 2020, at 6:15 AM, hvjunk <hvjunk@gmail.com> wrote:
>>> Question: should I delete them one by one, or  cheaper/beter/more-efficient to bundle them all together in batches?
>> 
>> The documentation recommends batches:
>> https://www.tarsnap.com/improve-speed.html#faster-delete
> 
> Yes, to quote that page: "Multiple archives can be deleted with
> the same command; this is usually faster (and never slower) than
> using individual delete commands”.

Thank you for confirming this.

>>> And if in batches: Does tarsnap recover gracefully from connectivity failures while in the process of deletion? Asking as, being >250ms away from the AWS region/zone, I’ve noticed a “hang” on one of my mass deletions, and not sure whether it was busy, or actually stuck and retrying
> 
> Three options that you might want to add to your delete commands:
>  -v             (to see which archive is currently being deleted)
>  --keep-going   (ignore an error from one archive when deleting
>                  multiple archives)
>  --archive-names    (to read a list of archive names from a file,
>                      instead of using multiple -f options)
> 
> The --keep-going is particularly useful for recovering if you
> cancel the tarsnap command.  If you have

Yes, already using it, but that —archive-names I wasn’t aware of, will try that.  How ever my issues starts with this:

Deleting 7 archives (I at present I’m in need for pruning many more) Takes around 3hours?
My traffic utilization also shot throw the roof, but I recall something that is how the dice rolls with tarsnap, so I’ll have to live with that part.

The issue with “recovery” is that tarsnap seems sometimes to be stuck doing that deletion, how can I see whether it’s still busy, or the server is slow, or the link is slow (don’t believe that, but it is from EU side, so we do have latency ;( ), as it being stuck and busy deleting seems to be the same “State” from what I’ve seen.

Oh, and going off course, while this multi day deletion is in progress I can’t do backups, so wonders about an interruption mechanism that is guaranteed.


time tarsnap -v -v -v -d --keep-going -ftracsdbprod01-2019-05-19_06-09-48 -ftracsdbprod01-2019-05-26_06-08-45 -ftracsdbprod01-2019-06-02_06-07-55 -ftracsdbprod01-2019-06-09_06-07-04 -ftracsdbprod01-2019-06-16_06-04-20 -ftracsdbprod01-2019-06-23_06-03-20 -ftracsdbprod01-2019-06-30_06-03-54
Deleting archive "tracsdbprod01-2019-05-19_06-09-48"
                                       Total size  Compressed size
All archives                      225965898333226   33116540894803
  (unique data)                       96202552048      15960528756
tracsdbprod01-2019-05-19_06-09-48     789490332813     118785513597
Deleted data                            446920883         85989389
Deleting archive "tracsdbprod01-2019-05-26_06-08-45"
                                       Total size  Compressed size
All archives                      225174121939237   32997494549562
  (unique data)                       94812088036      15733761124
tracsdbprod01-2019-05-26_06-08-45     791776393989     119046345241
Deleted data                           1390464012        226767632
Deleting archive "tracsdbprod01-2019-06-02_06-07-55"
                                       Total size  Compressed size
All archives                      224452185966072   32889377423202
  (unique data)                       94433885219      15654951606
tracsdbprod01-2019-06-02_06-07-55     721935973165     108117126360
Deleted data                            378202817         78809518
Deleting archive "tracsdbprod01-2019-06-09_06-07-04"
                                       Total size  Compressed size
All archives                      223727553550787   32780981949299
  (unique data)                       94015137980      15571887044
tracsdbprod01-2019-06-09_06-07-04     724632415285     108395473903
Deleted data                            418747239         83064562
Deleting archive "tracsdbprod01-2019-06-16_06-04-20"
                                       Total size  Compressed size
All archives                      223001051971806   32672339466735
  (unique data)                       93586043191      15487088049
tracsdbprod01-2019-06-16_06-04-20     726501578981     108642482564
Deleted data                            429094789         84798995
Deleting archive "tracsdbprod01-2019-06-23_06-03-20"
                                       Total size  Compressed size
All archives                      222272912587289   32563471853664
  (unique data)                       93205034739      15409849238
tracsdbprod01-2019-06-23_06-03-20     728139384517     108867613071
Deleted data                            381008452         77238811
Deleting archive "tracsdbprod01-2019-06-30_06-03-54"
                                       Total size  Compressed size
All archives                      221542862886540   32454351465027
  (unique data)                       93166985343      15400063912
tracsdbprod01-2019-06-30_06-03-54     730049700749     109120388637
Deleted data                             38049396          9785326

real    155m40.198s
user    4m1.124s
sys     0m5.632s

Attachment: signature.asc
Description: Message signed with OpenPGP