Which rsync parameters to backup huge .tar.gz file?
Hello,
my virtualization host server is producing an monthly "snapshot" of the OpenVZ VPS. The file is like vzdump1234.tar.gz and is around 250GB in size, this archive contents changing, i guess like 10% of the VPS files are new/modified. I want to transfer this archive monthly to remote server. Which rsync parameters are clever to be used in order a) least data transfer is used b) local and remote server space is very limited and storing double size of the archive might be issue. c) remote server is low CPU VPS d) both server's HDDs are quite a bottleneck of the overall performance Thank You |
Quote:
It's also odd your signature says "Linux Newbie"...when you've been registered and posting here for FOUR YEARS, and have asked about rsync many times: http://www.linuxquestions.org/questi...er-4175525376/ http://www.linuxquestions.org/questi...ge-4175503702/ http://www.linuxquestions.org/questi...ng-4175528225/ http://www.linuxquestions.org/questi...ze-4175538411/ http://www.linuxquestions.org/questi...nc-4175585087/ Read the "Question Guidelines", which you have been pointed to many, MANY times now. Thread reported. |
Transferring a compressed file will eliminate the benefit of rsync's delta transfer. Upon reaching the first changed byte, the entire remainder of the file will be different due to the compression, and thus will be transmitted in its entirety. At least rsync is intelligent enough to see the ".gz" extension and not attempt re-compression of the already compressed file.
|
Some thoughts only.
Maybe mount the final file like mounting an iso image and then using the directory structure let some program only move the differences. Unison or rsync. I think that then you could use the compression to transmit less after first sync. Jigdo is kind of what I'm thinking. |
gzip has a --rsyncable option nowadays, that is recommended if you want to rsync compressed files. otherwise see post #3
|
Quote:
Not to mention the fact they imply they don't have the disk space to store it. Backing up an entire VM container file with vzdump creates an image, but having multiple images is pointless. Take ONE image, then a delta rsync each day. Container fail? Restore the image via the bare metal restore utilities or vzrestore, and restore the rsync delta after you've got a working system. OP, you STILL need to read the "Question Guidelines" link, and start showing some effort of your own. Sorry, but after 4 years you're not a 'newbie' anymore, and having worked with rsync for at least 3, you should be able to figure out some options on your own. |
Quote:
|
Quote:
|
Quote:
And apparently, this has been an issue for three years now, going back to 2014: http://www.linuxquestions.org/questi...5/#post5284404 ...where the OP is (essentially), asking this same question. Didn't come back there to answer the follow-ups asked by wpeckham, either, and apparently the OP cannot post to the OpenVZ forums either, since they were banned from that forum. The openvz-diff-backups tool is in beta right now, specifically written to take deltas of existing PM's, and shovel it over SSH tunnels to different locations. The OP seems to have not looked. |
If the OP is out of space on source then I'll agree that compressing data to a file is useless. Only to ways to stop in time is snapshot by some means or live state. (wish I could remember that open sources linux live state program name)
Using highest compression levels of the selected compression program may help. Knowing what kind of data it is may yield a better compression program choice. |
Quote:
If the file is already 250GB compressed/gzipp'ed, it probably can't get a lot tighter...which takes us back to the OP saying there's not enough space on the target. The OP hasn't given any information, despite being asked, which seems to be a theme with the OP. |
Quote:
|
All times are GMT -5. The time now is 03:07 PM. |