rsyncing 150GB of data from USA to Europe takes a lot of time
rsyncing 150GB of data from USA to Europe takes a lot of time, close to 6-8 hours of waiting. I am wondering what do people do if they have to rsync like 150TB or a big movie / tube site, which would be even bigger. People dont wait around like 3 years for this to complete probably? It would be A LOT easier to physically ship the disks in a package or something like this.
I am doing something like this for the first time now, and I am wondering how this normally works? Are there any special services for it (like satellite maybe) or what is the situation with that. The same goes for data centers, all this is big, and consumes a lot of electricity and so on. It is all new and modern but it looks like a lot of things could be improved / done at this point - small data centers (how small it can be?), faster connection (like 100,000 faster for situations like this or more). It may be needed, I think, as the time goes by. But the original questions is the rsync speed. People dont wait like 3 years, so how does it work? |
They do not wait 3 years, they use faster connections than you.
|
Quote:
|
I haven't done this exact thing but I have downloaded a 120GB file over bittorrent and it completed in a few days. Rather than using rsync would it be possible to simply use wget or similar for the initial backup then rsync for incremental? I think rsync can also be piped throguh gzip but not done so myself so couldn't say how it's done or whether it's worth it.
|
Video is already compressed so trying to add compression with rsync (or ssh which is what rsync is traveling over) would only slow things down. If your data set is big enough, just put it on an encrypted, removable set of hard drives and ship the drives to the data center. If I recall correctly Netflix sends whole hardware units containing servers with populated drives to various geographic locations rather than transferring over the net. I cannot remember the buzzword they use to refer to the hardware units, but the disk set up in them is reportedly JBOD with FFS.
You have measured the transfer rate over the net. You also have the postal rates for several shipping options, including the time needed to get the drives to their destination. The math is fairly simple to see which one comes out ahead. For the initial population of the remote site, it will probably be the parcel service. There's an informative and possibly relevant backstory to the quote, "Never underestimate the bandwidth of a station wagon full of tapes hurtling down the highway." |
Good point with Netflix :)
|
Quote:
Some general information is here: http://queue.acm.org/detail.cfm?id=2933408 Looking around it seems that some of the data centers have import services for just such an occasion. The ones you are working with might have that available. |
I am uploading a 333MB database from home to the hosting now and it will take rougly 40+ minutes. My Internet plan is basic, so this would be better and it would not be that bad. I am feeling like in the Stone Age today :)
I dont want to sound like I am going to deep on this, but this is showing some important things here, and on a global scale. The cables are not fast enough for what is needed. I wasted like 8 hours today, how much time gets wasted if we factor all the people in, and also the ones who physically ship the files, instead of just sending them over the Net (like 10,000 or 100,000 or 1,000,000 faster than it is now). These speeds are not impossible at all, they will come up with something, I am sure. Or there is something like this already. I've seen Google Fiber on google.com, this looks better / good. Come to think of it, how many pcs / computers gets produced, how much electricity gets useds. I think that the electricity use doubles every four years or so (I am guessing here, but something like this, probably less than 8-10) and that all the computers / mobile phones (all these devices) consume more energy that all the planes combined (like airlines). This is still growing probably... I am not familiar with these topics, but it is all good points. Like how much $$$ economy wise moved to the IT sector now, all the jobs (like you look for a job and there is 1,000 IT offers, and like 100 something else). Things have changed... |
Quote:
An OC/48 is 2.488 Gbps....and if you have one, NO ONE BUT YOU uses it. Do the math there. |
@Adult... For the minimum package of any kind of ISP, waiting 40 minutes to upload 333 MB is not bad at all... Most ISPs are now on Fiber Optics cables or at least they have a backbone based on Fiber Optics but for 10 bucks they (rightfully) limit your upload speed... As for download speed, while higher that's also ussually limited by the ISP based on your package, not on the cables... In Romania we have 1 Gbps (it would take, in theory, 3 seconds to download 333 MB if the uploader can push it and your hardware can handle it) for just a bit more than 15 Euros.. And this is for HOME-use...
But even so, if you have to rsync daily large quantities of data as a business then be sure you're going to use a business option or even a partially dedicated line... If you're going to look, you'll see that solutions for your problems already exists, they we're already invented (and possibily re-invented) ... They just don't come at 10 or 15 Euros/Dollars.. |
Bandwidth costs money. Spend some.
Failing that. Buy a 1Tb USB Drive and mail it to Europe. When did common-sense take a backseat to Computing? Such is the life of a pornography host? |
Also, don't forget that this transfer might be going by satellite!
There's this pesky thing called "The Atlantic Ocean" in the way . . . |
That is funny. "
There's this pesky thing called "The Atlantic Ocean" in the way . . ." Might have to get your old AOL floppies and save it to physical media to send large amounts. Guess it can't hurt to look at modern compression for your type of data. If it is a database and you know typical record sizes then you may be able to consider compression program and options for best results. |
Quote:
Quote:
https://en.wikipedia.org/wiki/Optica...smission_rates Regular users usually don't upload that much data, so the typical 50 Mb down, 5 Mb up connections you get from many ISPs are more than enough. |
Quote:
Between hosting companies, as has been said, your speed may be dependent on a number of factors, such as what contracted bandwidth you have. If it's "speeds up to X Mb/s" then that's a maximum not a minimum, and may be burstable on a 95th percentile basis. You may find that if you try and transfer a large amount at a high speed that you exceed your burst limits and get throttled accordingly. There may be other bottlenecks rather than the pure connectivity element. If you're on shared hosting then you're (surprise, surprise) sharing the connection with all the others on the same box. If you're on a VPS then you're also sharing resources and connectivity with other machines on the box. If the underlying storage is something like RAID5 SATA then this will be a bottleneck compared to something like RAID1+0 SAS All of these things are factors. You mention you're transfering to OVH, are you transfering to their Canada DC or one of the ones in France? I recently migrated around 500Gb of data in 60,000 files to them from another hosting company and the speed was very acceptable. |
All times are GMT -5. The time now is 03:39 PM. |