Date: Fri, 21 Mar 2003 12:57:27 +0100 From: Alexander Haderer <alexander.haderer@charite.de> To: Greg 'groggy' Lehey <grog@FreeBSD.org> Cc: Maarten de Vries <mdv@unsavoury.net>, Dirk-Willem van Gulik <dirkx@webweaving.org>, freebsd-questions@FreeBSD.ORG Subject: Re: Three Terabyte Message-ID: <5.2.0.9.1.20030321113340.019d12a0@postamt1.charite.de> In-Reply-To: <20030320235600.GG60356@wantadilla.lemis.com> References: <5.2.0.9.1.20030320125711.019eb9c8@postamt1.charite.de> <20030320111436.N74106-100000@foem.leiden.webweaving.org> <20030320111436.N74106-100000@foem.leiden.webweaving.org> <5.2.0.9.1.20030320125711.019eb9c8@postamt1.charite.de>
next in thread | previous in thread | raw e-mail | index | archive | help
At 10:26 21.03.2003 +1030, Greg 'groggy' Lehey wrote: >On Thursday, 20 March 2003 at 13:13:18 +0100, Alexander Haderer wrote: > > At 12:53 20.03.2003 +0100, Maarten de Vries wrote: > >> This would be for backup. Data on about 50 webservers would be backed up > >> to it on a nightly basis. So performance wouldn't be important. > > > > Sure? Consider this: > > > > a. > > Filling 3TB with 1 Mbyte/s lasts more than 800 hours or 33 days. > >I do a nightly backup to disk. It's compressed (gzip), which is the >bottleneck. I get this sort of performance: > >dump -2uf - /home | gzip > /dump/wantadilla/2/home.gz > ... > DUMP: DUMP: 1254971 tape blocks > DUMP: finished in 217 seconds, throughput 5783 KBytes/sec > DUMP: level 2 dump on Thu Mar 20 21:01:31 2003 > >You don't normally fill up a backup disk at once, so this would be >perfectly adequate. I'd expect a system of the kind that Maarten's >talking about to be able to transfer at least 40 MB/s sequential at >the disk. That would mean he could backup over 1 TB in an 8 hour >period. Of course you are right. My note a. was meant as a more general hint to think about transfer rates when dealing with large files/filesystem. Maarten gave no details about how the webservers are connected with the backup server. I should have give more details of what I mean: When backing up 50 Webservers over network to one backup server the network may become a bottleneck. If you have to use encrypted connections (ssh) because the webservers are located elsewhere you need CPU power at server side for each connection. > > b. > > Using ssh + dump/cpio/tar needs CPU power for encryption, especially when > > multiple clients safe their data at the same time. > >You can share the compression across multiple machines. That's what >was happening in the example above. It is a good idea to do compression at the client side. As I understand your example /dump/wantadilla/2 is either a local dir or connected via NFS. The latter requires a local network if you don't want to do NFS mounts across the Internet. Is this right? with best regards Alexander -- ------------------------------------------------------------------ Alexander Haderer Charite Campus Virchow-Klinikum Tel. +49 30 - 450 557 182 Strahlenklinik und Poliklinik Fax. +49 30 - 450 557 117 Sekr. Prof. Felix Email alexander.haderer@charite.de Augustenburger Platz 1 www http://www.charite.de/rv/str/ 13353 Berlin - Germany ------------------------------------------------------------------ To Unsubscribe: send mail to majordomo@FreeBSD.org with "unsubscribe freebsd-questions" in the body of the message
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?5.2.0.9.1.20030321113340.019d12a0>