Date: Thu, 22 Oct 1998 23:19:28 +1000 From: "Hallam Oaks" <mlnn4@oaks.com.au> To: "freebsd-chat@FreeBSD.ORG" <freebsd-hackers@FreeBSD.ORG> Subject: Multi-terabyte disk farm Message-ID: <199810221320.XAA07482@mail.aussie.org>
next in thread | raw e-mail | index | archive | help
G'Day; I've a few things I'd like, If I may, to get opinions on. My current client (I'm a contractor) hired me to provide them with some assistance on a project that involves the digitising and non-real-time satellite transmission of television commercials. As part of this project, they want to store about 30,000 ads (they call them 'dubs') online for retrieval in a digital form. We estimate that we will need approximately 3-4 terabytes of storage to achieve this aim. Prior to my arrival two weeks ago, they had been planning to use a StorageTek Timberwolf library. This box has two DLT drives, a robot arm, and about 8 terabytes worth of tape slots. The requirement is for 24/7 availability, but speed is not an issue. Provided a dub can be retrieved onto a cache disk within an hour or so of it being requested from archive they'd be happy. Now, I know that this is a top-quality library (or so I have gathered from scanning the net) but when I heard how much the suppliers wanted for it I had to stop and think if there was a better way. I can't give exact figures but the box itself won't leave much change out of about US$100k, and that's with no software. Add the software on a Sun- based server with 100gb of cache and it goes to somewhere in the vicinity of US$375k (!). Needless to say my clients don't want to buy the software, but to write it internally. With current hard drive prices, I estimate we can put together a one terabyte disk farm for about US$60k (cost of media only), spread across several machines using hot swappable drive bays and dual SCSI buses per machine. We don't intend to use RAID unless there's strong advantages to it (I don't know a whole lot about it). We don't need striping or replication. One advantage of doing it via a distributed farm (I theorise) is that if one drive fries or one machine self-destructs, at least the rest of the system will still be working. A fried drive can be restored from one of the 25gb AIT backup tapes made of all the dubs. Secondly (and this is a major call I'm making), it won't work out cheaper for our estimated need of three terabytes unless the cost of HDD's keep on falling. We won't need full capacity until about two years has passed, meaning that we can start out with only a few hundred gig and scale it as we go, taking advantage of (hopefully) falling disk prices and increasing drive sizes. But, I'm stepping on one toe because one of the permanents (a hardware person) strongly favours using the TW. So I have to make darn sure I'm not blowing smoke out of my nether regions before I stick my neck out even further. My desire is to push for the farm because I believe it's better. (Going with the TW would actually be more profitable for me since one of the main reasons they hired me was to write the software to drive the flipping thing:). A farm ought to work out cheaper, be easier to work with (for the rest of the system software on a API level, rather than a hardware level), and could possibly be more reliable overall (being distributed). Additionally, we don't have to spend I-don't-know-how-long writing software to control it like we would with the library. OTOH the down sides are power, heat, extra administration costs, and probably a bunch of stuff I don't know yet (which is why I'm writing this ;). Needless to say I'm going to put my preferred solution as FreeBSD-based. Some of the criteria that I can't yet answer and would like feedback on are these - o is FreeBSD able to be made to recognise a new SCSI drive that wasn't present on boot ? i.e. a new drive is plugged into the hot bays. can it be recognised, formatted, and mounted by manual intervention ? o ditto if a drive fries. can it be taken out without the kernel getting too upset ? o is it feasable to automatically umount and spin down drives that haven't been accessed for a day or so ? typically, the older data (> 6 months) will be rarely, if ever, accessed before its two-year span expires and it's erased. o would the boot time of a system be dramatically prolonged by it having 500 or so gigabytes of SCSI drives hanging off its backside ? (I'm referring to a normal boot, with the drives having been properly unmounted. I don't even want to THINK about waiting on an fsck of 500 gigs of unclean disk ;). OTOH the size of the files is quite large, so it'd be feasable to use huge nodes. I've no particular objection to a reboot if need be to add/remove a drive, but if it took more than, say, 10 minutes, it'd be an issue I'd have to tackle with management. o we're thinking of using Seagate Elite 47gb drives. These are 5400 RPM units (speed isn't an issue to us). Does anyone have any opinions about these (good/bad/indifferent) or of previous members of that drive family ? o does anyone have an opinion as to whether it's safe to assume that drive prices will continue to fall as they have done over the past two years ? I guess my main interest, despite the above questions, is really to hear if others think this is a realistic goal to be aiming for, and if others- who-have-gone-before-me have struck major problems doing the same (or a similar) thing. regards, -- Chris To Unsubscribe: send mail to majordomo@FreeBSD.org with "unsubscribe freebsd-hackers" in the body of the message
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?199810221320.XAA07482>