Date: Wed, 6 Dec 2000 22:47:47 -0800 (PST) From: Matt Dillon <dillon@earth.backplane.com> To: "Matt Simerson" <mpsimerson@hostpro.com> Cc: "'A G F Keahan'" <ak@freenet.co.uk>, "'freebsd-hackers@freebsd.org'" <freebsd-hackers@FreeBSD.ORG> Subject: Re: RE: Optimal UFS parameters Message-ID: <200012070647.eB76llo07506@earth.backplane.com> References: <8D18712B2604D411A6BB009027F644980DD7B4@0SEA01EXSRV1>
next in thread | previous in thread | raw e-mail | index | archive | help
:That's because any "consensus" would be inappropriate for mass consumtion. :It really depends on a lot of fun things like the average file size and the :number of files that the drives will be storing. For example, a mail server :might want more inodes than a database server. The mail server will likely :have a lot of tiny files where the database server would have a collection :of much larger (a few k vs several mb's each). : :What makes you think the defaults are unreasonable? I set up a 300GB :filesystem a few months ago. I ran a few numbers, calculated my average file :size, compared it to the defaults and found they were very close to :reasonable. When I get a couple hundred gig's of data on there I'll know :better but I think my guess-timates are very good. : :Matt : :> -----Original Message----- :> From: A G F Keahan [mailto:ak@freenet.co.uk] :> Sent: Wednesday, December 06, 2000 7:53 PM :> To: freebsd-hackers@freebsd.org :> Subject: Optimal UFS parameters :> :> What parameters should I choose for a large (say, 60 or 80Gb) :> filesystem? I remember a while ago someone (phk?) conducted a survey, :> but nothing seems to have come of it. In the meantime, the capacity of :> an average hard drive has increased tenfold, and the defaults have :> become even less reasonable. :> :> What's the current consensus of opinion? :> :> newfs -b ????? -f ????? -c ????? Well, in general I think the defaults are a little overkill... but that may be a good thing. I don't recall us ever getting more then a handful of complaints about a filesystem running out of inodes. Running out of inodes is really annoying and it is best to avoid it. Still, unless your large partition is being used for something like, oh, /home in a multi-user environment, you can probably optimize the newfs parameters a bit to reduce fsck times and indirect block lookup overhead. The default filesystem parameters are: newfs -f 1024 -b 8192 -i 8192 -c 16 ... If you are not going to have a lot of tiny files I would recommend something like this: newfs -f 2048 -b 16384 -i 16384 -c 32 ... You can play with -c and -i, but for a production system the block size (-b) should be either 8192 (the default), or 16384. The filesystem buffer cache is only tuned well for those two sizes and going larger won't help anyway since the kernel already clusters adjacent blocks. Doubling -i from the default halves the number of inodes available. Doubling the cylinders per group reduces the number of allocation groups. If you reduce the number of groups too much your filesystems will become more prone to fragmentation, so don't go overboard. If you increase the number of bytes/inode (-i) too much the filesystem will not have enough inodes and you will run out. For a general purpose filesystem I would not go above -i 16384 -c 64. If the filesystem is going to house a big database (which has many fewer files), you can use a much larger -i but you still shouldn't go overboard with -c. -Matt To Unsubscribe: send mail to majordomo@FreeBSD.org with "unsubscribe freebsd-hackers" in the body of the message
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?200012070647.eB76llo07506>