Skip site navigation (1)Skip section navigation (2)
Date:      Thu, 9 May 2019 15:01:07 -0700
From:      Conrad Meyer <cem@freebsd.org>
To:        Rick Macklem <rmacklem@uoguelph.ca>
Cc:        "freebsd-fs@freebsd.org" <freebsd-fs@freebsd.org>
Subject:   Re: test hash functions for fsid
Message-ID:  <CAG6CVpU4zNtEAka40ArmwuxJ3kW-z32HCaw8o6Ep0d=7VhsQdg@mail.gmail.com>
In-Reply-To: <YQBPR0101MB226084001EB8D4756A3D81BEDD330@YQBPR0101MB2260.CANPRD01.PROD.OUTLOOK.COM>
References:  <YQBPR0101MB2260D82BAE348FB82902508CDD320@YQBPR0101MB2260.CANPRD01.PROD.OUTLOOK.COM> <CAG6CVpUER0ODF5mfdGFA-soSpGbh=qkjc4rKiJEdaOUaUaUmAw@mail.gmail.com> <YQBPR0101MB226068818D8C0D9591DD3D94DD330@YQBPR0101MB2260.CANPRD01.PROD.OUTLOOK.COM> <CAG6CVpWpduHKVU9Y_W01sZUcL6zT4uP5iNnF-4Lr3Bv_bt=jUg@mail.gmail.com> <YQBPR0101MB226084001EB8D4756A3D81BEDD330@YQBPR0101MB2260.CANPRD01.PROD.OUTLOOK.COM>

next in thread | previous in thread | raw e-mail | index | archive | help

Hi Rick,

On Thu, May 9, 2019 at 2:44 PM Rick Macklem <rmacklem@uoguelph.ca> wrote:
> I do now think I can dynamically size it based on how many file systems
> (which is how many structures) will be linked off the table, so this shouldn't be
> a problem.

Great!

> >Load factor is just number of valid entries divided by space in the
> >table.  So if your table has 256 spaces, load factor 0.8 would be
> >having about 205 valid entries.
>
> Hmm. That would suggest a large table size of about 100,000 for Peter's
> 72,000 file system case. It would result in a list length averaging about 1 entry,

Yeah, that's usually how hash tables are sized — aiming for a list
length of 1 (for chaining hash tables) or relatively short probe
sequence (for open addressing / embedded entry tables).  It's what
provides the various "O(1)" properties hash tables ostensibly have.

> but I think a linear search through 10-20 entries won't take long enough to be
> a problem for this case. (This happens whenever mountd receives a SIGHUP and each time a client does a mount.)

Yes, it might not be noticeable in this application either way.

> As noted in the last post, I was thinking along the lines of 10->20 as a target
> linked list length. (Or "table_size = num / L", where L is the target average list length.
> L = 10->20 would be roughly a load average of 10->20.)
>
> Does that sound reasonable? (Interested in hearing from anyone on this.)

Wikipedia discusses it a little bit:
https://en.wikipedia.org/wiki/Hash_table#Key_statistics

Facebook recently open sourced their general-purpose C++ hashtable
implementations and talk at length about various tradeoffs, including
load: https://code.fb.com/developer-tools/f14/?r=1  It might be
interesting to read.

Anyway, usually it's a number less than 1.  10-20x is unconventional.

> Yes. It would be the distribution of values and not their randomness that would
> matter for this case. Hopefully someone with a large # of UFS file systems will
> run the test program and post the results. To be honest, I doubt if anyone will
> create a server with enough UFS file systems for it to be important to hash their
> fsid well. It works fine for the small number of UFS file systems I have.)

Maybe pho@ will take that as a challenge.  You could create a lot of
UFS filesystems with mdconfig and enough RAM.

> Thanks for the comments, rick

Cheers,
Conrad



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?CAG6CVpU4zNtEAka40ArmwuxJ3kW-z32HCaw8o6Ep0d=7VhsQdg>