Date: Fri, 16 Oct 1998 00:46:29 -0400 (EDT) From: "John W. DeBoskey" <jwd@unx.sas.com> To: mike@smith.net.au (Mike Smith) Cc: freebsd-current@FreeBSD.ORG Subject: Re: -current NFS problem Message-ID: <199810160446.AAA23559@bb01f39.unx.sas.com> In-Reply-To: From Mike Smith at "Oct 14, 98 09:04:20 am"
next in thread | raw e-mail | index | archive | help
> > Perhaps this could be the problem with NFS "hanging" certain people all > > the time? (not the pine thing) The system spending way too much time > > inside the kernel transmitting NFS packets.... > > No. Lack of ACCESS caching makes us slow and eats the network (because > we are very good at generating/sending/receiving them). > > If there's someone out there that wants to work with the very best NFS > people in the business to sort out our problems, please let me know. > NetApp are keen to see our issues resolved (it will mean less angst > for them in the long run, as they have many FreeBSD-using customers). > > Right now, we are accumulating a bad NFS reputation. 8( > ... deleted for brevity... Hi, I have 50 266 Mhz pc's clustered around 3 netapp F630 Filers used as compile servers for an in-house distributed make facility. 85% of the traffic to the netapp(s) 'was' due to access calls. I do not recommend it for public consumption, but the following patch reduces the access overhead to less than 20%. diff nfs_vnops.c /nfs/locutus/usr/src/sys/nfs 273c273 < int v3 = NFS_ISV3(vp); --- > int v3 = 0; /* NFS_ISV3(vp); */ ie: tell the nfs_access() function we're always version 2, even though we're actually connected via 3. The side effect of this patch is to delay file access errors, but this is not a real problem in our environment. FWIW: currently, 28 pc's connected to 1 F630 netapp delivered the following performance (not a controlled test, just an overnight job): >>>STAT: 3149 .o : avg time 00:02 : avg speed 520/minute : 0 failed. ie: 6 minutes to compile 3149 files. Our usage of the filers is sequential read of c/h files, and then sequential write of the .o file. We do not do mmap'd or random access i/o and thus have seen no problems such as those reported on -current. Just my 0.02 John ps: BTW, I cannot amd mount the filers due to a problem with GARBAGE_ARGS being returned from clnt_call() in amfs_host.c\fetch_fhandle() regardless of whether I force a V2 or V3 mount protocol. This is a 'random' error which I can get to occurr approximately 1 out of 10 mounts. We wrote a small shell script to mount/umount the filers and interfaced this with amd via the type=program statement (being tested). So far, this has yet to fail... To Unsubscribe: send mail to majordomo@FreeBSD.org with "unsubscribe freebsd-current" in the body of the message
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?199810160446.AAA23559>