Date: Tue, 09 Apr 2013 12:19:29 -0400 From: Quartz <quartz@sneakertech.com> To: Tom Evans <tevans.uk@googlemail.com> Cc: FreeBSD FS <freebsd-fs@freebsd.org> Subject: Re: ZFS: Failed pool causes system to hang Message-ID: <51643F91.30704@sneakertech.com> In-Reply-To: <CAFHbX1%2BcdUtumDk3BB1jSE0sCuvVwsTNCAs8e=T0iML6WxHASw@mail.gmail.com> References: <2092374421.4491514.1365459764269.JavaMail.root@k-state.edu> <5163F03B.9060700@sneakertech.com> <CAFHbX1LO9OvbqyYYaob-7nQSA_dwQkMK7%2Bvn9c4QrXQuKvTCFA@mail.gmail.com> <51640BDB.1020403@sneakertech.com> <CAFHbX1%2BcdUtumDk3BB1jSE0sCuvVwsTNCAs8e=T0iML6WxHASw@mail.gmail.com>
next in thread | previous in thread | raw e-mail | index | archive | help
> Sorry, but you've not tested this. Your root is hanging off a > different controller to the others, but it is still using the same > ahci/cam stack. Is ahci/cam getting wedged, causing your root to get > wedged - irrespective of running on a different controller - or is ZFS > causing a deadlock. If I simulate failures by yanking the sata cable to various drives in the pool, I can disconnect any two (raidz2) at random and everything hums along just fine. Status tells me the pool is degraded and if I reconnect them I can resilver and whatnot with no problems. However if I have three drives yanked simultaneously is when everything goes to shit. I don't know the ahci/cam stack from a hole in the wall, but it seems to me that if it can gracefully handle two drives dropping out and coming back randomly, it ought to be able to handle three. I suppose it's possible that zfs itself is not the root cause of the problem, but one way or another there's some kind of interaction here, as I only experience an issue when the pool is no longer solvent. ______________________________________ it has a certain smooth-brained appeal
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?51643F91.30704>