From owner-freebsd-fs@FreeBSD.ORG Tue Dec 27 16:37:50 2011 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 1FFA61065680 for ; Tue, 27 Dec 2011 16:37:50 +0000 (UTC) (envelope-from freebsd-fs@m.gmane.org) Received: from lo.gmane.org (lo.gmane.org [80.91.229.12]) by mx1.freebsd.org (Postfix) with ESMTP id A6AF28FC0A for ; Tue, 27 Dec 2011 16:37:49 +0000 (UTC) Received: from list by lo.gmane.org with local (Exim 4.69) (envelope-from ) id 1Rfa27-0004W5-0Z for freebsd-fs@freebsd.org; Tue, 27 Dec 2011 17:37:47 +0100 Received: from dyn1243-120.vpn.ic.ac.uk ([129.31.243.120]) by main.gmane.org with esmtp (Gmexim 0.1 (Debian)) id 1AlnuQ-0007hv-00 for ; Tue, 27 Dec 2011 17:37:46 +0100 Received: from johannes by dyn1243-120.vpn.ic.ac.uk with local (Gmexim 0.1 (Debian)) id 1AlnuQ-0007hv-00 for ; Tue, 27 Dec 2011 17:37:46 +0100 X-Injected-Via-Gmane: http://gmane.org/ To: freebsd-fs@freebsd.org From: Johannes Totz Date: Tue, 27 Dec 2011 16:37:32 +0000 Lines: 36 Message-ID: References: <4EE764DA.4030206@brockmann-consult.de> Mime-Version: 1.0 Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit X-Complaints-To: usenet@dough.gmane.org X-Gmane-NNTP-Posting-Host: dyn1243-120.vpn.ic.ac.uk User-Agent: Mozilla/5.0 (Windows NT 6.1; WOW64; rv:8.0) Gecko/20111105 Thunderbird/8.0 In-Reply-To: Subject: Re: zpool failmode=continue X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 27 Dec 2011 16:37:50 -0000 On 13/12/2011 14:53, Johannes Totz wrote: > On 13/12/2011 14:44, Peter Maloney wrote: >> Are you using NFS or ZVOLs? > > Neither, see below. > >> My zfs hangs (all IO) if I go into the .zfs/snapshots directory over >> NFS. (planning to file a PR after I find a way to reproduce it reliably, >> but it depends on specific snapshots). My workaround is to mount >> /var/empty on top of the .zfs directory on the nfs client, and give >> nobody else access. Another workaround I thought of is to have another >> parent directory in the dataset, and share the 2nd level down which >> doesn't contain the .zfs directory. > > My pool is not exported to any clients. My situation is actually the > other way around, should have been more clear: the block device on which > I created the pool is a on the network. > It's kind of a crazy setup: > - sshfs to another (Linux) machine > - create big image file > - create pool from file vdev mounted via sshfs > Eventually the network drops out, zpool shows read and write errors, > fine so far. But all new io just hangs instead of failing with an error. After some observation, turns out that periodic/security/100.chksetuid makes all i/o die on the test pool. Is find doing something funny? As it does not even search around on the testpool (it's imported but not mounted) nor the sshfs (only ufs and zfs is searched) I don't have any clue as to what might go wrong... zpool status simply mentions read/write errors. I noticed this because when logging iostat to a file, i/o always stopped at 3am. But I can also trigger it by simply running 100.chksetuid. All the other stuff in daily and security is fine. Anybody has any idea what might cause it?