From owner-freebsd-fs@FreeBSD.ORG Wed Oct 14 22:02:34 2009 Return-Path: Delivered-To: freebsd-fs@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id D1FD2106568F; Wed, 14 Oct 2009 22:02:34 +0000 (UTC) (envelope-from grarpamp@gmail.com) Received: from mail-ew0-f218.google.com (mail-ew0-f218.google.com [209.85.219.218]) by mx1.freebsd.org (Postfix) with ESMTP id 404BD8FC0A; Wed, 14 Oct 2009 22:02:34 +0000 (UTC) Received: by ewy18 with SMTP id 18so262535ewy.43 for ; Wed, 14 Oct 2009 15:02:33 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:mime-version:received:date:message-id:subject :from:to:cc:content-type; bh=bOj3B2qWYfS9Qksn0n6pTLKZ3eMU71mmDxcbec7hk3A=; b=VdzmpkITIVLBfPUl9XTSI8zC8N+8wXIFUKuqb2HYb2RlCFdiV2hPD3/yGSuAS+U8SX 4Xuu3xURzfKobMFm3xe3Sff3JptB7Z/9+TROr2xYzEQF8tiPakH1qu/oTKs4lP+ThBNU bC90w3cLNNyzb19+te2UEjjiov3H0vwcEVB4E= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:date:message-id:subject:from:to:cc:content-type; b=d7fcLNedH3tGM3/m9lw/DydRPnMF9at0qAVtlaM0tKgyhBP8tPxHoisNgkbYyE7q1T gd3a/UMQuNrMpaGOoCu8PSBv6qOOaudiszsODIer8ZzwCc0LQtu+1hjJAWkT1xeFWbrm 451oZmJLihlnl+pQrWd7pkN2KpnkmXSSze1Q4= MIME-Version: 1.0 Received: by 10.211.130.15 with SMTP id h15mr8050293ebn.82.1255557753166; Wed, 14 Oct 2009 15:02:33 -0700 (PDT) Date: Wed, 14 Oct 2009 18:02:32 -0400 Message-ID: From: grarpamp To: freebsd-stable@freebsd.org Content-Type: text/plain; charset=ISO-8859-1 Cc: freebsd-fs@freebsd.org Subject: ZFS repeatable reboot 8.0-RC1 X-BeenThere: freebsd-fs@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Filesystems List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Wed, 14 Oct 2009 22:02:34 -0000 Hi. I'm running i386 on i386, single P4 cpu, 1GiB RAM. SiI 3114 -> SATA [single disk] -> GELI [AES-128] -> ZFS [sha256] Straight RELENG_8 as of cvsup Oct 12 14:49:00 aka 8.0-RC1 plus. ZFS pool is at v13, ZFS fs is at v3. Hardware seems stable. The only modification to config defaults is: loader.conf.local: vfs.zfs.arc_max=100663296 After boot -v, geli, zpool import, xf86, browser, etc my mem looks like: Mem: 33M Active, 22M Inact, 105M Wired, 676K Cache, 37M Buf, 827M Free When putting load on ZFS it usually grows to about: Mem: 95M Active, 22M Inact, 302M Wired, 468K Cache, 37M Buf, 569M Free Ls -l in one of the dirs takes 10min plus and I get: PID USERNAME PRI NICE SIZE RES STATE TIME WCPU COMMAND 11 root 171 ki31 0K 8K RUN 21:24 47.27% idle 1092 user 76 0 77328K 76116K zio->i 3:25 37.89% ls 802 root -8 - 0K 8K geli:w 1:42 8.98% g_eli[0] ad6 9 root -8 - 0K 128K arc_re 0:23 4.88% {arc_reclaim_thre} I did not watch these during rm. I have 1 parent dir holding 4 subdirs. The file count in each subdir is respectively: 256363, 254086, 256017, 178054 Two thirds of files are about 14KiB in size, not many are more than a few MiB nor less than 1KiB though a third are 1 byte. I issue rm -r and after maybe 30 seconds the machine reboots. No syslog, panic or console messages. Dmesg from the prior boot is still present in ram to prove kernel didn't emit any message. memtest86 passes. There are maybe 10 seconds of complete GUI hangup before the reboot occurs. I also see it when make release'ing. Usually during what I _think_ is distributeworld or rolling up the tarballs under /R. This is a big repeatable problem. How can I debug or fix it? Can someone else create some mega sized dirs as above and replicate? Thanks.