From owner-freebsd-current@FreeBSD.ORG Sat Jun 19 10:48:06 2004 Return-Path: Delivered-To: freebsd-current@freebsd.org Received: from mx1.FreeBSD.org (mx1.freebsd.org [216.136.204.125]) by hub.freebsd.org (Postfix) with ESMTP id 5FE8E16A4CE; Sat, 19 Jun 2004 10:48:06 +0000 (GMT) Received: from havoc.eusc.inter.net (havoc.eusc.inter.net [213.73.101.6]) by mx1.FreeBSD.org (Postfix) with ESMTP id EF8E143D5E; Sat, 19 Jun 2004 10:48:05 +0000 (GMT) (envelope-from msch@snafu.de) Received: from dial-76-070.de.inter.net ([213.73.76.70] helo=current.best-eng.de) by havoc.eusc.inter.net with esmtp (Exim 3.36 #3) id 1BbdNw-0001Df-00; Sat, 19 Jun 2004 12:47:45 +0200 Received: from current.best-eng.de (localhost.best-eng.de [127.0.0.1]) by current.best-eng.de (8.12.11/8.12.11) with ESMTP id i5JAliDb001987; Sat, 19 Jun 2004 12:47:44 +0200 (CEST) (envelope-from matthias@current.best-eng.de) Received: from localhost (localhost [[UNIX: localhost]]) by current.best-eng.de (8.12.11/8.12.11/Submit) id i5JAliVf001986; Sat, 19 Jun 2004 12:47:44 +0200 (CEST) (envelope-from matthias) From: Matthias Schuendehuette Organization: Micro$oft-free Zone To: freebsd-current@freebsd.org Date: Sat, 19 Jun 2004 12:47:43 +0200 User-Agent: KMail/1.6.2 MIME-Version: 1.0 Content-Disposition: inline Content-Type: text/plain; charset="iso-8859-15" Content-Transfer-Encoding: 7bit Message-Id: <200406191247.43937.msch@snafu.de> X-Mailman-Approved-At: Sat, 19 Jun 2004 12:07:10 +0000 Subject: gvinum looses drives X-BeenThere: freebsd-current@freebsd.org X-Mailman-Version: 2.1.1 Precedence: list Reply-To: msch@snafu.de List-Id: Discussions about the use of FreeBSD-current List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 19 Jun 2004 10:48:06 -0000 Hi Lukas et al, I decided to be brave and give geom_vinum a try... Here the typescript of the test with my comments inserted: ------------------------8><------------------------------- Script started on Sat Jun 19 12:25:24 2004 root@current - ~ 501 # gvinum gvinum -> list 5 drives: D testdrive State: up /dev/da4 A: 15/2063 MB (0%) D d1 State: up /dev/da0s3 A: 8/4133 MB (0%) D d2 State: up /dev/da1s1 A: 227/4353 MB (5%) D d3 State: up /dev/da2s1 A: 231/4356 MB (5%) D d4 State: up /dev/da3s2 A: 478/4604 MB (10%) 2 volumes: V mp3dev State: up Plexes: 1 Size: 2048 MB V raid5 State: up Plexes: 1 Size: 12 GB 2 plexes: P mp3dev.p0 C State: up Subdisks: 1 Size: 2048 MB P raid5.p0 R5 State: up Subdisks: 4 Size: 12 GB 5 subdisks: S mp3dev.p0.s0 State: up D: testdrive Size: 2048 MB S raid5.p0.s3 State: up D: d4 Size: 4125 MB S raid5.p0.s2 State: up D: d3 Size: 4125 MB S raid5.p0.s1 State: up D: d2 Size: 4125 MB S raid5.p0.s0 State: up D: d1 Size: 4125 MB gvinum -> quit # Well, looks good! All items found und "up 'n runnin' " # Now mount it! root@current - ~ 502 # mount /dev/gvinum/mp3dev /mp3dev mount: /dev/gvinum/mp3dev: Device not configured # Oops... what's that? root@current - ~ 503 # gvinum list 5 drives: D testdrive State: up /dev/da4 A: 15/2063 MB (0%) D d1 State: up /dev/da0s3 A: 8/4133 MB (0%) D d2 State: up /dev/da1s1 A: 227/4353 MB (5%) D d3 State: up /dev/da2s1 A: 231/4356 MB (5%) D d4 State: up /dev/da3s2 A: 478/4604 MB (10%) 2 volumes: V mp3dev State: down Plexes: 1 Size: 2048 MB V raid5 State: up Plexes: 1 Size: 12 GB 2 plexes: P mp3dev.p0 C State: down Subdisks: 1 Size: 2048 MB P raid5.p0 R5 State: up Subdisks: 4 Size: 12 GB 5 subdisks: S mp3dev.p0.s0 State: stale D: testdrive Size: 2048 MB S raid5.p0.s3 State: up D: d4 Size: 4125 MB S raid5.p0.s2 State: up D: d3 Size: 4125 MB S raid5.p0.s1 State: up D: d2 Size: 4125 MB S raid5.p0.s0 State: up D: d1 Size: 4125 MB # Hmm... 'testdrive' is 'up' but 'mp3dev.p0.s0' is 'stale'. As you see # below, geom_vinum lost 'testdrive' but tells still that it's 'up'... root@current - ~ 504 # mount /dev/gvinum/raid5 /raid/ mount: /dev/gvinum/raid5: Device not configured # Same here... root@current - ~ 505 # gvinum list 5 drives: D testdrive State: up /dev/da4 A: 15/2063 MB (0%) D d1 State: up /dev/da0s3 A: 8/4133 MB (0%) D d2 State: up /dev/da1s1 A: 227/4353 MB (5%) D d3 State: up /dev/da2s1 A: 231/4356 MB (5%) D d4 State: up /dev/da3s2 A: 478/4604 MB (10%) 2 volumes: V mp3dev State: down Plexes: 1 Size: 2048 MB V raid5 State: down Plexes: 1 Size: 12 GB 2 plexes: P mp3dev.p0 C State: down Subdisks: 1 Size: 2048 MB P raid5.p0 R5 State: down Subdisks: 4 Size: 12 GB 5 subdisks: S mp3dev.p0.s0 State: stale D: testdrive Size: 2048 MB S raid5.p0.s3 State: stale D: d4 Size: 4125 MB S raid5.p0.s2 State: stale D: d3 Size: 4125 MB S raid5.p0.s1 State: stale D: d2 Size: 4125 MB S raid5.p0.s0 State: stale D: d1 Size: 4125 MB # The drives are 'up' but the subdisks are 'stale' root@current - ~ 507 # tail -n 256 /var/log/messages Jun 19 12:24:54 current kernel: vinum: unloaded Jun 19 12:25:39 current kernel: FOO: sd raid5.p0.s3 is up Jun 19 12:25:39 current kernel: FOO: sd raid5.p0.s2 is up Jun 19 12:25:40 current kernel: FOO: sd raid5.p0.s1 is up Jun 19 12:25:40 current kernel: FOO: sd raid5.p0.s0 is up Jun 19 12:25:40 current kernel: FOO: sd mp3dev.p0.s0 is up Jun 19 12:26:47 current kernel: gvinum: lost drive 'testdrive' Jun 19 12:26:47 current kernel: FOO: sd mp3dev.p0.s0 is down Jun 19 12:26:47 current kernel: FOO: plex mp3dev.p0 is down Jun 19 12:26:47 current kernel: FOO: sd mp3dev.p0.s0 is stale Jun 19 12:26:47 current kernel: FOO: plex mp3dev.p0 is down Jun 19 12:27:28 current kernel: gvinum: lost drive 'd4' Jun 19 12:27:28 current kernel: FOO: sd raid5.p0.s3 is down Jun 19 12:27:28 current kernel: FOO: plex raid5.p0 is degraded Jun 19 12:27:28 current kernel: gvinum: lost drive 'd3' Jun 19 12:27:28 current kernel: FOO: sd raid5.p0.s2 is down Jun 19 12:27:28 current kernel: FOO: plex raid5.p0 is down Jun 19 12:27:28 current kernel: gvinum: lost drive 'd2' Jun 19 12:27:28 current kernel: FOO: sd raid5.p0.s1 is down Jun 19 12:27:28 current kernel: FOO: plex raid5.p0 is down Jun 19 12:27:28 current kernel: gvinum: lost drive 'd1' Jun 19 12:27:28 current kernel: FOO: sd raid5.p0.s0 is down Jun 19 12:27:28 current kernel: FOO: plex raid5.p0 is down Jun 19 12:27:28 current kernel: FOO: sd raid5.p0.s3 is stale Jun 19 12:27:28 current kernel: FOO: plex raid5.p0 is down Jun 19 12:27:28 current kernel: FOO: sd raid5.p0.s2 is stale Jun 19 12:27:28 current kernel: FOO: plex raid5.p0 is down Jun 19 12:27:28 current kernel: FOO: sd raid5.p0.s1 is stale Jun 19 12:27:28 current kernel: FOO: plex raid5.p0 is down Jun 19 12:27:28 current kernel: FOO: sd raid5.p0.s0 is stale Jun 19 12:27:28 current kernel: FOO: plex raid5.p0 is down root@current - ~ 508 # exit Script done on Sat Jun 19 12:28:19 2004 ------------------------8><------------------------------- So, I'm quite sure you need additional informations - please tell me what. BTW: Your fix of the unload-preventing geom-bug: will it be commited? -- Ciao/BSD - Matthias Matthias Schuendehuette , Berlin (Germany) PGP-Key at and ID: 0xDDFB0A5F