Skip site navigation (1)Skip section navigation (2)
Date:      Sat, 11 May 2013 10:44:39 +0200
From:      "Ronald Klop" <ronald-freebsd8@klop.yi.org>
To:        freebsd-fs@freebsd.org
Subject:   Re: Corrupted zpool import -f FAILS state FAULTED
Message-ID:  <op.wwwvwppi8527sy@ronaldradial>
In-Reply-To: <CAKYr3zzP%2BrVin5dAk38C=2HfGym_b3v6je_m-HUtL3eet2fxig@mail.gmail.com>
References:  <CAKYr3zz1gLZArACqdrzkr6APVMvom6y-80omghoo4nb1KMTrKA@mail.gmail.com> <518CFA05.6090706@gmail.com> <CAKYr3zzP%2BrVin5dAk38C=2HfGym_b3v6je_m-HUtL3eet2fxig@mail.gmail.com>

next in thread | previous in thread | raw e-mail | index | archive | help
On Fri, 10 May 2013 16:07:36 +0200, Outback Dingo <outbackdingo@gmail.com>  
wrote:

> On Fri, May 10, 2013 at 9:45 AM, Volodymyr Kostyrko  
> <c.kworr@gmail.com>wrote:
>
>> 09.05.2013 15:31, Outback Dingo:
>>
>>  ok zfsgurus, FreeBSD 9.1-STABLE box zpool import -f reports pool status
>>> Faulted, one of more devices contains corrupted data, however its  
>>> showing
>>> the guid as faulted in the poll, and not the actual disk device  
>>> /dev/daX,
>>> the pool is a single vdev 24 disk raidz3. Essentially the hardward
>>> platform
>>> is a dual node system, with 8 enclosures connected to 24 SAS drives  
>>> via 4
>>> LSI cards. I am not currently using geom_multipath, but the box is  
>>> zoned
>>> so
>>> that each node can see 50% of the drives,
>>> in case of Failure, carp kicks in and migrates "zpool import -af" the
>>> pools
>>> onto the other node. it seems as though somehow the pool is now seeing
>>> guid
>>> and not devices, not sure if they have switched devices ids due to a
>>> reboot.
>>>
>>
>> Am not a zfs guru, but I'll try to help.
>>
>> Any console log snippets are welcome. What does "showing the guid as
>> faulted in the pool" looks like.
>>
>> What are the guids for all partitions? Do they interlap for different
>> nodes?
>>
>> ZFS recognizes devices by tasting they vdev labels and not by their
>> logical location and naming. It can safely report any vdev location -  
>> but
>> it requires the same set vdevs to bring pool online.
>
>
> zdb shows valid data on the drives, no drives have been removed from the  
> box
> whats confusing is why its using guids and not devices daX is what  
> puzzles
> me
>  camcontrol devlist and dmesg clearly show the devices are there. The SAS
> bus is shared
> so both nodes with 2 LSI controllers each see all drives. We were  
> utilizing
> a failover script
> if nodeA dies, carp would kick the script to import the pool to nodeB,  
> both
> nodes are in the
> same chassis and see all the enclosures and all the drives

Are the machines configured the same? As in _exactly_ the same. Glabel  
modules, hint files, sysctls, etc.

Ronald.



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?op.wwwvwppi8527sy>