Date: Thu, 9 May 2019 11:41:12 +0200 From: Borja Marcos <borjam@sarenet.es> To: Michelle Sullivan <michelle@sorbs.net> Cc: Walter Parker <walterp@gmail.com>, freebsd-stable@freebsd.org Subject: Re: ZFS... Message-ID: <EF38E9E0-CB7A-48FB-B1FE-C2178CFEFEE9@sarenet.es> In-Reply-To: <7D18A234-E7BF-4855-BD51-4AE2253DB1E4@sorbs.net> References: <30506b3d-64fb-b327-94ae-d9da522f3a48@sorbs.net> <56833732-2945-4BD3-95A6-7AF55AB87674@sorbs.net> <3d0f6436-f3d7-6fee-ed81-a24d44223f2f@netfence.it> <17B373DA-4AFC-4D25-B776-0D0DED98B320@sorbs.net> <70fac2fe3f23f85dd442d93ffea368e1@ultra-secure.de> <70C87D93-D1F9-458E-9723-19F9777E6F12@sorbs.net> <CAGMYy3tYqvrKgk2c==WTwrH03uTN1xQifPRNxXccMsRE1spaRA@mail.gmail.com> <5ED8BADE-7B2C-4B73-93BC-70739911C5E3@sorbs.net> <d0118f7e-7cfc-8bf1-308c-823bce088039@denninger.net> <2e4941bf-999a-7f16-f4fe-1a520f2187c0@sorbs.net> <20190430102024.E84286@mulder.mintsol.com> <41FA461B-40AE-4D34-B280-214B5C5868B5@punkt.de> <20190506080804.Y87441@mulder.mintsol.com> <08E46EBF-154F-4670-B411-482DCE6F395D@sorbs.net> <33D7EFC4-5C15-4FE0-970B-E6034EF80BEF@gromit.dlib.vt.edu> <A535026E-F9F6-4BBA-8287-87EFD02CF207@sorbs.net> <a82bfabe-a8c3-fd9a-55ec-52530d4eafff@denninger.net> <a1b78a63-0ef1-af51-4e33-a9a97a257c8b@sorbs.net> <CAMPTd_A7RYJ12pFyY4TzbXct82kWfr1hcEkSpDg7bjP25xjJGA@mail.gmail.com> <d91cf5@sorbs.net> <7D18A234-E7BF-4855-BD51-4AE2253DB1E4@sorbs.net>
index | next in thread | previous in thread | raw e-mail
> On 9 May 2019, at 00:55, Michelle Sullivan <michelle@sorbs.net> wrote: > > > > This is true, but I am of the thought in alignment with the zFs devs this might not be a good idea... if zfs can’t work it out already, the best thing to do will probably be get everything off it and reformat… That’s true, I would rescue what I could and create the pool again but after testing the setup thoroughly. It would be worth to have a look at the excellent guide offered by the FreeNAS people. It’s full of excellent advice and a priceless list of “donts” such as SATA port multipliers, etc. > >> That sound not be hard to write if everything else on the disk has no >> issues. Don't you say in another message that the system is now returning >> 100's of drive errors. > > No, one disk in the 16 disk zRAID2 ... previously unseen but it could be the errors have occurred in the last 6 weeks... everytime I reboot it started resilvering, gets to 761M resilvered and then stops. That’s a really bad sign. It shouldn’t happen. >> How does that relate the statement =>Everything on >> the disk is fine except for a little bit of corruption in the freespace map? > > Well I think it goes through until it hits that little bit of corruption at stops it mounting... then stops again.. > > I’m seeing 100s of hard errors at the beginning of one of the drives.. they were reported in syslog but only just so could be a new thing. Could be previously undetected.. no way to know. As for disk monitoring, smartmontools can be pretty good although only as an indicator. I also monitor my systems using Orca (I wrote a crude “devilator” many years ago) and I gather disk I/O statistics using GEOM of which the read/write/delete/flush times are very valuable. An ailing disk can be returning valid data but become very slow due to retries. Borja.help
Want to link to this message? Use this
URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?EF38E9E0-CB7A-48FB-B1FE-C2178CFEFEE9>
