Date: Thu, 18 Oct 2012 09:41:41 +0200 From: "Ronald Klop" <ronald-freebsd8@klop.yi.org> To: freebsd-stable@freebsd.org, "Alexander Leidinger" <Alexander@leidinger.net> Subject: Re: daily run output misses zpool errors Message-ID: <op.wmc6brb18527sy@ronaldradial.versatec.local> In-Reply-To: <20121003220149.00007b0c@unknown> References: <20120925013438.4BC4213256@sjakie.klop.ws> <op.wk6zkzx2eclrs1@ronaldradial.versatec.local> <20121003220149.00007b0c@unknown>
next in thread | previous in thread | raw e-mail | index | archive | help
On Wed, 03 Oct 2012 22:01:49 +0200, Alexander Leidinger <Alexander@leidinger.net> wrote: > On Tue, 25 Sep 2012 14:56:49 +0200 "Ronald Klop" <ronald@klop.ws> wrote: > >> Hi, >> >> Below my daily report. And here my zpool status -x. It would be nice >> to see this error in my daily info. I am running with >> daily_show_info="NO", but this looks more severe than info. > > Just to make sure: you verified that you have > daily_status_zfs_enable=YES in periodic.conf? > > In the daily mail you provided I've seen several headings without > content, but I haven't seen the "Checking status of zfs pools:" part > which is supposed to show up when the zfs stats script is run. > > Bye, > Alexander. > Yes. My point is that as long as the pool is healthy the daily e-mail tells me that and when the pool is unhealthy it does not show me any info. I setup a test to reproduce this. I broke a mirror by dd-ing /dev/random over one of the md backing files. $ cat /etc/periodic.conf daily_show_info="NO" weekly_show_info="NO" monthly_show_info="NO" daily_status_zfs_enable="YES" daily_scrub_zfs_enable="YES" daily_status_smart_devices="AUTO" daily_clean_hoststat_enable="NO" daily_status_mail_rejects_enable="NO" daily_status_include_submit_mailq="NO" daily_submit_queuerun="NO" $ zpool status test pool: test state: ONLINE status: One or more devices could not be used because the label is missing or invalid. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Replace the device using 'zpool replace'. see: http://illumos.org/msg/ZFS-8000-4J scan: scrub repaired 0 in 0h0m with 0 errors on Wed Oct 17 14:26:43 2012 config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 mirror-0 ONLINE 0 0 0 md0 ONLINE 0 0 0 18290078248358455968 UNAVAIL 0 0 0 was /dev/md1 errors: No known data errors The daily mail before I broke the pool: --------------------------------------------------------------------- Removing stale files from /var/preserve: Cleaning out old system announcements: Removing stale files from /var/rwho: Backup passwd and group files: Verifying group file syntax: /etc/group is fine Backing up package db directory: Rotating accounting logs and gathering statistics: Checking status of zfs pools: NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT extern 298G 161G 137G 54% 1.00x ONLINE - tank 292G 215G 77.2G 73% 1.00x ONLINE - test 95.5M 5.32M 90.2M 5% 1.00x ONLINE - all pools are healthy Network interface status: Name Mtu Network Address Ipkts Ierrs Idrop Opkts Oerrs Coll Drop em0 1500 <Link#1> 00:21:70:46:6c:da 427203 0 0 321583 0 0 0 em0 1500 192.168.1.0 sjakie.home 368631 - - 322492 - - - em0 1500 192.168.1.36/ 192.168.1.36 64146 - - 0 - - - usbus 0 <Link#2> 0 0 0 0 0 0 0 usbus 0 <Link#3> 0 0 0 0 0 0 0 usbus 0 <Link#4> 0 0 0 0 0 0 0 usbus 0 <Link#5> 0 0 0 0 0 0 0 usbus 0 <Link#6> 0 0 0 0 0 0 0 usbus 0 <Link#7> 0 0 0 0 0 0 0 usbus 0 <Link#8> 0 0 0 0 0 0 0 usbus 0 <Link#9> 0 0 0 0 0 0 0 lo0 16384 <Link#10> 20857 0 0 20857 0 0 0 lo0 16384 localhost ::1 0 - - 0 - - - lo0 16384 fe80::1%lo0 fe80::1 0 - - 0 - - - lo0 16384 your-net localhost 34 - - 20857 - - - ipfw0 65536 <Link#11> 0 0 0 0 0 0 0 Security check: (output mailed separately) Checking for denied zone transfers (AXFR and IXFR): Scrubbing of zfs pools: skipping scrubbing of pool 'extern': last scrubbing is 20 days ago, threshold is set to 35 days skipping scrubbing of pool 'tank': last scrubbing is 4 days ago, threshold is set to 35 days skipping scrubbing of pool 'test': last scrubbing is 0 days ago, threshold is set to 35 days -- End of daily output -- The daily mail after I broke the pool: --------------------------------------------------------------------- Removing stale files from /var/preserve: Cleaning out old system announcements: Removing stale files from /var/rwho: Backup passwd and group files: Verifying group file syntax: /etc/group is fine Backing up package db directory: Rotating accounting logs and gathering statistics: Network interface status: Name Mtu Network Address Ipkts Ierrs Idrop Opkts Oerrs Coll Drop em0 1500 <Link#1> 00:21:70:46:6c:da 586075 0 0 443833 0 0 0 em0 1500 192.168.1.0 sjakie.home 493306 - - 445997 - - - em0 1500 192.168.1.36/ 192.168.1.36 98748 - - 0 - - - usbus 0 <Link#2> 0 0 0 0 0 0 0 usbus 0 <Link#3> 0 0 0 0 0 0 0 usbus 0 <Link#4> 0 0 0 0 0 0 0 usbus 0 <Link#5> 0 0 0 0 0 0 0 usbus 0 <Link#6> 0 0 0 0 0 0 0 usbus 0 <Link#7> 0 0 0 0 0 0 0 usbus 0 <Link#8> 0 0 0 0 0 0 0 usbus 0 <Link#9> 0 0 0 0 0 0 0 lo0 16384 <Link#10> 27243 0 0 27243 0 0 0 lo0 16384 localhost ::1 0 - - 0 - - - lo0 16384 fe80::1%lo0 fe80::1 0 - - 0 - - - lo0 16384 your-net localhost 34 - - 27243 - - - ipfw0 65536 <Link#11> 0 0 0 0 0 0 0 Security check: (output mailed separately) Checking for denied zone transfers (AXFR and IXFR): Scrubbing of zfs pools: skipping scrubbing of pool 'extern': last scrubbing is 21 days ago, threshold is set to 35 days skipping scrubbing of pool 'tank': last scrubbing is 5 days ago, threshold is set to 35 days skipping scrubbing of pool 'test': last scrubbing is 0 days ago, threshold is set to 35 days -- End of daily output --
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?op.wmc6brb18527sy>