Skip site navigation (1)Skip section navigation (2)
Date:      Wed, 19 Apr 2017 10:56:05 -0400
From:      Dan Langille <dan@langille.org>
To:        freebsd-fs@freebsd.org
Subject:   vdev state changed & zfs scrub
Message-ID:  <0030E8CC-66B2-4EBF-A63B-91CF8370D526@langille.org>

next in thread | raw e-mail | index | archive | help
I see this on more than one system:

Apr 19 03:12:22 slocum ZFS: vdev state changed, =
pool_guid=3D15387115135938424988 vdev_guid=3D3558867368789024889
Apr 19 03:12:22 slocum ZFS: vdev state changed, =
pool_guid=3D15387115135938424988 vdev_guid=3D3597532040953426928
Apr 19 03:12:22 slocum ZFS: vdev state changed, =
pool_guid=3D15387115135938424988 vdev_guid=3D8095897341669412185
Apr 19 03:12:22 slocum ZFS: vdev state changed, =
pool_guid=3D15387115135938424988 vdev_guid=3D15391662935041273970
Apr 19 03:12:22 slocum ZFS: vdev state changed, =
pool_guid=3D15387115135938424988 vdev_guid=3D8194939911233312160
Apr 19 03:12:22 slocum ZFS: vdev state changed, =
pool_guid=3D15387115135938424988 vdev_guid=3D4885020496131451443
Apr 19 03:12:22 slocum ZFS: vdev state changed, =
pool_guid=3D15387115135938424988 vdev_guid=3D14289732009384117747
Apr 19 03:12:22 slocum ZFS: vdev state changed, =
pool_guid=3D15387115135938424988 vdev_guid=3D7564561573692839552

zpool status output includes:

$ zpool status
  pool: system
 state: ONLINE
  scan: scrub in progress since Wed Apr 19 03:12:22 2017
        2.59T scanned out of 6.17T at 64.6M/s, 16h9m to go
        0 repaired, 41.94% done

The timing of the scrub is not coincidental.

Why is vdev status changing?

Thank you.

--=20
Dan Langille - BSDCan / PGCon
dan@langille.org





Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?0030E8CC-66B2-4EBF-A63B-91CF8370D526>