Skip site navigation (1)Skip section navigation (2)
Date:      Tue, 14 Oct 2014 13:30:27 +0200
From:      Mark Martinec <Mark.Martinec+freebsd@ijs.si>
To:        freebsd-fs@freebsd.org, freebsd-stable@freebsd.org
Subject:   Re: zpool import hangs when out of space - Was: zfs pool import hangs on [tx->tx_sync_done_cv]
Message-ID:  <543D0953.1070604@ijs.si>
In-Reply-To: <138CF459AA0B41EB8CB4E11B3DE932CF@multiplay.co.uk>
References:  <54372173.1010100@ijs.si> <644FA8299BF848E599B82D2C2C298EA7@multiplay.co.uk> <54372EBA.1000908@ijs.si> <DE7DD7A94E9B4F1FBB3AFF57EDB47C67@multiplay.co.uk> <543731F3.8090701@ijs.si> <543AE740.7000808@ijs.si> <A5BA41116A7F4B23A9C9E469C4146B99@multiplay.co.uk> <CAHM0Q_N%2BC=3qgUnyDkEugOFcL=J8gBjbTg8v45Vz3uT=e=Fn2g@mail.gmail.com> <6E01BBEDA9984CCDA14F290D26A8E14D@multiplay.co.uk> <CAHM0Q_OpV2sAQQAH6Cj_=yJWAOt8pTPWQ-m45JSiXDpBwT6WTA@mail.gmail.com> <E2E24A91B8B04C2DBBBC7E029A12BD05@multiplay.co.uk> <CAHM0Q_Oeka25-kdSDRC2evS1R8wuQ0_XgbcdZCjS09aXJ9_WWQ@mail.gmail.com> <14ADE02801754E028D9A0EAB4A16527E@multiplay.co.uk> <543C3C47.4010208@ijs.si> <E3C3C359999140B48943A0E1A04F83A9@multiplay.co.uk> <CAHM0Q_O7LNBiQAEjygANa%2B0rqm9cywjTPbNXabB4TePfEHAZsA@mail.gmail.com> <A7771879317F4194A1D5E4921CD33593@multiplay.co.uk> <138CF459AA0B41EB8CB4E11B3DE932CF@multiplay.co.uk>

next in thread | previous in thread | raw e-mail | index | archive | help
On 10/14/2014 13:19, Steven Hartland wrote:
> Well interesting issue I left this pool alone this morning literally doing
> nothing, and its now out of space.
> zpool list
> NAME       SIZE  ALLOC   FREE   FRAG  EXPANDSZ    CAP  DEDUP  HEALTH
> ALTROOT
> sys1boot  3.97G  3.97G   190K     0%         -    99%  1.00x  ONLINE  -
> sys1copy  3.97G  3.97G     8K     0%         -    99%  1.00x  ONLINE  -
>
> There's something very wrong here as nothing has been accessing the pool.
>
>   pool: zfs
> state: ONLINE
> status: One or more devices are faulted in response to IO failures.
> action: Make sure the affected devices are connected, then run 'zpool
> clear'.
>    see: http://illumos.org/msg/ZFS-8000-HC
>   scan: none requested
> config:
>
>         NAME        STATE     READ WRITE CKSUM
>         zfs         ONLINE       0     2     0
>           md1       ONLINE       0     0     0
>
> I tried destroying the pool and ever that failed, presumably because
> the pool has suspended IO.

That's exactly how trouble started here. Got the
   "One or more devices are faulted in response to IO failures"
on all three small cloned boot pools one day, out of the blue.
There was no activity there, except for periodic snapshoting
every 10 minutes.

   Mark



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?543D0953.1070604>