Skip site navigation (1)Skip section navigation (2)
Date:      Sat, 21 Nov 2020 22:11:23 -0500
From:      mike tancsa <mike@sentex.net>
To:        Mateusz Guzik <mjguzik@gmail.com>, Allan Jude <allanjude@freebsd.org>
Cc:        Philip Paeps <philip@freebsd.org>, "Bjoern A. Zeeb" <bz@freebsd.org>, netperf-admin@freebsd.org, netperf-users@freebsd.org, Josh Paetzel <jpaetzel@FreeBSD.org>
Subject:   Re: zoo reboot Friday Nov 20 14:00 UTC
Message-ID:  <2115dda4-bd6a-fe90-ffd9-2d53c67f830f@sentex.net>
In-Reply-To: <949305ed-c248-1ee1-2c53-552f2c732dbc@sentex.net>
References:  <1f8e49ff-e3da-8d24-57f1-11f17389aa84@sentex.net> <CAGudoHH=H4Xok5HG3Hbw7S=6ggdsi%2BN4zHirW50cmLGsLnhd4g@mail.gmail.com> <270b65c0-8085-fe2f-cf4f-7a2e4c17a2e8@sentex.net> <CAGudoHFLy2dxBMGd2AJZ6q6zBsU%2Bn8uLXLSiFZ1QGi_qibySVg@mail.gmail.com> <a716e874-d736-d8d5-9c45-c481f6b3dee7@sentex.net> <CAGudoHELFz7KyzQmRN8pCbgLQXPgCdHyDAQ4pzFLF%2BYswcP87A@mail.gmail.com> <163d1815-fc4a-7987-30c5-0a21e8383c93@sentex.net> <CAGudoHF3c1e2DFSAtyjMpcrbfzmMV5x6kOA_5BT5jyoDyKEHsA@mail.gmail.com> <a1ef98c6-e734-1760-f0cb-a8d31c6acc18@sentex.net> <CAGudoHE%2BxjHdBQAD3cAL84=k-kHDsZNECBGNNOn2LsStL5A7Dg@mail.gmail.com> <f9a074b9-17d3-dcfd-5559-a00e1ac75c07@sentex.net> <c01e037b-bb3a-72de-56dc-335097bb7159@freebsd.org> <CAGudoHF=oqqwt_S07PqYBC71HFR4dW5_bEJ=Lt=JWUvEg5-Jxw@mail.gmail.com> <5a46fa23-b09f-86c2-0cef-a9fbb248f2ec@freebsd.org> <CAGudoHH=LTOEaARFKvkvJ2C4ntk1WbzFTjNhSZ%2B1O=Q2m2kP9Q@mail.gmail.com> <adc30bdf-e485-964a-1c1b-0f2fe3ede704@sentex.net> <949305ed-c248-1ee1-2c53-552f2c732dbc@sentex.net>

next in thread | previous in thread | raw e-mail | index | archive | help
Just a quick update for tonight. I have everything restored (I think)
from the morning of the 20th.  I need to sort out how the home
directories were for some users.  I am just building world/kernel with

WITHOUT_LLVM_ASSERTIONS=yes
WITH_MALLOC_PRODUCTION=yes
KERNCONF=GENERIC-NODEBUG

and then will continue tomorrow afternoon with a reboot as well as try
and sort out everyone's home directory location.  But all the data seems
to be there anyways


    ---Mike

On 11/21/2020 12:47 PM, mike tancsa wrote:
> OK, the new zoo is booting off a pair of 500G SSDs we donated.  I am
> restoring to the raidz array tank
>
>  pigz -d -c zroot-.0.gz | zfs recv -vF tank/old
>
> mdtancsa@zoo2:~ % zpool status
>   pool: tank
>  state: ONLINE
> config:
>
>         NAME        STATE     READ WRITE CKSUM
>         tank        ONLINE       0     0     0
>           raidz1-0  ONLINE       0     0     0
>             ada1p1  ONLINE       0     0     0
>             ada3p1  ONLINE       0     0     0
>             ada4p1  ONLINE       0     0     0
>             ada5p1  ONLINE       0     0     0
>             ada6p1  ONLINE       0     0     0
>
> errors: No known data errors
>
>   pool: zooroot
>  state: ONLINE
> config:
>
>         NAME        STATE     READ WRITE CKSUM
>         zooroot     ONLINE       0     0     0
>           mirror-0  ONLINE       0     0     0
>             ada7p4  ONLINE       0     0     0
>             ada8p4  ONLINE       0     0     0
>
> errors: No known data errors
>
> it seems to be working so far
>
> root@zoo2:/home/mdtancsa # zfs list -t snapshot
> NAME                USED  AVAIL     REFER  MOUNTPOINT
> tank/old@HOURLY30     0B      -      141K  -
> tank/old@HOURLY40     0B      -      141K  -
> tank/old@HOURLY50     0B      -      141K  -
> tank/old@HOURLY00     0B      -      141K  -
> tank/old@HOURLY10     0B      -      141K  -
> tank/old@HOURLY20     0B      -      141K  -
> tank/old@prev-1       0B      -      141K  -
> tank/old@1            0B      -      141K  -
> tank/old@2            0B      -      141K  -
> tank/old@3            0B      -      141K  -
> tank/old@4            0B      -      141K  -
> tank/old@5            0B      -      141K  -
> tank/old@6            0B      -      141K  -
> tank/old@0            0B      -      141K  -
> root@zoo2:/home/mdtancsa #
>
> I imagine it will take a while
>
> After the "level 0" is done,
>
> pigz -d -c zroot-.1.gz | zfs recv -v tank/old
>
> Unfortunately, I set up these backup scripts many years ago before I had
> a sense of zfs and saw it all through the lens of dump/restore :(  It
> was one of those, "I should get to fixing the backup soon" :(
>
>     ---Mike
>
> On 11/21/2020 11:18 AM, mike tancsa wrote:
>> Just going to reinstall now. I will boot from 2 new SSDs and then use 4
>> 4TB in RAIDZ
>>
>> On 11/21/2020 12:47 AM, Mateusz Guzik wrote:
>>> root@zoo2:/home/mjg #  zdb -l /dev/gptid/db15e826-1a9c-11eb-8d25-0cc47a1f2fa0
>>> ------------------------------------
>>> LABEL 0
>>> ------------------------------------
>>>     version: 5000
>>>     name: 'zroot'
>>>     state: 0
>>>     txg: 40630433
>>>     pool_guid: 11911329414887727775
>>>     errata: 0
>>>     hostid: 3594518197
>>>     hostname: 'zoo2.sentex.ca'
>>>     top_guid: 7321270789669113643
>>>     guid: 9170931574354766059
>>>     vdev_children: 4
>>>     vdev_tree:
>>>         type: 'mirror'
>>>         id: 3
>>>         guid: 7321270789669113643
>>>         metaslab_array: 26179
>>>         metaslab_shift: 32
>>>         ashift: 9
>>>         asize: 482373533696
>>>         is_log: 0
>>>         create_txg: 40274122
>>>         children[0]:
>>>             type: 'disk'
>>>             id: 0
>>>             guid: 9170931574354766059
>>>             path: '/dev/gptid/db15e826-1a9c-11eb-8d25-0cc47a1f2fa0'
>>>             whole_disk: 1
>>>             create_txg: 40274122
>>>         children[1]:
>>>             type: 'disk'
>>>             id: 1
>>>             guid: 4871900363652985181
>>>             path: '/dev/mfid1p2'
>>>             whole_disk: 1
>>>             create_txg: 40274122
>>>     features_for_read:
>>>         com.delphix:hole_birth
>>>         com.delphix:embedded_data
>>>     labels = 0 1 2 3
>>>
>>>
>>> On 11/21/20, Allan Jude <allanjude@freebsd.org> wrote:
>>>> On 2020-11-20 21:56, Mateusz Guzik wrote:
>>>>> root@zoo2:/home/mjg # zpool import
>>>>>    pool: zroot
>>>>>      id: 11911329414887727775
>>>>>   state: FAULTED
>>>>> status: The pool metadata is corrupted.
>>>>>  action: The pool cannot be imported due to damaged devices or data.
>>>>> 	The pool may be active on another system, but can be imported using
>>>>> 	the '-f' flag.
>>>>>    see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-72
>>>>>  config:
>>>>>
>>>>> 	zroot                                           FAULTED  corrupted data
>>>>> 	  mirror-0                                      DEGRADED
>>>>> 	    replacing-0                                 DEGRADED
>>>>> 	      1517819109053923011                       OFFLINE
>>>>> 	      ada0p3                                    ONLINE
>>>>> 	    ada1                                        ONLINE
>>>>> 	  mirror-1                                      ONLINE
>>>>> 	    ada3p3                                      ONLINE
>>>>> 	    ada4p3                                      ONLINE
>>>>> 	  mirror-2                                      ONLINE
>>>>> 	    ada5p3                                      ONLINE
>>>>> 	    ada6p3                                      ONLINE
>>>>> 	  mirror-3                                      ONLINE
>>>>> 	    gptid/db15e826-1a9c-11eb-8d25-0cc47a1f2fa0  ONLINE
>>>>> 	    gptid/d98a2545-1a9c-11eb-8d25-0cc47a1f2fa0  ONLINE
>>>>>
>>>>>
>>>>> On 11/21/20, Allan Jude <allanjude@freebsd.org> wrote:
>>>>>> On 2020-11-20 18:05, mike tancsa wrote:
>>>>>>> OK. Although looks like I will have to pull it in from backups now :(
>>>>>>>
>>>>>>>
>>>>>>> root@zoo2:/home/mdtancsa # zpool import -f -R /mnt zroot
>>>>>>> cannot import 'zroot': I/O error
>>>>>>>         Destroy and re-create the pool from
>>>>>>>         a backup source.
>>>>>>> root@zoo2:/home/mdtancsa #
>>>>>>>
>>>>>>> all the disks are there :(  Not sure why its not importing ?
>>>>>>>
>>>>>> Can you get the output of just:
>>>>>>
>>>>>> zpool import
>>>>>>
>>>>>> To try to see what the issue might be
>>>>>>
>>>>>> --
>>>>>> Allan Jude
>>>>>>
>>>> The special vdev appears to be being see as just a plain mirror vdev,
>>>> that is odd.
>>>>
>>>> zdb -l /dev/gptid/db15e826-1a9c-11eb-8d25-0cc47a1f2fa0
>>>>
>>>>
>>>> --
>>>> Allan Jude
>>>>



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?2115dda4-bd6a-fe90-ffd9-2d53c67f830f>