Current-Users archive

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]

Re: kern/54289 hosed my RAID. Recovery possible?



On 2019-08-15 00:03, jdbaker%consolidated.net@localhost wrote:
The SiI3214 SATALink card suffers from the identify problem in netbsd-9
and -current (PR kern/54289).

Booting a netbsd-9 kernel, the drives failed to identify which caused
RAIDframe to mark the 4 drives on that card (of 8) in my RAID as FAILED. Rebooting netbsd-8, the drives identify properly, but are still marked as
FAILED.

Is there any way to unmark them so the raid will configure and recover?
Normally 'raidctl -C' is used during first time configuration. Could it
be used to force configuration, ignoring the FAILED status? Would the RAID
be recoverable with parity rebuild afterwards?

This seems to have worked. The disks not being correctly identified/attached under netbsd-9 apparently had them recorded as failed on the components that
did attach (on the machine's on-board intel ahcisata ports).  Rebooting
netbsd-8, although the drives identified and attached properly, they were
still considered failed components.

Being a multiple-disk failure is usually fatal to a RAID, but the components weren't actually failed. Un-configuring with 'raidctl -u' then forcing a config with 'raidctl -C /path/to/config' did not show any fatal errors and subsequent 'raidctl -s' showed all component labels (w/serial number) intact.
Parity rewrite took a long time.

Afterwards, 'gpt show raid0d' and 'dkctl raid0d listwedges' showed things to be intact that far. Rebooting the machine, the RAID properly autoconfigured. 'fsck' reported the filesystem as clean (since it never got mounted after the
failed reboot into netbsd-9).  An 'fsck -f' run is in progress.


John D. Baker


Home | Main Index | Thread Index | Old Index