Current-Users archive

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]

Re: RAIDframe question



	hello.  If you reboot again, the raid2 will probably look as you
expect.  The general procedure for disk replacement is;

1.  raidctl -a /dev/newdisk raidset

2.  raidctl -F /dev/baddisk raidset (fails the bad disk, uses  the spare and reconstructs to it)

3.  Raid is left with a used_spare, but all is wel.  

4.  Reboot.  All components become optimal.

	It has long been my desire that once a spare is used, it get
automatically promoted to optimal without the interveening reboot.  I
probably could have made this change with Greg's blessing, but I never did
the work.

Hope that helps.
-Brian

On Jun 16, 12:18am, Greywolf wrote:
} Subject: Re: RAIDframe question
} I don't know what I did to get that volume to recover but ripping
} it apart and placing the good component first on reconfiguration
} produced a good volume on a rebuild.  As I recall it looked a lot like this:
} 
} Components:
}           component0: failed
}             /dev/wd1c: optimal
} Spares:
}             /dev/wd0c: spare
} component0 status is: failed. skipping label
} Component label for /dev/wd1c:
}    Row: 0, Column: 1, Num Rows: 1, Num Columns: 2
}    Version: 2, Serial Number: 1984, Mod Counter: 7232
}    Clean: No, Status: 0
}    sectPerSU: 128, SUsPerPU: 4, SUsPerRU: 1
}    Queue size: 120, blocksize: 512, numBlocks: 976772992
}    RAID Level: 1
}    Autoconfig: Yes
}    Root partition: No
}    Last configured as: raid1
} /dev/wd0c status is: spare.  Skipping label.
} Reconstruction is 100% complete.
} Parity Re-write is 100% complete.
} Copyback is 100% complete.
} 
} On the other hand, I have the following showing up after
} a rebuild (different volume, "raid2", mirrored 2TB disks):
} 
} Components:
}             /dev/dk0: optimal
}           component1: spared
} Spares:
}             /dev/dk1: used_spare
} Component label for /dev/dk0:
}    Row: 0, Column: 0, Num Rows: 1, Num Columns: 2
}    Version: 2, Serial Number: 3337, Mod Counter: 468
}    Clean: No, Status: 0
}    sectPerSU: 128, SUsPerPU: 1, SUsPerRU: 1
}    Queue size: 100, blocksize: 512, numBlocks: 3907028992
}    RAID Level: 1
}    Autoconfig: Yes
}    Root partition: No
}    Last configured as: raid2
} component1 status is: spared.  Skipping label.
} Component label for /dev/dk1:
}    Row: 0, Column: 1, Num Rows: 1, Num Columns: 2
}    Version: 2, Serial Number: 3337, Mod Counter: 468
}    Clean: No, Status: 0
}    sectPerSU: 128, SUsPerPU: 1, SUsPerRU: 1
}    Queue size: 100, blocksize: 512, numBlocks: 3907028992
}    RAID Level: 1
}    Autoconfig: Yes
}    Root partition: No
}    Last configured as: raid2
} Parity status: clean
} Reconstruction is 100% complete.
} Parity Re-write is 100% complete.
} Copyback is 100% complete.
} 
} I've been thru enough different results it's hard to tell whether that is sane;
} I would have expected /dev/dk1 to have shifted up to 'optimal' and component1 to
} have vanished.
} 
} On Sat, Jun 13, 2020 at 11:48 PM Martin Husemann <martin%duskware.de@localhost> wrote:
} >
} > On Sat, Jun 13, 2020 at 09:44:35PM -0700, Greywolf wrote:
} > > raidctl -a /dev/wd0c raid1
} > >
} > > raidctl -F component0 raid1
} >
} > I would have expected that to work. What is the raidctl status output
} > after the -a ?
} >
} > Martin
} 
} 
} 
} -- 
} --*greywolf;
>-- End of excerpt from Greywolf




Home | Main Index | Thread Index | Old Index