NetBSD-Users archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
Re: wedges and raidframe recovery
On Jun 14, 2010, at 6:38 38PM, Greg Oster wrote:
> On Mon, 14 Jun 2010 18:34:01 -0400
> Steven Bellovin <smb%cs.columbia.edu@localhost> wrote:
>
>> I'm setting up three 2TB disks in a RAID-5 array, on amd64/5.1_rc2.
>> During testing, I ran into what I suspect is a serious problem.
>>
>> Because of the size of the disks, as best I can tell I have to use
>> wedges instead of disklabel. After poking around and fighting it a
>> bit, I managed to set up dk0, dk1, and dk2 as the three wedges, each
>> of which is all of one disk. I then successfully configured the
>> raid1 device across those three dk devices, and configured a fourth
>> wedge, dk3, to describe all of raid1. I was then able to build a
>> file system:
>>
>> # raidctl -s raid1
>> Components:
>> /dev/dk0: optimal
>> /dev/dk1: optimal
>> /dev/dk2: optimal
>> No spares.
>> Component label for /dev/dk0:
>> Row: 0, Column: 0, Num Rows: 1, Num Columns: 3
>> Version: 2, Serial Number: 2010061200, Mod Counter: 187
>> Clean: No, Status: 0
>> sectPerSU: 64, SUsPerPU: 1, SUsPerRU: 1
>> Queue size: 100, blocksize: 512, numBlocks: 3907028992
>> RAID Level: 5
>> Autoconfig: Yes
>> Root partition: No
>> Last configured as: raid1
>> Component label for /dev/dk1:
>> Row: 0, Column: 1, Num Rows: 1, Num Columns: 3
>> Version: 2, Serial Number: 2010061200, Mod Counter: 187
>> Clean: No, Status: 0
>> sectPerSU: 64, SUsPerPU: 1, SUsPerRU: 1
>> Queue size: 100, blocksize: 512, numBlocks: 3907028992
>> RAID Level: 5
>> Autoconfig: Yes
>> Root partition: No
>> Last configured as: raid1
>> Component label for /dev/dk2:
>> Row: 0, Column: 2, Num Rows: 1, Num Columns: 3
>> Version: 2, Serial Number: 2010061200, Mod Counter: 187
>> Clean: No, Status: 0
>> sectPerSU: 64, SUsPerPU: 1, SUsPerRU: 1
>> Queue size: 100, blocksize: 512, numBlocks: 3907028992
>> RAID Level: 5
>> Autoconfig: Yes
>> Root partition: No
>> Last configured as: raid1
>> Parity status: clean
>> Reconstruction is 100% complete.
>> Parity Re-write is 100% complete.
>> Copyback is 100% complete.
>> # df
>> Filesystem 1024-blocks Used Avail %Cap Mounted on
>> /dev/raid0a 302652118 29210740 258308774 10% /
>> kernfs 1 1 0 100% /kern
>> ptyfs 1 1 0 100% /dev/pts
>> procfs 4 4 0 100% /proc
>> /dev/dk3 3787858122 2 3598465214 0% /shared
>>
>> During some testing involving removing cables, dk2 was perceived as
>> "failed" by the RAIDfrom code. Ah -- a perfect opportunity to test
>> recovery. The problem is that I couldn't make it work; no matter
>> what I tried, I could not induce the kernel to start recovery on that
>> wedge. dmesg showed complaints about being unable to open the
>> device, with 16 -- EBUSY -- as the error code.
>
> What was the raidctl command you used? I'd have expected
>
> raidctl -R /dev/dk2 raid1
>
> to have worked. (i.e. to rebuild that 'disk' in place...)
Yes, that's what I tried, without success. It silently did nothing, even with
-v.
--Steve Bellovin, http://www.cs.columbia.edu/~smb
Home |
Main Index |
Thread Index |
Old Index