registry team mailing list archive
-
registry team
-
Mailing list archive
-
Message #06028
[Bug 292302] Re: Raidset stays inactive due to wrong # of devices
I am using Ubuntu 10.04 Kernel 2.6.32-22-generic
Asus K8V SE Deluxe BIOS: AMIBIOS Version 08.00.09 ID: A0058002
Promise-Controller deaktivated
4 IDE-harddisks (Samsung SP1634N) configured as RAID-0 connected via the VIA VT8237 controller
All harddisks are show in BIOS identicaly
I created the RAID with the pratitioning tool included on the Ubuntu 10.04 64Bit minimal Instalation-CD
The system worked fine for two weeks or so
after changing the /etc/fstab by adding
tmpfs /tempFileSystem tmpfs noexec,default,noatime,nodiratime 0 0
and removing the line that seeks for a floppy
/dev/fd0 /media/floppy auto rw,noauto,user,sync 0 0 # This line is a example because I can't read my harddisk files...
the system hangs after the message
JDB: barrier-based sync failed on md1-8 - disabling barriers
(The message before the grub-screen that the system can't find a floppy
still appears (floppy controler in BIOS is deactivated))
I can switch to tty7 and back to tty1 but not to the once between (I have not diabled them, tty7 show just a blinking cursor)
If I add an usb-cd-rom it is found and a message is printed out -> the system does not hang totaly
I can not connect via ssh (don't know if i configured the ssh-deamon yet)
If I hit ctrl-alt-del the system shuts down
I can look to some earlyer messages by using shift-pageup. There is a mesage: "md0: unknown partition table", but i think to remember that this massage has been there all the time since installation
And two lines later comes:
EXT4-fs (md1): mounted filesystem with ordered data mode
Begin: Running /scripts/local-bottom ...
Done.
Done.
Begin: Running /scripts/init-bottom ...
Done.
The grub entry is:
normal:
recordfail
insmod raid
insmod mdraid
insmod ext2
set rot='(md1)'
search --no-floppy --fs-uuid --set 88d5917f-fdb3-4673-a3bc-82e29469467a
linux /boot/vmlinuz-2.6.32-22-generic root=UUID=88d5917f-fdb3-4673-a3bc-82e29469467a ro splash
initrd /boot/initrd.img-2.6.32-22-generic
covery:
recordfail
insmod raid
insmod mdraid
insmod ext2
set rot='(md1)'
search --no-floppy --fs-uuid --set 88d5917f-fdb3-4673-a3bc-82e29469467a
echo 'Linux 2.6.32-22-generic wird geladen ...'
linux /boot/vmlinuz-2.6.32-22-generic root=UUID=88d5917f-fdb3-4673-a3bc-82e29469467a ro single
echo ' Initiale Ramdisk wird geladen ...'
initrd /boot/initrd.img-2.6.32-22-generic
I tried following:
-Booting via "Recovery mode" -> nothing changed
-Booting with added bootoption nodmraid gives me the message "dmraid-activate: WARNING: dmraid disable by boot option" -> hangs after the same message
-Booting Ubuntu 10.04 Kernel 2.6.32-21-generic Live-CD
fdisk -l
Disk /dev/sda: 160.0 GB, 160041885696 bytes
255 heads, 63 sectors/track, 19457 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x3870bf41
Device Boot Start End Blocks Id System
/dev/sda1 * 1 244 1951744 fd Linux raid autodetect
Partition 1 does not end on cylinder boundary.
/dev/sda2 244 19458 154336257 5 Extended
/dev/sda5 244 19458 154336256 fd Linux raid autodetect
Disk /dev/sdb: 160.0 GB, 160041885696 bytes
255 heads, 63 sectors/track, 19457 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0xeee62945
Device Boot Start End Blocks Id System
/dev/sdb1 1 244 1951744 fd Linux raid autodetect
Partition 1 does not end on cylinder boundary.
/dev/sdb2 244 19458 154336257 5 Extended
/dev/sdb5 244 19458 154336256 fd Linux raid autodetect
Disk /dev/sdc: 160.0 GB, 160041885696 bytes
255 heads, 63 sectors/track, 19457 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x3898d9fa
Device Boot Start End Blocks Id System
/dev/sdc1 1 244 1951744 fd Linux raid autodetect
Partition 1 does not end on cylinder boundary.
/dev/sdc2 244 19458 154336257 5 Extended
/dev/sdc5 244 19458 154336256 fd Linux raid autodetect
Disk /dev/sdd: 160.0 GB, 160041885696 bytes
255 heads, 63 sectors/track, 19457 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x386237a8
Device Boot Start End Blocks Id System
/dev/sdd1 * 1 244 1951744 fd Linux raid autodetect
Partition 1 does not end on cylinder boundary.
/dev/sdd2 244 19458 154336257 5 Extended
/dev/sdd5 244 19458 154336256 fd Linux raid autodetect
dmraid -r
/dev/sdb: pdc, "pdc_icfhhggf-0", stripe, ok, 312499968 sectors, data@ 0
dmraid -V
dmraid version: 1.0.0.rc16 (2009.09.16) shared
dmraid library version: 1.0.0.rc16 (2009.09.16)
device-mapper version: 4.15.0
dmraid -b
/dev/sdd: 312581808 total, "S0BKJ10YC31771"
/dev/sdc: 312581808 total, "S0BKJ1FYC02391"
/dev/sdb: 312581808 total, "S016J10X454363"
/dev/sda: 312581808 total, "S0BKJ10YC31770"
dmraid -ay
ERROR: pdc: wrong # of devices in RAID set "pdc_icfhhggf-0" [1/2] on /dev/sdb
ERROR: keeping degraded mirror set "pdc_icfhhggf"
RAID set "pdc_icfhhggf-0" was not activateddmraid -tay
ERROR: pdc: wrong # of devices in RAID set "pdc_icfhhggf-0" [1/2] on /dev/sdb
ERROR: keeping degraded mirror set "pdc_icfhhggf"
pdc_icfhhggf-0: 0 312499968 linear /dev/zero 0
ERROR: no mapping possible for RAID set pdc_icfhhggf
I am wondering why there ist [1/2] in the line
ERROR: pdc: wrong # of devices in RAID set "pdc_icfhhggf-0" [1/2] on /dev/sdb
what does it mean? should it not be [1/4]?
dmraid -ay -v -d
DEBUG: not isw at 160041884672
DEBUG: isw trying hard coded -2115 offset.
DEBUG: not isw at 160040802816
DEBUG: not isw at 160041884672
DEBUG: isw trying hard coded -2115 offset.
DEBUG: not isw at 160040802816
DEBUG: not isw at 160041884672
DEBUG: isw trying hard coded -2115 offset.
DEBUG: not isw at 160040802816
DEBUG: not isw at 160041884672
DEBUG: isw trying hard coded -2115 offset.
DEBUG: not isw at 160040802816
DEBUG: _find_set: searching pdc_icfhhggf-0
DEBUG: _find_set: not found pdc_icfhhggf-0
DEBUG: _find_set: searching pdc_icfhhggf
DEBUG: _find_set: not found pdc_icfhhggf
DEBUG: _find_set: searching pdc_icfhhggf-0
DEBUG: _find_set: not found pdc_icfhhggf-0
DEBUG: checking pdc device "/dev/sdb"
ERROR: pdc: wrong # of devices in RAID set "pdc_icfhhggf-0" [1/2] on /dev/sdb
DEBUG: set status of set "pdc_icfhhggf-0" to 2
ERROR: keeping degraded mirror set "pdc_icfhhggf"
RAID set "pdc_icfhhggf-0" was not activated
DEBUG: freeing devices of RAID set "pdc_icfhhggf-0"
DEBUG: freeing device "pdc_icfhhggf-0", path "/dev/sdb"
DEBUG: freeing devices of RAID set "pdc_icfhhggf"
Do I have to repair something? How?
How can I mount the filesystem, using a live cd, so that I can change back the fstab?
Are the changes in fstab the reason for that problem? Or maybe a automaticaly run of the fschk?
--
Raidset stays inactive due to wrong # of devices
https://bugs.launchpad.net/bugs/292302
You received this bug notification because you are a member of Registry
Administrators, which is the registrant for Debian.