system stops booting after message JDB: barrier-based sync failed on md1-8 - disabling barriers ... wrong # of devices in RAID
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
dmraid (Ubuntu) |
Invalid
|
Undecided
|
Unassigned |
Bug Description
Binary package hint: dmraid
I am using Ubuntu 10.04 Kernel 2.6.32-22-generic
Asus K8V SE Deluxe BIOS: AMIBIOS Version 08.00.09 ID: A0058002
Promise-Controller deaktivated
4 IDE-harddisks (Samsung SP1634N) configured as RAID-0 connected via the VIA VT8237 controller
All harddisks are show in BIOS identicaly
I created the RAID with the pratitioning tool included on the Ubuntu 10.04 64Bit minimal Instalation-CD
The system worked fine for two weeks or so
after changing the /etc/fstab by adding
tmpfs /tempFileSystem tmpfs noexec,
and removing the line that seeks for a floppy
/dev/fd0 /media/floppy auto rw,noauto,user,sync 0 0 # This line is a example because I can't read my harddisk files...
the system hangs after the message
JDB: barrier-based sync failed on md1-8 - disabling barriers
(The message before the grub-screen that the system can't find a floppy still appears (floppy controler in BIOS is deactivated))
I can switch to tty7 and back to tty1 but not to the once between (I have not diabled them, tty7 show just a blinking cursor)
If I add an usb-cd-rom it is found and a message is printed out -> the system does not hang totaly
I can not connect via ssh (don't know if i configured the ssh-deamon yet)
If I hit ctrl-alt-del the system shuts down
I can look to some earlyer messages by using shift-pageup. There is a mesage: "md0: unknown partition table", but i think to remember that this massage has been there all the time since installation
And two lines later comes:
EXT4-fs (md1): mounted filesystem with ordered data mode
Begin: Running /scripts/
Done.
Done.
Begin: Running /scripts/
Done.
The grub entry is:
normal:
recordfail
insmod raid
insmod mdraid
insmod ext2
set rot='(md1)'
search --no-floppy --fs-uuid --set 88d5917f-
linux /boot/vmlinuz-
initrd /boot/initrd.
covery:
recordfail
insmod raid
insmod mdraid
insmod ext2
set rot='(md1)'
search --no-floppy --fs-uuid --set 88d5917f-
echo 'Linux 2.6.32-22-generic wird geladen ...'
linux /boot/vmlinuz-
echo ' Initiale Ramdisk wird geladen ...'
initrd /boot/initrd.
I tried following:
-Booting via "Recovery mode" -> nothing changed
-Booting with added bootoption nodmraid gives me the message "dmraid-activate: WARNING: dmraid disable by boot option" -> hangs after the same message
-Booting Ubuntu 10.04 Kernel 2.6.32-21-generic Live-CD
fdisk -l
Disk /dev/sda: 160.0 GB, 160041885696 bytes
255 heads, 63 sectors/track, 19457 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x3870bf41
Device Boot Start End Blocks Id System
/dev/sda1 * 1 244 1951744 fd Linux raid autodetect
Partition 1 does not end on cylinder boundary.
/dev/sda2 244 19458 154336257 5 Extended
/dev/sda5 244 19458 154336256 fd Linux raid autodetect
Disk /dev/sdb: 160.0 GB, 160041885696 bytes
255 heads, 63 sectors/track, 19457 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0xeee62945
Device Boot Start End Blocks Id System
/dev/sdb1 1 244 1951744 fd Linux raid autodetect
Partition 1 does not end on cylinder boundary.
/dev/sdb2 244 19458 154336257 5 Extended
/dev/sdb5 244 19458 154336256 fd Linux raid autodetect
Disk /dev/sdc: 160.0 GB, 160041885696 bytes
255 heads, 63 sectors/track, 19457 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x3898d9fa
Device Boot Start End Blocks Id System
/dev/sdc1 1 244 1951744 fd Linux raid autodetect
Partition 1 does not end on cylinder boundary.
/dev/sdc2 244 19458 154336257 5 Extended
/dev/sdc5 244 19458 154336256 fd Linux raid autodetect
Disk /dev/sdd: 160.0 GB, 160041885696 bytes
255 heads, 63 sectors/track, 19457 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x386237a8
Device Boot Start End Blocks Id System
/dev/sdd1 * 1 244 1951744 fd Linux raid autodetect
Partition 1 does not end on cylinder boundary.
/dev/sdd2 244 19458 154336257 5 Extended
/dev/sdd5 244 19458 154336256 fd Linux raid autodetect
dmraid -r
/dev/sdb: pdc, "pdc_icfhhggf-0", stripe, ok, 312499968 sectors, data@ 0
dmraid -V
dmraid version: 1.0.0.rc16 (2009.09.16) shared
dmraid library version: 1.0.0.rc16 (2009.09.16)
device-mapper version: 4.15.0
dmraid -b
/dev/sdd: 312581808 total, "S0BKJ10YC31771"
/dev/sdc: 312581808 total, "S0BKJ1FYC02391"
/dev/sdb: 312581808 total, "S016J10X454363"
/dev/sda: 312581808 total, "S0BKJ10YC31770"
dmraid -ay
ERROR: pdc: wrong # of devices in RAID set "pdc_icfhhggf-0" [1/2] on /dev/sdb
ERROR: keeping degraded mirror set "pdc_icfhhggf"
RAID set "pdc_icfhhggf-0" was not activateddmraid -tay
ERROR: pdc: wrong # of devices in RAID set "pdc_icfhhggf-0" [1/2] on /dev/sdb
ERROR: keeping degraded mirror set "pdc_icfhhggf"
pdc_icfhhggf-0: 0 312499968 linear /dev/zero 0
ERROR: no mapping possible for RAID set pdc_icfhhggf
I am wondering why there ist [1/2] in the line
ERROR: pdc: wrong # of devices in RAID set "pdc_icfhhggf-0" [1/2] on /dev/sdb
what does it mean? should it not be [1/4]?
dmraid -ay -v -d
DEBUG: not isw at 160041884672
DEBUG: isw trying hard coded -2115 offset.
DEBUG: not isw at 160040802816
DEBUG: not isw at 160041884672
DEBUG: isw trying hard coded -2115 offset.
DEBUG: not isw at 160040802816
DEBUG: not isw at 160041884672
DEBUG: isw trying hard coded -2115 offset.
DEBUG: not isw at 160040802816
DEBUG: not isw at 160041884672
DEBUG: isw trying hard coded -2115 offset.
DEBUG: not isw at 160040802816
DEBUG: _find_set: searching pdc_icfhhggf-0
DEBUG: _find_set: not found pdc_icfhhggf-0
DEBUG: _find_set: searching pdc_icfhhggf
DEBUG: _find_set: not found pdc_icfhhggf
DEBUG: _find_set: searching pdc_icfhhggf-0
DEBUG: _find_set: not found pdc_icfhhggf-0
DEBUG: checking pdc device "/dev/sdb"
ERROR: pdc: wrong # of devices in RAID set "pdc_icfhhggf-0" [1/2] on /dev/sdb
DEBUG: set status of set "pdc_icfhhggf-0" to 2
ERROR: keeping degraded mirror set "pdc_icfhhggf"
RAID set "pdc_icfhhggf-0" was not activated
DEBUG: freeing devices of RAID set "pdc_icfhhggf-0"
DEBUG: freeing device "pdc_icfhhggf-0", path "/dev/sdb"
DEBUG: freeing devices of RAID set "pdc_icfhhggf"
Do I have to repair something? How?
How can I mount the filesystem, using a live cd, so that I can change back the fstab?
Are the changes in fstab the reason for that problem? Or maybe a automaticaly run of the fschk?
It sounds like you are trying to use both mdadm software raid, and dmraid fake hardware raid, at the same time. You can not do that. You should go into your bios and delete the raid array there.