IBM DS3400 Will Not Bring Up Second Path

Bug #824790 reported by Joseph Salisbury
6
This bug affects 1 person
Affects Status Importance Assigned to Milestone
multipath-tools (Ubuntu)
Invalid
Undecided
Unassigned

Bug Description

System configuration is an IBM HS21 blade connected to an IBM DS3400 SAN.
The system is running Ubuntu 10.10 with multipath-tools version: 0.4.8-14ubuntu4.10.10.2
The Ubuntu OS is installed on an internal disk and not the SAN.

multipathd> show paths
hcil dev dev_t pri dm_st chk_st next_check
0:0:0:0 sda 8:0 0 [undef][ghost] [orphan]
1:0:0:0 sdb 8:16 3 [undef][ready] [orphan]
2:1:1:0 sdc 8:32 1 [undef][ready] [orphan]

It appears everything is working but multipathd cant define the maps, a multipath -v 6 shows

"360080e50001b5f4c00003d284e37e864: domap (0) failure for create/reload map"

That is the uuid of the path associated with /sda and /sdb but it cant seem to create it.

libdevmapper: ioctl/libdm-iface.c(1740): dm reload 360080e50001b5f4c00003d284e37e864 OF [16384]
libdevmapper: ioctl/libdm-iface.c(1757): device-mapper: reload ioctl failed: Invalid argument
libdevmapper: ioctl/libdm-iface.c(1740): dm remove 360080e50001b5f4c00003d284e37e864 NF [16384]
libdevmapper: libdm-common.c(799): 360080e50001b5f4c00003d284e37e864: Stacking NODE_DEL (replaces other stacked ops)
libdevmapper: ioctl/libdm-iface.c(1740): dm info 360080e50001b5f4c00003d284e37e864 NF [16384]
360080e50001b5f4c00003d284e37e864: domap (0) failure for create/reload map
360080e50001b5f4c00003d284e37e864: remove multipath map
sda: orphaned
sdb: orphaned

The SAN is on devices sda and sdb. The internal root disk is sdc. The device sda is reporting I/O errors in syslog because multipath cannot bring up the path.

Revision history for this message
Peter Petrakis (peter-petrakis) wrote :

@Joe

Please attach the latest multipath.conf and lvm.conf responsible for this
stream of messages. If you could move over that logs and other stuff
from salesforce into this bug that would be ideal. Thanks.

Revision history for this message
Joseph Salisbury (jsalisbury) wrote :

multipath.conf

Revision history for this message
Joseph Salisbury (jsalisbury) wrote :

syslog

Revision history for this message
Joseph Salisbury (jsalisbury) wrote :

The lvm.conf file has not been modified from what ships with Ubuntu.

Revision history for this message
Peter Petrakis (peter-petrakis) wrote :

So the paths not showing up with a config file error, a few well
placed regexp and a proper blacklist to filter out the internal disks
solved that. Also added scsi_dh_rdac to initial ramdisk. Now the paths
are up, and stay up. However when we tried testing failover using
a scsi "delete" of the primary member. The secondary never became
active.

We asked the customer to check with any additional SAN side
config, like automatic lun transfer, is necessary. Depending on how
that turns it we may actually have a real failover bug.

Working multipath config, taken almost verbatim from the IBM
documentation: Section 5-10.
ftp://ftp.software.ibm.com/systems/support/system_x_pdf/94y8402.pdf

# Canonical Customer Support: 08/12/11
# Version 1.2

blacklist {
  devnode "^(ram|raw|loop|fd|md|dm-|sr|scd|st)[0-9]*"
  devnode "^hd[a-z]"
  # Ignore internal storage from the LSI RAID
  device {
    vendor LSILOGIC
    product .*
  }
}

devices {
  device {
    vendor "IBM"
    product ".*"
    path_grouping_policy group_by_prio
    path_checker rdac
    getuid_callout "/lib/udev/scsi_id -g -u -d /dev/%n"
    prio_callout "/sbin/mpath_prio_rdac /dev/%n"
    hardware_handler "1 rdac"
    failback immediate
# rr_weight uniform
    no_path_retry 30
    rr_min_io 100
    features "2 pg_init_retries 50"
  }
}

Changed in multipath-tools (Ubuntu):
assignee: nobody → Ubuntu Storage Development Team (ubuntu-storage-dev)
Changed in multipath-tools (Ubuntu):
status: New → Confirmed
Revision history for this message
Peter Petrakis (peter-petrakis) wrote :

In an addition to the MP issues, LVM was misconfigured. We attempt
to repair this on one of the blades but the config file just wouldn't *take*.
I suspect there was something else going, especially since we didn't
even have direct access to the console and was forced to relay
directives to the customer to execute.

I set up a mock test environment to test the the filtering by
PCI device.

root@kickseed:/dev/disk/by-path# ls -l
total 0
lrwxrwxrwx 1 root root 8 2011-08-15 10:04 pci-0000:00:00.9-scsi-0:0:0:2 -> /dev/sdc
lrwxrwxrwx 1 root root 9 2011-08-15 10:04 pci-0000:00:00.9-scsi-0:0:0:2-part1 -> /dev/sdc1
lrwxrwxrwx 1 root root 8 2011-08-15 10:04 pci-0000:00:00.9-scsi-0:0:0:3 -> /dev/sdd
lrwxrwxrwx 1 root root 9 2011-08-15 10:04 pci-0000:00:00.9-scsi-0:0:0:3-part1 -> /dev/sdd1
lrwxrwxrwx 1 root root 8 2011-08-15 10:02 pci-0000:00:1f.2-scsi-0:0:0:0 -> /dev/sda
lrwxrwxrwx 1 root root 9 2011-08-15 10:03 pci-0000:00:1f.2-scsi-0:0:0:0-part1 -> /dev/sda1
lrwxrwxrwx 1 root root 8 2011-08-15 10:03 pci-0000:00:1f.2-scsi-0:0:0:1 -> /dev/sdb
lrwxrwxrwx 1 root root 9 2011-08-15 10:03 pci-0000:00:1f.2-scsi-0:0:0:1-part1 -> /dev/sdb1

Then, I commented out the existing "filter" directive on /etc/lvm/lvm.conf and tested:
    filter = ["a|/dev/disk/by-path/pci-0000:00:00.9-scsi.*|", "r/.*/" ]

and

    filter = ["a|/dev/disk/by-path/pci-0000:00:1f.2-scsi.*|", "r/.*/" ]

lvmdiskscan confirms the filter directives.

root@kickseed:/dev/disk/by-path# lvmdiskscan
  /dev/disk/by-path/pci-0000:00:1f.2-scsi-0:0:0:0-part1 [ 7.45 GiB]
  /dev/disk/by-path/pci-0000:00:1f.2-scsi-0:0:0:1-part1 [ 10.00 GiB]
  0 disks
  2 partitions
  0 LVM physical volume whole disks
  0 LVM physical volumes

root@kickseed:/dev/disk/by-path# lvmdiskscan
  /dev/disk/by-path/pci-0000:00:00.9-scsi-0:0:0:2-part1 [ 10.00 GiB]
  /dev/disk/by-path/pci-0000:00:00.9-scsi-0:0:0:3-part1 [ 10.00 GiB]
  0 disks
  2 partitions
  0 LVM physical volume whole disks
  0 LVM physical volumes

I'm going to leave this part of the effort to customer support. There are
no bugs in LVM filtering, simply execution issues.

Changed in multipath-tools (Ubuntu):
status: Confirmed → Invalid
Changed in multipath-tools (Ubuntu):
assignee: Ubuntu Storage Development Team (ubuntu-storage-dev) → nobody
To post a comment you must log in.
This report contains Public information  
Everyone can see this information.

Other bug subscribers

Remote bug watches

Bug watches keep track of this bug in other bug trackers.