Comment 5 for bug 959218

Revision history for this message
ITec (itec) wrote :

Hi Martin,

I could verify your fix.

I downgraded all my packages manually to "Ubuntu:12.04/precise".
Then I added "Ubuntu:12.04/precise-proposed" to /etc/apt/sources.list.

# rmdir /var/run/lvm

# apt-get update
# apt-get install lvm2

# dpkg -l clvm lvm2
ii clvm 2.02.66-4ubuntu7.1
ii lvm2 2.02.66-4ubuntu7.1

On the 1st node:

# /etc/init.d/clvm start
 * Starting Cluster LVM Daemon clvm [fail]
 * Activating all VGs
  6 logical volume(s) in volume group "lvm" now active
# ps fax | grep clvm
 9152 ? SLsl 0:00 /usr/sbin/clvmd -T20

On the 2nd node:

# /etc/init.d/clvm start
 * Starting Cluster LVM Daemon clvm [ OK ]
 * Activating all VGs
  6 logical volume(s) in volume group "lvm" now active
# ps fax | grep clvm
26964 ? SLsl 0:00 /usr/sbin/clvmd -T20

On the 1st node:
# /etc/init.d/clvm restart
 * Deactivating VG : [ OK ]
 * Stopping Cluster LVM Daemon clvm [ OK ]
 * Starting Cluster LVM Daemon clvm [ OK ]
 * Activating all VGs
  6 logical volume(s) in volume group "lvm" now active

I suspect "vgscan" in "do_start()" to return with rc=5 when clvm is only running on one node.

# vgscan
  clvmd not running on node tuxtest2
  Unable to obtain global lock.

But clvm does still start, even without clvm on the other node.
When started on the 2nd node no "[fail]" is shown.
When restarted on the 1st node "[fail]" is shown as well.

The problem "/etc/init.d/clvm status exitcode always 0" reported in:
https://bugs.launchpad.net/ubuntu/+source/lvm2/+bug/988881
still persists.

Best regards
Christian