[lvs-users] Problems with ping_group

Tim.Einmahl at kba.de Tim.Einmahl at kba.de
Wed Aug 29 12:32:12 BST 2007


Hi all,

we have problems regarding heartbeat using the ping_group-statement on
Redhat Enterprise Linux 3 and 4.

Using heartbeat-1.2.3-2.rhel.3.0.rpm, everything worked well, ping_group
included.

But with heartbeat-1.2.3.cvs.20050927-1.rh.el.um.1.rpm and also
with the latest version heartbeat-1.2.5-1.rhel3.1.rpm from
Ultramonkey.org
the ping_group does not work and is declared dead although every node
in this group is alive and the icmp-messages are sent and received
(verified with tcpdump).


ha.cf:
ping_group servergroup xx.xx.xx.1 xx.xx.xx.2 xx.xx.xx.3 xx.xx.xx.4


ha-log:
info: **************************
info: Configuration validated. Starting heartbeat 1.2.5
info: heartbeat: version 1.2.5
info: Heartbeat generation: 13
info: ucast: write socket priority set to IPTOS_LOWDELAY on bond0
info: ucast: bound send socket to device: bond0
info: ucast: bound receive socket to device: bond0
info: ucast: started on port 694 interface bond0 to xx.xx.xx.xx
info: ucast: write socket priority set to IPTOS_LOWDELAY on bond1
info: ucast: bound send socket to device: bond1
info: ucast: bound receive socket to device: bond1
info: ucast: started on port 694 interface bond1 to xx.xx.xx.xx
info: ping group heartbeat started.
info: pid 4490 locked in memory.
info: Local status now set to: 'up'
info: pid 4493 locked in memory.
info: pid 4499 locked in memory.
info: pid 4494 locked in memory.
info: pid 4498 locked in memory.
info: pid 4495 locked in memory.
info: pid 4497 locked in memory.
info: pid 4496 locked in memory.
info: Link web1.test.de:bond0 up.
info: Status update for node web1.test.de: status active
info: Link web1.test.de:bond1 up.
info: Running /etc/ha.d/rc.d/status status
WARN: node servergroup: is dead
info: Local status now set to: 'active'
info: Starting child client "/usr/lib/heartbeat/ipfail" (1001,104)
info: Starting "/usr/lib/heartbeat/ipfail" as uid 1001  gid 104 (pid
4532)
info: remote resource transition completed.
info: remote resource transition completed.
info: Local Resource acquisition completed. (none)
info: Initial resource acquisition complete (T_RESOURCES(them))
info: Running /etc/ha.d/rc.d/status status


I read about the problem in the archives, but I thought the
problem was fixed with the latest version, which doesn't seem
to be the case.

Thanks for your help in advance

Best regards

Tim





More information about the lvs-users mailing list