velmurugan murugesan | 14 Dec 14:15 2015

CIB not supported: validator 'pacemaker-2.0', release '3.0.9'


I am new to HA.

I am facing following error while crm configure

ERROR: CIB not supported: validator 'pacemaker-2.0', release '3.0.9'
ERROR: You may try the upgrade command

Please help me resolve this issue.

Linux-HA mailing list is closing down.
Please subscribe to users <at> instead.
Linux-HA <at>

Lorenz Vanthillo | 10 Nov 15:55 2015

HA masters using pacemaker and virtual IP

I want to install OpenShift V3 using the advanced method. So I will create an environment with 2 masters and 3 nodes.
The masters need to be HA. So I want a pacemaker between them.
All my hosts are Amazon EC2 instances and they're using CentOS as OS.

The hostnames of my master are:

Their public IP's:

Their private IP's:

So they are in different subnets (public). Is it possbile to configure a virtual IP for them, and how?

Linux-HA mailing list is closing down.
Please subscribe to users <at> instead.
Linux-HA <at>

Dejan Bucar | 28 Oct 16:21 2015

Download of Cluster Glue package?


The download link for cluster glue has stopped working, Is it down permanently?


Linux-HA mailing list is closing down.
Please subscribe to users <at> instead.
Linux-HA <at>

Cristiano Coltro | 28 Oct 09:32 2015

ORACLE 12 and SLES HAE (Sles 11sp3)

most of the SLES 11 sp3 with HAE are migrating Oracle Db.
The migration will be from Oracle 11 to Oracle 12

They have verified that the Oracles cluster resources actually supports  
- Oracle 10.2 and 11.2 
Command used: usando il comando *crm ra info ocf:heartbeat:SAPDatabase*
So seems they are out of support.
So I would like to know which version of cluster/SO/Agent supports Oracle 12.
AFAIK agents are tipically included in rpm.
# rpm -qf /usr/lib/ocf/resource.d/heartbeat/SAPDatabase
and there are NOT updates about the channel.

Any Idea on that?


Cristiano Coltro
Premium Support Engineer

mail: cristiano.coltro <at>
phone +39 02 36634936
mobile +39 3351435589


(Continue reading)

Karthik | 30 Oct 11:02 2015

Pacemaker 10-15% CPU.

  We are using Pacemaker to manage the services that run on a node, as part 
of a service management framework, and manage the nodes running the services 
as a cluster.  One service will be running as 1+1 and other services with be 

  During our testing, we see that the pacemaker processes are taking about 
10-15% of the CPU.  We would like to know if this is normal and could the 
CPU utilization be minimised.  

Sample Output of most used CPU process in a Active Manager is

189      15766 30.4  0.0  94616 12300 ?        Ss   18:01  48:15 
189      15770 28.9  0.0 118320 20276 ?        Ss   18:01  45:53 
root     15768  2.6  0.0  76196  3420 ?        Ss   18:01   4:12 
root     15767 15.5  0.0  95380  5764 ?        Ss   18:01  24:33 

189      15766 30.5  0.0  94616 12300 ?        Ss   18:01  49:58 
189      15770 29.0  0.0 122484 20724 ?        Rs   18:01  47:29 
root     15768  2.6  0.0  76196  3420 ?        Ss   18:01   4:21 
root     15767 15.5  0.0  95380  5764 ?        Ss   18:01  25:25 
(Continue reading)

J. Echter | 23 Sep 14:38 2015

Cluster for HA VM's serving our local network


i was using this guide to 
set up my cluster for some services, all works pretty good.

I decided to use this cluster as a HA vm provider for my network.

I have a little, maybe silly, question.

The guide tells me to disable qemu default network, like this:

>       Disable the 'qemu' Bridge
> By default, libvirtd <> creates a bridge 
> called virbr0 designed to connect virtual machines to the first eth0 
> interface. Our system will not need this, so we will remove it now.
> If libvirtd has started, skip to the next step. If you haven't started 
> libvirtd yet, you can manually disable the bridge by blanking out the 
> config file.
> cat  /dev/null>/etc/libvirt/qemu/networks/default.xml
i skipped the step to create the bridge device, as it was not needed for 
my belongings.

> vim  /etc/sysconfig/network-scripts/ifcfg-vbr2
> # Internet-Facing Network - Bridge
> DEVICE="vbr2"
(Continue reading)

Dustinta Cristian | 4 Sep 09:13 2015

Add ipv6 address on Hearbeat

I am using Heartbeat 2.1.3 on Solaris 10 OS. I already have configured the hearbeat service  , but I want
to configure Hearbeat on dualStack(ipv4-ipv6). As I said I already have configured the ip-s(ipv4) and I
was wondering there is an easy way to add an ipv6 address without major impacts(without any
reconfigurations etc)? I found /opt/heartbeat/etc/ha.d/resource.d/IPv6addr and I saw the usage of
this script: $0 <ip-address> $LEGAL_ACTIONS. So basically I should just run this script like:
/opt/heartbeat/etc/ha.d/resource.d/IPv6addr 2620:0:60:b008::87f7:a394 start ? Or there are
some more steps to configure an ipv6 address?

Linux-HA mailing list is closing down.
Please subscribe to users <at> instead.
Linux-HA <at>
Ben Collins | 23 Aug 19:55 2015

MySQL slaves don't come back up with slave config

I’ve configured 6 nodes as mysql master/slave using this config:

primitive p_mysql ocf:heartbeat:mysql \
        params socket="/var/run/mysqld/mysqld.sock" replication_user="slave"
replication_passwd=“XXXXX" test_user="test_user" test_passwd="test_pass" \
        op start interval="0" timeout="120s" \
        op stop interval="0" timeout="120s" \
        op monitor timeout="30s" interval="30s" role="Master" OCF_CHECK_LEVEL="10" \
        op monitor timeout="30s" interval="60s" role="Slave" OCF_CHECK_LEVEL="10"
primitive p_mysql-ip ocf:heartbeat:IPaddr \
        params ip="" \
        op monitor interval="1s" timeout="20s" \
        op start interval="0" timeout="20s" \
        op stop interval="0" timeout="20s" \
        meta is-managed="true" resource-stickiness="500"
ms cl_mysql p_mysql
colocation co_ip-on-mysql inf: p_mysql-ip cl_mysql:Master

On the initial setup, everything looks good. The slaves are all reporting proper status. However, if I
reboot one of the slaves, even though it is reported in crm status as a slave, the mysql server shows that
slave status is not configured or started on that node and log shows:

Aug 23 08:44:35 [1204] app5       lrmd:     info: log_execute: 	executing - rsc:p_mysql action:start call_id:99
mysql(p_mysql)[1562]:	2015/08/23_08:44:35 INFO: MySQL is not running
mysql(p_mysql)[1562]:	2015/08/23_08:44:35 INFO: Creating PID dir: /var/run/mysqld
mysql(p_mysql)[1562]:	2015/08/23_08:44:35 INFO: MySQL is not running
mysql(p_mysql)[1562]:	2015/08/23_08:44:37 INFO: MySQL is not running
mysql(p_mysql)[1562]:	2015/08/23_08:44:41 INFO: No MySQL master present - clearing replication state
mysql(p_mysql)[1562]:	2015/08/23_08:44:41 ERROR: check_slave invoked on an instance that is not a
replication slave.
(Continue reading)

MOULLE, ALAIN | 21 Aug 15:27 2015

Question around resources constraints (pacemaker on RHE7.1)


I can't find a way to configure constraints in pacemaker so that with these resources:


with current colocation constraints :
Res2 with Res1
Res3 with Res2

Res4 with Res1
Res5 with Res4

and current order symmetrical constraints :
Res2 after Res1
Res3 after Res2

Res4 after Res1
Res5 after Res4

and migration-threshold=1 on all resources .

What I want it that if I have a failure for example on Res3  is that all the 5 Ressources are migrated.

Is there a solution ?

(Continue reading)

Muhammad Sharfuddin | 5 Jul 18:13 2015

file system resource becomes inaccesible when any of the node goes down

SLES 11 SP 3 + online updates(pacemaker-1.1.11- 

Its a dual primary drbd cluster, which mounts a file system resource on 
both the cluster nodes simultaneously(file system type is ocfs2).

Whenever any of the nodes goes down, the file system(/sharedata) become 
inaccessible for exact 35 seconds on the other (surviving/online) node, 
and then become available again on the online node.

Please help me understand why the node which survives or remains online 
unable to access the file system resource(/sharedata) for 35 seconds ? 
and how can I fix the cluster so that file system remains accessible on 
the surviving node without any interruption/delay(as in my case of about 
35 seconds)

By inaccessible, I meant to say that running "ls -l /sharedata" and "df 
/sharedata" does not return any output and does not return the prompt 
back on the online node for exact 35 seconds once the other node becomes 

e.g "node1" got offline somewhere around  01:37:15, and then /sharedata 
file system was inaccessible during 01:37:35 and 01:38:18 on the online 
node i.e "node2".

/var/log/messages on node2, when node1 went offline:
Jul  5 01:37:26 node2 kernel: [  675.255865] drbd r0: PingAck did not 
arrive in time.
Jul  5 01:37:26 node2 kernel: [  675.255886] drbd r0: peer( Primary -> 
Unknown ) conn( Connected -> NetworkFailure ) pdsk( UpToDate -> DUnknown )
(Continue reading)

Eden | 12 May 12:49 2015

Creating A resource agent

Hi all,
I am trying to write my own resource agent service,
I have corosync + pacemaker + lcmc

When I load the service it shows that it's running, Then after a few seconds
it stops,
I checked with a default resource agent (Dummy, sshd etc) and they work.

I'm running centos6.6 and these are the packages version:

When I look in ps I can see it actually starts the server but then it fall

This is my resource agent:


# Initialization:
: ${OCF_FUNCTIONS=${OCF_ROOT}/resource.d/heartbeat/.ocf-shellfuncs}
: ${__OCF_ACTION=$1}

(Continue reading)