Ben Collins | 23 Aug 19:55 2015

MySQL slaves don't come back up with slave config

I’ve configured 6 nodes as mysql master/slave using this config:

primitive p_mysql ocf:heartbeat:mysql \
        params socket="/var/run/mysqld/mysqld.sock" replication_user="slave"
replication_passwd=“XXXXX" test_user="test_user" test_passwd="test_pass" \
        op start interval="0" timeout="120s" \
        op stop interval="0" timeout="120s" \
        op monitor timeout="30s" interval="30s" role="Master" OCF_CHECK_LEVEL="10" \
        op monitor timeout="30s" interval="60s" role="Slave" OCF_CHECK_LEVEL="10"
primitive p_mysql-ip ocf:heartbeat:IPaddr \
        params ip="10.10.10.191" \
        op monitor interval="1s" timeout="20s" \
        op start interval="0" timeout="20s" \
        op stop interval="0" timeout="20s" \
        meta is-managed="true" resource-stickiness="500"
ms cl_mysql p_mysql
colocation co_ip-on-mysql inf: p_mysql-ip cl_mysql:Master

On the initial setup, everything looks good. The slaves are all reporting proper status. However, if I
reboot one of the slaves, even though it is reported in crm status as a slave, the mysql server shows that
slave status is not configured or started on that node and log shows:

Aug 23 08:44:35 [1204] app5       lrmd:     info: log_execute: 	executing - rsc:p_mysql action:start call_id:99
mysql(p_mysql)[1562]:	2015/08/23_08:44:35 INFO: MySQL is not running
mysql(p_mysql)[1562]:	2015/08/23_08:44:35 INFO: Creating PID dir: /var/run/mysqld
mysql(p_mysql)[1562]:	2015/08/23_08:44:35 INFO: MySQL is not running
mysql(p_mysql)[1562]:	2015/08/23_08:44:37 INFO: MySQL is not running
mysql(p_mysql)[1562]:	2015/08/23_08:44:41 INFO: No MySQL master present - clearing replication state
mysql(p_mysql)[1562]:	2015/08/23_08:44:41 ERROR: check_slave invoked on an instance that is not a
replication slave.
(Continue reading)

MOULLE, ALAIN | 21 Aug 15:27 2015
Picon

Question around resources constraints (pacemaker on RHE7.1)

Hi

I can't find a way to configure constraints in pacemaker so that with these resources:

Res1
Res2
Res3
Res4
Res5

with current colocation constraints :
Res2 with Res1
Res3 with Res2

Res4 with Res1
Res5 with Res4

and current order symmetrical constraints :
Res2 after Res1
Res3 after Res2

Res4 after Res1
Res5 after Res4

and migration-threshold=1 on all resources .

What I want it that if I have a failure for example on Res3  is that all the 5 Ressources are migrated.

Is there a solution ?

(Continue reading)

Muhammad Sharfuddin | 5 Jul 18:13 2015
Picon

file system resource becomes inaccesible when any of the node goes down

SLES 11 SP 3 + online updates(pacemaker-1.1.11-0.8.11.70 
openais-1.1.4-5.22.1.7)

Its a dual primary drbd cluster, which mounts a file system resource on 
both the cluster nodes simultaneously(file system type is ocfs2).

Whenever any of the nodes goes down, the file system(/sharedata) become 
inaccessible for exact 35 seconds on the other (surviving/online) node, 
and then become available again on the online node.

Please help me understand why the node which survives or remains online 
unable to access the file system resource(/sharedata) for 35 seconds ? 
and how can I fix the cluster so that file system remains accessible on 
the surviving node without any interruption/delay(as in my case of about 
35 seconds)

By inaccessible, I meant to say that running "ls -l /sharedata" and "df 
/sharedata" does not return any output and does not return the prompt 
back on the online node for exact 35 seconds once the other node becomes 
offline.

e.g "node1" got offline somewhere around  01:37:15, and then /sharedata 
file system was inaccessible during 01:37:35 and 01:38:18 on the online 
node i.e "node2".

/var/log/messages on node2, when node1 went offline:
Jul  5 01:37:26 node2 kernel: [  675.255865] drbd r0: PingAck did not 
arrive in time.
Jul  5 01:37:26 node2 kernel: [  675.255886] drbd r0: peer( Primary -> 
Unknown ) conn( Connected -> NetworkFailure ) pdsk( UpToDate -> DUnknown )
(Continue reading)

Eden | 12 May 12:49 2015

Creating A resource agent

Hi all,
I am trying to write my own resource agent service,
I have corosync + pacemaker + lcmc

When I load the service it shows that it's running, Then after a few seconds
it stops,
I checked with a default resource agent (Dummy, sshd etc) and they work.

I'm running centos6.6 and these are the packages version:
corosync-1.4.7-1.el6.x86_64
corosynclib-1.4.7-1.el6.x86_64
pacemaker-1.1.12-4.el6.x86_64
pacemaker-cli-1.1.12-4.el6.x86_64
pacemaker-cluster-libs-1.1.12-4.el6.x86_64
pacemaker-libs-1.1.12-4.el6.x86_64

When I look in ps I can see it actually starts the server but then it fall
again,

This is my resource agent:

#!/bin/bash

#######################################################################
# Initialization:
: ${OCF_FUNCTIONS=${OCF_ROOT}/resource.d/heartbeat/.ocf-shellfuncs}
. ${OCF_FUNCTIONS}
: ${__OCF_ACTION=$1}

#######################################################################
(Continue reading)

Picon

Re: Linux-HA Digest, Vol 138, Issue 1

If your operative system is the same in both, you can install pacemaker
without problems.

Victor Acosta

2015-05-07 14:00 GMT-04:00 <linux-ha-request <at> lists.linux-ha.org>:

> Send Linux-HA mailing list submissions to
>         linux-ha <at> lists.linux-ha.org
>
> To subscribe or unsubscribe via the World Wide Web, visit
>         http://lists.linux-ha.org/mailman/listinfo/linux-ha
> or, via email, send a message with subject or body 'help' to
>         linux-ha-request <at> lists.linux-ha.org
>
> You can reach the person managing the list at
>         linux-ha-owner <at> lists.linux-ha.org
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of Linux-HA digest..."
>
>
> Today's Topics:
>
>    1. Doubt in HA configuration (Rajesh S)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
(Continue reading)

Rajesh S | 7 May 15:09 2015

Doubt in HA configuration

Hi Team,

            I want to do HA between two different machines one is HP
workshop and another one is VMware virtual machine but both of the device's
are in different IP series as 192.168.90.102 and 192.168.90.130 .Does HA is
possible or not.Please advice me.

With Regards,
RAJESH S.
_______________________________________________
Linux-HA mailing list is closing down.
Please subscribe to users <at> clusterlabs.org instead.
http://clusterlabs.org/mailman/listinfo/users
_______________________________________________
Linux-HA <at> lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha

Richard B Winters | 22 Apr 19:44 2015

SONAME bump required; cluster-glue

Hello,

Recently we've prepared the 1.0.12 release of cluster-glue for Debian
Linux; and we found when generating symbols that in version 1.0.7 of
libplumb there are reports of dropped symbols, creating a need to bump
the SONAME.

dpkg-gensymbols: warning: some new symbols appeared in the symbols file:
see diff output below
dpkg-gensymbols: warning: some symbols or patterns disappeared in the
symbols file: see diff output below
dpkg-gensymbols: warning: libplumb-tmp.symbols doesn't match completely
libplumb-tmp.symbols
--- libplumb-tmp.symbols (libplumb_1.0.7_amd64)
+++ dpkg-gensymbols4HL2JV	2015-04-22 13:43:32.547196065 -0400
 <at>  <at>  -1,13 +1,13  <at>  <at> 
 libplumb.so.2 libplumb #MINVER#
- ChildLogIPCMessage <at> Base 1.0.3
+#MISSING: 1.0.7# ChildLogIPCMessage <at> Base 1.0.3
  CompressFuncs <at> Base 1.0.3
  DisableProcLogging <at> Base 1.0.3
- DumpCircularBuffer <at> Base 1.0.3
- EmptyCircularBuffer <at> Base 1.0.3
+#MISSING: 1.0.7# DumpCircularBuffer <at> Base 1.0.3
+#MISSING: 1.0.7# EmptyCircularBuffer <at> Base 1.0.3
  EnableProcLogging <at> Base 1.0.3
  FT_strings <at> Base 1.0.3
  ForEachProc <at> Base 1.0.3
- FreeChildLogIPCMessage <at> Base 1.0.3
+#MISSING: 1.0.7# FreeChildLogIPCMessage <at> Base 1.0.3
(Continue reading)

Robert.Koeppl | 19 Apr 16:01 2015

AUTO: Robert Koeppl ist außer Haus. Robert Koeppl is out of office (Rückkehr am 25.04.2015)


Ich kehre zurück am 25.04.2015.

Hinweis: Dies ist eine automatische Antwort auf Ihre Nachricht  "[Linux-HA]
Ping nodes and cluster rules" gesendet am 17.04.2015 14:58:11.

Diese ist die einzige Benachrichtigung, die Sie empfangen werden, während
diese Person abwesend ist.

_______________________________________________
Linux-HA mailing list is closing down.
Please subscribe to users <at> clusterlabs.org instead.
http://clusterlabs.org/mailman/listinfo/users
_______________________________________________
Linux-HA <at> lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha

Daniel Krambrock | 17 Apr 15:11 2015
Picon

CRIT: Emergency,Shutdown: Master Control process died.

Hi.

I have got a backtrace of a dieing MCP:

root <at> s4a:/coredump# gdb /usr/lib/heartbeat/heartbeat
core.1429243494.heartbeat.12806_0-0
GNU gdb (GDB) 7.4.1-debian
Copyright (C) 2012 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later 
<http://gnu.org/licenses/gpl.html>
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law. Type "show copying"
and "show warranty" for details.
This GDB was configured as "x86_64-linux-gnu".
For bug reporting instructions, please see:
<http://www.gnu.org/software/gdb/bugs/>...
Reading symbols from /usr/lib/heartbeat/heartbeat...(no debugging symbols
found)...done.
[New LWP 12806]

warning: Can't read pathname for load map: Input/output error.
[Thread debugging using libthread_db enabled]
Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1".
Core was generated by `heartbeat: master control pro'.
Program terminated with signal 24, CPU time limit exceeded.
#0 0x00007f22de197629 in vfprintf () from /lib/x86_64-linux-gnu/libc.so.6

(gdb) bt
#0 0x00007f22de197629 in vfprintf () from /lib/x86_64-linux-gnu/libc.so.6
#1 0x00007f22de1bd999 in vsprintf () from /lib/x86_64-linux-gnu/libc.so.6
(Continue reading)

Adam Błaszczykowski | 17 Apr 14:58 2015
Picon

Ping nodes and cluster rules

Hello,
I am using two nodes in cluster with corosync 2.3.4 and pacemaker 1.1.12
and each node has access to 2 ping nodes.
I would like to know if it is possible to set following cluster rules:

Rule 1 - do nothing with resources if 1 of 2 ping nodes are unreachable for
one cluster node
Rule 2 - move resources if 2 of 2 ping nodes are unreachable for one
cluster node
Rule 3 - do nothing with resources if all ping nodes are unreachable for
both cluster nodes

NOTE: In my current configuration resources are exported when I lose access
to all ping nodes on both cluster nodes. I would like to change this
behavior as described in "Rule 3"

Thank you in advance !

Best Regards
Adam Blaszczykowski
_______________________________________________
Linux-HA mailing list is closing down.
Please subscribe to users <at> clusterlabs.org instead.
http://clusterlabs.org/mailman/listinfo/users
_______________________________________________
Linux-HA <at> lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha

Heiko Finzel | 1 Apr 16:44 2015
Picon

heartbeat custom health checks

Hello there,

is it possible to configure custom healthchecks for heartbeat such like testing 
a HTTP request or if a process is running?
Unfortunately I can't seem to find any hint or howto on this topic.
Any help appreciated.
_______________________________________________
Linux-HA mailing list
Linux-HA <at> lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Gmane