Alan Evangelista | 23 Oct 15:49 2014
Picon

Problems building fence-agents from source

Hi.

I'm trying to build fence-agents from source (master branch) on CentOS 6.5.
I already installed the following rpm packages (dependencies): autoconf,
automake, gcc, libtool, nss, nss-devel. When I tried to run ./autogen.sh,
I got:

configure.ac:162: error: possibly undefined macro: AC_PYTHON_MODULE
       If this token and others are legitimate, please use m4_pattern_allow.
       See the Autoconf documentation.

I then run

$ autoreconf --install

and autogen worked. Then, I have a problem running ./configure:

./configure: line 18284: syntax error near unexpected token `suds,'
./configure: line 18284: `AC_PYTHON_MODULE(suds, 1)'

I never had this problem before with earlier fence-agents versions.
Am I missing something or is there an issue with upstream code?

RPM dependencies versions:
  autoconf-2.63-5.1.el6.noarch
  automake-1.11.1-4.el6.noarch
  libtool-2.2.6-15.5.el6.x86_64

Regards,
Alan Evangelista
(Continue reading)

Sunhux G | 22 Oct 10:44 2014
Picon

Rhel BootLoader, Single-user mode password & Interactive Boot in a Cloud environment

We run cloud service & our vCenter is not accessible to our tenants
and their IT support; so I would say console access is not feasible
unless the tenant/customer IT come to our DC.

If the following 3 hardenings are done our tenant/customer RHEL
Linux VM,  what's the impact to the tenant's sysadmin & IT operation?


a) CIS 1.5.3 Set Boot Loader Password :
    if this password is set, when tenant reboot (shutdown -r)
    their VM each time, will it prompt for the bootloader
    password at console?  If so, is there any way the tenant,
    could still get their VM booted up if they have no access
    to vCenter's console?

b) CIS 1.5.4 Require Authentication for Single-User Mode :
    Does Linux allow ssh access while in single-user mode &
    can this 'single-user mode password' be entered via an
    ssh session (without access to console), assuming certain
    'terminal' service is started up / running while in single
    user mode

c) CIS 1.5.5 Disable Interactive Boot :
    what's the general consensus on this? Disable or enable?
    Our corporate hardening guide does not mention this item.
    So if the tenant wishes to boot up step by step (ie pausing
    at each startup script), they can't do it?

Feel free to add any other impacts that anyone can think of

Lastly, how do people out there grant console access to their
tenants in Cloud environment without security compromise
(I mean without granting vCenter access) : I heard that we can
customize vCenter to grant limited access of vCenter to 
tenants, is this so?


Sun

--

-- 
Linux-cluster mailing list
Linux-cluster <at> redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster
Neale Ferguson | 14 Oct 21:40 2014
Picon

Re: Permission denied

Yeah, I noted I was looking at the wrong lockspace. The gfs2 lockspace in
this cluster is vol1. Once I corrected at what I was looking at, I think I
solved my problem: I believe the problem is an endian thing. In
set_rcom_status:

        rs->rs_flags = cpu_to_le32(flags)

However, in receive_rcom_status() flags are checked:

        if (!(rs->rs_flags & DLM_RSF_NEED_SLOTS)) {

But it should be:

        if (!(le32_to_cpu(rs->rs_flags) & DLM_RSF_NEED_SLOTS)) {

I made this change and now the gfs2 volume is being mounted correctly on
both nodes. I¹ve repeated it a number of times and it¹s kept working.

Neale

On 10/14/14, 3:20 PM, "David Teigland" <teigland <at> redhat.com> wrote:

>clvmd is a userland lockspace and does not use lockspace_ops or slots/jids
>like a gfs2 (kernel) lockspace.
>
>To debug the dlm/gfs2 control mechanism, which assigns gfs2 a jid based on
>dlm slots, enable the fs_info() lines in gfs2/lock_dlm.c.  (Make sure that
>you're not somehow running gfs_controld on these nodes; we quit using that
>in RHEL7.)

--

-- 
Linux-cluster mailing list
Linux-cluster <at> redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Thomas Meier | 13 Oct 21:10 2014
Picon
Picon

Fencing issues with fence_apc_snmp (APC Firmware 6.x)

Hi

When configuring PDU fencing in my 2-node-cluster I ran into some problems with
the fence_apc_snmp agent. Turning a node off works fine, but
fence_apc_snmp then exits with error.

When I do this manually (from node2):

   fence_apc_snmp -a node1 -n 1 -o off

the output of the command is not an expected:

   Success: Powered OFF

but in my case:

   Returned 2: Error in packet.
   Reason: (genError) A general failure occured
   Failed object: .1.3.6.1.4.1.318.1.1.4.4.2.1.3.21

When I check the PDU, the port is without power, so this part works.
But it seems that the fence agent can't read the status of the PDU
and then exits with error. The same seems to happen when fenced 
is calling the agent. The agent also exits with an error and fencing can't succeed
and the cluster hangs.

>From the logfile: 

    fenced[2100]: fence node1 dev 1.0 agent fence_apc_snmp result: error from agent

My Setup: - CentOS 6.5 with fence-agents-3.1.5-35.el6_5.4.x86_64 installed. 
          - APC AP8953 PDU with firmware 6.1
          - 2-node-cluster based on https://alteeve.ca/w/AN!Cluster_Tutorial_2
          - fencing agents in use: fence_ipmilan (working) and fence_apc_snmp

I did some recherche, and for me it looks like that my fence-agents package is too old for my APC firmware.

I've already found the fence-agents repo: https://git.fedorahosted.org/cgit/fence-agents.git/

Here https://git.fedorahosted.org/cgit/fence-agents.git/commit/?id=55ccdd79f530092af06eea5b4ce6a24bd82c0875
it says: "fence_apc_snmp: Add support for firmware 6.x"

I've managed to build fence-agents-4.0.11.tar.gz on a CentOS 6.5 test box, but my build
of fence_apc_snmp doesn't work.

It gives:

[root <at> box1]# fence_apc_snmp -v -a node1 -n 1 -o status
Traceback (most recent call last):
  File "/usr/sbin/fence_apc_snmp", line 223, in <module>
    main()
  File "/usr/sbin/fence_apc_snmp", line 197, in main
    options = check_input(device_opt, process_input(device_opt))
  File "/usr/share/fence/fencing.py", line 705, in check_input
    logging.getLogger().addHandler(logging.StreamHandler(stream=sys.stderr))
TypeError: __init__() got an unexpected keyword argument 'stream'

I'd really like to see if a patched fence_apc_snmp agent fixes my problem, and if so,
install the right version of fence_apc_snmp on the cluster without breaking things,
but I'm a bit clueless how to build me a working version. 

Maybe you have some tips?

Thanks in advance

Thomas

--

-- 
Linux-cluster mailing list
Linux-cluster <at> redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Neale Ferguson | 13 Oct 17:20 2014
Picon

Permission denied

I reported last week that I was getting permission denied when pcs was
starting a gfs2 resource. I thought it was due to the resource being
defined incorrectly, but it doesn¹t appear to be the case. On rare
occasions the mount works but most of the time one node gets it mounted
but the other gets denied. I¹ve enabled a number of logging options and
done straces on both sides but I¹m not getting anywhere.

My cluster looks like:

# pcs resource show
 Clone Set: dlm-clone [dlm]
   Started: [ rh7cn1.devlab.sinenomine.net rh7cn2.devlab.sinenomine.net ]
 Resource Group: apachegroup
   VirtualIP	(ocf::heartbeat:IPaddr2):	Started
   Website	(ocf::heartbeat:apache):	Started
   httplvm	(ocf::heartbeat:LVM):	Started
   http_fs	(ocf::heartbeat:Filesystem):	Started
 Clone Set: clvmd-clone [clvmd]
   Started: [ rh7cn1.devlab.sinenomine.net rh7cn2.devlab.sinenomine.net ]
 Clone Set: clusterfs-clone [clusterfs]
   Started: [ rh7cn1.devlab.sinenomine.net ]
   Stopped: [ rh7cn2.devlab.sinenomine.net ]

The gfs2 resource is defined:

# pcs resource show clusterfs
 Resource: clusterfs (class=ocf provider=heartbeat type=Filesystem)
 Attributes: device=/dev/vg_cluster/ha_lv directory=/mnt/gfs2-demo
fstype=gfs2 options=noatime
  Operations: start interval=0s timeout=60 (clusterfs-start-timeout-60)
              stop interval=0s timeout=60 (clusterfs-stop-timeout-60)
              monitor interval=10s on-fail=fence
(clusterfs-monitor-interval-10s)

When the mount is attempted on node 2 the log contains:

Oct 13 11:10:42 rh7cn2 kernel: GFS2: fsid=rh7cluster:vol1: Trying to join
cluster "lock_dlm", "rh7cluster:vol1"
Oct 13 11:10:42 rh7cn2 corosync[47978]: [QB    ]
ipc_setup.c:handle_new_connection:485 IPC credentials authenticated
(47978-48271-30)
Oct 13 11:10:42 rh7cn2 corosync[47978]: [QB    ]
ipc_shm.c:qb_ipcs_shm_connect:294 connecting to client [48271]
Oct 13 11:10:42 rh7cn2 corosync[47978]: [QB    ]
ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672;
rb->word_size:263168
Oct 13 11:10:42 rh7cn2 corosync[47978]: message repeated 2 times: [[QB
] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672;
rb->word_size:263168]
Oct 13 11:10:42 rh7cn2 corosync[47978]: [MAIN  ]
ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 13 11:10:42 rh7cn2 corosync[47978]: [CPG   ]
cpg.c:cpg_lib_init_fn:1532 lib_init_fn: conn=0x2ab16a953a0,
cpd=0x2ab16a95a64
Oct 13 11:10:42 rh7cn2 corosync[47978]: [CPG   ]
cpg.c:message_handler_req_exec_cpg_procjoin:1349 got procjoin message from
cluster node 0x2 (r(0) ip(172.17.16.148) ) for pid 48271
Oct 13 11:10:43 rh7cn2 kernel: GFS2: fsid=rh7cluster:vol1: Joined cluster.
Now mounting FS...
Oct 13 11:10:43 rh7cn2 corosync[47978]: [CPG   ]
cpg.c:message_handler_req_lib_cpg_leave:1617 got leave reques
t on 0x2ab16a953a0Oct 13 11:10:43 rh7cn2 corosync[47978]: [CPG   ]
cpg.c:message_handler_req_exec_cpg_procleave:1365 got proclea
ve message from cluster node 0x2 (r(0) ip(172.17.16.148) ) for pid 48271
Oct 13 11:10:43 rh7cn2 corosync[47978]: [CPG   ]
cpg.c:message_handler_req_lib_cpg_finalize:1655 cpg finalize for
conn=0x2ab16a953a0
Oct 13 11:10:43 rh7cn2 dlm_controld[48271]: 251492 cpg_dispatch error 9

Is the ³leave request² symptomatic or causal? If the latter, why is it
generated? 
On other other side:
Oct 13 11:10:41 rh7cn1 corosync[10423]: [QUORUM]
vsf_quorum.c:message_handler_req_lib_quorum_getquorate:395 got quorate
request on 0x2ab0e33c8b0
Oct 13 11:10:41 rh7cn1 corosync[10423]: [QUORUM]
vsf_quorum.c:message_handler_req_lib_quorum_getquorate:395 got quorate
request on 0x2ab0e33c8b0
Oct 13 11:10:42 rh7cn1 corosync[10423]: [CPG   ]
cpg.c:message_handler_req_exec_cpg_procjoin:1349 got procjoin message from
cluster node 0x2 (r(0) ip(172.17.16.148) ) for pid 48271
Oct 13 11:10:43 rh7cn1 kernel: GFS2: fsid=rh7cluster:vol1.0: recover
generation 6 doneOct 13 11:10:43 rh7cn1 corosync[10423]: [CPG   ]
cpg.c:message_handler_req_exec_cpg_procleave:1365 got proclea
ve message from cluster node 0x2 (r(0) ip(172.17.16.148) ) for pid
48271Oct 13 11:10:43 rh7cn1 kernel: GFS2: fsid=rh7cluster:vol1.0: recover
generation 7 done

dlm_tool dump shows:

251469 dlm:ls:vol1 conf 2 1 0 memb 1 2 join 2 left
251469 vol1 add_change cg 6 joined nodeid 2
251469 vol1 add_change cg 6 counts member 2 joined 1 remove 0 failed 0
251469 vol1 stop_kernel cg 6
251469 write "0" to "/sys/kernel/dlm/vol1/control"
251469 vol1 check_ringid done cluster 43280 cpg 1:43280
251469 vol1 check_fencing done
251469 vol1 send_start 1:6 counts 5 2 1 0 0
251469 vol1 receive_start 1:6 len 80
251469 vol1 match_change 1:6 matches cg 6
251469 vol1 wait_messages cg 6 need 1 of 2
251469 vol1 receive_start 2:1 len 80
251469 vol1 match_change 2:1 matches cg 6
251469 vol1 wait_messages cg 6 got all 2
251469 vol1 start_kernel cg 6 member_count 2
251469 dir_member 1
251469 set_members mkdir
"/sys/kernel/config/dlm/cluster/spaces/vol1/nodes/2"
251469 write "1" to "/sys/kernel/dlm/vol1/control"
251469 vol1 prepare_plocks
251469 vol1 set_plock_data_node from 1 to 1
251469 vol1 send_all_plocks_data 1:6
251469 vol1 send_all_plocks_data 1:6 0 done
251469 vol1 send_plocks_done 1:6 counts 5 2 1 0 0 plocks_data 0
251469 vol1 receive_plocks_done 1:6 flags 2 plocks_data 0 need 0 save 0
251470 dlm:ls:vol1 conf 1 0 1 memb 1 join left 2
251470 vol1 add_change cg 7 remove nodeid 2 reason leave
251470 vol1 add_change cg 7 counts member 1 joined 0 remove 1 failed 0
251470 vol1 stop_kernel cg 7
251470 write "0" to "/sys/kernel/dlm/vol1/control"
251470 vol1 purged 0 plocks for 2
251470 vol1 check_ringid done cluster 43280 cpg 1:43280
251470 vol1 check_fencing done
251470 vol1 send_start 1:7 counts 6 1 0 1 0
251470 vol1 receive_start 1:7 len 76
251470 vol1 match_change 1:7 matches cg 7
251470 vol1 wait_messages cg 7 got all 1
251470 vol1 start_kernel cg 7 member_count 1
251470 dir_member 2
251470 dir_member 1
251470 set_members rmdir
"/sys/kernel/config/dlm/cluster/spaces/vol1/nodes/2"
251470 write "1" to "/sys/kernel/dlm/vol1/control"
251470 vol1 prepare_plocks

I would appreciate any debugging suggestions. I¹ve straced
dlm_controld/corosync but not gained much clarity.

Neale

--

-- 
Linux-cluster mailing list
Linux-cluster <at> redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Neale Ferguson | 3 Oct 21:32 2014
Picon

gfs2 resource not mounting

Using the same two-node configuration I described in an earlier post this forum, I'm having problems
getting a gfs2 resource started on one of the nodes. The resource in question:

 Resource: clusterfs (class=ocf provider=heartbeat type=Filesystem)
  Attributes: device=/dev/vg_cluster/ha_lv directory=/mnt/gfs2-demo fstype=gfs2 options=noatime 
  Operations: start interval=0s timeout=60 (clusterfs-start-timeout-60)
              stop interval=0s timeout=60 (clusterfs-stop-timeout-60)
              monitor interval=10s on-fail=fence (clusterfs-monitor-interval-10s)

pcs status shows:

Clone Set: dlm-clone [dlm]
     Started: [ rh7cn1.devlab.sinenomine.net rh7cn2.devlab.sinenomine.net ]
 Clone Set: clvmd-clone [clvmd]
     Started: [ rh7cn1.devlab.sinenomine.net rh7cn2.devlab.sinenomine.net ]
 Clone Set: clusterfs-clone [clusterfs]
     Started: [ rh7cn1.devlab.sinenomine.net ]
     Stopped: [ rh7cn2.devlab.sinenomine.net ]

Failed actions:
    clusterfs_start_0 on rh7cn2.devlab.sinenomine.net 'unknown error' (1): call=46, status=complete,
last-rc-change='Fri Oct  3 14:41:26 2014', queued=4702ms, exec=0ms

Using pcs resource debug-start I see:

Operation start for clusterfs:0 (ocf:heartbeat:Filesystem) returned 1
 >  stderr: INFO: Running start for /dev/vg_cluster/ha_lv on /mnt/gfs2-demo
 >  stderr: mount: permission denied
 >  stderr: ERROR: Couldn't mount filesystem /dev/vg_cluster/ha_lv on /mnt/gfs2-demo

The log on the node shows - 

Oct  3 14:57:37 rh7cn2 kernel: GFS2: fsid=rh7cluster:vol1: Trying to join cluster "lock_dlm", "rh7cluster:vol1"
Oct  3 14:57:38 rh7cn2 kernel: GFS2: fsid=rh7cluster:vol1: Joined cluster. Now mounting FS...
Oct  3 14:57:38 rh7cn2 dlm_controld[5857]: 1564 cpg_dispatch error 9

On the other node - 

Oct  3 15:09:47 rh7cn1 kernel: GFS2: fsid=rh7cluster:vol1.0: recover generation 14 done
Oct  3 15:09:48 rh7cn1 kernel: GFS2: fsid=rh7cluster:vol1.0: recover generation 15 done

I'm assuming I didn't define the gfs2 resource such that it could be used concurrently by both nodes. Here's
the cib.xml definition for it:

      <clone id="clusterfs-clone">
        <primitive class="ocf" id="clusterfs" provider="heartbeat" type="Filesystem">
          <instance_attributes id="clusterfs-instance_attributes">
            <nvpair id="clusterfs-instance_attributes-device" name="device" value="/dev/vg_cluster/ha_lv"/>
            <nvpair id="clusterfs-instance_attributes-directory" name="directory" value="/mnt/gfs2-demo"/>
            <nvpair id="clusterfs-instance_attributes-fstype" name="fstype" value="gfs2"/>
            <nvpair id="clusterfs-instance_attributes-options" name="options" value="noatime"/>
          </instance_attributes>
          <operations>
            <op id="clusterfs-start-timeout-60" interval="0s" name="start" timeout="60"/>
            <op id="clusterfs-stop-timeout-60" interval="0s" name="stop" timeout="60"/>
            <op id="clusterfs-monitor-interval-10s" interval="10s" name="monitor" on-fail="fence"/>
          </operations>
        </primitive>
        <meta_attributes id="clusterfs-clone-meta">
          <nvpair id="clusterfs-interleave" name="interleave" value="true"/>
        </meta_attributes>
      </clone>

-------------------------------

Unrelated (I believe) to the above, I also note the following messages in /var/log/messages which appear
to be related to pacemaker and http (another resource I have defined):

Oct  3 15:05:06 rh7cn2 systemd: pacemaker.service: Got notification message from PID 6036, but reception
only permitted for PID 5575

I'm running systemd-208-11.el7_0.2. A bugzilla search matches with one report but the fix was put into -11.

Neale

--

-- 
Linux-cluster mailing list
Linux-cluster <at> redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Digimer | 3 Oct 19:56 2014
Picon

Re: clvmd issues

On 03/10/14 12:57 PM, manish vaidya wrote:
> First i apologise for late reply , delay due to i cannot believe ,any
> response from site , I am a newcomer , already , i had posted this
> problem on many online forums , but they didn't give any response
>
> Thank all , for taking my problem seriously
>
> ** response from you
>
> are you using clvmd? if your answer is = yes, you need to be sure, you pv
>
> is visibile to your cluster nodes
>
> *** i am using clvmd & When use pvscan command cluster hangs
>
> I want to reproduce this situation again for perfection , such as when i
> try to run pvcreate command in cluster , message should come lock from
> node2 & node3 , I have created new cluster , this new cluster is working
> fine ,
> How to do This? any setting in lvm.conf

Can you share your setup please?

What kind of cluster? What version? What is the configuration file? Was 
there anything interesting in the system logs? etc.

-- 
Digimer
Papers and Projects: https://alteeve.ca/w/
What if the cure for cancer is trapped in the mind of a person without 
access to education?

--

-- 
Linux-cluster mailing list
Linux-cluster <at> redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Daniel Dehennin | 3 Oct 16:35 2014

cLVM unusable on quorated cluster

Hello,

I'm trying to setup pacemaker+corosync on Debian Wheezy to access a SAN
for an OpenNebula cluster.

As I'm new to cluster world, I have hard time figuring why sometime
things get really wrong and where I must look to find answers.

My OpenNebula frontend, running in a VM, does not manage to run the
resources and my syslog has a lot of:

#+begin_src
ocfs2_controld: Unable to open checkpoint "ocfs2:controld": Object does not exist
#+end_src

When this happens, other nodes have problem:

#+begin_src
root <at> nebula3:~# LANG=C vgscan
  cluster request failed: Host is down
  Unable to obtain global lock.
#+end_src

But things looks fin in “crm_mon”:

#+begin_src
root <at> nebula3:~# crm_mon -1
============
Last updated: Fri Oct  3 16:25:43 2014
Last change: Fri Oct  3 14:51:59 2014 via cibadmin on nebula1
Stack: openais
Current DC: nebula3 - partition with quorum
Version: 1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff
5 Nodes configured, 5 expected votes
32 Resources configured.
============

Node quorum: standby
Online: [ nebula3 nebula2 nebula1 ]
OFFLINE: [ one ]

 Stonith-nebula3-IPMILAN    (stonith:external/ipmi):    Started nebula2
 Stonith-nebula2-IPMILAN    (stonith:external/ipmi):    Started nebula3
 Stonith-nebula1-IPMILAN    (stonith:external/ipmi):    Started nebula2
 Clone Set: ONE-Storage-Clone [ONE-Storage]
     Started: [ nebula1 nebula3 nebula2 ]
     Stopped: [ ONE-Storage:3 ONE-Storage:4 ]
 Quorum-Node    (ocf::heartbeat:VirtualDomain): Started nebula3
 Stonith-Quorum-Node   (stonith:external/libvirt):   Started nebula3
#+end_src

I don't know how to interpret dlm_tool informations:

#+begin_src
root <at> nebula3:~# dlm_tool ls -n
dlm lockspaces
name          CCB10CE8D4FF489B9A2ECB288DACF2D7
id            0x09250e49
flags         0x00000008 fs_reg
change        member 3 joined 1 remove 0 failed 0 seq 2,2
members       1189587136 1206364352 1223141568 
all nodes
nodeid 1189587136 member 1 failed 0 start 1 seq_add 1 seq_rem 0 check none
nodeid 1206364352 member 1 failed 0 start 1 seq_add 2 seq_rem 0 check none
nodeid 1223141568 member 1 failed 0 start 1 seq_add 1 seq_rem 0 check none

name          clvmd
id            0x4104eefa
flags         0x00000000 
change        member 3 joined 0 remove 1 failed 0 seq 4,4
members       1189587136 1206364352 1223141568 
all nodes
nodeid 1172809920 member 0 failed 0 start 0 seq_add 3 seq_rem 4 check none
nodeid 1189587136 member 1 failed 0 start 1 seq_add 1 seq_rem 0 check none
nodeid 1206364352 member 1 failed 0 start 1 seq_add 2 seq_rem 0 check none
nodeid 1223141568 member 1 failed 0 start 1 seq_add 1 seq_rem 0 check none
#+end_src

1412340044 dlm_controld 3.0.12 started
1412340044 found /dev/misc/dlm-control minor 58
1412340044 found /dev/misc/dlm-monitor minor 57
1412340044 found /dev/misc/dlm_plock minor 56
1412340044 /dev/misc/dlm-monitor fd 11
1412340044 /sys/kernel/config/dlm/cluster/comms: opendir failed: 2
1412340044 /sys/kernel/config/dlm/cluster/spaces: opendir failed: 2
1412340044 totem/rrp_mode = 'none'
1412340044 set protocol 0
1412340044 group_mode 3 compat 0
1412340044 setup_cpg_daemon 13
1412340044 dlm:controld conf 2 1 0 memb 1189587136 1223141568 join 1223141568 left
1412340044 run protocol from nodeid 1189587136
1412340044 daemon run 1.1.1 max 1.1.1 kernel run 1.1.1 max 1.1.1
1412340044 plocks 15
1412340044 plock cpg message size: 104 bytes
1412340044 Processing membership 22676
1412340044 Adding address ip(192.168.231.70) to configfs for node 1189587136
1412340044 set_configfs_node 1189587136 192.168.231.70 local 0
1412340044 Added active node 1189587136: born-on=22628, last-seen=22676, this-event=22676, last-event=0
1412340044 Adding address ip(192.168.231.71) to configfs for node 1206364352
1412340044 set_configfs_node 1206364352 192.168.231.71 local 0
1412340044 Added active node 1206364352: born-on=22632, last-seen=22676, this-event=22676, last-event=0
1412340044 Adding address ip(192.168.231.72) to configfs for node 1223141568
1412340044 set_configfs_node 1223141568 192.168.231.72 local 1
1412340044 Added active node 1223141568: born-on=22636, last-seen=22676, this-event=22676, last-event=0
1412340044 dlm:controld conf 3 1 0 memb 1189587136 1206364352 1223141568 join 1206364352 left
1412340045 client connection 5 fd 16
1412340047 uevent: add <at> /kernel/dlm/clvmd
1412340047 kernel: add <at>  clvmd
1412340047 uevent: online <at> /kernel/dlm/clvmd
1412340047 kernel: online <at>  clvmd
1412340047 dlm:ls:clvmd conf 2 1 0 memb 1189587136 1223141568 join 1223141568 left
1412340047 clvmd add_change cg 1 joined nodeid 1223141568
1412340047 clvmd add_change cg 1 we joined
1412340047 clvmd add_change cg 1 counts member 2 joined 1 remove 0 failed 0
1412340047 clvmd check_fencing done
1412340047 clvmd check_quorum disabled
1412340047 clvmd check_fs none registered
1412340047 clvmd send_start cg 1 flags 1 data2 0 counts 0 2 1 0 0
1412340047 clvmd receive_start 1189587136:2 len 80
1412340047 clvmd match_change 1189587136:2 matches cg 1
1412340047 clvmd wait_messages cg 1 need 1 of 2
1412340047 clvmd receive_start 1223141568:1 len 80
1412340047 clvmd match_change 1223141568:1 matches cg 1
1412340047 clvmd wait_messages cg 1 got all 2
1412340047 clvmd start_kernel cg 1 member_count 2
1412340047 write "1090842362" to "/sys/kernel/dlm/clvmd/id"
1412340047 set_members mkdir "/sys/kernel/config/dlm/cluster/spaces/clvmd/nodes/1189587136"
1412340047 set_members mkdir "/sys/kernel/config/dlm/cluster/spaces/clvmd/nodes/1223141568"
1412340047 write "1" to "/sys/kernel/dlm/clvmd/control"
1412340047 write "0" to "/sys/kernel/dlm/clvmd/event_done"
1412340047 clvmd set_plock_ckpt_node from 0 to 1189587136
1412340047 clvmd receive_plocks_stored 1189587136:2 flags a sig 0 need_plocks 1
1412340047 clvmd match_change 1189587136:2 matches cg 1
1412340047 clvmd retrieve_plocks
1412340047 clvmd retrieve_plocks first 0 last 0 r_count 0 p_count 0 sig 0
1412340047 uevent: add <at> /devices/virtual/misc/dlm_clvmd
1412340047 dlm:ls:clvmd conf 3 1 0 memb 1189587136 1206364352 1223141568 join 1206364352 left
1412340047 clvmd add_change cg 2 joined nodeid 1206364352
1412340047 clvmd add_change cg 2 counts member 3 joined 1 remove 0 failed 0
1412340047 clvmd stop_kernel cg 2
1412340047 write "0" to "/sys/kernel/dlm/clvmd/control"
1412340047 clvmd check_fencing done
1412340047 clvmd check_quorum disabled
1412340047 clvmd check_fs none registered
1412340047 clvmd send_start cg 2 flags 2 data2 0 counts 1 3 1 0 0
1412340047 clvmd receive_start 1206364352:1 len 84
1412340047 clvmd match_change 1206364352:1 matches cg 2
1412340047 clvmd wait_messages cg 2 need 2 of 3
1412340047 clvmd receive_start 1189587136:3 len 84
1412340047 clvmd match_change 1189587136:3 matches cg 2
1412340047 clvmd wait_messages cg 2 need 1 of 3
1412340047 clvmd receive_start 1223141568:2 len 84
1412340047 clvmd match_change 1223141568:2 matches cg 2
1412340047 clvmd wait_messages cg 2 got all 3
1412340047 clvmd start_kernel cg 2 member_count 3
1412340047 dir_member 1223141568
1412340047 dir_member 1189587136
1412340047 set_members mkdir "/sys/kernel/config/dlm/cluster/spaces/clvmd/nodes/1206364352"
1412340047 write "1" to "/sys/kernel/dlm/clvmd/control"
1412340047 clvmd set_plock_ckpt_node from 1189587136 to 1189587136
1412340047 clvmd receive_plocks_stored 1189587136:3 flags a sig 0 need_plocks 0
1412340049 uevent: add <at> /kernel/dlm/CCB10CE8D4FF489B9A2ECB288DACF2D7
1412340049 kernel: add <at>  CCB10CE8D4FF489B9A2ECB288DACF2D7
1412340049 uevent: online <at> /kernel/dlm/CCB10CE8D4FF489B9A2ECB288DACF2D7
1412340049 kernel: online <at>  CCB10CE8D4FF489B9A2ECB288DACF2D7
1412340049 dlm:ls:CCB10CE8D4FF489B9A2ECB288DACF2D7 conf 2 1 0 memb 1189587136 1223141568 join
1223141568 left
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 add_change cg 1 joined nodeid 1223141568
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 add_change cg 1 we joined
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 add_change cg 1 counts member 2 joined 1 remove 0 failed 0
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 check_fencing done
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 check_quorum disabled
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 check_fs done
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 send_start cg 1 flags 1 data2 0 counts 0 2 1 0 0
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 receive_start 1223141568:1 len 80
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 match_change 1223141568:1 matches cg 1
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 wait_messages cg 1 need 1 of 2
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 receive_start 1189587136:2 len 80
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 match_change 1189587136:2 matches cg 1
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 wait_messages cg 1 got all 2
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 start_kernel cg 1 member_count 2
1412340049 write "153423433" to "/sys/kernel/dlm/CCB10CE8D4FF489B9A2ECB288DACF2D7/id"
1412340049 set_members mkdir "/sys/kernel/config/dlm/cluster/spaces/CCB10CE8D4FF489B9A2ECB288DACF2D7/nodes/1189587136"
1412340049 set_members mkdir "/sys/kernel/config/dlm/cluster/spaces/CCB10CE8D4FF489B9A2ECB288DACF2D7/nodes/1223141568"
1412340049 write "1" to "/sys/kernel/dlm/CCB10CE8D4FF489B9A2ECB288DACF2D7/control"
1412340049 write "0" to "/sys/kernel/dlm/CCB10CE8D4FF489B9A2ECB288DACF2D7/event_done"
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 set_plock_ckpt_node from 0 to 1189587136
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 receive_plocks_stored 1189587136:2 flags a sig 0
need_plocks 1
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 match_change 1189587136:2 matches cg 1
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 retrieve_plocks
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 retrieve_plocks first 0 last 0 r_count 0 p_count 0 sig 0
1412340049 dlm:ls:CCB10CE8D4FF489B9A2ECB288DACF2D7 conf 3 1 0 memb 1189587136 1206364352 1223141568
join 1206364352 left
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 add_change cg 2 joined nodeid 1206364352
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 add_change cg 2 counts member 3 joined 1 remove 0 failed 0
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 stop_kernel cg 2
1412340049 write "0" to "/sys/kernel/dlm/CCB10CE8D4FF489B9A2ECB288DACF2D7/control"
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 check_fencing done
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 check_quorum disabled
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 check_fs done
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 send_start cg 2 flags 2 data2 0 counts 1 3 1 0 0
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 receive_start 1206364352:1 len 84
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 match_change 1206364352:1 matches cg 2
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 wait_messages cg 2 need 2 of 3
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 receive_start 1189587136:3 len 84
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 match_change 1189587136:3 matches cg 2
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 wait_messages cg 2 need 1 of 3
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 receive_start 1223141568:2 len 84
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 match_change 1223141568:2 matches cg 2
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 wait_messages cg 2 got all 3
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 start_kernel cg 2 member_count 3
1412340049 dir_member 1223141568
1412340049 dir_member 1189587136
1412340049 set_members mkdir "/sys/kernel/config/dlm/cluster/spaces/CCB10CE8D4FF489B9A2ECB288DACF2D7/nodes/1206364352"
1412340049 write "1" to "/sys/kernel/dlm/CCB10CE8D4FF489B9A2ECB288DACF2D7/control"
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 set_plock_ckpt_node from 1189587136 to 1189587136
1412340049 CCB10CE8D4FF489B9A2ECB288DACF2D7 receive_plocks_stored 1189587136:3 flags a sig 0
need_plocks 0
1412340173 Processing membership 22680
1412340173 Adding address ip(192.168.231.68) to configfs for node 1156032704
1412340173 set_configfs_node 1156032704 192.168.231.68 local 0
1412340173 Added active node 1156032704: born-on=0, last-seen=22680, this-event=22680, last-event=22676
1412340173 Skipped active node 1189587136: born-on=22628, last-seen=22680, this-event=22680, last-event=22676
1412340173 Skipped active node 1206364352: born-on=22632, last-seen=22680, this-event=22680, last-event=22676
1412340173 Skipped active node 1223141568: born-on=22636, last-seen=22680, this-event=22680, last-event=22676
1412340294 Processing membership 22684
1412340294 Skipped active node 1156032704: born-on=22680, last-seen=22684, this-event=22684, last-event=22680
1412340294 Adding address ip(192.168.231.69) to configfs for node 1172809920
1412340294 set_configfs_node 1172809920 192.168.231.69 local 0
1412340294 Added active node 1172809920: born-on=0, last-seen=22684, this-event=22684, last-event=22680
1412340294 Skipped active node 1189587136: born-on=22628, last-seen=22684, this-event=22684, last-event=22680
1412340294 Skipped active node 1206364352: born-on=22632, last-seen=22684, this-event=22684, last-event=22680
1412340294 Skipped active node 1223141568: born-on=22636, last-seen=22684, this-event=22684, last-event=22680
1412340439 dlm:controld conf 4 1 0 memb 1172809920 1189587136 1206364352 1223141568 join 1172809920 left
1412340443 dlm:ls:clvmd conf 4 1 0 memb 1172809920 1189587136 1206364352 1223141568 join 1172809920 left
1412340443 clvmd add_change cg 3 joined nodeid 1172809920
1412340443 clvmd add_change cg 3 counts member 4 joined 1 remove 0 failed 0
1412340443 clvmd stop_kernel cg 3
1412340443 write "0" to "/sys/kernel/dlm/clvmd/control"
1412340443 clvmd check_fencing done
1412340443 clvmd check_quorum disabled
1412340443 clvmd check_fs none registered
1412340443 clvmd send_start cg 3 flags 2 data2 0 counts 2 4 1 0 0
1412340443 clvmd receive_start 1206364352:2 len 88
1412340443 clvmd match_change 1206364352:2 matches cg 3
1412340443 clvmd wait_messages cg 3 need 3 of 4
1412340443 clvmd receive_start 1223141568:3 len 88
1412340443 clvmd match_change 1223141568:3 matches cg 3
1412340443 clvmd wait_messages cg 3 need 2 of 4
1412340443 clvmd receive_start 1172809920:1 len 88
1412340443 clvmd match_change 1172809920:1 matches cg 3
1412340443 clvmd wait_messages cg 3 need 1 of 4
1412340443 clvmd receive_start 1189587136:4 len 88
1412340443 clvmd match_change 1189587136:4 matches cg 3
1412340443 clvmd wait_messages cg 3 got all 4
1412340443 clvmd start_kernel cg 3 member_count 4
1412340443 dir_member 1206364352
1412340443 dir_member 1223141568
1412340443 dir_member 1189587136
1412340443 set_members mkdir "/sys/kernel/config/dlm/cluster/spaces/clvmd/nodes/1172809920"
1412340443 write "1" to "/sys/kernel/dlm/clvmd/control"
1412340443 clvmd set_plock_ckpt_node from 1189587136 to 1189587136
1412340443 clvmd receive_plocks_stored 1189587136:4 flags a sig 0 need_plocks 0
1412340447 dlm:ls:clvmd conf 3 0 1 memb 1189587136 1206364352 1223141568 join left 1172809920
1412340447 clvmd add_change cg 4 remove nodeid 1172809920 reason 2
1412340447 clvmd add_change cg 4 counts member 3 joined 0 remove 1 failed 0
1412340447 clvmd stop_kernel cg 4
1412340447 write "0" to "/sys/kernel/dlm/clvmd/control"
1412340447 clvmd check_fencing done
1412340447 clvmd check_quorum disabled
1412340447 clvmd check_fs none registered
1412340447 clvmd send_start cg 4 flags 2 data2 0 counts 3 3 0 1 0
1412340447 clvmd receive_start 1223141568:4 len 84
1412340447 clvmd match_change 1223141568:4 matches cg 4
1412340447 clvmd wait_messages cg 4 need 2 of 3
1412340447 clvmd receive_start 1189587136:5 len 84
1412340447 clvmd match_change 1189587136:5 matches cg 4
1412340447 clvmd wait_messages cg 4 need 1 of 3
1412340447 clvmd receive_start 1206364352:3 len 84
1412340447 clvmd match_change 1206364352:3 matches cg 4
1412340447 clvmd wait_messages cg 4 got all 3
1412340447 clvmd start_kernel cg 4 member_count 3
1412340447 dir_member 1172809920
1412340447 dir_member 1206364352
1412340447 dir_member 1223141568
1412340447 dir_member 1189587136
1412340447 set_members rmdir "/sys/kernel/config/dlm/cluster/spaces/clvmd/nodes/1172809920"
1412340447 write "1" to "/sys/kernel/dlm/clvmd/control"
1412340447 clvmd set_plock_ckpt_node from 1189587136 to 1189587136
1412340447 clvmd receive_plocks_stored 1189587136:5 flags a sig 0 need_plocks 0
1412340448 dlm:controld conf 3 0 1 memb 1189587136 1206364352 1223141568 join left 1172809920
1412340448 dlm:controld conf 3 0 1 memb 1189587136 1206364352 1223141568 join left 1172809920
1412340507 Processing membership 22688
1412340507 Skipped active node 1156032704: born-on=22680, last-seen=22688, this-event=22688, last-event=22684
1412340507 del_configfs_node rmdir "/sys/kernel/config/dlm/cluster/comms/1172809920"
1412340507 Removed inactive node 1172809920: born-on=22684, last-seen=22684, this-event=22688, last-event=22684
1412340507 Skipped active node 1189587136: born-on=22628, last-seen=22688, this-event=22688, last-event=22684
1412340507 Skipped active node 1206364352: born-on=22632, last-seen=22688, this-event=22688, last-event=22684
1412340507 Skipped active node 1223141568: born-on=22636, last-seen=22688, this-event=22688, last-event=22684
1412340532 Processing membership 22692
1412340532 Skipped active node 1156032704: born-on=22680, last-seen=22692, this-event=22692, last-event=22688
1412340532 Adding address ip(192.168.231.69) to configfs for node 1172809920
1412340532 set_configfs_node 1172809920 192.168.231.69 local 0
1412340532 Added active node 1172809920: born-on=22684, last-seen=22692, this-event=22692, last-event=22688
1412340532 Skipped active node 1189587136: born-on=22628, last-seen=22692, this-event=22692, last-event=22688
1412340532 Skipped active node 1206364352: born-on=22632, last-seen=22692, this-event=22692, last-event=22688
1412340532 Skipped active node 1223141568: born-on=22636, last-seen=22692, this-event=22692, last-event=22688
1412340570 Processing membership 22696
1412340570 Skipped active node 1156032704: born-on=22680, last-seen=22696, this-event=22696, last-event=22692
1412340570 del_configfs_node rmdir "/sys/kernel/config/dlm/cluster/comms/1172809920"
1412340570 Removed inactive node 1172809920: born-on=22692, last-seen=22692, this-event=22696, last-event=22692
1412340570 Skipped active node 1189587136: born-on=22628, last-seen=22696, this-event=22696, last-event=22692
1412340570 Skipped active node 1206364352: born-on=22632, last-seen=22696, this-event=22696, last-event=22692
1412340570 Skipped active node 1223141568: born-on=22636, last-seen=22696, this-event=22696, last-event=22692

Is there any documentation on troubleshooting DLM/cLVM?

Regards.

-- 
Daniel Dehennin
Récupérer ma clef GPG: gpg --recv-keys 0xCC1E9E5B7A6FE2DF
Fingerprint: 3E69 014E 5C23 50E8 9ED6  2AAD CC1E 9E5B 7A6F E2DF
--

-- 
Linux-cluster mailing list
Linux-cluster <at> redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster
Neale Ferguson | 2 Oct 21:30 2014
Picon

Fencing of node

After creating simple two node cluster, one node is being fenced continually. I'm running pacemaker
(1.1.10-29) with two nodes and the following corosync.conf:

totem {
version: 2
secauth: off
cluster_name: rh7cluster
transport: udpu
}

nodelist {
  node {
        ring0_addr: rh7cn1.devlab.sinenomine.net
        nodeid: 1
       }
  node {
        ring0_addr: rh7cn2.devlab.sinenomine.net
        nodeid: 2
       }
}

quorum {
provider: corosync_votequorum
two_node: 1
}

logging {
to_syslog: yes
}

Starting the cluster shows:

Oct  2 15:17:47 rh7cn1 kernel: dlm: connect from non cluster node

In the logs of both nodes. Both nodes then try and bring up resources (dlm, clvmd, and a cluster fs). 

Just prior to a node being fence, both nodes show the following

# pcs resource show
 Clone Set: dlm-clone [dlm]
     Started: [ rh7cn1.devlab.sinenomine.net rh7cn2.devlab.sinenomine.net ]
 Clone Set: clvmd-clone [clvmd]
     clvmd	(ocf::heartbeat:clvm):	FAILED 
     Started: [ rh7cn2.devlab.sinenomine.net ]
 Clone Set: clusterfs-clone [clusterfs]
     Started: [ rh7cn2.devlab.sinenomine.net ]
     Stopped: [ rh7cn1.devlab.sinenomine.net ]

Shortly after there is a clvmd timeout message in one of the logs and then that node gets fenced. I had added
the high-availability firewalld service to both nodes.

Running crm_simulate -SL -VV shows:

 warning: unpack_rsc_op: 	Processing failed op start for clvmd:1 on rh7cn1.devlab.sinenomine.net:
unknown error (1)

Current cluster status:
Online: [ rh7cn1.devlab.sinenomine.net rh7cn2.devlab.sinenomine.net ]

 ZVMPOWER	(stonith:fence_zvm):	Started rh7cn2.devlab.sinenomine.net 
 Clone Set: dlm-clone [dlm]
     Started: [ rh7cn1.devlab.sinenomine.net rh7cn2.devlab.sinenomine.net ]
 Clone Set: clvmd-clone [clvmd]
     clvmd	(ocf::heartbeat:clvm):	FAILED rh7cn1.devlab.sinenomine.net 
     Started: [ rh7cn2.devlab.sinenomine.net ]
 Clone Set: clusterfs-clone [clusterfs]
     Started: [ rh7cn2.devlab.sinenomine.net ]
     Stopped: [ rh7cn1.devlab.sinenomine.net ]

 warning: common_apply_stickiness: 	Forcing clvmd-clone away from rh7cn1.devlab.sinenomine.net
after 1000000 failures (max=1000000)
 warning: common_apply_stickiness: 	Forcing clvmd-clone away from rh7cn1.devlab.sinenomine.net
after 1000000 failures (max=1000000)
Transition Summary:
 * Stop    clvmd:1	(rh7cn1.devlab.sinenomine.net)

Executing cluster transition:
 * Pseudo action:   clvmd-clone_stop_0
 * Resource action: clvmd           stop on rh7cn1.devlab.sinenomine.net
 * Pseudo action:   clvmd-clone_stopped_0
 * Pseudo action:   all_stopped

Revised cluster status:
 warning: unpack_rsc_op: 	Processing failed op start for clvmd:1 on rh7cn1.devlab.sinenomine.net:
unknown error (1)
Online: [ rh7cn1.devlab.sinenomine.net rh7cn2.devlab.sinenomine.net ]

 ZVMPOWER	(stonith:fence_zvm):	Started rh7cn2.devlab.sinenomine.net 
 Clone Set: dlm-clone [dlm]
     Started: [ rh7cn1.devlab.sinenomine.net rh7cn2.devlab.sinenomine.net ]
 Clone Set: clvmd-clone [clvmd]
     Started: [ rh7cn2.devlab.sinenomine.net ]
     Stopped: [ rh7cn1.devlab.sinenomine.net ]
 Clone Set: clusterfs-clone [clusterfs]
     Started: [ rh7cn2.devlab.sinenomine.net ]
     Stopped: [ rh7cn1.devlab.sinenomine.net ]

With RHEL 6 I would use a qdisk but this has been replaced by corosync_votequorum.

This is my first RHEL 7 HA cluster so I'm at the beginning of my learning. Any pointers as to what I should look
at or what I need to read?

Neale

--

-- 
Linux-cluster mailing list
Linux-cluster <at> redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Ferenc Wagner | 22 Sep 10:24 2014
Picon

ordering scores and kinds

Hi,

http://clusterlabs.org/doc/en-US/Pacemaker/1.1/html/Pacemaker_Explained/s-resource-ordering.html
says that optional ordering is achieved by setting the "kind" attribute
to "Optional".  However, the next section
http://clusterlabs.org/doc/en-US/Pacemaker/1.1/html/Pacemaker_Explained/_advisory_ordering.html
says that advisory ordering is achieved by setting the "score" attribute
to 0.  Is there any difference between an optional and an advisory
ordering constraint?  How do nonzero score values influence cluster
behaviour, if at all?  Or is the kind attribute intended to replace all
score settings on ordering constraints?
-- 
Thanks,
Feri.

--

-- 
Linux-cluster mailing list
Linux-cluster <at> redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Kaisar Ahmed Khan | 21 Sep 08:06 2014
Picon

GFS2 mount problem


 Dear All,

 I have been experiencing a problem for long time in GFS2 with three node cluster.

Short brief about my scenario
All three nodes in a Host with KVM technology.  storage accessing by iSCSI on all three nodes.
One 50GB LUN initiated on all three nodes , and configured GFS2 file system .
GFS file system mounted at all three nodes persistently by fstab.

Problem is:
When I reboot/ fence any machine , I found GFS2 file system not mounted . it got  mounted after  applying # mount –a Command .
 
What possible cause of this problem. ?

Thanks
Kaisar


 
--

-- 
Linux-cluster mailing list
Linux-cluster <at> redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Gmane