[Linux-ha-jp] PacemakerのSTONITHデバイスのWARNINGにつきまして

アーカイブの一覧に戻る

ogaya****@yahoo***** ogaya****@yahoo*****
2017年 10月 2日 (月) 15:00:28 JST


お世話になっております。小笠原と申します。


現在以下の構成で環境を構築しようとしています。

  ・VMware上の仮想環境
  ・CentOS7
  ・pacemaker 1.1.16
  ・corosync 2.4.2
  ・DRBD 8.4

STONITHの設定をいれて、動作確認を行ったところ
リソースのstop失敗やインターコネクト障害でSTONITHが発動しました。
そのためSTONITHの設定は出来ていると思ったのですが pcs status コマンド
でステータスを確認すると以下のメッセージが表示されます。

"WARNING: no stonith devices and stonith-enabled is not false"

何かの設定が足りないのでしょうか。

pcs status の結果と pcs config の結果は以下です。
※IPアドレスはマスクしています

# pcs status
Cluster name: mycluster
WARNING: no stonith devices and stonith-enabled is not false <====このメッセージ
Stack: corosync
Current DC: host01 (version 1.1.16-1.el7-94ff4df) - partition with quorum
Last updated: Mon Oct  2 14:21:07 2017
Last change: Fri Sep 29 10:54:24 2017 by root via crm_attribute on host01

2 nodes configured
11 resources configured

Online: [ host01 host02 ]

Full list of resources:

 Master/Slave Set: RES_MySQLData [RES_DRBD]
     Masters: [ host01 ]
     Slaves: [ host02 ]
 Resource Group: res_grp
     RES_FS     (ocf::heartbeat:Filesystem):    Started host01
     RES_MySQLSERVICE   (ocf::heartbeat:mysql): Started host01
     ClusterIP  (ocf::heartbeat:IPaddr2):       Started host01
 Resource Group: grpStonith1
     prmHelper1 (stonith:external/stonith-helper):      Started host02
     prmLibvirt1        (stonith:external/libvirt):     Started host02
 Resource Group: grpStonith2
     prmHelper2 (stonith:external/stonith-helper):      Started host01
     prmLibvirt2        (stonith:external/libvirt):     Started host01
 Clone Set: clone_ping [RES_PING]
     Started: [ host01 host02 ]

PCSD Status:
  host01: Online
  host02: Online

Daemon Status:
  corosync: active/disabled
  pacemaker: active/disabled
  pcsd: active/enabled

-------------------------------------------------------------

# pcs config
Cluster Name: mycluster
Corosync Nodes:
 host01 host02
Pacemaker Nodes:
 host01 host02

Resources:
 Master: RES_MySQLData
  Meta Attrs: master-max=1 master-node-max=1 clone-max=2 clone-node-max=1 notify=true
  Resource: RES_DRBD (class=ocf provider=linbit type=drbd)
   Attributes: drbd_resource=r0
   Operations: start interval=0s timeout=240 (RES_DRBD-start-interval-0s)
               promote interval=0s timeout=90 (RES_DRBD-promote-interval-0s)
               demote interval=0s timeout=90 (RES_DRBD-demote-interval-0s)
               stop interval=0s timeout=100 (RES_DRBD-stop-interval-0s)
               monitor interval=30s (RES_DRBD-monitor-interval-30s)
 Group: res_grp
  Resource: RES_FS (class=ocf provider=heartbeat type=Filesystem)
   Attributes: device=/dev/drbd0 directory=/var/lib/mysql fstype=ext4
   Operations: start interval=0s timeout=60 (RES_FS-start-interval-0s)
               monitor interval=20 timeout=40 (RES_FS-monitor-interval-20)
               stop on-fail=fence interval=0s timeout=60 (RES_FS-stop-interval-0s)
  Resource: RES_MySQLSERVICE (class=ocf provider=heartbeat type=mysql)
   Attributes: config=/etc/my.cnf binary=/usr/sbin/mysqld user=mysql datadir=/var/lib/mysql/data pid=/var/lib/mysql/mysqld.pid socket=/var/lib/mysql/mysql.sock log=/var/log/mysql/mysqld.err
   Meta Attrs: migration-threshold=3
   Operations: start interval=0s timeout=120s (RES_MySQLSERVICE-start-interval-0s)
               monitor interval=20s timeout=30s on-fail=restart (RES_MySQLSERVICE-monitor-interval-20s)
               stop interval=0s timeout=120s on-fail=fence (RES_MySQLSERVICE-stop-interval-0s)
  Resource: ClusterIP (class=ocf provider=heartbeat type=IPaddr2)
   Attributes: ip=1NN.NN.242.253 cidr_netmask=22
   Operations: start interval=0s timeout=20s (ClusterIP-start-interval-0s)
               monitor interval=30s (ClusterIP-monitor-interval-30s)
               stop interval=0s timeout=20s on-fail=fence (ClusterIP-stop-interval-0s)
 Group: grpStonith1
  Resource: prmHelper1 (class=stonith type=external/stonith-helper)
   Attributes: pcmk_reboot_retries=1 hostlist=host01 dead_check_target="1NN.NN.239.47 1NN.NN.235.8 1NN.NN.243.226" standby_check_command="/usr/sbin/crm_resource -r ClusterIP -W | grep -qi `hostname`" run_online_check=yes
   Operations: start on-fail=restart interval=0s timeout=20s (prmHelper1-start-0s)
               monitor on-fail=restart interval=3600s timeout=20s (prmHelper1-monitor-3600s)
               stop on-fail=ignore interval=0s timeout=15 (prmHelper1-stop-0s)
  Resource: prmLibvirt1 (class=stonith type=external/libvirt)
   Attributes: hostlist=host01 hypervisor_uri=vpx://1NN.NN.239.13/TEST/production/TDC-TST001/esx02.vmware.tk.ad.jp?no_verify=1
   Operations: start on-fail=restart interval=0s timeout=60s (prmLibvirt1-start-0s)
               monitor on-fail=restart interval=3600s timeout=60s (prmLibvirt1-monitor-3600s)
               stop on-fail=ignore interval=0s timeout=60s (prmLibvirt1-stop-0s)
 Group: grpStonith2
  Resource: prmHelper2 (class=stonith type=external/stonith-helper)
   Attributes: pcmk_reboot_retries=1 hostlist=host02 dead_check_target="1NN.NN.239.48 1NN.NN.243.227" standby_check_command="/usr/sbin/crm_resource -r ClusterIP -W | grep -qi `hostname`" run_online_check=yes
   Operations: start on-fail=restart interval=0s timeout=20s (prmHelper2-start-0s)
               monitor on-fail=restart interval=3600s timeout=20s (prmHelper2-monitor-3600s)
               stop on-fail=ignore interval=0s timeout=15 (prmHelper2-stop-0s)
  Resource: prmLibvirt2 (class=stonith type=external/libvirt)
   Attributes: hostlist=host02 hypervisor_uri=vpx://1NN.NN.239.13/TEST/production/TDC-TST001/esx02.vmware.tk.ad.jp?no_verify=1
   Operations: start on-fail=restart interval=0s timeout=60s (prmLibvirt2-start-0s)
               monitor on-fail=restart interval=3600s timeout=60s (prmLibvirt2-monitor-3600s)
               stop on-fail=ignore interval=0s timeout=60s (prmLibvirt2-stop-0s)
 Clone: clone_ping
  Resource: RES_PING (class=ocf provider=pacemaker type=ping)
   Attributes: name=default_ping_set host_list=1NN.NN.239.254 multiplier=100 attempts=30 timeout=2 debug=true
   Operations: start on-fail=restart interval=0s timeout=60s (RES_PING-start-0s)
               monitor on-fail=restart interval=10s timeout=60s (RES_PING-monitor-10s)
               stop on-fail=ignore interval=0s timeout=60s (RES_PING-stop-0s)

Stonith Devices:
Fencing Levels:

 Node: host01
  Level 1 - prmHelper1
  Level 2 - prmLibvirt1
 Node: host02
  Level 1 - prmHelper2
  Level 2 - prmLibvirt2
Location Constraints:
  Resource: RES_MySQLData
    Enabled on: host01 (score:100) (id:location-RES_MySQLData-host01-100)
  Resource: grpStonith1
    Disabled on: host01 (score:-INFINITY) (id:location-grpStonith1-host01--INFINITY)
  Resource: grpStonith2
    Disabled on: host02 (score:-INFINITY) (id:location-grpStonith2-host02--INFINITY)
  Resource: res_grp
    Constraint: res_grp-location
      Rule: score=-INFINITY boolean-op=or  (id:res_grp-location-rule)
        Expression: not_defined default_ping_set  (id:res_grp-location-rule-expression)
        Expression: default_ping_set lt 100  (id:res_grp-location-rule-expression-0)
Ordering Constraints:
  promote RES_MySQLData then start res_grp (kind:Mandatory) (id:order-RES_MySQLData-res_grp-mandatory)
  clone_ping then res_grp (score:0) (non-symmetrical) (id:rsc_order-clone_ping-res_grp-1)
Colocation Constraints:
  res_grp with RES_MySQLData (score:INFINITY) (with-rsc-role:Master) (id:colocation-res_grp-RES_MySQLData-INFINITY)

Resources Defaults:
 resource-stickiness: 100
Operations Defaults:
 No defaults set

Cluster Properties:
 cluster-infrastructure: corosync
 cluster-name: mycluster
 dc-version: 1.1.16-1.el7-94ff4df
 have-watchdog: false
 stonith-enabled: true

----------------------------------------------------------------------

また、pcs stonith list の結果と stonith -L の結果は以下です。

# pcs stonith list
Error: No stonith agents available. Do you have fence agents installed?

# stonith -L
apcmaster
apcmastersnmp
apcsmart
baytech
bladehpi
cyclades
external/drac5
external/dracmc-telnet
external/ec2
external/hetzner
external/hmchttp
external/ibmrsa
external/ibmrsa-telnet
external/ipmi
external/ippower9258
external/kdumpcheck
external/libvirt
external/nut
external/rackpdu
external/riloe
external/stonith-helper
external/vcenter
external/vmware
external/xen0
external/xen0-ha
ibmhmc
ipmilan
meatware
nw_rpc100s
rcd_serial
rps10
suicide
wti_mpc
wti_nps

よろしくお願いいたします。





Linux-ha-japan メーリングリストの案内
アーカイブの一覧に戻る