ogaya****@yahoo*****
ogaya****@yahoo*****
2017年 10月 2日 (月) 15:00:28 JST
お世話になっております。小笠原と申します。 現在以下の構成で環境を構築しようとしています。 ・VMware上の仮想環境 ・CentOS7 ・pacemaker 1.1.16 ・corosync 2.4.2 ・DRBD 8.4 STONITHの設定をいれて、動作確認を行ったところ リソースのstop失敗やインターコネクト障害でSTONITHが発動しました。 そのためSTONITHの設定は出来ていると思ったのですが pcs status コマンド でステータスを確認すると以下のメッセージが表示されます。 "WARNING: no stonith devices and stonith-enabled is not false" 何かの設定が足りないのでしょうか。 pcs status の結果と pcs config の結果は以下です。 ※IPアドレスはマスクしています # pcs status Cluster name: mycluster WARNING: no stonith devices and stonith-enabled is not false <====このメッセージ Stack: corosync Current DC: host01 (version 1.1.16-1.el7-94ff4df) - partition with quorum Last updated: Mon Oct 2 14:21:07 2017 Last change: Fri Sep 29 10:54:24 2017 by root via crm_attribute on host01 2 nodes configured 11 resources configured Online: [ host01 host02 ] Full list of resources: Master/Slave Set: RES_MySQLData [RES_DRBD] Masters: [ host01 ] Slaves: [ host02 ] Resource Group: res_grp RES_FS (ocf::heartbeat:Filesystem): Started host01 RES_MySQLSERVICE (ocf::heartbeat:mysql): Started host01 ClusterIP (ocf::heartbeat:IPaddr2): Started host01 Resource Group: grpStonith1 prmHelper1 (stonith:external/stonith-helper): Started host02 prmLibvirt1 (stonith:external/libvirt): Started host02 Resource Group: grpStonith2 prmHelper2 (stonith:external/stonith-helper): Started host01 prmLibvirt2 (stonith:external/libvirt): Started host01 Clone Set: clone_ping [RES_PING] Started: [ host01 host02 ] PCSD Status: host01: Online host02: Online Daemon Status: corosync: active/disabled pacemaker: active/disabled pcsd: active/enabled ------------------------------------------------------------- # pcs config Cluster Name: mycluster Corosync Nodes: host01 host02 Pacemaker Nodes: host01 host02 Resources: Master: RES_MySQLData Meta Attrs: master-max=1 master-node-max=1 clone-max=2 clone-node-max=1 notify=true Resource: RES_DRBD (class=ocf provider=linbit type=drbd) Attributes: drbd_resource=r0 Operations: start interval=0s timeout=240 (RES_DRBD-start-interval-0s) promote interval=0s timeout=90 (RES_DRBD-promote-interval-0s) demote interval=0s timeout=90 (RES_DRBD-demote-interval-0s) stop interval=0s timeout=100 (RES_DRBD-stop-interval-0s) monitor interval=30s (RES_DRBD-monitor-interval-30s) Group: res_grp Resource: RES_FS (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/drbd0 directory=/var/lib/mysql fstype=ext4 Operations: start interval=0s timeout=60 (RES_FS-start-interval-0s) monitor interval=20 timeout=40 (RES_FS-monitor-interval-20) stop on-fail=fence interval=0s timeout=60 (RES_FS-stop-interval-0s) Resource: RES_MySQLSERVICE (class=ocf provider=heartbeat type=mysql) Attributes: config=/etc/my.cnf binary=/usr/sbin/mysqld user=mysql datadir=/var/lib/mysql/data pid=/var/lib/mysql/mysqld.pid socket=/var/lib/mysql/mysql.sock log=/var/log/mysql/mysqld.err Meta Attrs: migration-threshold=3 Operations: start interval=0s timeout=120s (RES_MySQLSERVICE-start-interval-0s) monitor interval=20s timeout=30s on-fail=restart (RES_MySQLSERVICE-monitor-interval-20s) stop interval=0s timeout=120s on-fail=fence (RES_MySQLSERVICE-stop-interval-0s) Resource: ClusterIP (class=ocf provider=heartbeat type=IPaddr2) Attributes: ip=1NN.NN.242.253 cidr_netmask=22 Operations: start interval=0s timeout=20s (ClusterIP-start-interval-0s) monitor interval=30s (ClusterIP-monitor-interval-30s) stop interval=0s timeout=20s on-fail=fence (ClusterIP-stop-interval-0s) Group: grpStonith1 Resource: prmHelper1 (class=stonith type=external/stonith-helper) Attributes: pcmk_reboot_retries=1 hostlist=host01 dead_check_target="1NN.NN.239.47 1NN.NN.235.8 1NN.NN.243.226" standby_check_command="/usr/sbin/crm_resource -r ClusterIP -W | grep -qi `hostname`" run_online_check=yes Operations: start on-fail=restart interval=0s timeout=20s (prmHelper1-start-0s) monitor on-fail=restart interval=3600s timeout=20s (prmHelper1-monitor-3600s) stop on-fail=ignore interval=0s timeout=15 (prmHelper1-stop-0s) Resource: prmLibvirt1 (class=stonith type=external/libvirt) Attributes: hostlist=host01 hypervisor_uri=vpx://1NN.NN.239.13/TEST/production/TDC-TST001/esx02.vmware.tk.ad.jp?no_verify=1 Operations: start on-fail=restart interval=0s timeout=60s (prmLibvirt1-start-0s) monitor on-fail=restart interval=3600s timeout=60s (prmLibvirt1-monitor-3600s) stop on-fail=ignore interval=0s timeout=60s (prmLibvirt1-stop-0s) Group: grpStonith2 Resource: prmHelper2 (class=stonith type=external/stonith-helper) Attributes: pcmk_reboot_retries=1 hostlist=host02 dead_check_target="1NN.NN.239.48 1NN.NN.243.227" standby_check_command="/usr/sbin/crm_resource -r ClusterIP -W | grep -qi `hostname`" run_online_check=yes Operations: start on-fail=restart interval=0s timeout=20s (prmHelper2-start-0s) monitor on-fail=restart interval=3600s timeout=20s (prmHelper2-monitor-3600s) stop on-fail=ignore interval=0s timeout=15 (prmHelper2-stop-0s) Resource: prmLibvirt2 (class=stonith type=external/libvirt) Attributes: hostlist=host02 hypervisor_uri=vpx://1NN.NN.239.13/TEST/production/TDC-TST001/esx02.vmware.tk.ad.jp?no_verify=1 Operations: start on-fail=restart interval=0s timeout=60s (prmLibvirt2-start-0s) monitor on-fail=restart interval=3600s timeout=60s (prmLibvirt2-monitor-3600s) stop on-fail=ignore interval=0s timeout=60s (prmLibvirt2-stop-0s) Clone: clone_ping Resource: RES_PING (class=ocf provider=pacemaker type=ping) Attributes: name=default_ping_set host_list=1NN.NN.239.254 multiplier=100 attempts=30 timeout=2 debug=true Operations: start on-fail=restart interval=0s timeout=60s (RES_PING-start-0s) monitor on-fail=restart interval=10s timeout=60s (RES_PING-monitor-10s) stop on-fail=ignore interval=0s timeout=60s (RES_PING-stop-0s) Stonith Devices: Fencing Levels: Node: host01 Level 1 - prmHelper1 Level 2 - prmLibvirt1 Node: host02 Level 1 - prmHelper2 Level 2 - prmLibvirt2 Location Constraints: Resource: RES_MySQLData Enabled on: host01 (score:100) (id:location-RES_MySQLData-host01-100) Resource: grpStonith1 Disabled on: host01 (score:-INFINITY) (id:location-grpStonith1-host01--INFINITY) Resource: grpStonith2 Disabled on: host02 (score:-INFINITY) (id:location-grpStonith2-host02--INFINITY) Resource: res_grp Constraint: res_grp-location Rule: score=-INFINITY boolean-op=or (id:res_grp-location-rule) Expression: not_defined default_ping_set (id:res_grp-location-rule-expression) Expression: default_ping_set lt 100 (id:res_grp-location-rule-expression-0) Ordering Constraints: promote RES_MySQLData then start res_grp (kind:Mandatory) (id:order-RES_MySQLData-res_grp-mandatory) clone_ping then res_grp (score:0) (non-symmetrical) (id:rsc_order-clone_ping-res_grp-1) Colocation Constraints: res_grp with RES_MySQLData (score:INFINITY) (with-rsc-role:Master) (id:colocation-res_grp-RES_MySQLData-INFINITY) Resources Defaults: resource-stickiness: 100 Operations Defaults: No defaults set Cluster Properties: cluster-infrastructure: corosync cluster-name: mycluster dc-version: 1.1.16-1.el7-94ff4df have-watchdog: false stonith-enabled: true ---------------------------------------------------------------------- また、pcs stonith list の結果と stonith -L の結果は以下です。 # pcs stonith list Error: No stonith agents available. Do you have fence agents installed? # stonith -L apcmaster apcmastersnmp apcsmart baytech bladehpi cyclades external/drac5 external/dracmc-telnet external/ec2 external/hetzner external/hmchttp external/ibmrsa external/ibmrsa-telnet external/ipmi external/ippower9258 external/kdumpcheck external/libvirt external/nut external/rackpdu external/riloe external/stonith-helper external/vcenter external/vmware external/xen0 external/xen0-ha ibmhmc ipmilan meatware nw_rpc100s rcd_serial rps10 suicide wti_mpc wti_nps よろしくお願いいたします。