千家信息网

Redhat 6配置RHCS实现双机HA群集

发表于:2025-02-21 作者:千家信息网编辑
千家信息网最后更新 2025年02月21日,最近测试了RedHat 6.5上RHCS,搭建了一个双机HA群集,在此将配置过程和测试过程分享给大家,主要包括节点配置、群集管理服务器配置、群集创建与配置、群集测试等内容。一、测试环境计算机名操作系统
千家信息网最后更新 2025年02月21日Redhat 6配置RHCS实现双机HA群集


最近测试了RedHat 6.5上RHCS,搭建了一个双机HA群集,在此将配置过程和测试过程分享给大家,主要包括节点配置、群集管理服务器配置、群集创建与配置、群集测试等内容。


一、测试环境

计算机名

操作系统

IP地址

群集IP

安装的软件包

HAmanager

RedHat 6.5

192.168.10.150

-

luciiscsi target(用于仲裁盘)

node1

RedHat 6.5

192.168.10.104


192.168.10.103

High Availability、httpd

node2

RedHat 6.5

192.168.10.105

High Availability、httpd


二、节点配置

1、在三台机分别配置hosts互相解析

[root@HAmanager ~]# cat /etc/hosts

192.168.10.104 node1 node1.localdomain

192.168.10.105 node2 node2.localdomain

192.168.10.150 HAmanager HAmanager.localdomain


[root@node1 ~]# cat /etc/hosts

192.168.10.104 node1 node1.localdomain

192.168.10.105 node2 node2.localdomain

192.168.10.150 HAmanager HAmanager.localdomain


[root@node2 ~]# cat /etc/hosts

192.168.10.104 node1 node1.localdomain

192.168.10.105 node2 node2.localdomain

192.168.10.150 HAmanager HAmanager.localdomain


2、在三台机分别配置SSH互信

[root@HAmanager ~]# ssh-keygen -t rsa

[root@HAmanager ~]# ssh-copy-id -i node1


[root@node1 ~]# ssh-keygen -t rsa

[root@node1 ~]# ssh-copy-id -i node2


[root@node2 ~]# ssh-keygen -t rsa

[root@node2 ~]# ssh-copy-id -i node1



3、两个节点关闭NetworkManager和acpid服务

[root@node1 ~]# service NetworkManager stop

[root@node1 ~]# chkconfig NetworkManager off

[root@node1 ~]# service acpid stop

[root@node1 ~]# chkconfig acpid off


[root@node2 ~]# service NetworkManager stop

[root@node2 ~]# chkconfig NetworkManager off

[root@node2 ~]# service acpid stop

[root@node2 ~]# chkconfig acpid off


4、两个节点配置本地yum源

[root@node1 ~]# cat/etc/yum.repos.d/rhel6.5.repo

[Server]

name=base

baseurl=file:///mnt/

enabled=1

gpgcheck=0

[HighAvailability]

name=base

baseurl=file:///mnt/HighAvailability

enabled=1

gpgcheck=0


[root@node2 ~]# cat/etc/yum.repos.d/rhel6.5.repo

[Server]

name=base

baseurl=file:///mnt/

enabled=1

gpgcheck=0

[HighAvailability]

name=base

baseurl=file:///mnt/HighAvailability

enabled=1

gpgcheck=0


5、两个节点分别安装群集软件包

[root@node1 ~]# yum groupinstall 'High Availability' -y

Installed:

ccs.x86_64 0:0.16.2-69.el6 cman.x86_640:3.0.12.1-59.el6

omping.x86_64 0:0.0.4-1.el6 rgmanager.x86_640:3.0.12.1-19.el6

Dependency Installed:

cifs-utils.x86_640:4.8.1-19.el6 clusterlib.x86_64 0:3.0.12.1-59.el6

corosync.x86_640:1.4.1-17.el6 corosynclib.x86_64 0:1.4.1-17.el6

cyrus-sasl-md5.x86_640:2.1.23-13.el6_3.1 fence-agents.x86_64 0:3.1.5-35.el6

fence-virt.x86_640:0.2.3-15.el6 gnutls-utils.x86_64 0:2.8.5-10.el6_4.2

ipmitool.x86_64 0:1.8.11-16.el6 keyutils.x86_640:1.4-4.el6

libevent.x86_640:1.4.13-4.el6 libgssglue.x86_64 0:0.1-11.el6

libibverbs.x86_640:1.1.7-1.el6 librdmacm.x86_64 0:1.0.17-1.el6

libtirpc.x86_640:0.2.1-6.el6_4 libvirt-client.x86_64 0:0.10.2-29.el6

lm_sensors-libs.x86_640:3.1.1-17.el6 modcluster.x86_640:0.16.2-28.el6

nc.x86_64 0:1.84-22.el6 net-snmp-libs.x86_64 1:5.5-49.el6

net-snmp-utils.x86_641:5.5-49.el6 nfs-utils.x86_641:1.2.3-39.el6

nfs-utils-lib.x86_640:1.1.5-6.el6 numactl.x86_640:2.0.7-8.el6

oddjob.x86_64 0:0.30-5.el6 openais.x86_64 0:1.1.1-7.el6

openaislib.x86_640:1.1.1-7.el6 perl-Net-Telnet.noarch 0:3.03-11.el6

pexpect.noarch 0:2.3-6.el6 python-suds.noarch0:0.4.1-3.el6

quota.x86_64 1:3.17-20.el6 resource-agents.x86_640:3.9.2-40.el6

ricci.x86_64 0:0.16.2-69.el6 rpcbind.x86_640:0.2.0-11.el6

sg3_utils.x86_64 0:1.28-5.el6 tcp_wrappers.x86_640:7.6-57.el6

telnet.x86_64 1:0.17-47.el6_3.1 yajl.x86_64 0:1.0.7-3.el6

Complete!

[root@node2 ~]# yum groupinstall 'High Availability' -y

Installed:

ccs.x86_64 0:0.16.2-69.el6 cman.x86_640:3.0.12.1-59.el6

omping.x86_64 0:0.0.4-1.el6 rgmanager.x86_640:3.0.12.1-19.el6

Dependency Installed:

cifs-utils.x86_640:4.8.1-19.el6 clusterlib.x86_64 0:3.0.12.1-59.el6

corosync.x86_640:1.4.1-17.el6 corosynclib.x86_64 0:1.4.1-17.el6

cyrus-sasl-md5.x86_640:2.1.23-13.el6_3.1 fence-agents.x86_64 0:3.1.5-35.el6

fence-virt.x86_640:0.2.3-15.el6 gnutls-utils.x86_64 0:2.8.5-10.el6_4.2

ipmitool.x86_64 0:1.8.11-16.el6 keyutils.x86_640:1.4-4.el6

libevent.x86_640:1.4.13-4.el6 libgssglue.x86_64 0:0.1-11.el6

libibverbs.x86_640:1.1.7-1.el6 librdmacm.x86_64 0:1.0.17-1.el6

libtirpc.x86_640:0.2.1-6.el6_4 libvirt-client.x86_64 0:0.10.2-29.el6

lm_sensors-libs.x86_640:3.1.1-17.el6 modcluster.x86_640:0.16.2-28.el6

nc.x86_64 0:1.84-22.el6 net-snmp-libs.x86_64 1:5.5-49.el6

net-snmp-utils.x86_641:5.5-49.el6 nfs-utils.x86_641:1.2.3-39.el6

nfs-utils-lib.x86_640:1.1.5-6.el6 numactl.x86_640:2.0.7-8.el6

oddjob.x86_64 0:0.30-5.el6 openais.x86_64 0:1.1.1-7.el6

openaislib.x86_640:1.1.1-7.el6 perl-Net-Telnet.noarch 0:3.03-11.el6

pexpect.noarch 0:2.3-6.el6 python-suds.noarch0:0.4.1-3.el6

quota.x86_64 1:3.17-20.el6 resource-agents.x86_640:3.9.2-40.el6

ricci.x86_64 0:0.16.2-69.el6 rpcbind.x86_640:0.2.0-11.el6

sg3_utils.x86_64 0:1.28-5.el6 tcp_wrappers.x86_640:7.6-57.el6

telnet.x86_64 1:0.17-47.el6_3.1 yajl.x86_64 0:1.0.7-3.el6

Complete!


6、两个节点分别启动群集服务

[root@node1 ~]# service ricci start

[root@node1 ~]# chkconfig ricci on

[root@node1 ~]# chkconfig cman on

[root@node1 ~]# chkconfig rgmanager on


[root@node2 ~]# service ricci start

[root@node2 ~]# chkconfig ricci on

[root@node2 ~]# chkconfig cman on

[root@node2 ~]# chkconfig rgmanager on


7、两个节点分别配置ricci密码

[root@node1 ~]# passwd ricci

New password:

BAD PASSWORD: it is too short

BAD PASSWORD: is too simple

Retype new password:

passwd: all authentication tokens updated successfully.


[root@node2 ~]# passwd ricci

New password:

BAD PASSWORD: it is too short

BAD PASSWORD: is too simple

Retype new password:

passwd: all authentication tokens updated successfully.


8、两个节点分别安装httpd服务,方便后面测试应用的高可用性

[root@node1 ~]# yum -y install httpd


[root@node1 ~]# echo "This is Node1" > /var/www/html/index.html

[root@node2 ~]# yum -y install httpd

[root@node2 ~]# echo "This is Node2" > /var/www/html/index.html

二、群集管理服务器配置

1、在群集管理服务器安装luci软件包

[root@HAmanager ~]#yum -y install luci

Installed:

luci.x86_64 0:0.26.0-48.el6

Dependency Installed:

TurboGears2.noarch 0:2.0.3-4.el6

python-babel.noarch 0:0.9.4-5.1.el6

python-beaker.noarch 0:1.3.1-7.el6

python-cheetah.x86_64 0:2.4.1-1.el6

python-decorator.noarch 0:3.0.1-3.1.el6

python-decoratortools.noarch 0:1.7-4.1.el6

python-formencode.noarch 0:1.2.2-2.1.el6

python-genshi.x86_64 0:0.5.1-7.1.el6

python-mako.noarch 0:0.3.4-1.el6

python-markdown.noarch 0:2.0.1-3.1.el6

python-markupsafe.x86_64 0:0.9.2-4.el6

python-myghty.noarch 0:1.1-11.el6

python-nose.noarch 0:0.10.4-3.1.el6

python-paste.noarch 0:1.7.4-2.el6

python-paste-deploy.noarch 0:1.3.3-2.1.el6

python-paste-script.noarch 0:1.7.3-5.el6_3

python-peak-rules.noarch 0:0.5a1.dev-9.2582.1.el6

python-peak-util-addons.noarch 0:0.6-4.1.el6

python-peak-util-assembler.noarch 0:0.5.1-1.el6

python-peak-util-extremes.noarch 0:1.1-4.1.el6

python-peak-util-symbols.noarch 0:1.0-4.1.el6

python-prioritized-methods.noarch 0:0.2.1-5.1.el6

python-pygments.noarch 0:1.1.1-1.el6

python-pylons.noarch 0:0.9.7-2.el6

python-repoze-tm2.noarch 0:1.0-0.5.a4.el6

python-repoze-what.noarch 0:1.0.8-6.el6

python-repoze-what-pylons.noarch 0:1.0-4.el6

python-repoze-who.noarch 0:1.0.18-1.el6

python-repoze-who-friendlyform.noarch 0:1.0-0.3.b3.el6

python-repoze-who-testutil.noarch 0:1.0-0.4.rc1.el6

python-routes.noarch 0:1.10.3-2.el6

python-setuptools.noarch 0:0.6.10-3.el6

python-sqlalchemy.noarch 0:0.5.5-3.el6_2

python-tempita.noarch 0:0.4-2.el6

python-toscawidgets.noarch 0:0.9.8-1.el6

python-transaction.noarch 0:1.0.1-1.el6

python-turbojson.noarch 0:1.2.1-8.1.el6

python-weberror.noarch 0:0.10.2-2.el6

python-webflash.noarch 0:0.1-0.2.a9.el6

python-webhelpers.noarch 0:0.6.4-4.el6

python-webob.noarch 0:0.9.6.1-3.el6

python-webtest.noarch 0:1.2-2.el6

python-zope-filesystem.x86_64 0:1-5.el6

python-zope-interface.x86_64 0:3.5.2-2.1.el6

python-zope-sqlalchemy.noarch 0:0.4-3.el6

Complete!

[root@HAmanager ~]#


2、启动luci服务

[root@HAmanager ~]# service luci start

Adding following auto-detected host IDs (IP addresses/domain names), corresponding to `HAmanager.localdomain' address, to the configuration of self-managed certificate `/var/lib/luci/etc/cacert.config' (you can change them by editing `/var/lib/luci/etc/cacert.config', removing the generated certificate `/var/lib/luci/certs/host.pem' and restarting luci):

(none suitable found, you can still do it manually as mentioned above)


Generating a 2048 bit RSA private key

writing new private key to '/var/lib/luci/certs/host.pem'

Starting saslauthd: [ OK ]

Start luci...

Point your web browser to https://HAmanager.localdomain:8084 (or equivalent) to access luci

[root@HAmanager ~]# chkconfig luci on

江健龙的技术博客http://jiangjianlong.blog.51cto.com/3735273/1931499


三、创建和配置群集

1、使用浏览器访问HA的web管理界面https://192.168.10.150:8084


2、创建群集并添加节点至群集中


3、添加vCenter为fence设备


4、查找节点的虚拟机UUID

[root@node1 ~]# fence_vmware_soap -a 192.168.10.91 -z -l administrator@vsphere.local -p P@ssw0rd -o list

node1,564df192-7755-9cd6-8a8b-45d6d74eabbb

node2,564df4ed-cda1-6383-bbf5-f99807416184


5、两个节点添加fence方法和实例




6、查看fence设备状态

[root@node1 ~]# fence_vmware_soap -a 192.168.10.91 -z -l administrator@vsphere.local -p P@ssw0rd -o status

Status: ON


7、测试fence设备

[root@node2 ~]# fence_check

fence_check run at Tue May 23 09:41:30 CST 2017 pid: 3455

Testing node1.localdomain method 1: success

Testing node2.localdomain method 1: success


8、创建故障域


9、添加群集资源,分别添加IP地址和脚本为群集资源


10、创建群集服务组并添加已有的资源


11、配置仲裁盘,在HAmanager服务器安装iSCSI target服务并创建一块100M的共享磁盘给两个节点

[root@HAmanager ~]#yum install scsi-target-utils -y

[root@HAmanager ~]#dd if=/dev/zero of=/iSCSIdisk/100m.img bs=1M seek=100 count=0

[root@HAmanager ~]#vi /etc/tgt/targets.conf

backing-store /iSCSIdisk/100m.img

initiator-address 192.168.10.104 #for node1

initiator-address 192.168.10.105 #for node2

[root@HAmanager ~]#service tgtd start

[root@HAmanager ~]#chkconfig tgtd on

[root@HAmanager ~]#tgt-admin -show

Target 1: iqn.2016-08.disk.rh7:disk100m

System information:

Driver: iscsi

State: ready

I_T nexus information:

LUN information:

LUN: 0

Type: controller

SCSI ID: IET 00010000

SCSI SN: beaf10

Size: 0 MB, Block size: 1

Online: Yes

Removable media: No

Prevent removal: No

Readonly: No

Backing store type: null

Backing store path: None

Backing store flags:

LUN: 1

Type: disk

SCSI ID: IET 00010001

SCSI SN: beaf11

Size: 105 MB, Block size: 512

Online: Yes

Removable media: No

Prevent removal: No

Readonly: No

Backing store type: rdwr

Backing store path: /sharedisk/100m.img

Backing store flags:

Account information:

ACL information:

192.168.10.104

192.168.10.105

[root@HAmanager ~]#


12、两个节点安装iscsi-initiator-utils并登录iscsi目标

[root@node1 ~]# yum install iscsi-initiator-utils

[root@node1 ~]# chkconfig iscsid on

[root@node1 ~]# iscsiadm -m discovery -t sendtargets -p 192.168.10.150

[root@node1 ~]# iscsiadm -m node

[root@node1 ~]# iscsiadm -m node -T iqn.2016-08.disk.rh7:disk100m --login


[root@node2 ~]# yum install iscsi-initiator-utils

[root@node2 ~]# chkconfig iscsid on

[root@node2 ~]# iscsiadm -m discovery -t sendtargets -p 192.168.10.150

[root@node2 ~]# iscsiadm -m node

[root@node2 ~]# iscsiadm -m node -T iqn.2016-08.disk.rh7:disk100m --login


13、在节点一将共享磁盘/dev/sdb创建分区sdb1

[root@node1 ~]# fdisk /dev/sdb

然后创建成sdb1

[root@node1 ~]# partprobe /dev/sdb1


14、在节点一将sdb1创建成仲裁盘

[root@node1 ~]# mkqdisk -c /dev/sdb1 -l testqdisk

mkqdisk v3.0.12.1

Writing new quorum disk label 'testqdisk' to /dev/sdb1.

WARNING: About to destroy all data on /dev/sdb1; proceed [N/y] ? y

Initializing status block for node 1...

Initializing status block for node 2...

Initializing status block for node 3...

Initializing status block for node 4...

Initializing status block for node 5...

Initializing status block for node 6...

Initializing status block for node 7...

Initializing status block for node 8...

Initializing status block for node 9...

Initializing status block for node 10...

Initializing status block for node 11...

Initializing status block for node 12...

Initializing status block for node 13...

Initializing status block for node 14...

Initializing status block for node 15...

Initializing status block for node 16...

[root@node1 ~]#


[root@node1 ~]# mkqdisk -L

mkqdisk v3.0.12.1

/dev/block/8:17:

/dev/disk/by-id/scsi-1IET_00010001-part1:

/dev/disk/by-path/ip-192.168.10.150:3260-iscsi-iqn.2016-08.disk.rh7:disk100m-lun-1-part1:

/dev/sdb1:

Magic: eb7a62c2

Label: testqdisk

Created: Mon May 22 22:52:01 2017

Host: node1.localdomain

Kernel Sector Size: 512

Recorded Sector Size: 512


[root@node1 ~]#


15、在节点二查看仲裁盘,也正常识别

[root@node2 ~]# partprobe /dev/sdb1

[root@node2 ~]# mkqdisk -L

mkqdisk v3.0.12.1

/dev/block/8:17:

/dev/disk/by-id/scsi-1IET_00010001-part1:

/dev/disk/by-path/ip-192.168.10.150:3260-iscsi-iqn.2016-08.disk.rh7:disk100m-lun-1-part1:

/dev/sdb1:

Magic: eb7a62c2

Label: testqdisk

Created: Mon May 22 22:52:01 2017

Host: node1.localdomain

Kernel Sector Size: 512

Recorded Sector Size: 512


16、配置群集使用该仲裁盘


17、重启群集,使仲裁盘生效

[root@node1 ~]# ccs -h node1 --stopall

node1 password:

Stopped node2.localdomain

Stopped node1.localdomain


[root@node1 ~]# ccs -h node1 --startall

Started node2.localdomain

Started node1.localdomain

[root@node1 ~]#


18、查看群集状态

[root@node1 ~]# clustat

Cluster Status for TestCluster2 @ Mon May 22 23:48:27 2017

Member Status: Quorate


Member Name ID Status

------ ---- ---- ------

node1.localdomain 1 Online, Local, rgmanager

node2.localdomain 2 Online, rgmanager

/dev/block/8:17 0 Online, Quorum Disk


Service Name Owner (Last) State

------- ---- ----- ------ -----

service:TestServGrp node1.localdomain started

[root@node1 ~]#


19、查看群集节点状态

[root@node1 ~]# ccs_tool lsnode

Cluster name: icpl_cluster, config_version: 21

Nodename Votes Nodeid Fencetype

node1.localdomain 1 1 vcenter_fence

node2.localdomain 1 2 vcenter_fence


20、查看群集节点同步状态

[root@node1 ~]# ccs -h node1 --checkconf

All nodes in sync.


21、使用群集IP访问web服务

江健龙的技术博客http://jiangjianlong.blog.51cto.com/3735273/1931499


四、群集故障转移测试

1、关闭主节点,故障自动转移功能正常

[root@node1 ~]#poweroff

[root@node1 ~]#tail-f /var/log/messages

May 23 10:29:26 node1 modclusterd: shutdown succeeded

May 23 10:29:26 node1 rgmanager[2125]: Shutting down

May 23 10:29:26 node1 rgmanager[2125]: Shutting down

May 23 10:29:26 node1 rgmanager[2125]:Stopping service service:TestServGrp

May 23 10:29:27 node1 rgmanager[2125]: [ip] Removing IPv4 address 192.168.10.103/24 from eth0

May 23 10:29:36 node1rgmanager[2125]: Service service:TestServGrp is stopped

May 23 10:29:36 node1 rgmanager[2125]: Disconnecting from CMAN

May 23 10:29:52 node1 rgmanager[2125]: Exiting

May 23 10:29:53 node1 ricci:shutdown succeeded

May 23 10:29:54 node1 oddjobd: oddjobd shutdown succeeded

May 23 10:29:54 node1 saslauthd[2315]:server_exit : master exited: 2315



[root@node2 ~]#tail-f /var/log/messages

May 23 10:29:45 node2 rgmanager[2130]: Member 1 shutting down

May 23 10:29:45 node2 rgmanager[2130]: Starting stopped service service:TestServGrp

May 23 10:29:45 node2 rgmanager[5688]: [ip] Adding IPv4 address 192.168.10.103/24 to eth0

May 23 10:29:49 node2 rgmanager[2130]: Service service:TestServGrp started

May 23 10:30:06 node2 qdiskd[1480]: Node 1 shutdown

May 23 10:30:06 node2 corosync[1437]: [QUORUM Members[1]: 2

May 23 10:30:06 node2 corosync[1437]: [TOTEM ] A processor joined or left the membership

and a new membership was formed.

May 23 10:30:06 node2 corosync[1437]: [CPG ] chosen downlist: sender r(0) ip (192.168.10.105) :

members(old:2 left:1)

May 23 10:30:06 node2 corosync[1437]: [MAIN ] Completed service synchronization, ready to

provide service

May 23 10:30:06 node2 kernel: dlm: closing connection to node 1

May 23 10:30:06 node2 qdiskd[1480]: Assuming master role


[root@node2 ~]# clustat

Cluster Status for TestCluster2 @ Mon May 22 23:48:27 2017

Member Status: Quorate


Member Name ID Status

------ ---- ---- ------

node1.localdomain 1 Online, Local, rgmanager

node2.localdomain 2 Online, rgmanager

/dev/block/8:17 0 Online, Quorum Disk


Service Name Owner (Last) State

------- ---- ----- ------ -----

service:TestServGrp node2.localdomain started

[root@node2 ~]#


2、停掉主节点应用服务,故障自动转移功能正常

[root@node2 ~]# /etc/init.d/httpd stop

[root@node2 ~]#tail-f /var/log/messages

May 23 11:14:02 node2 rgmanager[11264]: [script] Executing /etc/init.d/httpd status

May 23 11:14:02 node2 rgmanager[11289]: [script] script:icpl: status of /etc/init.d/httpd failed (returned 3)

May 23 11:14:02 node2 rgmanager[2127]: status on script "httpd" returned 1 (generic error)

May 23 11:14:02 node2 rgmanager[2127]: Stopping service service:TestServGrp

May 23 11:14:03 node2 rgmanager[11320]: [script] Executing /etc/init.d/httpd stop

May 23 11:14:03 node2 rgmanager[11384]: [ip] Removing IPv4 address 192.168.10.103/24 from eth0

May 23 11:14:08 node2 ricci[11416]: Executing '/usr/bin/virsh nodeinfo'

May 23 11:14:08 node2 ricci[11418]: Executing '/usr/libexec/ricci/ricci-worker -f /var/lib/ricci/queue/2116732044'

May 23 11:14:09 node2 ricci[11422]: Executing '/usr/libexec/ricci/ricci-worker -f /var/lib/ricci/queue/1193918332'

May 23 11:14:13 node2 rgmanager[2127]: Service service:TestServGrp is recovering

May 23 11:14:17 node2 rgmanager[2127]: Service service:TestServGrp is now running on member 1


[root@node1 ~]#tail-f /var/log/messages

May 23 11:14:20 node1 rgmanager[2130]: Recovering failed service service:TestServGrp

May 23 11:14:20 node1 rgmanager[13006]: [ip] Adding IPv4 address 192.168.10.103/24 to eth0

May 23 11:14:24 node1 rgmanager[13092]: [script] Executing /etc/init.d/httpd start

May 23 11:14:24 node1 rgmanager[2130]: Service service:TestServGrp started

May 23 11:14:58 node1 rgmanager[13280]: [script] Executing /etc/init.d/httpd status



[root@node1 ~]# clustat

Cluster Status for TestCluster2 @ Mon May 22 23:48:27 2017

Member Status: Quorate


Member Name ID Status

------ ---- ---- ------

node1.localdomain 1 Online, Local, rgmanager

node2.localdomain 2 Online, rgmanager

/dev/block/8:17 0 Online, Quorum Disk


Service Name Owner (Last) State

------- ---- ----- ------ -----

service:TestServGrp node1.localdomain started

[root@node1 ~]#


3、停掉主节点网络服务,故障自动转移功能正常

[root@node1 ~]#service network stop


[root@node2 ~]#tail-f /var/log/messages

May 23 22:11:16 node2 qdiskd[1480]: Assuming master role

May 23 22:11:17 node2 qdiskd[1480]: Writing eviction notice for node 1

May 23 22:11:17 node2 corosync[1437]: [TOTEM ] A processor failed, forming new configuration.

May 23 22:11:18 node2 qdiskd[1480]: Node 1 evicted

May 23 22:11:19 node2 corosync[1437]: [QUORUM] Members[1]: 2

May 23 22:11:19 node2 corosync[1437]: [TOTEM ] A processor joined or left the membership and a new membership was formed.

May 23 22:11:19 node2 corosync[1437]: [CPG ] chosen downlist: sender r(0) ip(192.168.10.105) ; members(old:2 left:1)

May 23 22:11:19 node2 corosync[1437]: [MAIN ] Completed service synchronization, ready to provide service.

May 23 22:11:19 node2 kernel: dlm: closing connection to node 1

May 23 22:11:19 node2 rgmanager[2131]: State change: node1.localdomain DOWN

May 23 22:11:19 node2 fenced[1652]: fencing node1.localdomain

May 23 22:11:58 node2 fenced[1652]: fence node1.localdomain success

May 23 22:11:59 node2 rgmanager[2131]: Taking over service service:TestServGrp from down member node1.localdomain

May 23 22:11:59 node2 rgmanager[6145]: [ip] Adding IPv4 address 192.168.10.103/24 to eth0

May 23 22:12:03 node2 rgmanager[6234]: [script] Executing /etc/init.d/httpd start

May 23 22:12:03 node2 rgmanager[2131]: Service service:TestServGrp started

May 23 22:12:35 node2 corosync[1437]: [TOTEM ] A processor joined or left the membership and a new membership was formed.

May 23 22:12:35 node2 corosync[1437]: [QUORUM] Members[2]: 1 2

May 23 22:12:35 node2 corosync[1437]: [QUORUM] Members[2]: 1 2

May 23 22:12:35 node2 corosync[1437]: [CPG ] chosen downlist: sender r(0) ip(192.168.10.105) ; members(old:1 left:0)

May 23 22:12:35 node2 corosync[1437]: [MAIN ] Completed service synchronization, ready to provide service.

May 23 22:12:41 node2 rgmanager[6425]: [script] Executing /etc/init.d/httpd status

May 23 22:12:43 node2 qdiskd[1480]: Node 1 shutdown

May 23 22:12:55 node2 kernel: dlm: got connection from 1

May 23 22:13:08 node2 rgmanager[2131]: State change: node1.localdomain UP


[root@node2 ~]# clustat

Cluster Status for TestCluster2 @ Mon May 22 23:48:27 2017

Member Status: Quorate


Member Name ID Status

------ ---- ---- ------

node1.localdomain 1 Online, Local, rgmanager

node2.localdomain 2 Online, rgmanager

/dev/block/8:17 0 Online, Quorum Disk


Service Name Owner (Last) State

------- ---- ----- ------ -----

service:TestServGrp node2.localdomain started

[root@node2 ~]#

附:RHCS名词解释

1 分布式集群管理器(CMAN,Cluster Manager

管理集群成员,了解成员之间的运行状态。

2 分布式锁管理器(DLM,Distributed Lock Manager

每一个节点都运行了一个后台进程DLM,当用记操作一个元数据时,会通知其它节点,只能读取这个元数据。

3 配置文件管理(CCS,Cluster Configuration System)

主要用于集群配置文件管理,用于配置文件的同步。每个节点运行了CSS后台进程。当发现配置文件(/etc/cluster/cluster.conf)变化后,马上将此变化传播到其它节点上去。

4.fence设备(fence)

工作原理:当主机异常,务机会调用fence设备,然后将异常主机重启,当fence设备操作成功后,返回信息给备机,备机接到fence设备的消息后,接管主机的服务和资源。

5、Conga集群管理软件:

Conga由两部分组成:luci和ricci,luci是跑在集群管理服务器上的服务,而ricci则是跑在各集群节点上的服务,luci也可以装在节点上。集群的管理和配置由这两个服务进行通信,可以使用Conga的web界面来管理RHCS集群。


6、高可用性服务管理(rgmanager)

提供节点服务监控和服务故障转移功能,当一个节点服务出现故障时,将服务转移到另一个健康节点。

0