操作目的:
基于Corosync和NFS服务器实现MySQL的高可用
规划:
准备工作:根据规划图,设置三台主机的主机名和地址
(这里不再给出具体过程)
一、配置nfs服务器
#fdisk /dev/sda
创建一个大小为20G 类型为8e的分区 本文为/dev/sda5
#partprobe /dev/sda
#pvcreate /dev/sda5
#vgcreate myvg /dev/sda5
#lvcreate -L 10G -n mydata myvg
#mke2fs -j -L MYDATA /dev/myvg/mydata
#mkdir /mydata
#vim /etc/fstab
添加
LABEL=MYDATA /mydata ext3 defaults 0 0
#mount -a
添加用户
# groupadd -g 306 -r mysql
# useradd -g mysql -r -u 306 -s /sbin/nologin mysql
# id mysql
uid=306(mysql) gid=306(mysql) groups=306(mysql) context=root:system_r:unconfined_t:SystemLow-SystemHigh
让mysql用户对/mydata目录有读写的权限
# chown -R mysql:mysql /mydata/
通过nfs共享/mydata
#vim /etc/exports
/mydata 172.16.220.11(rw,no_root_squash) 172.16.220.12(rw,no_root_squash)
#service nfs start
# showmount -e localhost
Export list for localhost:
/mydata 172.16.220.12,172.16.220.11
二、配置mysql的高可用
1 配置node1 node2的时间同步和ssh互联
node1:
#hwclock -s
#ssh-keygen -i rsa
#vim /etc/hosts
172.16.220.11 node1
172.16.220.12 node2
# ssh-copy-id -i .ssh/id_rsa.pub root@node2
# ssh node2 'ifconfig'
node2:
#hwclock -s
#ssh-keygen -i rsa
#vim /etc/hosts
172.16.220.11 node1
172.16.220.12 node2
# ssh-copy-id -i .ssh/id_rsa.pub root@node1
# ssh node1 'ifconfig'
2 添加用户
node1 node2:
# groupadd -r -g 306 mysql
# useradd -g mysql -u 306 -r mysql
# mkdir /mydata
# mount -t nfs 172.16.220.22:/mydata /mydata/
# ls /mydata/
lost+found
验证mysql用户是否对/mydata目录有写的权限
# su - mysql
su: warning: cannot change directory to /home/mysql: No such file or directory
-bash-3.2$ cd /mydata
-bash-3.2$ mkdir data //创建这个目录,作为mysql的数据目录
-bash-3.2$ ls
data lost+found
3 安装mysql
node1 :
安装 mysql-5.5.22-linux2.6-i686.tar.gz
# tar xf mysql-5.5.22-linux2.6-i686.tar.gz -C /usr/local/
# cd /usr/local/
# ln -sv mysql-5.5.22-linux2.6-i686 mysql
#cd mysql
初始化mysql
#chown -R mysql:mysql .
# scripts/mysql_install_db --user=mysql --datadir=/mydata/data/
提供配置文件
# cp /usr/local/mysql/support-files/my-large.cnf /etc/my.cnf
#vim /etc/my.cnf
[mysqld]中修改
thread_concurrency = 2
datadir = /mydata/data
提供脚本
# cp /usr/local/mysql/support-files/mysql.server /etc/init.d/mysqld
# chkconfig --add mysqld
# service mysqld start
Starting MySQL........ [ OK ]
# /usr/local/mysql/bin/mysql 连上mysql
停止服务
#service mysqld stop
# chkconfig mysqld off
# chkconfig --list mysqld //保证都是off
node2:
安装 mysql-5.5.22-linux2.6-i686.tar.gz
# tar xf mysql-5.5.22-linux2.6-i686.tar.gz -C /usr/local/
# cd /usr/local/
# ln -sv mysql-5.5.22-linux2.6-i686 mysql
#cd mysql
#chown -R root:mysql
# cp /usr/local/mysql/support-files/my-large.cnf /etc/my.cnf
#vim /etc/my.cnf
[mysqld]中修改
thread_concurrency = 2
datadir = /mydata/data
# cp /usr/local/mysql/support-files/mysql.server /etc/init.d/mysqld
# chkconfig --add mysqld
# service mysqld start
Starting MySQL........ [ OK ]
# /usr/local/mysql/bin/mysql 连上mysql
#service mysqld stop
# chkconfig mysqld off
# chkconfig --list mysqld //保证都是off
到此两个mysql服务器可以实现nfs的共享存储了
4 添加集群服务
准备的包:
corosync-1.2.7-1.1.el5.i386.rpm
corosynclib-1.2.7-1.1.el5.i386.rpm
pacemaker-1.1.5-1.1.el5.i386.rpm
pacemaker-libs-1.1.5-1.1.el5.i386.rpm
heartbeat-3.0.3-2.3.el5.i386.rpm
heartbeat-libs-3.0.3-2.3.el5.i386.rpm
cluster-glue-1.0.6-1.6.el5.i386.rpm
cluster-glue-libs-1.0.6-1.6.el5.i386.rpm
perl-TimeDate-1.16-5.el5.noarch.rpm
libesmtp-1.0.4-5.el5.i386.rpm
resource-agents-1.0.4-1.1.el5.i386.rpm
下载到node1 node2 的/root目录下
4.1安装corosync
node1 node2:
#yum -y --nogpgcheck localinstall *.rpm
4.2 corosync的配置文件
node1:
# cd /etc/corosync/corosync.conf.example /etc/corosync/corosync.conf
#vim /etc/corosync.conf
修改内容如下:
secauth: on
bindnetaddr: 172.16.0.0
mcastaddr: 226.99.1.1
to_logfile: yes
to_syslog: no
接着编辑添加如下内容:
service {
ver: 0
name: pacemaker
use_mgmtd: yes
}
aisexec {
user: root
group: root
}
生成节点间通信时用到的认证密钥文件:
# corosync-keygen //会自动的在/etc/corosync目录下生成一个key文件,是二进制格式的,权限为400
将corosync和authkey复制至node2:
# scp -p authkey corosync.conf node2:/etc/corosync/
分别为两个节点创建corosync生成的日志所在的目录:
# mkdir /var/log/cluster
# ssh node2 'mkdir /var/log/cluster'
启动服务
# service corosync start
# ssh node2 '/etc/init.d/corosync start'
查看工作状态
# crm status
============
Last updated: Tue Aug 21 09:07:33 2012
Stack: openais
Current DC: node1 - partition with quorum
Version: 1.1.5-1.1.el5-01e86afaaa6d4a8c4836f68df80ababd6ca3902f
2 Nodes configured, 2 expected votes
0 Resources configured.
============
Online: [ node1 node2 ]
三、配置集群资源
node1:(集群资源的配置,在一个节点上完成即可)
1 禁用stonith
# crm configure property stonith-enabled=false
2 修改忽略quorum不能满足的集群状态检查:
# crm configure property no-quorum-policy=ignore
3 为资源指定默认黏性值:
# crm configure rsc_defaults resource-stickiness=100
4 配置资源(资源的添加,在哪个节点上都可以)
1)添加访问mysql的Ip
# crm configure primitive myip ocf:heartbeat:IPaddr params ip=172.16.220.21
2)添加挂载目录的资源
现在node1 node2上卸载挂载的目录
#umount /mydata
#crm
crm(live)configure# primitive mynfs ocf:heartbeat:Filesystem params device="172.16.220.22:/mydata" directory="/mydata" fstype="nfs" op start timeout=60s op stop timeout=60s
资源的工作状态
# crm status
============
Last updated: Tue Aug 21 09:46:30 2012
Stack: openais
Current DC: node1 - partition with quorum
Version: 1.1.5-1.1.el5-01e86afaaa6d4a8c4836f68df80ababd6ca3902f
2 Nodes configured, 2 expected votes
2 Resources configured.
============
Online: [ node1 node2 ]
myip (ocf::heartbeat:IPaddr): Started node1
mynfs (ocf::heartbeat:Filesystem): Started node2
此时可以测试:
mynfs资源在node2上,查看即可
# ls /mydata/data
ibdata1 ib_logfile1 mysql-bin.000001 mysql-bin.000003 node1.err performance_schema
ib_logfile0 mysql mysql-bin.000002 mysql-bin.index node2.err test
有数据文件了,此时在node1上是没有/data目录的
让node2变为被动节点
#crm node standby
# crm status
============
Last updated: Tue Aug 21 09:51:52 2012
Stack: openais
Current DC: node1 - partition with quorum
Version: 1.1.5-1.1.el5-01e86afaaa6d4a8c4836f68df80ababd6ca3902f
2 Nodes configured, 2 expected votes
2 Resources configured.
============
Node node2: standby
Online: [ node1 ]
myip (ocf::heartbeat:IPaddr): Started node1
mynfs (ocf::heartbeat:Filesystem): Started node1
此时node1上查看
#ls /mydata/data
ibdata1 ib_logfile1 mysql-bin.000001 mysql-bin.000003 node1.err performance_schema
ib_logfile0 mysql mysql-bin.000002 mysql-bin.index node2.err test
3)添加mysqld服务
(mysqld 一定要和myip mynfs在一起,并且mysqld要晚于mynfs和myip)
# crm
crm(live)# configure
crm(live)configure# primitive mysqld lsb:mysqld
定义排列约束
crm(live)configure# colocation mysqld_with_mynfs_with_myip inf: mysqld mynfs myip
定义顺序约束
crm(live)configure# order mysqld_after_myip mandatory: myip mysqld
crm(live)configure# order mysqld_after_mynfs mandatory: mynfs mysqld:start
定义完顺序之后,查看xml
crm(live)configure# show xml 要确保是order是如下内容
<rsc_order first="mynfs" id="mysqld_after_mynfs" score="INFINITY" then="mysqld" then-action="start"/>
<rsc_order first="myip" id="mysqld_after_myip" score="INFINITY" then="mysqld"/>
提交配置
crm(live)configure# commit
#crm status
Online: [ node1 node2 ]
myip (ocf::heartbeat:IPaddr): Started node1
mynfs (ocf::heartbeat:Filesystem): Started node1
mysqld (lsb:mysqld): Started node1
此时可以测试以下,mysqld在node1,就在node1上测试
#/usr/local/mysql/bin/mysql
mysql> show databases;
+--------------------+
| Database |
+--------------------+
| information_schema |
| mysql |
| performance_schema |
| test |
+--------------------+ 证明资源生效了
建立一个远程连接的用户
mysql> grant all on *.* to root@'%' identified by 'redhat';
mysql> flush privileges;
远程连接172.16.220.21的mysql服务如图
操作过程中遇到的问题
1 添加mynfs资源时
Failed actions:
mynfs_start_0 (node=node1, call=5, rc=5, status=complete): not installed
mynfs_start_0 (node=node2, call=4, rc=5, status=complete): not installed
第一次使用
# crm configure primitive mynfs ocf:heartbeat:Filesystem params device="172.16.220.22/mydata" directory="/mydata" fstype="nfs" op start timeout=60s op stop timeout=60s
来添加的,出现了如上错误
第二次使用
#crm
crm(live)configure# primitive mynfs ocf:heartbeat:Filesystem params device="172.16.220.22:/mydata" directory="/mydata" fstype="nfs" op start timeout=60s op stop timeout=60s
工作正常了 诡异啊
2 添加进了mgmtd服务,但是5560服务还是没有启动。解决一下
3 可以远程连接所在节点的mysql,但是不能通过vip连接,怎么原因?
解决:地址冲突,其他主机占用了这个vip(一定要保证vip,虚拟ip没有被其他主机占用)
©著作权归作者所有:来自51CTO博客作者不凡一夏的原创作品,如需转载,请注明出处,否则将追究法律责任
corosyncLinux集群
共同學習,寫下你的評論
評論加載中...
作者其他優質文章