MySQL数据库高可用MHA
MHA概述
1、简介
主要是为了解决msater只有一台,若master出现故障时就使用MHA解决单点故障问题
一套优秀的MySQL高可用环境下故障切换和主从复制的软件
MySQL故障过程中,MHA能做到0-30秒内自动完成故障切换
2、MHA的组成(性能、组件)
MHA Manager (管理节点)
可以借助于node组件,实现健康检查(ssh、主从复制)、控制故障切换
MHA Node 组件(数据节点)
部署在所有MySQL服务器内;主要作用:
定时汇报给主manager节点,mysql服务的状态
监控本地mysql服务功能
主从复制架构:主从复制(一主多从) ; MHA架构(多主多从) ; MMM架构(双主多从)
3、特点
自动故障切换过程中,MHA试图从宕机的主服务器上保存二进制日志,最大程度的保证数据不丢失
自动故障切换过程的本质
故障切换(本质VIP漂移)
自动(使用脚本触发自动)
使用半同步复制,可以大大降低数据丢失的风险
搭建MySQL-MHA实验
设备准备:一台manager,一台master和两台slave做主从复制
修改三台mysql服务器的配置文件,开启日志,并做mysql命令和节点恢复命令的软连接,然后其后台启动
master:
[root@master ~]# vim /etc/my.cnf
[mysqld]
server-id = 10
log_bin = master-bin
log-slave-update = true
[root@master ~]# ln -s /usr/local/mysql/bin/mysql /usr/sbin
[root@master ~]# ln -s /usr/local/mysql/bin/mysqlbinlog /usr/sbin
slave1:
[root@slave1 ~]# vim /etc/my.cnf
server-id = 12
log_bin = master-bin
relay-log = relay-log-bin
relay-log-index = slave-relay-bin.index
[root@slave1 ~]# ln -s /usr/local/mysql/bin/mysql /usr/sbin
[root@slave1 ~]# ln -s /usr/local/mysql/bin/mysqlbinlog /usr/sbin
slave2:
[root@slave2 opt]# vim /etc/my.cnf
server-id = 13
log_bin = master-bin
relay-log = relay-log-bin
relay-log-index = slave-log-bin.index
[root@slave2 opt]# ln -s /usr/local/mysql/bin/mysql /usr/sbin/
[root@slave2 opt]# ln -s /usr/local/mysql/bin/mysqlbinlog /usr/sbin
在所有数据库节点上授权两个用户,一个是从库同步使用,另一个是manager使用
grant replication slave on *.* to 'myslave'@'192.168.8.%' identified by '123456';
grant all privileges on *.* to 'mha'@'192.168.8.%' identified by 'manager';
grant all privileges on *.* to 'mha'@'master' identified by 'manager';
grant all privileges on *.* to 'mha'@'slave1' identified by 'manager';
grant all privileges on *.* to 'mha'@'slave2' identified by 'manager';
mster:
mysql> show master status;
+-------------------+----------+--------------+------------------+-------------------+
| File | Position | Binlog_Do_DB | Binlog_Ignore_DB | Executed_Gtid_Set |
+-------------------+----------+--------------+------------------+-------------------+
| master-bin.000003 | 154 | | | |
+-------------------+----------+--------------+------------------+-------------------+
1 row in set (0.00 sec)
slave1:
mysql> change master to master_host='192.168.8.151',master_user='myslave',master_password='123456',master_log_file='masster-bin.000003',master_log_pos=154;
Query OK, 0 rows affected, 2 warnings (0.01 sec)
mysql> start slave;
Query OK, 0 rows affected (0.00 sec)
mysql> show slave status \G
…………………………
Slave_IO_Running: Yes
Slave_SQL_Running: Yes
…………………………
slave2:
mysql> change master to master_host='192.168.8.151',master_user='myslave',master_password='123456',master_log_file='masster-bin.000003',master_log_pos=154;
Query OK, 0 rows affected, 2 warnings (0.01 sec)
mysql> start slave;
Query OK, 0 rows affected (0.00 sec)
mysql> show slave status \G
…………………………
Slave_IO_Running: Yes
Slave_SQL_Running: Yes
…………………………
在所有的服务器上安装MHA依赖环境,在所有服务器上安装node组件
yum install epel-release --nogpgcheck -y
yum install -y perl-DBD-MySQL \
perl-Config-Tiny \
perl-Log-Dispatch \
perl-Parallel-ForkManager \
perl-ExtUtils-CBuilder \
perl-ExtUtils-MakeMaker \
perl-CPAN
tar zxvf /opt/mha4mysql-node-0.57.tar.gz
cd mha4mysql-node-0.57
perl Makefile.PL
make && make install
在MHA节点上安装manager
tar zxvf /opt/mha4mysql-manager-0.57.tar.gz
cd mha4mysql-manager-0.57
perl Makefile.PL
make && make install
##安装后在/usr/local/bin目录下会生成以下几个工具
[root@localhost mha4mysql-manager-0.57]# cd /usr/local/bin
[root@master bin]l#ls
在所有服务器上配置无密码认证
ssh-keygen -t rsa
ssh-copy-id 192.168.118.200
ssh-copy-id 192.168.118.100
ssh-copy-id 192.168.118.50
ssh-copy-id 192.168.118.55
第一次配置需要在master节点上手动开启虚拟IP
/sbin/ifconfig ens33:1 192.168.118.100
配置MHA节点
在MHA节点上复制相关脚本到/usr/local/bin目录
[root@manager ~]# cp -ra /opt/mha4mysql-manager-0.57/samples/scripts/ /usr/local/bin
[root@manager ~]# cd /usr/local/bin/scripts/
[root@manager scripts]# ls
master_ip_failover master_ip_online_change power_manager send_report
[root@manager scripts]# vim /usr/bin/master_ip_failover #删除原有模板,添加内容
#!/usr/bin/env perl
use strict;
use warnings FATAL => 'all';
use Getopt::Long;
my (
$command, $ssh_user, $orig_master_host, $orig_master_ip,
$orig_master_port, $new_master_host, $new_master_ip, $new_master_port
);
my $vip = '192.168.8.151'; #浮动IP
my $brdc = '192.168.8.255'; #广播地址
my $ifdev = 'ens33'; #使用的网卡为ens33
my $key = '1'; #国际序列号
my $ssh_start_vip = "/sbin/ifconfig ens33:$key $vip"; #使用ifoconfig命令将其启动,同时设置浮动地址
my $ssh_stop_vip = "/sbin/ifconfig ens33:$key down"; #可以使用ifconfig命令将其down掉(关闭)
my $exit_code = 0; #正常退出(返回状态码)
#my $ssh_start_vip = "/usr/sbin/ip addr add $vip/24 brd $brdc dev $ifdev label $ifdev:$key;/usr/sbin/arping -q -A -c 1 -I $ifdev $vip;iptables -F;";
#my $ssh_stop_vip = "/usr/sbin/ip addr del $vip/24 dev $ifdev label $ifdev:$key";
GetOptions(
'command=s' => \$command,
'ssh_user=s' => \$ssh_user,
'orig_master_host=s' => \$orig_master_host,
'orig_master_ip=s' => \$orig_master_ip,
'orig_master_port=i' => \$orig_master_port,
'new_master_host=s' => \$new_master_host,
'new_master_ip=s' => \$new_master_ip,
'new_master_port=i' => \$new_master_port,
);
exit &main();
sub main {
print "\n\nIN SCRIPT TEST====$ssh_stop_vip==$ssh_start_vip===\n\n";
if ( $command eq "stop" || $command eq "stopssh" ) {
my $exit_code = 1;
eval {
print "Disabling the VIP on old master: $orig_master_host \n";
&stop_vip();
$exit_code = 0;
};
if ($@) {
warn "Got Error: $@\n";
exit $exit_code;
}
exit $exit_code;
}
elsif ( $command eq "start" ) {
my $exit_code = 10;
eval {
print "Enabling the VIP - $vip on the new master - $new_master_host \n";
&start_vip();
$exit_code = 0;
};
if ($@) {
warn $@;
exit $exit_code;
}
exit $exit_code;
}
elsif ( $command eq "status" ) {
print "Checking the Status of the script.. OK \n";
exit 0;
}
else {
&usage();
exit 1;
}
}
sub start_vip() {
`ssh $ssh_user\@$new_master_host \" $ssh_start_vip \"`;
}
# A simple system call that disable the VIP on the old_master
sub stop_vip() {
`ssh $ssh_user\@$orig_master_host \" $ssh_stop_vip \"`;
}
sub usage {
print
"Usage: master_ip_failover --command=start|stop|stopssh|status --orig_master_host=host --orig_master_ip=ip --orig_master_port=port --new_master_host=host --new_master_ip=ip --new_master_port=port\n";
}
创建MHA软件目录并拷贝配置文件
mkdir /etc/masterha
cp /root/mha4mysql-manager-0.57/samples/conf/app1.cnf /etc/masterha/
vim /etc/masterha/app1.cnf
[server default]
#日志文件
manager_log=/var/log/masterha/app1/manager.log
#工作目录
manager_workdir=/var/log/masterha/app1
#二进制文件
master_binlog_dir=/usr/local/mysql/data ##这里需要跟master配置文件内的datadir的路径保持一致;不一致无法保存同步二进制文件
#故障转移切换的工具
master_ip_failover_script=/usr/local/bin/master_ip_failover
#在线切换VIP工具管理
master_ip_online_change_script=/usr/local/bin/master_ip_online_change
#以下是密码账号的管理配置
password=manager ##此处密码为之前创建监控用户的密码
user=mha ###设置监控用户
ping_interval=1 ##设置监控主库,发送ping包的事件间隔,默认为3秒,尝试三次没有回应的时候自动进性failove
remote_workdir=/tmp ##设置远端的MySQL在发生切换时binlog的保存位置
repl_password=123456 ##设置复制用户的密码;从和主授权的密码
repl_user=myslave ##设置复制用户的账户
secondary_check_script=/usr/local/bin/masterha_secondary_check -s 192.168.35.10 -s 192.168.35.20 #设置检查从服务器的脚本
shutdown_script=""
ssh_user=root ##需要进行ssh检测,用户是root;前期做免交互的原因
[server1]
hostname=192.168.8.153
port=3306
[server2]
candidate_master=1 ##设置候选msater,如果设置了此参数,发送主从切换后,会将此库提升为主库
hostname=192.168.8.150
check_repl_delay=0 ##关闭复制的检查,默认情况下如果一个slave落后master 100M的relay-logs(中继日志)的话,MHA将不会选择该slave作为新的master,设置为0的话会忽略此规则
port=3306
[server3]
hostname=192.168.8.152
port=3306
测试无密码认证
如果正常会输出successfully
masterha_check_ssh -conf=/etc/masterha/app1.cnf
测试主从复制
[root@manager ~]# masterha_check_repl -conf=/etc/masterha/app1.cnf
…………………………
Tue Jul 20 22:29:37 2021 - [info] Alive Servers:
Tue Jul 20 22:29:37 2021 - [info] 192.168.8.151(192.168.8.151:3306)
Tue Jul 20 22:29:37 2021 - [info] 192.168.8.152(192.168.8.152:3306)
Tue Jul 20 22:29:37 2021 - [info] 192.168.8.153(192.168.8.153:3306)
…………………………
启动MHA
[root@manager scripts]# nohup masterha_manager --conf=/etc/masterha/app1.cnf --remove_dead_master_conf --ignore_last_failover < /dev/null > /var/log/masterha/app1/manager.log 2>&1 &
[1] 78897
#查看到当前的master节点是谁
[root@manager bin]# masterha_check_status --conf=/etc/masterha/app1.cnf
app1 (pid:78897) is running(0:PING_OK), master:192.168.8.151
总结
MHA主要是为了解决msater只有一台,若master出现故障时就使用MHA解决单点故障问题
组件:
MHA Manager (管理节点)
可以借助于node组件,实现健康检查(ssh、主从复制)、控制故障切换
MHA Node 组件(数据节点):
部署在所有MySQL服务器内;主要作用:
定时汇报给主manager节点,mysql服务的状态
监控本地mysql服务功能
配置步骤
实验环境:四台服务器,一台master,二台slave,一台MHA(在MHA节点上安装manager组件 )