MySQL CLUSTER(集群)系列 (2.另类在线增加节点-online hotplugin)

一、介绍 这篇文档旨在介绍如何设计MySQL集群,制做集群模板,有效规避mysql限制。(此文为二管理,四存储,八数据的模块制作) 安装环境及软件包: vmware workstation 5.5.3 mysql-5.2.3-falcon-alpha.tar.gz gentoo 2006.1 (每机单网卡多ip) Server1: 192.168.1.111 (ndb_mgmd, id=1) Server1: 192.168.1.112 (ndbd ,id=3) server1: 192.168.1.113 (ndbd,id=4) Server2: 192.168.1.110 (ndb_mgmd,id=2) server2: 192.168.1.109 (ndbd,id=5) server2: 192.168.1.108 (ndbd,id=6) 二、在Server1和Server2上安装MySQL 以下步骤需要在Server1和Server2上各做一次 # mv mysql-5.2.3-falcon-alpha.tar.gz /tmp/package # cd /tmp/package # groupadd mysql # useradd -g mysql mysql # tar -zxvf mysql-5.2.3-falcon-alpha.tar.gz # rm -f mysql-5.2.3-falcon-alpha.tar.gz # mv mysql-5.2.3-falcon-alpha mysql # cd mysql # ./configure –prefix=/usr –with-extra-charsets=complex –with-plugin-ndbcluster –with-plugin-partition –with-plugin-innobase # make && make install #ln -s /usr/libexec/ndbd /usr/bin #ln -s /usr/libexec/ndb_mgmd /usr/bin #ln -s /usr/libexec/ndb_cpcd /usr/bin #ln -s /usr/libexec/mysqld /usr/bin #ln -s /usr/libexec/mysqlmanager /usr/bin #mysql_install_db –user=mysql 三、安装并配置节点 以下步骤需要在Server1和Server2上各做一次 1.配置管理节点配置文件: # mkdir /var/lib/mysql-cluster # cd /var/lib/mysql-cluster # vi config.ini 在config.ini中添加如下内容: [ndbd default] NoOfReplicas= 4 (此文有四个存储节点) MaxNoOfConcurrentOperations= 10000 DataMemory= 80M IndexMemory= 24M TimeBetweenWatchDogCheck= 30000 DataDir= /var/lib/mysql-cluster MaxNoOfOrderedIndexes= 512 [ndb_mgmd default] DataDir= /var/lib/mysql-cluster [ndb_mgmd] Id=1 HostName= 192.168.1.111 [ndb_mgmd] Id=2 HostName= 192.168.1.110 [ndbd] Id= 3 HostName= 192.168.1.112 [ndbd] Id= 4 HostName= 192.168.1.113 [ndbd] Id= 5 HostName= 192.168.1.109 [ndbd] Id= 6 HostName= 192.168.1.108 [mysqld] [mysqld] [mysqld] [mysqld] [mysqld] [mysqld] [mysqld] [mysqld] 共八个mysqld定义 [tcp default] PortNumber= 63132 2.配置通用my.cnf文件,mysqld及ndbd,ndb_mgmd均使用此文件. # vi /etc/my.cnf 在my.cnf中添加如下内容: [mysqld] default-storage-engine=ndbcluster 避免在sql语句中还要加入ENGINE=NDBCLUSTER。 ndbcluster ndb-connectstring=192.168.1.111,192.168.1.110 [ndbd] connect-string=192.168.1.111,192.168.1.110 [ndb_mgm] connect-string=192.168.1.111,192.168.1.110 [ndb_mgmd] config-file=/var/lib/mysql-cluster/config.ini [mysql_cluster] ndb-connectstring= 192.168.1.111,192.168.1.110 保存退出后,启动管理节点Server1为: # ndb_mgmd –ndb_nodeid=1 启动管理节点Server2为: # ndb_mgmd –ndb_nodeid=2 四、初始化集群 在Server1中 # ndbd –bind_address=192.168.1.112 –nodeid=3 –initial #ndbd –bind_address=192.168.1.113–nodeid=4 –initial 在Server2中 #ndbd –bind_address=192.168.1.109 –nodeid=5 –initial #ndbd –bind_address=192.168.1.108 –nodeid=6 –initial 注:只有在第一次启动ndbd时或者对config.ini进行改动后才需要使用–initial参数! 五、检查工作状态 在任意一台机子上启动管理终端: # ndb_mgm 键入show命令查看当前工作状态:(下面是一个状态输出示例) – NDB Cluster — Management Client – ndb_mgm> show Connected to Management Server at: 192.168.1.111:1186 Cluster Configuration ——————— [ndbd(NDB)] 4 node(s) id=3 @192.168.1.111 (Version: 5.2.3, Nodegroup: 0, Master) id=4 @192.168.1.110 (Version: 5.2.3, Nodegroup: 0) id=5 @192.168.1.109 (Version: 5.2.3, Nodegroup: 0) id=6 @192.168.1.108 (Version: 5.2.3, Nodegroup: 0) [ndb_mgmd(MGM)] 2 node(s) id=1 @192.168.1.111 (Version: 5.2.3) id=2 @192.168.1.110 (Version: 5.2.3) [mysqld(API)] 8 node(s) id=7 (not connected, accepting connect from any host) id=8 (not connected, accepting connect from any host) id=9 (not connected, accepting connect from any host) id=10 (not connected, accepting connect from any host) id=11 (not connected, accepting connect from any host) id=12 (not connected, accepting connect from any host) id=13 (not connected, accepting connect from any host) id=14 (not connected, accepting connect from any host) ndb_mgm> 如果上面没有问题,现在开始加入mysqld(API): 注意,这篇文档对于MySQL并没有设置root密码,推荐你自己设置Server1和Server2的MySQL root密码。 在Server1 中: #mysqld_safe –ndb_nodeid=7 –user=mysql & 在Server2 中: #mysqld_safe –ndb_nodeid=8 –user=mysql & # ndb_mgm -e show 信息如下: Connected to Management Server at: 192.168.1.111:1186 Cluster Configuration ——————— [ndbd(NDB)] 4 node(s) id=3 @192.168.1.111 (Version: 5.2.3, Nodegroup: 0, Master) id=4 @192.168.1.110 (Version: 5.2.3, Nodegroup: 0) id=5 @192.168.1.109 (Version: 5.2.3, Nodegroup: 0) id=6 @192.168.1.108 (Version: 5.2.3, Nodegroup: 0) [ndb_mgmd(MGM)] 2 node(s) id=1 @192.168.1.111 (Version: 5.2.3) id=2 @192.168.1.110 (Version: 5.2.3) [mysqld(API)] 4 node(s) id=5 @192.168.1.111 (Version: 5.2.3) id=6 @192.168.1.110 (Version: 5.2.3) id=7 (not connected, accepting connect from any host) id=8 (not connected, accepting connect from any host) id=9 (not connected, accepting connect from any host) id=10 (not connected, accepting connect from any host) id=11 (not connected, accepting connect from any host) id=12 (not connected, accepting connect from any host) ok,可以测试了: 在Server1 中 # /usr/local/mysql/bin/mysql -u root -p >create database aa; > use aa; > CREATE TABLE ctest (i INT) ; > INSERT INTO ctest () VALUES (1); > SELECT * FROM ctest; 应该可以看到1 row returned信息(返回数值1)。 如果上述正常,则换到Server2,观察效果。如果成功,则在Server2中执行INSERT再换回到Server1观察是否工作正常。 如果都没有问题,那么恭喜成功!至此模板制作完成。 注意:在模板制作完成的过程中,测试过程完成了对集群的初次读写,若跳过测试(跳过读写),在集群未协同工作前进行下面的操作,则进行下面的操作毫无意义。因为,集群初始后只在/var/lib/mysql-cluster/下建了几个空目录,还没有正常协同工作. 六、备份/var/lib/mysql-cluster/下的相应ndb_*_fs,留做以后使用。 七、原理: ndbd –initial初始集群结构,在官方文档中增加节点要bakcup后restore要initial的主要是存储节点,管理节点和数据节点都可用集群restart实现. 而存储节点特点是只要一个有效,则整个集群可以正常工作,因此,在模板完成后,未使用的存储节点可以当做故障节点,当启用时,存储节点可以通过其内部达到同步. 缺点,在集群启动时,每个未使用存储节点默认等待启用时间为60秒,造成集群启动时间加长,可以能通过设置参数设小,不推荐。 八、 实现: 数据节点不用说了,管理节点两个也够用了,主要是存储节点了。 根据上面的步骤搭建一台服务器,拷贝相应的ndb_*_fs到/var/lib/mysql-cluster/下,机器ip符合config.ini中的定义,好了,可以启用了,简单吧。 对于现有设计,可以按官方文档增加修改节点的工作时,参考此文档,预留相应的空闲节点,为以后的工作增加便利。 MySQL CLUSTER(集群)系列 (3.lvs+keeplived+mysql cluster) 一、介绍 这篇文档旨在介绍如何结合lvs+keepalived实现mysql cluster的高可用及负载均衡。此文是在1文(双机高可用)基础上增加lvs+keepalived,稍加修改就可适合更多结点的集群. 安装环境及软件包: vmware workstation 5.5.3 mysql-5.2.3-falcon-alpha.tar.gz gentoo 2006.1 ipvsadm-1.24.tar.gz keepalived-1.1.13.tar.gz linux-2.6.20.3.tar.bz2 iproute2-2.6.15-060110.tar.gz Server1: 192.168.1.111 (ndb_mgmd, id=1) Server2: 192.168.1.110 (ndb_mgmd,id=2) 二~六步参见1文(双机高可用) 以下步骤需要在Server1和Server2上各做一次 七、内核linux-2.6.20.3.tar.bz2安装 # tar xvjf linux-2.6.20.3.tar.bz2 -C /usr/src #cd /usr/src/linux-2.6.20.3 #zcat /proc/config.gz .config #make menuconfig 选择 · Network packet filtering framework (Netfilter) —> 后在 [ ] TCP: MD5 Signature Option support (RFC2385) (EXPERIMENTAL) 下出现 IP: Virtual Server Configuration —> 关天netfilter内的配置及Virtual Server的配置根据自己的需要选择。 选择 · IP: advanced router Choose IP: FIB lookup algorithm (choose FIB_HASH if unsure) (FIB_HASH) —> · IP: policy routing # make all && make modules_install && make install #vi /boot/grub.conf 加入 title=2.6.20.3 kernel /vmlinuz-2.6.20.3 root=/你的根设备 #reboot (以新内核启动系统) 八、安装ipvsadm和keepalived #tar -zxvf ipvsadm-1.24.tar.gz -C /tmp/package # cd /tmp/package/ipvsadm-1.24 # make && make install #tar -zxvf keepalived-1.1.13.tar.gz -C /tmp/package #cd /tmp/package/keepalived-1.1.13 #vi keepalived/vrrp/vrrp_arp.c 将 26 #include 27 28 /* local includes */ 29 #include “vrrp_arp.h” 30 #include “memory.h” 31 #include “utils.h” 修改为 26 /* local includes */ 27 #include “vrrp_arp.h” 28 #include “memory.h” 29 #include “utils.h” 30#include 31 就是将#include 这行移到下面. #./configure –prefix=/usr –with-kernel-dir=/usr/src/linux-2.6.20.3 #make && make install #vi /etc/init.d/keepalived 加入以下内容 #!/sbin/runscript # Copyright 1999-2004 Gentoo Foundation # Distributed under the terms of the GNU General Public License v2 # $Header: /var/cvsroot/gentoo-x86/sys-cluster/keepalived/files/init-keepalived,v 1.3 2004/07/15 00:55:17 agriffis Exp $ depend() { use logger need net } checkconfig() { if [ ! -e /etc/keepalived/keepalived.conf ] ; then eerror “You need an /etc/keepalived/keepalived.conf file to run keepalived” return 1 fi } start() { checkconfig || return 1 ebegin “Starting Keepalived” start-stop-daemon –start –quiet –pidfile /var/run/keepalived.pid \ –startas /usr/sbin/keepalived eend $? } stop() { ebegin “Stopping Keepalived” start-stop-daemon –stop –quiet –pidfile /var/run/keepalived.pid eend $? } 此为gentoo的keepalived的脚本. #chmod 755 /etc/init.d/keepalived #rc-update add keepalived default #vi /etc/keepalived/keepalived.conf 加入 ! Configuration File for keepalived global_defs { router_id mysql_cluster } vrrp_sync_group VG1 { (此处是ha部分) group { VI_1 } } vrrp_instance VI_1 { state MASTER interface eth0 lvs_sync_daemon_interface eth0 virtual_router_id 1 (此处server1为1,server2为2) priority 150 advert_int 1 authentication { auth_type PASS auth_pass mysqlcluster } virtual_ipaddress { 192.168.1.120 } } virtual_server 192.168.1.120 3306 { (此处定义负载均衡部分,使用DR方式) delay_loop 6 lvs_sched wlc lvs_method DR persistence_timeout 60 ha_suspend protocol TCP real_server 192.168.1.110 3306 { weight 1 TCP_CHECK { connect_timeout 10 } } real_server 192.168.1.111 3306 { weight 1 TCP_CHECK { connect_timeout 10 } } } 九,启动 #/etc/init.d/keepalived start #ip addr list (未安装iproute2 ,无此命今,可以使用emerge iproute2安装,注意emerge是gentoo的命今) 出现类似下面的信息 eth0: mtu 1500 qdisc pfifo_fast qlen 1000 link/ether 00:0c:29:6f:f9:21 brd ff:ff:ff:ff:ff:ff inet 192.168.1.111/24 brd 192.168.1.255 scope global eth0 inet 192.168.1.120/32 scope global eth0 (此行表明虚拟的ip生效了) inet6 fe80::20c:29ff:fe6f:f921/64 scope link valid_lft forever preferred_lft forever #tail /var/log/messages 可以查看更多信息. 类似如下 Keepalived: Starting Keepalived v1.1.13 (03/26,2007) Keepalived_healthcheckers: Using LinkWatch kernel netlink reflector… Keepalived_healthcheckers: Registering Kernel netlink reflector Keepalived_healthcheckers: Registering Kernel netlink command channel Keepalived_healthcheckers: Configuration is using : 9997 Bytes Keepalived: Starting Healthcheck child process, pid=27738 Keepalived_vrrp: Using LinkWatch kernel netlink reflector… Keepalived_vrrp: Registering Kernel netlink reflector Keepalived_vrrp: Registering Kernel netlink command channel Keepalived_vrrp: Registering gratutious ARP shared channel Keepalived_vrrp: Configuration is using : 36549 Bytes Keepalived: Starting VRRP child process, pid=27740 Keepalived_healthcheckers: Activating healtchecker for service [192.168.1.110:3306] Keepalived_healthcheckers: Activating healtchecker for service [192.168.1.111:3306] IPVS: sync thread started: state = MASTER, mcast_ifn = eth0, syncid = 2 Keepalived_vrrp: VRRP_Instance(VI_1) Transition to MASTER STATE Keepalived_vrrp: VRRP_Instance(VI_1) Entering MASTER STATE Keepalived_vrrp: VRRP_Group(VG1) Syncing instances to MASTER state Keepalived_vrrp: Netlink: skipping nl_cmd msg… 十、结束语 此文档旨在从mysql cluster应用角度考虑如何更好的使用mysql及linux和相关工具.文中有不到及错误的地方敬请不吝指正。 超微服务器集群方案,已经在全球被广泛应用, 本文出自 51CTO.COM技术博客

你可能感兴趣的:(vmware,linux,mysql,SQL Server,配置管理)