二.Hadoop(HA)集群的搭建
2.1 配置详细
主机名 | IP | NameNode | DataNode | Year | Zookeeper | JournalNode |
mast1 | 192.168.177.131 | 是 | 是 | 否 | 是 | 是 |
mast2 | 192.168.177.132 | 是 | 是 | 否 | 是 | 是 |
mast3 | 192.168.177.133 | 否 | 是 | 是 | 是 | 是 |
2.2 安装jdk
(省略)安装jdk和配置环境变量
2.2 SSH免登录
(省略),参考:http://eksliang.iteye.com/blog/2187265
2.4 Zookeeper集群搭建
(省略),参考,http://eksliang.iteye.com/blog/2107002,这是我的solr集群部署,也是使用zookeeper进行管理,zookeeper这里步骤跟操作一模一样,最后我的zoo.cfg文件如下所示
[hadoop@Mast1 conf]$ cat zoo.cfg # The number of milliseconds of each tick tickTime=2000 # The number of ticks that the initial # synchronization phase can take initLimit=10 # The number of ticks that can pass between # sending a request and getting an acknowledgement syncLimit=5 # the directory where the snapshot is stored. dataDir=/home/hadoop/zookeeper/data dataLogDir=/home/hadoop/zookeeper/datalog # the port at which the clients will connect clientPort=2181 server.1=mast1:2888:3888 server.2=mast2:2888:3888 server.3=mast3:2888:3888
2.5配置Hadoop配置文件
先配置mast1这台机器,配置后了后,将配置环境,复制到mast2、mast3上面即可!
hadoop2.0的配置存放在~/etc/hadoop目录下面,
- core.xml
‘ fs.defaultFS hdfs://ns hadoop.tmp.dir /home/hadoop/workspace/hdfs/temp io.file.buffer.size 4096 ha.zookeeper.quorum mast1:2181,mast2:2181,mast3:2181
- hdfs-site.xml
dfs.nameservices ns dfs.ha.namenodes.ns nn1,nn2 dfs.namenode.rpc-address.ns.nn1 mast1:9000 dfs.namenode.http-address.ns.nn1 mast1:50070 dfs.namenode.rpc-address.ns.nn2 mast2:9000 dfs.namenode.http-address.ns.nn2 mast2:50070 dfs.namenode.shared.edits.dir qjournal://mast1:8485;mast2:8485;mast3:8485/ns dfs.journalnode.edits.dir /home/hadoop/workspace/journal dfs.ha.automatic-failover.enabled true dfs.client.failover.proxy.provider.ns org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider dfs.ha.fencing.methods sshfence dfs.ha.fencing.ssh.private-key-files /home/hadoop/.ssh/id_rsa dfs.namenode.name.dir file:///home/hadoop/workspace/hdfs/name dfs.datanode.data.dir file:///home/hadoop/workspace/hdfs/data dfs.replication 2 dfs.webhdfs.enabled true
- mapred-site.xml
mapreduce.framework.name yarn
- yarn-site.xml
yarn.nodemanager.aux-services mapreduce_shuffle yarn.resourcemanager.hostname mast3
- slaves
[hadoop@Mast1 hadoop]$ cat slaves mast1 mast2 mast3
- 修改JAVA_HOME
分别在文件hadoop-env.sh和yarn-env.sh中添加JAVA_HOME配置
#export JAVA_HOME=${JAVA_HOME} --原来 export JAVA_HOME=/usr/local/java/jdk1.7.0_67
虽然默认配置了${JAVA_HOME}的环境变量,但是hadoop启动时,会提示找不到,没有办法,指定绝对路径,这个是必须的。
- 配置hadoop的环境变量,参考我的配置
[hadoop@Mast1 hadoop]$ vim ~/.bash_profile export HADOOP_HOME="/home/hadoop/hadoop-2.5.2" export PATH=$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib"
- 将配置复制到mast2、mast3
scp -r ~/.bash_profile hadoop@mast2:/home/hadoop/ scp -r ~/.bash_profile hadoop@mast3:/home/hadoop/ scp -r $HADOOP_HOME/etc/hadoop hadoop@mast2:/home/hadoop/hadoop-2.5.2/etc/ scp -r $HADOOP_HOME/etc/hadoop hadoop@mast3:/home/hadoop/hadoop-2.5.2/etc/
至此Hadoop的配置完毕,接下来就是启动集群了
三.集群的启动
3.1 启动zookeeper集群
分别在mast1、mast2、mast3上执行如下命令启动zookeeper集群;
[hadoop@Mast1 bin]$ sh zkServer.sh start
验证集群zookeeper集群是否启动,分别在mast1、mast2、mast3上执行如下命令验证zookeeper集群是否启动,集群启动成功,有两个follower节点跟一个leader节点;
[hadoop@Mast1 bin]$ sh zkServer.sh status JMX enabled by default Using config: /home/hadoop/zookeeper/zookeeper-3.3.6/bin/../conf/zoo.cfg Mode: follower
3.2 启动journalnode集群
在mast1上执行如下命令完成JournalNode集群的启动
[hadoop@Mast1 hadoop-2.5.2]$ sbin/hadoop-daemons.sh start journalnode
执行jps命令,可以查看到JournalNode的java进程pid
3.3 格式化zkfc,让在zookeeper中生成ha节点
在mast1上执行如下命令,完成格式化
hdfs zkfc –formatZK
(注意,这条命令最好手动输入,直接copy执行有可能会有问题,当时部署时我是蛋疼了许久)
格式成功后,查看zookeeper中可以看到
[zk: localhost:2181(CONNECTED) 1] ls /hadoop-ha [ns]
3.4 格式化hdfs
hadoop namenode –format
(注意,这条命令最好手动输入,直接copy执行有可能会有问题)
3.5 启动NameNode
首先在mast1上启动active节点,在mast1上执行如下命令
[hadoop@Mast1 hadoop-2.5.2]$ sbin/hadoop-daemon.sh start namenode
在mast2上同步namenode的数据,同时启动standby的namenod,命令如下
#把NameNode的数据同步到mast2上 [hadoop@Mast2 hadoop-2.5.2]$ hdfs namenode –bootstrapStandby #启动mast2上的namenode作为standby [hadoop@Mast2 hadoop-2.5.2]$ sbin/hadoop-daemon.sh start namenode
3.6 启动启动datanode
在mast1上执行如下命令
[hadoop@Mast1 hadoop-2.5.2]$ sbin/hadoop-daemons.sh start datanode
3.7 启动year
在作为资源管理器上的机器上启动,我这里是mast3,执行如下命令完成year的启动
[hadoop@Mast3 hadoop-2.5.2]$ sbin/start-yarn.sh
3.8 启动ZKFC
在mast1上执行如下命令,完成ZKFC的启动
[hadoop@Mast1 hadoop-2.5.2]$ sbin/hadoop-daemons.sh start zkfc
全部启动完后分别在mast1,mast2,mast3上执行jps是可以看到下面这些进程的
#mast1上的java PID进程 [hadoop@Mast1 hadoop-2.5.2]$ jps 2837 NodeManager 3054 DFSZKFailoverController 4309 Jps 2692 DataNode 2173 QuorumPeerMain 2551 NameNode 2288 JournalNode #mast2上的java PID进程 [hadoop@Mast2 ~]$ jps 2869 DFSZKFailoverController 2353 DataNode 2235 JournalNode 4522 Jps 2713 NodeManager 2591 NameNode 2168 QuorumPeerMain #mast3上的java PID进程 [hadoop@Mast3 ~]$ jps 2167 QuorumPeerMain 2337 JournalNode 3506 Jps 2457 DataNode 2694 NodeManager 2590 ResourceManager
四.测试HA的高可用性
启动后mast1的namenode和mast2的namenode如下所示:
此时在mast1上执行如下命令关闭mast1上的namenode
[hadoop@Mast1 hadoop-2.5.2]$ sbin/hadoop-daemon.sh stop namenode
再次查看mast1上的namenode,发现自动切换为active了!证据如下:
来自:http://www.iteye.com/news/30739