[root@hadoop201 ~]# rm -rf anaconda-ks.cfg install.log install.log.syslog
弹窗 | 操作 |
---|---|
1 | 下一步 |
2 | 克隆自:虚拟机中的当前状态 |
3 | 创建完整克隆 |
4 | 虚拟机命名,选择存储位置 |
5 | 完成 |
删除第一个eth0,改第二个eth0为eth1SUBSYSTEM=="net", ACTION=="add", DRIVERS=="?*", ATTR{address}=="00:0c:29:2e:99:a1", ATTR{type}=="1", KERNEL=="eth*", NAME=="eth0"~
DEVICE=eth0
HWADDR=00:0c:29:b9:4d:21
TYPE=Ethernet
UUID=73c15477-5d31-4956-8be7-2809e1d204db
ONBOOT=yes
NM_CONTROLLED=yes
BOOTPROTO=static
IPADDR=192.168.1.208
GATEWAY=192.168.1.3
DNS1=192.168.1.3
|- 当前主机名:hostname
hadoop206.cevent.com
|- 修改网络主机配置:vi /etc/sysconfig/network
NETWORKING=yes
HOSTNAME=hadoop208.cevent.com
|- 修改主机配置:vi /etc/hosts
127.0.0.1
localhost localhost.localdomain localhost4 localhost4.localdomain4
::1
localhost localhost.localdomain localhost6 localhost6.localdomain6
192.168.1.201 hadoop201.cevent.com
192.168.1.202 hadoop202.cevent.com
192.168.1.203 hadoop203.cevent.com
…
(1)sync
(2)reboot
路径:C:\Windows\System32\drivers\etc\hosts
# Copyright (c) 1993-2009 Microsoft Corp.
#
# This is a sample HOSTS file used by
Microsoft TCP/IP for Windows.
#
# This file contains the mappings of IP
addresses to host names. Each
# entry should be kept on an individual
line. The IP address should
# be placed in the first column followed by
the corresponding host name.
# The IP address and the host name should
be separated by at least one
# space.
#
# Additionally, comments (such as these)
may be inserted on individual
# lines or following the machine name
denoted by a '#' symbol.
#
# For example:
#
#
102.54.94.97
rhino.acme.com # source server
#
38.25.63.10 x.acme.com # x client host
# localhost name resolution is handled
within DNS itself.
# 127.0.0.1 localhost
# ::1 localhost
127.0.0.1 www.vmix.com
192.30.253.112 github.com
151.101.88.249 github.global.ssl.fastly.net
127.0.0.1
www.xmind.net
192.168.1.201 hadoop201.cevent.com
192.168.1.202 hadoop202.cevent.com
192.168.1.203 hadoop203.cevent.com
192.168.1.204 hadoop204.cevent.com
192.168.1.205 hadoop205.cevent.com
192.168.1.206 hadoop206.cevent.com
192.168.1.207 hadoop207.cevent.com
192.168.1.208 hadoop208.cevent.com
192.168.1.209 hadoop209.cevent.com
192.168.1.210 hadoop210.cevent.com
192.168.1.211 hadoop211.cevent.com
192.168.1.212 hadoop212.cevent.com
192.168.1.1 windows10.microdone.cn
[cevent@hadoop207 ~]$ sudo ssh
hadoop209.cevent.com
[sudo] password for cevent:
ssh: Could not resolve hostname
hadoop209.cevent.com: Name or service not known
修改hosts: vi /etc/hosts
127.0.0.1
localhost localhost.localdomain localhost4 localhost4.localdomain4
::1
localhost localhost.localdomain localhost6 localhost6.localdomain6
192.168.1.201 hadoop201.cevent.com
192.168.1.202 hadoop202.cevent.com
192.168.1.203 hadoop203.cevent.com
192.168.1.204 hadoop204.cevent.com
192.168.1.205 hadoop205.cevent.com
192.168.1.206 hadoop206.cevent.com
192.168.1.207 hadoop207.cevent.com
192.168.1.208 hadoop208.cevent.com
192.168.1.209 hadoop209.cevent.com
(1)ssh登录A
[cevent@hadoop210 ~]$ cd .ssh/
[cevent@hadoop210 .ssh]$ ll
总用量 8
-rw-------. 1 cevent cevent 409 3月 24 09:40 authorized_keys
-rw-r–r--. 1 cevent cevent 1248 4月
28 17:16 known_hosts
[cevent@hadoop207 .ssh]$ ssh
hadoop208.cevent.com
The authenticity of host
‘hadoop208.cevent.com (192.168.1.208)’ can’t be established.
RSA key fingerprint is
fe:07:91:9c:00:5d:09:2c:48:bb:d5:53:9f:09:6c:34.
Are you sure you want to continue
connecting (yes/no)? yes
Warning: Permanently added
‘hadoop208.cevent.com,192.168.1.208’ (RSA) to the list of known hosts.
[email protected]’s password:
Last login: Mon Jun 15 16:06:58 2020 from
192.168.1.1
[cevent@hadoop208 ~]$ exit
logout
(2)ssh登录B
[cevent@hadoop207 ~]$ ssh
hadoop209.cevent.com
The authenticity of host ‘hadoop209.cevent.com
(192.168.1.209)’ can’t be established.
RSA key fingerprint is
fe:07:91:9c:00:5d:09:2c:48:bb:d5:53:9f:09:6c:34.
Are you sure you want to continue
connecting (yes/no)? yes
Warning: Permanently added
‘hadoop209.cevent.com,192.168.1.209’ (RSA) to the list of known hosts.
[email protected]’s password:
Last login: Mon Jun 15 17:01:11 2020 from
192.168.1.1
[cevent@hadoop209 ~]$ exit
logout
(3)查看登录记录
[cevent@hadoop207 ~]$ cd
[cevent@hadoop207 ~]$ ls -al
总用量 188
drwx------. 26 cevent cevent 4096 6月 15 17:03 .
drwxr-xr-x.
3 root root 4096 4月 30 09:15 …
drwxrwxr-x.
2 cevent cevent 4096 5月
7 10:43 .abrt
-rw-------.
1 cevent cevent 17232 6月 15 17:05 .bash_history
-rw-r–r--.
1 cevent cevent 18 5月
11 2016 .bash_logout
-rw-r–r--.
1 cevent cevent 176 5月
11 2016 .bash_profile
-rw-r–r--.
1 cevent cevent 124 5月
11 2016 .bashrc
drwxrwxr-x.
2 cevent cevent 4096 5月
7 17:32 .beeline
drwxr-xr-x.
3 cevent cevent 4096 5月
7 10:43 .cache
drwxr-xr-x.
5 cevent cevent 4096 5月
7 10:43 .config
drwx------.
3 cevent cevent 4096 5月
7 10:43 .dbus
-rw-------.
1 cevent cevent 16 5月
7 10:43 .esd_auth
drwx------.
4 cevent cevent 4096 6月
14 22:13 .gconf
drwx------.
2 cevent cevent 4096 6月
14 23:14 .gconfd
drwxr-xr-x.
5 cevent cevent 4096 5月
7 10:43 .gnome2
drwxrwxr-x.
3 cevent cevent 4096 5月
7 10:43 .gnote
drwx------.
2 cevent cevent 4096 6月
14 22:13 .gnupg
-rw-rw-r–.
1 cevent cevent 195 6月
14 22:13 .gtk-bookmarks
drwx------.
2 cevent cevent 4096 5月
7 10:43 .gvfs
-rw-rw-r–.
1 cevent cevent 589 6月
12 13:42 .hivehistory
-rw-------.
1 cevent cevent 620 6月
14 22:13 .ICEauthority
-rw-r–r--.
1 cevent cevent 874 6月
14 23:14 .imsettings.log
drwxr-xr-x.
3 cevent cevent 4096 5月
7 10:43 .local
drwxr-xr-x.
4 cevent cevent 4096 3月
13 01:51 .mozilla
-rw-------.
1 cevent cevent 214 5月
7 13:37 .mysql_history
drwxr-xr-x.
2 cevent cevent 4096 5月
7 10:43 .nautilus
drwx------.
2 cevent cevent 4096 5月 7 10:43 .pulse
-rw-------.
1 cevent cevent 256 5月
7 10:43 .pulse-cookie
drwx------.
2 cevent cevent 4096 4月
30 15:20 .ssh
…
[cevent@hadoop207 ~]$ cd .ssh/
[cevent@hadoop207 .ssh]$ ll
总用量 16
-rw-------. 1 cevent cevent 409 4月 30 15:20 authorized_keys
-rw-------. 1 cevent cevent 1675 4月
30 15:20 id_rsa
-rw-r–r--. 1 cevent cevent 409 4月 30 15:20 id_rsa.pub
-rw-r–r--. 1 cevent cevent 832 6月 15 17:08 known_hosts
查看登录记录
[cevent@hadoop207 .ssh]$ vi known_hosts
[cevent@hadoop207 .ssh]$ ssh-keygen -t rsa
Generating public/private rsa key pair.
Enter file in which to save the key
(/home/cevent/.ssh/id_rsa): cevent
Enter passphrase (empty for no passphrase):
Enter same passphrase again:
Your identification has been saved in
cevent.
Your public key has been saved in
cevent.pub.
The key fingerprint is:
1c:5a:1a:d2:e4:3b:fe:36:39:df:95:b3:75:85:0f:af
cevent@hadoop207.cevent.com
The key's randomart image is:
+--[ RSA 2048]----+
|
. |
|
+ |
|
. + o |
|
. B . . |
|
= S o .|
|
. . =.|
|
. . + =|
|
.= . . =.|
|
..+. . E |
+-----------------+
(6)复制key-id(本机也需要配)
[cevent@hadoop207 .ssh]$ ssh-copy-id
hadoop208.cevent.com
[cevent@hadoop207 .ssh]$ ssh-copy-id
hadoop209.cevent.com
[cevent@hadoop207 .ssh]$ ssh
hadoop208.cevent.com
Last login: Mon Jun 15 18:23:00 2020 from
hadoop207.cevent.com
[cevent@hadoop208 ~]$ exit
logout
Connection to hadoop208.cevent.com closed.
[cevent@hadoop207 .ssh]$ ssh
hadoop209.cevent.com
Last login: Mon Jun 15 17:08:11 2020 from
hadoop207.cevent.com
[cevent@hadoop209 ~]$ exit
logout
8.查看其他虚拟机无秘钥登录
[cevent@hadoop208 ~]$ cd
[cevent@hadoop208 ~]$ ls -al
[cevent@hadoop208 ~]$ cd .ssh/
[cevent@hadoop208 .ssh]$ ll
总用量 8
-rw-------. 1 cevent cevent 818 6月 15 18:24 authorized_keys
-rw-r--r--. 1 cevent cevent 1664 6月
16 17:56 known_hosts
[cevent@hadoop208 .ssh]$ vi authorized_keys
ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEAyRWmouFLr4b6orfQrWqtlukJ1orkeYcqM6orH1JMOMo1BLYY/Bop2ZIAHahUxGgUdRlZ5mFkGHut1A7h/VdGLBdegyZXQwxwl6Cx67XIsQWRwUgZWXbujg+qV9irPE7WvxF5e3FvJGfbmh7boPk+q/Hsk6rgZ3k9qrEDx4vhv7eL+Ostt2D8tV4UbRReUNl3yI9bt4P/S7ARBpelFtB4p9drbGSxjtH0sNKBHiDAcAV+MOSLz21+WlYr2x58jAZc37UXi/qYfosgc+u5GJ88z/kEI+1YqXBX11FFiRWZcI2aiLweri5WtHH0hoEZGrTXBeShuWQzegx9/E0upPlsfw==
cevent@hadoop202.cevent.com
ssh-rsa
AAAAB3NzaC1yc2EAAAABIwAAAQEA0Fe9XV0baD7RPiGkIf+ZMoMPOaCF445aAvaJdGt8wuegkxJjqPMTAop79xcA7AY/vFS7PjpllM162t/lVoCozGHK1iOfElObiLo6+pxBcwfVYnEUlzAz/L0Ngpss54Eb48xOq068gcKcDAZrNbdtxDkTgzHFttcWpB7j++gRXrfB9O9HxKcRObu16sBM8tLmLF4M+tvxTC/Ko/amnrOvi3/AyCtxH1sRumqUiu9buDJAFAgV1Y+s7CR7GORGIkDkmHr9e3O5UMpNXTgnfIaCPdNzn6qRTUM/Sb5KAkkMBb3MY5NgbOPDvFwkPlG8xcFS5Ua8Arq58n8kwa2dyy94kQ==
cevent@hadoop207.cevent.com
(1)新建文件
[cevent@hadoop207 module]$ vim rsync.txt
kakaxi
(2)执行同步
[cevent@hadoop207 module]$ rsync -rvl
rsync.txt [email protected]:/opt/module/
10.xsync同步任务文件
(1)在/usr/local/bin目录下创建xsync文件,文件内容如下:
#!/bin/bash
#1 获取输入参数的个数,如果没有参数,直接退出,$#获取参数个数
pcount=$#
if((pcount==0)); then
#如果没有参数,返回no args,退出,程序关闭
echo no args;
exit;
fi
#2 获取文件名称:$1获取第一个参数,basename+ 路径1/路径2,获取路径最后一个值名称
p1=$1
fname=basename $p1
echo fname=$fname
#3 获取上级目录到:绝对路径
#获取文件路径:$(dirname $p1)
#cd -P 进入绝对路径 pwd获取路径
pdir=cd -P $(dirname $p1); pwd
echo pdir=$pdir
#4 获取当前用户名称
user=whoami
#5 循环
for((host=207; host<210; host++)); do
#echo p d i r / pdir/ pdir/fname u s e r @ h a d o o p user@hadoop user@hadoophost:$pdir
echo --------------- hadoop$host.cevent.com ----------------
#拼接路径用户
rsync -rvl p d i r / pdir/ pdir/fname u s e r @ h a d o o p user@hadoop user@hadoophost.cevent.com:$pdir
done
[cevent@hadoop210 bin]$ chmod 777 xsync 修改文件权限
[cevent@hadoop210 bin]$ ll
总用量 4
-rwxrwxrwx. 1 cevent cevent 842 6月
28 13:47 xsync
[cevent@hadoop207 module]$ vim xsync.txt
this is a xsync test!
[cevent@hadoop207 module]$ xsync xsync.txt
fname=xsync.txt
pdir=/opt/module
--------------- hadoop207.cevent.com
----------------
sending incremental file list
sent 32 bytes received 12 bytes 29.33 bytes/sec
total size is 23 speedup is 0.52
--------------- hadoop208.cevent.com
----------------
sending incremental file list
xsync.txt
sent 98 bytes received 31 bytes 258.00 bytes/sec
total size is 23 speedup is 0.18
--------------- hadoop209.cevent.com
----------------
sending incremental file list
xsync.txt
sent 98 bytes received 31 bytes 258.00 bytes/sec
total size is 23 speedup is 0.18
(1)在/usr/local/bin目录下创建xcall文件,文件内容如下:
#!/bin/bash
#$#获取参数个数
#$@ 获取所有参数
pcount=$#
if((pcount==0));then
echo no args;
exit;
fi
echo
-------------localhost.cevent.com----------
$@
for((host=207; host<210; host++)); do
echo ----------hadoop$host.cevent.com---------
ssh hadoop$host.cevent.com $@
done
(2)修改脚本 xcall 具有执行权限
[cevent@hadoop207 bin]$ sudo chmod a+x
xcall
[cevent@hadoop207 bin]$ xcall rm -rf
/opt/module/rsync.txt
-------------localhost.cevent.com----------
----------hadoop207.cevent.com---------
----------hadoop208.cevent.com---------
----------hadoop209.cevent.com---------
[cevent@hadoop207 ~]$ cat
/opt/module/hadoop-2.7.2/etc/hadoop/core-site.xml
"1.0"
encoding="UTF-8"?>
-stylesheet
type="text/xsl" href="configuration.xsl"?>
<!--
Licensed under the Apache License, Version 2.0 (the
"License");
you
may not use this file except in compliance with the License.
You
may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS"
BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See
the License for the specific language governing permissions and
limitations under the License. See accompanying LICENSE file.
-->
<!-- Put site-specific property
overrides in this file. -->
<!-- 指定HDFS中NameNode地址 ,设置的hadoop207.cevent.com=hostname -->
fs.defaultFS</name>
hdfs://hadoop207.cevent.com:8020</value>
</property>
<!-- 指定tmp数据存储位置 -->
hadoop.tmp.dir</name>
/opt/module/hadoop-2.7.2/data/tmp</value>
</property>
</configuration>
[cevent@hadoop207 ~]$ cat
/opt/module/hadoop-2.7.2/etc/hadoop/hdfs-site.xml
"1.0"
encoding="UTF-8"?>
-stylesheet
type="text/xsl" href="configuration.xsl"?>
<!--
Licensed under the Apache License, Version 2.0 (the
"License");
you
may not use this file except in compliance with the License.
You
may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS"
BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See
the License for the specific language governing permissions and
limitations under the License. See accompanying LICENSE file.
-->
<!-- Put site-specific property
overrides in this file. -->
<!-- 指定HDFS副本的数量:主从复制 -->
dfs.replication</name>
1</value>
</property>
dfs.namenode.secondary.http-address</name>
hadoop207.cevent.com:50090</value>
</property>
</configuration>
[cevent@hadoop207 ~]$ vim
/opt/module/hadoop-2.7.2/etc/hadoop/slaves
hadoop207.cevent.com
hadoop208.cevent.com
hadoop209.cevent.com
[cevent@hadoop207 ~]$ cat
/opt/module/hadoop-2.7.2/etc/hadoop/yarn-site.xml
"1.0"?>
<!--
Licensed under the Apache License, Version 2.0 (the
"License");
you
may not use this file except in compliance with the License.
You
may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS"
BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See
the License for the specific language governing permissions and
limitations under the License. See accompanying LICENSE file.
-->
<!-- Site specific YARN configuration properties -->
<!-- reducer获取数据的方式 -->
yarn.nodemanager.aux-services</name>
mapreduce_shuffle</value>
</property>
<!-- 指定YARN的ResourceManager的地址 -->
yarn.resourcemanager.hostname</name>
hadoop207.cevent.com</value>
</property>
</configuration>
[cevent@hadoop207 ~]$ cat /opt/module/hadoop-2.7.2/etc/hadoop/mapred-site.xml
"1.0"?>
-stylesheet
type="text/xsl" href="configuration.xsl"?>
<!--
Licensed under the Apache License, Version 2.0 (the
"License");
you
may not use this file except in compliance with the License.
You
may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS"
BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See
the License for the specific language governing permissions and
limitations under the License. See accompanying LICENSE file.
-->
<!-- Put site-specific property
overrides in this file. -->
<!-- 指定mr运行在yarn上 -->
mapreduce.framework.name</name>
yarn</value>
</property>
</configuration>
[cevent@hadoop207 module]$ xsync
hadoop-2.7.2/
[cevent@hadoop207 module]$ xcall cat
/opt/module/hadoop-2.7.2/etc/hadoop/slaves
-------------localhost.cevent.com----------
hadoop207.cevent.com
hadoop208.cevent.com
hadoop209.cevent.com
----------hadoop207.cevent.com---------
hadoop207.cevent.com
hadoop208.cevent.com
hadoop209.cevent.com
----------hadoop208.cevent.com---------
hadoop207.cevent.com
hadoop208.cevent.com
hadoop209.cevent.com
----------hadoop209.cevent.com---------
hadoop207.cevent.com
hadoop208.cevent.com
hadoop209.cevent.com
[cevent@hadoop210 hadoop]$ vim /opt/module/hadoop-2.7.2/etc/hadoop/mapred-env.sh
expor
tHADOOP_JOB_HISTORYSERVER_HEAPSIZE=1000
# Licensed to the Apache Software
Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information
regarding copyright ownership.
# The ASF licenses this file to You under
the Apache License, Version 2.0
# (the "License"); you may not
use this file except in compliance with
# the License. You may obtain a copy of the License at
#
#
http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or
agreed to in writing, software
# distributed under the License is
distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY
KIND, either express or implied.
# See the License for the specific language
governing permissions and
# limitations under the License.
# export
JAVA_HOME=/home/y/libexec/jdk1.6.0/
export
JAVA_HOME=/opt/module/jdk1.7.0_79
expor
tHADOOP_JOB_HISTORYSERVER_HEAPSIZE=1000
export HADOOP_MAPRED_ROOT_LOGGER=INFO,RFA
#export HADOOP_JOB_HISTORYSERVER_OPTS=
#export HADOOP_MAPRED_LOG_DIR=""
# Where log files are stored.
$HADOOP_MAPRED_HOME/logs by default.
#export HADOOP_JHS_LOGGER=INFO,RFA # Hadoop
JobSummary logger.
#export HADOOP_MAPRED_PID_DIR= # The pid
files are stored. /tmp by default.
#export HADOOP_MAPRED_IDENT_STRING= #A
string representing this instance of hadoop. $USER by default
#export HADOOP_MAPRED_NICENESS= #The
scheduling priority for daemons. Defaults to 0.
[cevent@hadoop210 hadoop-2.7.2]$ rm -rf logs/
[cevent@hadoop210 hadoop-2.7.2]$ bin/hdfs namenode -format 格式化namenode
[cevent@hadoop210 hadoop-2.7.2]$ sbin/start-dfs.sh
Starting namenodes on
[hadoop210.cevent.com]
hadoop210.cevent.com: starting namenode,
logging to /opt/module/hadoop-2.7.2/logs/hadoop-cevent-namenode-hadoop210.cevent.com.out
hadoop210.cevent.com: starting datanode,
logging to
/opt/module/hadoop-2.7.2/logs/hadoop-cevent-datanode-hadoop210.cevent.com.out
hadoop211.cevent.com: starting datanode,
logging to
/opt/module/hadoop-2.7.2/logs/hadoop-cevent-datanode-hadoop211.cevent.com.out
hadoop212.cevent.com: starting datanode,
logging to
/opt/module/hadoop-2.7.2/logs/hadoop-cevent-datanode-hadoop212.cevent.com.out
Starting secondary namenodes
[hadoop210.cevent.com]
hadoop210.cevent.com: starting
secondarynamenode, logging to
/opt/module/hadoop-2.7.2/logs/hadoop-cevent-secondarynamenode-hadoop210.cevent.com.out
[cevent@hadoop210 hadoop-2.7.2]$ jps
18471 DataNode
18338 NameNode
18638 SecondaryNameNode
18760 Jps
[cevent@hadoop210 hadoop-2.7.2]$ sbin/start-yarn.sh
starting yarn daemons
starting resourcemanager, logging to
/opt/module/hadoop-2.7.2/logs/yarn-cevent-resourcemanager-hadoop210.cevent.com.out
hadoop211.cevent.com: starting nodemanager,
logging to
/opt/module/hadoop-2.7.2/logs/yarn-cevent-nodemanager-hadoop211.cevent.com.out
hadoop210.cevent.com: starting nodemanager,
logging to
/opt/module/hadoop-2.7.2/logs/yarn-cevent-nodemanager-hadoop210.cevent.com.out
hadoop212.cevent.com: starting nodemanager,
logging to
/opt/module/hadoop-2.7.2/logs/yarn-cevent-nodemanager-hadoop212.cevent.com.out
[cevent@hadoop210 hadoop-2.7.2]$ jps
19202 Jps
18471 DataNode
18338 NameNode
18638 SecondaryNameNode
18811 ResourceManager
18919 NodeManager
[cevent@hadoop210 opt]$ mkdir soft
mkdir: 无法创建目录"soft": 权限不够
[cevent@hadoop210 opt]$ sudo vi /etc/sudoers
## Next comes the main part: which users
can run what software on
## which machines (the sudoers file can
be shared between multiple
## systems).
## Syntax:
##
##
user MACHINE=COMMANDS
##
## The COMMANDS section may have other
options added to it.
##
## Allow root to run any commands
anywhere
root ALL=(ALL) ALL
cevent ALL=(ALL) ALL
[cevent@hadoop210 opt]$ sudo chown -R cevent:cevent /opt/
[cevent@hadoop207 zookeeper-3.4.10]$ tar
-zxvf zookeeper-3.4.10.tar.gz -C /opt/module/
[cevent@hadoop207 soft]$ cd /opt/module/
[cevent@hadoop207 zookeeper-3.4.10]$ ll
总用量 1592
drwxr-xr-x.
2 cevent cevent 4096 3月
23 2017 bin
-rw-rw-r--.
1 cevent cevent 84725 3月
23 2017 build.xml
drwxr-xr-x.
2 cevent cevent 4096 6月
16 22:38 conf
drwxr-xr-x. 10 cevent cevent 4096 3月 23 2017 contrib
drwxrwxr-x.
3 cevent cevent 4096 6月
16 22:34 data
drwxr-xr-x.
2 cevent cevent 4096 3月
23 2017 dist-maven
drwxr-xr-x.
6 cevent cevent 4096 3月
23 2017 docs
-rw-rw-r--.
1 cevent cevent 1709 3月
23 2017 ivysettings.xml
-rw-rw-r--.
1 cevent cevent 5691 3月
23 2017 ivy.xml
drwxr-xr-x.
4 cevent cevent 4096 3月
23 2017 lib
-rw-rw-r--.
1 cevent cevent 11938 3月
23 2017 LICENSE.txt
-rw-rw-r--.
1 cevent cevent 3132 3月
23 2017 NOTICE.txt
-rw-rw-r--.
1 cevent cevent 1770 3月
23 2017 README_packaging.txt
-rw-rw-r--.
1 cevent cevent 1585 3月
23 2017 README.txt
drwxr-xr-x.
5 cevent cevent 4096 3月
23 2017 recipes
drwxr-xr-x.
8 cevent cevent 4096 3月
23 2017 src
-rw-rw-r--.
1 cevent cevent 1456729 3月 23 2017 zookeeper-3.4.10.jar
-rw-rw-r--.
1 cevent cevent 819 3月
23 2017 zookeeper-3.4.10.jar.asc
-rw-rw-r--.
1 cevent cevent 33 3月
23 2017 zookeeper-3.4.10.jar.md5
-rw-rw-r--.
1 cevent cevent 41 3月
23 2017 zookeeper-3.4.10.jar.sha1
[cevent@hadoop207 zookeeper-3.4.10]$ mv
conf/zoo_sample.cfg zoo.cfg
[cevent@hadoop207 zookeeper-3.4.10]$ mv zoo.cfg conf/
[cevent@hadoop207 zookeeper-3.4.10]$ mkdir
-p data/zkData
/opt/module/zookeeper-3.4.5/data/zkData
[cevent@hadoop207 zookeeper-3.4.10]$ vim
conf/zoo.cfg
# The number of snapshots to retain in
dataDir
#autopurge.snapRetainCount=3
# Purge task interval in hours
# Set to "0" to disable auto
purge feature
# The number of milliseconds of each tick
tickTime=2000
# The number of ticks that the initial
# synchronization phase can take
initLimit=10
# The number of ticks that can pass between
# sending a request and getting an
acknowledgement
syncLimit=5
# the directory where the snapshot is
stored.
# do not use /tmp for storage, /tmp here is
just
# example sakes.
dataDir=/opt/module/zookeeper-3.4.10/data/zkData
# the port at which the clients will
connect
clientPort=2181
# the maximum number of client connections.
# increase this if you need to handle more
clients
#maxClientCnxns=60
#
# Be sure to read the maintenance section
of the
# administrator guide before turning on
autopurge.
#
#
http://zookeeper.apache.org/doc/current/zookeeperAdmin.html#sc_maintenance
#
# The number of snapshots to retain in
dataDir
#autopurge.snapRetainCount=3
# Purge task interval in hours
# Set to "0" to disable auto
purge feature
#autopurge.purgeInterval=1
【执行同步】
[cevent@hadoop210 module]$ xsync zookeeper-3.4.10/
[cevent@hadoop207/208/209分别配置相应id zookeeper-3.4.10]$ vim
data/zkData/myid
207
208
209
[cevent@hadoop207 zookeeper-3.4.10]$
bin/zkServer.sh start
ZooKeeper JMX enabled by default
Using config:
/opt/module/zookeeper-3.4.10/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED
[cevent@hadoop207 zookeeper-3.4.10]$ jps
6134 QuorumPeerMain
6157 Jps
[cevent@hadoop207 zookeeper-3.4.10]$
bin/zkCli.sh
Connecting to localhost:2181
【报错】
[cevent@hadoop207 zookeeper-3.4.10]$
bin/zkServer.sh status
ZooKeeper JMX enabled by default
Using config:
/opt/module/zookeeper-3.4.10/bin/../conf/zoo.cfg
Error contacting service. It is probably
not running.
【解决】
|-关闭防火墙
暂时关闭防火墙:(立即生效,开机重启,会重新打开)
service iptables stop
永久关闭防火墙(关机重启才会生效)
chkconfig iptables off
重新解压zookeeper,在data/zkData下新建zookeeper_server.pid=进程ID
[cevent@hadoop207 zookeeper-3.4.10]$ jps
5449 QuorumPeerMain
5763 Jps
[cevent@hadoop207 zookeeper-3.4.10]$ cd
data/zkData/
[cevent@hadoop207 zkData]$ ll
总用量 4
drwxrwxr-x. 2 cevent cevent 4096 6月
17 11:54 version-2
[cevent@hadoop207 zkData]$ vim
zookeeper_server.pid
5449
【启动成功】
[cevent@hadoop207 zookeeper-3.4.10]$
bin/zkServer.sh status
ZooKeeper JMX enabled by default
Using config: /opt/module/zookeeper-3.4.10/bin/../conf/zoo.cfg
Mode: standalone
[zk: localhost:2181(CONNECTED) 1] ll
ZooKeeper -server host:port cmd args
connect host:port
get path [watch]
ls path [watch]
set path data [version]
rmr path
delquota [-n|-b] path
quit
printwatches on|off
create [-s] [-e] path data acl
stat path [watch]
close
ls2 path [watch]
history
listquota path
setAcl path acl
getAcl path
sync path
redo cmdno
addauth scheme auth
delete path [version]
setquota -n|-b val path
[zk: localhost:2181(CONNECTED) 2] quit
Quitting...
2020-06-17 12:00:36,210 [myid:] - INFO [main:ZooKeeper@684] - Session:
0x172c06c50490000 closed
2020-06-17 12:00:36,211 [myid:] - INFO [main-EventThread:ClientCnxn$EventThread@519]
- EventThread shut down for session: 0x172c06c50490000
[cevent@hadoop207 zookeeper-3.4.10]$ vim
conf/zoo.cfg
__>最后一行添加如下:
##################cluster#################
server.207=hadoop207.cevent.com:2888:3888
server.208=hadoop208.cevent.com:2888:3888
server.209=hadoop209.cevent.com:2888:3888
【分发】
[cevent@hadoop210 zookeeper-3.4.10]$ xsync conf/zoo.cfg
fname=zoo.cfg
pdir=/opt/module/zookeeper-3.4.10/conf
--------------- hadoop210.cevent.com ----------------
sending incremental file list
sent 30 bytes received 12
bytes 84.00 bytes/sec
total size is 1118 speedup is
26.62
--------------- hadoop211.cevent.com ----------------
sending incremental file list
zoo.cfg
sent 495 bytes received 43
bytes 358.67 bytes/sec
total size is 1118 speedup is
2.08
--------------- hadoop212.cevent.com ----------------
sending incremental file list
zoo.cfg
sent 495 bytes received 43
bytes 1076.00 bytes/sec
total size is 1118
speedup is 2.08
[cevent@hadoop207 hadoop-2.7.2]$
sbin/start-dfs.sh
[cevent@hadoop207 hadoop-2.7.2]$ sbin/start-yarn.sh
依次启动zookeeper,全部启动查看zkServer.sh status
[cevent@hadoop207 zookeeper-3.4.10]$
bin/zkServer.sh status
ZooKeeper JMX enabled by default
Using config: /opt/module/zookeeper-3.4.10/bin/../conf/zoo.cfg
Mode: follower
[cevent@hadoop208 zookeeper-3.4.10]$
bin/zkServer.sh status
ZooKeeper JMX enabled by default
Using config: /opt/module/zookeeper-3.4.10/bin/../conf/zoo.cfg
Mode: leader
[cevent@hadoop210 soft]$ tar -zxvf apache-hive-1.2.1-bin.tar.gz -C /opt/module/
[cevent@hadoop210 module]$ mv apache-hive-1.2.1-bin/ hive-1.2.1
[cevent@hadoop210 module]$ ll
总用量 16
drwxr-xr-x. 12 cevent cevent 4096 6月
28 15:21 hadoop-2.7.2
drwxrwxr-x.
8 cevent cevent 4096 6月 28 16:23 hive-1.2.1
drwxr-xr-x.
8 cevent cevent 4096 3月 24 09:14 jdk1.7.0_79
drwxr-xr-x. 11 cevent cevent 4096 6月
28 15:50 zookeeper-3.4.10
[cevent@hadoop210 module]$ vim /etc/profile
#JAVA_HOME
export JAVA_HOME=/opt/module/jdk1.7.0_79
export PATH=$PATH:$JAVA_HOME/bin
#HADOOP_HOME
export
HADOOP_HOME=/opt/module/hadoop-2.7.2
export PATH=$PATH:$HADOOP_HOME/bin
export PATH=$PATH:$HADOOP_HOME/sbin
#ZOOKEEPER_HOME
export
ZOOKEEPER_HOME=/opt/module/zookeeper-3.4.10
export PATH=$PATH:$ZOOKEEPER_HOME/bin
#HIVE_HOME
export HIVE_HOME=/opt/module/hive-1.2.1
export PATH=$PATH:$HIVE_HOME/bin
~
# Set HADOOP_HOME to point to a specific
hadoop install directory
# HADOOP_HOME=${bin}/../../hadoop
HADOOP_HOME=/opt/module/hadoop-2.7.2
# Hive Configuration Directory can be
controlled by:
# export HIVE_CONF_DIR=
export HIVE_CONF_DIR=/opt/module/hive-1.2.1/conf
# Folder containing extra ibraries
required for hive compilation/execution can be controlled by:
# export HIVE_AUX_JARS_PATH=
[cevent@hadoop210 hadoop-2.7.2]$ jps
24482 SecondaryNameNode
24646 ResourceManager
24761 NodeManager
25108 QuorumPeerMain
25142 Jps
24303 DataNode
24163 NameNode
[cevent@hadoop210 hive-1.2.1]$ hive
Logging initialized using configuration in
jar:file:/opt/module/hive-1.2.1/lib/hive-common-1.2.1.jar!/hive-log4j.properties
hive> create
database cevent01;
OK
Time taken: 0.954 seconds
hive> use
cevent01;
OK
Time taken: 0.062 seconds
hive> create
table c_student(id int,name string);
OK
Time taken: 0.567 seconds
hive> show
tables;
OK
c_student
Time taken: 0.194 seconds, Fetched: 1
row(s)
hive>
[cevent@hadoop207 soft]$ rpm -qa | grep mysql 查询默认安装的mysql依赖
mysql-libs-5.1.73-7.el6.x86_64
[cevent@hadoop207 soft]$ rpm -qa | grep mysql |
xargs sudo rpm -e --nodeps 直接删除-e 强制卸载--nodeps(忽略依赖关系)
Xargs标准输入流 作为参数 传送到管道 | grep 进行rpm删除
[sudo] password for cevent:
[cevent@hadoop207 soft]$ rpm -ivh MySQL-server-5.6.24-1.el6.x86_64.rpm 安装mySQL的服务端server
error: can't create transaction lock on
/var/lib/rpm/.rpm.lock (权限不够)
[cevent@hadoop207 soft]$ sudo rpm -ivh MySQL-server-5.6.24-1.el6.x86_64.rpm
2020-05-07 13:17:14 5020 [Note] InnoDB:
Shutdown completed; log sequence number 1625987
A RANDOM PASSWORD HAS BEEN SET FOR THE MySQL root USER ! 生成了一个随机密码在root下
You will find that password in '/root/.mysql_secret'.
You must change that password on your first
connect,
no other statement but 'SET PASSWORD' will be
accepted.
See the manual for the semantics of the
'password expired' flag.
Also, the account for the anonymous user
has been removed.
In addition, you can run:
/usr/bin/mysql_secure_installation 重新设置密码
which will also give you the option of
removing the test database.
This is strongly recommended for production
servers.
See the manual for more instructions.
Please report any problems at
http://bugs.mysql.com/
[cevent@hadoop210 soft]$ sudo cat /root/.mysql_secret
# The random password set for the root user
at Sun Jun 28 16:43:41 2020 (local time): zaSF9FWPlTD1C6Mv
[cevent@hadoop207 soft]$ sudo rpm -ivh MySQL-client-5.6.24-1.el6.x86_64.rpm 安装mySQL客户端
Preparing...
########################################### [100%]
1:MySQL-client
########################################### [100%]
[cevent@hadoop207 soft]$ sudo service mysql start 启动mySQL服务
Starting MySQL....... [确定]
[cevent@hadoop207 soft]$ mysql_secure_installation 重新设置密码
Enter current password for root (enter for
none):
ERROR 1045 (28000): Access denied for user
'root'@'localhost' (using password: YES)
Enter current password for root (enter for none): 输入random密码
OK, successfully used password, moving
on...
Setting the root password ensures that
nobody can log into the MySQL
root user without the proper authorisation.
You already have a root password set, so
you can safely answer 'n'.
Change the root password? [Y/n] y
New password:
Re-enter new password:
Password updated successfully!
Reloading privilege tables..
...
Success!
By default, a MySQL installation has an
anonymous user, allowing anyone
to log into MySQL without having to have a
user account created for
them.
This is intended only for testing, and to make the installation
go a bit smoother. You should remove them before moving into a
production environment.
Remove anonymous users? [Y/n] y
...
Success!
Normally, root should only be allowed to
connect from 'localhost'. This
ensures that someone cannot guess at the
root password from the network.
Disallow root login remotely? [Y/n] y
...
Success!
By default, MySQL comes with a database
named 'test' that anyone can
access.
This is also intended only for testing, and should be removed
before moving into a production
environment.
Remove test database and access to it? [Y/n] t
-
Dropping test database...
...
Success!
-
Removing privileges on test database...
...
Success!
Reloading the privilege tables will ensure
that all changes made so far
will take effect immediately.
Reload privilege tables now? [Y/n] y
...
Success!
All done!
If you've completed all of the above steps, your MySQL
installation should now be secure.
Thanks for using MySQL!
Cleaning up...
[cevent@hadoop207 soft]$ mysql -uroot -pcevent
进入mysql修改访问权限
Warning: Using a password on the command
line interface can be insecure.
Welcome to the MySQL monitor. Commands end with ; or \g.
Your MySQL connection id is 13
Server version: 5.6.24 MySQL Community
Server (GPL)
Copyright (c) 2000, 2015, Oracle and/or its
affiliates. All rights reserved.
Oracle is a registered trademark of Oracle
Corporation and/or its
affiliates. Other names may be trademarks
of their respective
owners.
Type 'help;' or '\h' for help. Type '\c' to
clear the current input statement.
mysql> show
databases; 显示数据库
ERROR 1064 (42000): You have an error in
your SQL syntax; check the manual that corresponds to your MySQL server version
for the right syntax to use near 'database' at line 1
mysql> show
databases;
+--------------------+
| Database |
+--------------------+
| information_schema |
| mysql |
| performance_schema |
+--------------------+
3 rows in set (0.00 sec)
mysql> use mysql
Reading table information for completion of
table and column names
You can turn off this feature to get a
quicker startup with -A
Database changed
mysql> show tables;
+---------------------------+
| Tables_in_mysql |
+---------------------------+
| columns_priv |
| db |
| event |
| func |
| general_log |
| help_category |
| help_keyword |
| help_relation |
| help_topic |
| innodb_index_stats |
| innodb_table_stats |
| ndb_binlog_index |
| plugin |
| proc |
| procs_priv |
| proxies_priv |
| servers |
| slave_master_info |
| slave_relay_log_info |
| slave_worker_info |
| slow_log |
| tables_priv |
| time_zone |
| time_zone_leap_second |
| time_zone_name |
| time_zone_transition |
| time_zone_transition_type |
| user |
+---------------------------+
28 rows in set (0.00 sec)
mysql> select user,host,password from
user;
+------+-----------+-------------------------------------------+
| user | host | password |
+------+-----------+-------------------------------------------+
| root | localhost |
*076AB7BA096BCDF76A8F0BD7987FFFCCD05C05F7 |
| root | 127.0.0.1 |
*076AB7BA096BCDF76A8F0BD7987FFFCCD05C05F7 |
| root | ::1 |
*076AB7BA096BCDF76A8F0BD7987FFFCCD05C05F7 |
+------+-----------+-------------------------------------------+
3 rows in set (0.00 sec)
mysql> delete from user where
host="127.0.0.1" or host="::1";
Query OK, 2 rows affected (0.00 sec)
mysql> select user,root,password from
user;
ERROR 1054 (42S22): Unknown column 'root'
in 'field list'
mysql> select user,host,password from
user;
+------+-----------+-------------------------------------------+
| user | host | password |
+------+-----------+-------------------------------------------+
| root | localhost |
*076AB7BA096BCDF76A8F0BD7987FFFCCD05C05F7 |
+------+-----------+-------------------------------------------+
1 row in set (0.00 sec)
mysql> update user set
host="%";
Query OK, 1 row affected (0.00 sec)
Rows matched: 1 Changed: 1
Warnings: 0
mysql> select
user,host,password from user; 查询当前用户(%可远程访问)
+------+------+-------------------------------------------+
| user | host | password |
+------+------+-------------------------------------------+
| root | % | *076AB7BA096BCDF76A8F0BD7987FFFCCD05C05F7
|
+------+------+-------------------------------------------+
1 row in set (0.00 sec)
mysql> flush
privileges; 刷新已配置的权限
Query OK, 0 rows affected (0.00 sec)
mysql> quit; 退出
Bye
[cevent@hadoop210 soft]$ cd
/opt/module/hive-1.2.1/
[cevent@hadoop210 hive-1.2.1]$ cd conf/
[cevent@hadoop210 conf]$ vim hive-site.xml
"1.0"?>
-stylesheet
type="text/xsl" href="configuration.xsl"?>
<!--mySQL数据库地址-->
javax.jdo.option.ConnectionURL</name>
jdbc:mysql://hadoop210.cevent.com:3306/metastore?createDatabaseIfNotExist=true</value>
JDBC connect string for a JDBC
metastore</description>
</property>
javax.jdo.option.ConnectionDriverName</name>
com.mysql.jdbc.Driver</value>
Driver class name for a JDBC
metastore</description>
</property>
<!--mySQL数据库访问用户名及密码-->
javax.jdo.option.ConnectionUserName</name>
root</value>
username to use against metastore
database</description>
</property>
javax.jdo.option.ConnectionPassword</name>
cevent</value>
password to use against metastore
database</description>
</property>
</configuration>
[cevent@hadoop207 conf]$ cd /opt/soft/
[cevent@hadoop207 soft]$ ll
总用量 507012
-rw-r--r--. 1 cevent cevent 92834839 4月 30 15:53
apache-hive-1.2.1-bin.tar.gz
-rw-r--r--. 1 cevent cevent 197657687 4月
30 13:23 hadoop-2.7.2.tar.gz
-rw-r--r--. 1 cevent cevent 153512879 4月
30 13:21 jdk-7u79-linux-x64.gz
-rw-r--r--. 1 cevent cevent 18509960 4月 30 15:56
MySQL-client-5.6.24-1.el6.x86_64.rpm
-rw-r--r--. 1 cevent cevent 872303 4月 30 15:56
mysql-connector-java-5.1.27-bin.jar
-rw-r--r--. 1 cevent cevent 55782196 4月 30 15:56
MySQL-server-5.6.24-1.el6.x86_64.rpm
[cevent@hadoop207 soft]$ cp mysql-connector-java-5.1.27-bin.jar
/opt/module/hive-1.2.1/lib/ 将connector拷贝到hive/lib下
[cevent@hadoop207 opt]$ cd module/hive-1.2.1/
[cevent@hadoop207 hive-1.2.1]$ ll
总用量 504
drwxrwxr-x. 3 cevent cevent 4096 4月 30 15:59 bin
drwxrwxr-x. 2 cevent cevent 4096 5月 7 13:49 conf
-rw-rw-r--. 1 cevent cevent 21062 4月 30 16:44 derby.log
drwxrwxr-x. 4 cevent cevent 4096 4月 30 15:59 examples
drwxrwxr-x. 7 cevent cevent 4096 4月 30 15:59 hcatalog
drwxrwxr-x. 4 cevent cevent 4096 5月 7 13:51 lib
-rw-rw-r--. 1 cevent cevent 24754 4月 30 2015 LICENSE
drwxrwxr-x. 5 cevent cevent 4096 4月 30 16:44 metastore_db
-rw-rw-r--. 1 cevent cevent 397 6月 19 2015 NOTICE
-rw-rw-r--. 1 cevent cevent 4366 6月 19 2015 README.txt
-rw-rw-r--. 1 cevent cevent 421129 6月
19 2015 RELEASE_NOTES.txt
drwxrwxr-x. 3 cevent cevent 4096 4月 30 15:59 scripts
[cevent@hadoop207 hive-1.2.1]$ hive
Logging initialized using configuration in
jar:file:/opt/module/hive-1.2.1/lib/hive-common-1.2.1.jar!/hive-log4j.properties
hive> create
database cevent01; 创建数据库
OK
Time taken: 1.22 seconds
hive> use
cevent01; 进入数据库
OK
Time taken: 0.048 seconds
hive> create
table student(id int,name string); 创建表
OK
Time taken: 1.212 seconds
hive> insert into ta
table tables tablesample tan(
hive> insert
into table student values(2020,"cevent"); 插入表
Query ID =
cevent_20200507135624_60770fe8-ce63-4514-be6f-960f1fafae70
Total jobs = 3
Launching Job 1 out of 3
Number of reduce tasks is set to 0 since
there's no reduce operator
Starting Job = job_1588830813660_0001,
Tracking URL = http://hadoop207.cevent.com:8088/proxy/application_1588830813660_0001/
Kill Command =
/opt/module/hadoop-2.7.2/bin/hadoop job
-kill job_1588830813660_0001
Hadoop job information for Stage-1: number
of mappers: 1; number of reducers: 0
2020-05-07 13:56:47,121 Stage-1 map =
0%, reduce = 0%
2020-05-07 13:56:57,747 Stage-1 map =
100%, reduce = 0%, Cumulative CPU 2.0
sec
MapReduce Total cumulative CPU time: 2 seconds
0 msec
Ended Job = job_1588830813660_0001
Stage-4 is selected by condition resolver.
Stage-3 is filtered out by condition
resolver.
Stage-5 is filtered out by condition
resolver.
Moving data to:
hdfs://hadoop207.cevent.com:8020/user/hive/warehouse/cevent01.db/student/.hive-staging_hive_2020-05-07_13-56-24_509_3865710950866541111-1/-ext-10000
Loading data to table cevent01.student
Table cevent01.student stats: [numFiles=1,
numRows=1, totalSize=12, rawDataSize=11]
MapReduce Jobs Launched:
Stage-Stage-1: Map: 1 Cumulative CPU: 2.0 sec HDFS Read: 3675 HDFS Write: 84 SUCCESS
Total MapReduce CPU Time Spent: 2 seconds 0
msec
OK
Time taken: 35.948 seconds
hive> select
* from student; 查询表
OK
2020
cevent
Time taken: 0.24 seconds, Fetched: 1 row(s)
hive> drop
table student; 删除表
OK
hive> drop
database cevent01; 删除数据库
OK
Time taken: 0.214 seconds
hive> [cevent@hadoop210 hive-1.2.1]$ ll
总用量 504
drwxrwxr-x. 3 cevent cevent 4096 6月 28 16:23 bin
drwxrwxr-x. 2 cevent cevent 4096 6月 28 16:56 conf
-rw-rw-r--. 1 cevent cevent 21061 6月 28 16:35 derby.log
drwxrwxr-x. 4 cevent cevent 4096 6月 28 16:23 examples
drwxrwxr-x. 7 cevent cevent 4096 6月 28 16:23 hcatalog
drwxrwxr-x. 4 cevent cevent 4096 6月 28 16:58 lib
-rw-rw-r--. 1 cevent cevent 24754 4月 30 2015 LICENSE
drwxrwxr-x. 5 cevent cevent 4096 6月 28 16:35 metastore_db
-rw-rw-r--. 1 cevent cevent 397 6月 19 2015 NOTICE
-rw-rw-r--. 1 cevent cevent 4366 6月 19 2015 README.txt
-rw-rw-r--. 1 cevent cevent 421129 6月
19 2015 RELEASE_NOTES.txt
drwxrwxr-x. 3 cevent cevent 4096 6月 28 16:23 scripts
[cevent@hadoop210 hive-1.2.1]$ bin/hiveserver2
OK
[cevent@hadoop210 hive-1.2.1]$ bin/beeline
Beeline version 1.2.1 by Apache Hive
beeline> !connect
jdbc:hive2://hadoop210.cevent.com:10000
Connecting to
jdbc:hive2://hadoop210.cevent.com:10000
Enter username for
jdbc:hive2://hadoop210.cevent.com:10000: cevent
Enter password for
jdbc:hive2://hadoop210.cevent.com:10000: ******
Connected to: Apache Hive (version 1.2.1)
Driver: Hive JDBC (version 1.2.1)
Transaction isolation:
TRANSACTION_REPEATABLE_READ
0:
jdbc:hive2://hadoop210.cevent.com:10000> show databases;
+----------------+--+
| database_name |
+----------------+--+
| default |
+----------------+--+
1 row selected (1.524 seconds)
[cevent@hadoop210 conf]$ tar -zxvf apache-flume-1.7.0-bin.tar.gz -C /opt/module/
[cevent@hadoop210 soft]$ cd /opt/module/
[cevent@hadoop210 module]$ ll
总用量 20
drwxrwxr-x. 7 cevent cevent 4096 6月 28 17:24 apache-flume-1.7.0-bin
drwxr-xr-x. 12 cevent cevent 4096 6月 28 15:21 hadoop-2.7.2
drwxrwxr-x. 9 cevent cevent 4096 6月 28 16:35 hive-1.2.1
drwxr-xr-x. 8 cevent cevent 4096 3月 24 09:14 jdk1.7.0_79
drwxr-xr-x. 11 cevent cevent 4096 6月 28 15:50 zookeeper-3.4.10
[cevent@hadoop210 module]$ mv apache-flume-1.7.0-bin/ flume-1.7.0
[cevent@hadoop210 module]$ ll
总用量 20
drwxrwxr-x. 7 cevent cevent 4096 6月 28 17:24 flume-1.7.0
drwxr-xr-x. 12 cevent cevent 4096 6月 28 15:21 hadoop-2.7.2
drwxrwxr-x. 9 cevent cevent 4096 6月 28 16:35 hive-1.2.1
drwxr-xr-x. 8 cevent cevent 4096 3月 24 09:14 jdk1.7.0_79
drwxr-xr-x. 11 cevent cevent 4096 6月 28 15:50 zookeeper-3.4.10
[cevent@hadoop210 module]$ cd flume-1.7.0/conf/
[cevent@hadoop210 conf]$ ll
总用量 16
-rw-r--r--. 1 cevent cevent 1661 9月 26 2016 flume-conf.properties.template
-rw-r--r--. 1 cevent cevent 1455 9月 26 2016 flume-env.ps1.template
-rw-r--r--. 1 cevent cevent 1565 9月 26 2016 flume-env.sh.template
-rw-r--r--. 1 cevent cevent 3107 9月 26 2016 log4j.properties
[cevent@hadoop210 conf]$ mv flume-env.sh.template flume-env.sh
[cevent@hadoop210 conf]$ vim flume-env.sh
# Licensed to the Apache Software
Foundation (ASF) under one
# or more contributor license
agreements. See the NOTICE file
# distributed with this work for
additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License,
Version 2.0 (the
# "License"); you may not use
this file except in compliance
# with the License. You may obtain a copy of the License at
#
#
http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or
agreed to in writing, software
# distributed under the License is
distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY
KIND, either express or implied.
# See the License for the specific
language governing permissions and
# limitations under the License.
# If this file is placed at FLUME_CONF_DIR/flume-env.sh,
it will be sourced
# during Flume startup.
# Enviroment variables can be set here.
# export
JAVA_HOME=/usr/lib/jvm/java-6-sun
export JAVA_HOME=/opt/module/jdk1.7.0_79
# Give Flume more memory and
pre-allocate, enable remote monitoring via JMX
# export JAVA_OPTS="-Xms100m
-Xmx2000m -Dcom.sun.management.jmxremote"
# Let Flume write raw event data and
configuration information to its log files for debugging
# purposes. Enabling these flags is not
recommended in production,
# as it may result in logging sensitive
user information or encryption secrets.
# export JAVA_OPTS="$JAVA_OPTS
-Dorg.apache.flume.log.rawdata=true -Dorg.apache.flume.log.printconfig=true
"
# Note that the Flume conf directory is
always included in the classpath.
#FLUME_CLASSPATH=""
[cevent@hadoop210 yum.repos.d]$ sudo vi CentOS-Base.repo
[sudo] password for cevent:
#contrib - packages by Centos Users
[contrib]
name=CentOS-$releasever - Contrib -
163.com
baseurl=http://mirrors.163.com/centos/$releasever/contrib/$basearch/
#mirrorlist=http://mirrorlist.centos.org/?release=$releasever&arch=$basearch&repo=contrib
gpgcheck=1
enabled=0
# CentOS-Base.repo
#
# The mirror system uses the connecting
IP address of the client and the
# update status of each mirror to pick
mirrors that are updated to and
# geographically close to the
client. You should use this for CentOS
updates
# unless you are manually picking other
mirrors.
#
# If the mirrorlist= does not work for
you, as a fall back you can try the
# remarked out baseurl= line instead.
#
#
[base]
name=CentOS-$releasever - Base - 163.com
baseurl=http://mirrors.163.com/centos/$releasever/os/$basearch/
#mirrorlist=file:///mnt/cdrom
gpgcheck=1
enable=1
gpgkey=http://mirror.centos.org/centos/RPM-GPG-KEY-CentOS-6
#released updates
[updates]
name=CentOS-$releasever - Updates - 163.com
baseurl=http://mirrors.163.com/centos/$releasever/updates/$basearch/
#mirrorlist=http://mirrorlist.centos.org/?release=$releasever&arch=$basearch&repo=updates
gpgcheck=1
gpgkey=http://mirror.centos.org/centos/RPM-GPG-KEY-CentOS-6
#additional packages that may be useful
[extras]
name=CentOS-$releasever - Extras - 163.com
baseurl=http://mirrors.163.com/centos/$releasever/extras/$basearch/
#mirrorlist=http://mirrorlist.centos.org/?release=$releasever&arch=$basearch&repo=extras
gpgcheck=1
gpgkey=http://mirror.centos.org/centos/RPM-GPG-KEY-CentOS-6
#additional packages that extend functionality of existing packages
[centosplus]
name=CentOS-$releasever - Plus - 163.com
baseurl=http://mirrors.163.com/centos/$releasever/centosplus/$basearch/
#mirrorlist=http://mirrorlist.centos.org/?release=$releasever&arch=$basearch&repo=centosplus
gpgcheck=1
enabled=0
gpgkey=http://mirror.centos.org/centos/RPM-GPG-KEY-CentOS-6
#contrib - packages by Centos Users
[contrib]
name=CentOS-$releasever - Contrib - 163.com
baseurl=http://mirrors.163.com/centos/$releasever/contrib/$basearch/
#mirrorlist=http://mirrorlist.centos.org/?release=$releasever&arch=$basearch&repo=contrib
gpgcheck=1
enabled=0
gpgkey=http://mirror.centos.org/centos/RPM-GPG-KEY-CentOS-6
[cevent@hadoop210 yum.repos.d]$ sudo yum install -y nc
已加载插件:fastestmirror, refresh-packagekit, security
设置安装进程
Loading mirror speeds from cached
hostfile
base
| 3.7 kB 00:00
base/primary_db
| 4.7 MB 00:07
extras
| 3.4 kB 00:00
updates
| 3.4 kB 00:00
解决依赖关系
[cevent@hadoop207 ~]$
cd /opt/module/apache-flume-1.7.0/
[cevent@hadoop207 apache-flume-1.7.0]$ ll
总用量 148
drwxr-xr-x.
2 cevent cevent 4096 6月 11
13:35 bin
-rw-r--r--.
1 cevent cevent 77387 10月 11 2016
CHANGELOG
drwxr-xr-x.
2 cevent cevent 4096 6月 11
13:42 conf
-rw-r--r--.
1 cevent cevent 6172 9月 26
2016 DEVNOTES
-rw-r--r--.
1 cevent cevent 2873 9月 26
2016 doap_Flume.rdf
drwxr-xr-x. 10 cevent cevent 4096 10月 13 2016 docs
drwxrwxr-x.
2 cevent cevent 4096 6月 11
13:35 lib
-rw-r--r--.
1 cevent cevent 27625 10月 13 2016
LICENSE
-rw-r--r--.
1 cevent cevent 249 9月 26
2016 NOTICE
-rw-r--r--.
1 cevent cevent 2520 9月 26
2016 README.md
-rw-r--r--.
1 cevent cevent 1585 10月 11 2016 RELEASE-NOTES
drwxrwxr-x.
2 cevent cevent 4096 6月 11
13:35 tools
[cevent@hadoop207 apache-flume-1.7.0]$ mkdir job
[cevent@hadoop207 apache-flume-1.7.0]$ ll
总用量 152
drwxr-xr-x.
2 cevent cevent 4096 6月 11
13:35 bin
-rw-r--r--.
1 cevent cevent 77387 10月 11 2016
CHANGELOG
drwxr-xr-x.
2 cevent cevent 4096 6月 11
13:42 conf
-rw-r--r--.
1 cevent cevent 6172 9月 26
2016 DEVNOTES
-rw-r--r--.
1 cevent cevent 2873 9月 26
2016 doap_Flume.rdf
drwxr-xr-x. 10 cevent cevent 4096 10月 13 2016 docs
drwxrwxr-x.
2 cevent cevent 4096 6月 11
16:52 job
drwxrwxr-x.
2 cevent cevent 4096 6月 11
13:35 lib
-rw-r--r--.
1 cevent cevent 27625 10月 13 2016
LICENSE
-rw-r--r--.
1 cevent cevent 249 9月 26
2016 NOTICE
-rw-r--r--.
1 cevent cevent 2520 9月 26
2016 README.md
-rw-r--r--.
1 cevent cevent 1585 10月 11 2016 RELEASE-NOTES
drwxrwxr-x.
2 cevent cevent 4096 6月 11
13:35 tools
[cevent@hadoop207 apache-flume-1.7.0]$ vim
job/flume-netcat-logger.conf
#
Name the components on this agent
a1.sources
= r1
a1.sinks
= k1
a1.channels
= c1
#
Describe/configure the source
a1.sources.r1.type
= netcat
a1.sources.r1.bind
= localhost
a1.sources.r1.port
= 44444
#
Describe the sink
a1.sinks.k1.type
= logger
#
Use a channel which buffers events in memory
a1.channels.c1.type
= memory
a1.channels.c1.capacity
= 1000
a1.channels.c1.transactionCapacity
= 100
#
Bind the source and sink to the channel
a1.sources.r1.channels
= c1
a1.sinks.k1.channel
= c1
~
"job/flume-netcat-logger.conf" [新] 22L, 495C 已写入
[cevent@hadoop207 apache-flume-1.7.0]$ sudo netstat -nlp | grep 44444 查询端口是否被占用netstat
-nlp | grep
[sudo] password for cevent:
[cevent@hadoop207 apache-flume-1.7.0]$ ll bin
总用量 36
-rwxr-xr-x. 1 cevent cevent 12387 9月 26
2016 flume-ng
-rw-r--r--. 1 cevent cevent 936 9月 26 2016 flume-ng.cmd
-rwxr-xr-x. 1 cevent cevent 14176 9月 26
2016 flume-ng.ps1
[cevent@hadoop207 apache-flume-1.7.0]$ bin/flume-ng agent --name a1 --conf conf/ --conf-file job/flume-netcat-logger.conf
启动bin 设置flume name
配置文件 自定义配置文件路径
Info: Sourcing environment configuration script
/opt/module/apache-flume-1.7.0/conf/flume-env.sh
Info: Including Hadoop libraries found via
(/opt/module/hadoop-2.7.2/bin/hadoop) for HDFS access
Info: Including Hive libraries found via
(/opt/module/hive-1.2.1) for Hive access
+ exec /opt/module/jdk1.7.0_79/bin/java -Xmx20m
-cp
'/opt/module/apache-flume-1.7.0/conf:/opt/module/apache-flume-1.7.0/lib/*:/opt/module/hadoop-2.7.2/etc/hadoop:/opt/module/hadoop-2.7.2/share/hadoop/common/lib/*:/opt/module/hadoop-2.7.2/share/hadoop/common/*:/opt/module/hadoop-2.7.2/share/hadoop/hdfs:/opt/module/hadoop-2.7.2/share/hadoop/hdfs/lib/*:/opt/module/hadoop-2.7.2/share/hadoop/hdfs/*:/opt/module/hadoop-2.7.2/share/hadoop/yarn/lib/*:/opt/module/hadoop-2.7.2/share/hadoop/yarn/*:/opt/module/hadoop-2.7.2/share/hadoop/mapreduce/lib/*:/opt/module/hadoop-2.7.2/share/hadoop/mapreduce/*:/opt/module/hadoop-2.7.2/contrib/capacity-scheduler/*.jar:/opt/module/hive-1.2.1/lib/*'
-Djava.library.path=:/opt/module/hadoop-2.7.2/lib/native
org.apache.flume.node.Application --name a1 --conf-file
job/flume-netcat-logger.conf
SLF4J: Class path contains multiple SLF4J
bindings.
SLF4J: Found binding in [jar:file:/opt/module/apache-flume-1.7.0/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/opt/module/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings
for an explanation.
[cevent@hadoop207 apache-flume-1.7.0]$ cd conf/
[cevent@hadoop207 conf]$ ll
总用量 16
-rw-r--r--. 1 cevent cevent 1661 9月 26 2016 flume-conf.properties.template
-rw-r--r--. 1 cevent cevent 1455 9月 26 2016 flume-env.ps1.template
-rw-r--r--. 1 cevent cevent 1563 6月 11 13:41 flume-env.sh
-rw-r--r--. 1 cevent cevent 3107 9月 26 2016 log4j.properties
[cevent@hadoop207 conf]$ vim
log4j.properties
#
# Licensed to the Apache Software
Foundation (ASF) under one
# or more contributor license
agreements. See the NOTICE file
# distributed with this work for
additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License,
Version 2.0 (the
# "License"); you may not use
this file except in compliance
# with the License. You may obtain a copy of the License at
#
#
http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or
agreed to in writing,
# software distributed under the License
is distributed on an
# "AS IS" BASIS, WITHOUT
WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions
and limitations
# under the License.
#
# Define some default values that can be
overridden by system properties.
#
# For testing, it may also be convenient
to specify
# -Dflume.root.logger=DEBUG,console when
launching flume.
#flume.root.logger=DEBUG,console
# 默认输出日志在logfile,而不是console
flume.root.logger=INFO,LOGFILE
#flume.log.dir=./logs
flume.log.dir=/opt/module/apache-flume-1.7.0/loggers
flume.log.file=flume.log
log4j.logger.org.apache.flume.lifecycle =
INFO
log4j.logger.org.jboss = WARN
log4j.logger.org.mortbay = INFO
log4j.logger.org.apache.avro.ipc.NettyTransceiver
= WARN
log4j.logger.org.apache.hadoop = INFO
log4j.logger.org.apache.hadoop.hive =
ERROR
# Define the root logger to the system
property "flume.root.logger".
log4j.rootLogger=${flume.root.logger}
# 将所有日志放到flume.log.dir一个目录内 Stock log4j rolling file
appender
# Default log rotation configuration
log4j.appender.LOGFILE=org.apache.log4j.RollingFileAppender
log4j.appender.LOGFILE.MaxFileSize=100MB
log4j.appender.LOGFILE.MaxBackupIndex=10
log4j.appender.LOGFILE.File=${flume.log.dir}/${flume.log.file}
log4j.appender.LOGFILE.layout=org.apache.log4j.PatternLayout
log4j.appender.LOGFILE.layout.ConversionPattern=%d{dd
MMM yyyy HH:mm:ss,SSS} %-5p [%t] (%C.%M:%L) %x - %m%n
# Warning: If you enable the following
appender it will fill up your disk if you don't have a cleanup job!
# This uses the updated rolling file
appender from log4j-extras that supports a reliable time-based rolling
policy.
# See
http://logging.apache.org/log4j/companions/extras/apidocs/org/apache/log4j/rolling/TimeBasedRollingPolicy.html
# 每天日志按天滚动 Add "DAILY" to flume.root.logger
above if you want to use this
log4j.appender.DAILY=org.apache.log4j.rolling.RollingFileAppender
log4j.appender.DAILY.rollingPolicy=org.apache.log4j.rolling.TimeBasedRollingPolicy
log4j.appender.DAILY.rollingPolicy.ActiveFileName=${flume.log.dir}/${flume.log.file}
log4j.appender.DAILY.rollingPolicy.FileNamePattern=${flume.log.dir}/${flume.log.file}.%d{yyyy-MM-dd}
log4j.appender.DAILY.layout=org.apache.log4j.PatternLayout
log4j.appender.DAILY.layout.ConversionPattern=%d{dd
MMM yyyy HH:mm:ss,SSS} %-5p [%t] (%C.%M:%L) %x - %m%n
# 控制台输出console
# Add "console" to
flume.root.logger above if you want to use this
log4j.appender.console=org.apache.log4j.ConsoleAppender
log4j.appender.console.target=System.err
log4j.appender.console.layout=org.apache.log4j.PatternLayout
[cevent@hadoop207 apache-flume-1.7.0]$ sudo netstat -nlp | grep 44444 查看端口是否被占用netstat -nlp | grep 44444
[sudo] password for cevent:
[cevent@hadoop207 apache-flume-1.7.0]$ ll bin
总用量 36
-rwxr-xr-x. 1 cevent cevent 12387 9月 26 2016 flume-ng
-rw-r--r--. 1 cevent cevent 936 9月 26 2016 flume-ng.cmd
-rwxr-xr-x. 1 cevent cevent 14176 9月 26 2016 flume-ng.ps1
[cevent@hadoop207 apache-flume-1.7.0]$ bin/flume-ng agent --name a1 --conf conf/ --conf-file
job/flume-netcat-logger.conf
启动agent flume名 配置文件 自定义日志配置
Info: Sourcing environment configuration
script /opt/module/apache-flume-1.7.0/conf/flume-env.sh
Info: Including Hadoop libraries found
via (/opt/module/hadoop-2.7.2/bin/hadoop) for HDFS access
Info: Including Hive libraries found via
(/opt/module/hive-1.2.1) for Hive access
+ exec /opt/module/jdk1.7.0_79/bin/java
-Xmx20m -cp
'/opt/module/apache-flume-1.7.0/conf:/opt/module/apache-flume-1.7.0/lib/*:/opt/module/hadoop-2.7.2/etc/hadoop:/opt/module/hadoop-2.7.2/share/hadoop/common/lib/*:/opt/module/hadoop-2.7.2/share/hadoop/common/*:/opt/module/hadoop-2.7.2/share/hadoop/hdfs:/opt/module/hadoop-2.7.2/share/hadoop/hdfs/lib/*:/opt/module/hadoop-2.7.2/share/hadoop/hdfs/*:/opt/module/hadoop-2.7.2/share/hadoop/yarn/lib/*:/opt/module/hadoop-2.7.2/share/hadoop/yarn/*:/opt/module/hadoop-2.7.2/share/hadoop/mapreduce/lib/*:/opt/module/hadoop-2.7.2/share/hadoop/mapreduce/*:/opt/module/hadoop-2.7.2/contrib/capacity-scheduler/*.jar:/opt/module/hive-1.2.1/lib/*'
-Djava.library.path=:/opt/module/hadoop-2.7.2/lib/native
org.apache.flume.node.Application --name a1 --conf-file job/flume-netcat-logger.conf
SLF4J: Class path contains multiple SLF4J
bindings.
SLF4J: Found binding in
[jar:file:/opt/module/apache-flume-1.7.0/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/opt/module/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See
http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
[cevent@hadoop207 apache-flume-1.7.0]$ cd conf/
[cevent@hadoop207 conf]$ ll
总用量 16
-rw-r--r--. 1 cevent cevent 1661 9月 26 2016 flume-conf.properties.template
-rw-r--r--. 1 cevent cevent 1455 9月 26 2016 flume-env.ps1.template
-rw-r--r--. 1 cevent cevent 1563 6月 11 13:41 flume-env.sh
-rw-r--r--. 1 cevent cevent 3107 9月 26 2016 log4j.properties
[cevent@hadoop207 conf]$ vim log4j.properties 编辑日志配置
#
# Licensed to the Apache Software
Foundation (ASF) under one
# or more contributor license
agreements. See the NOTICE file
# distributed with this work for
additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License,
Version 2.0 (the
# "License"); you may not use
this file except in compliance
# with the License. You may obtain a copy of the License at
#
#
http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or
agreed to in writing,
# software distributed under the License
is distributed on an
# "AS IS" BASIS, WITHOUT
WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions
and limitations
# under the License.
#
# Define some default values that can be
overridden by system properties.
#
# For testing, it may also be convenient
to specify
# -Dflume.root.logger=DEBUG,console when
launching flume.
#flume.root.logger=DEBUG,console
# cevent:default output log to logfile,not console
flume.root.logger=INFO,LOGFILE
#flume.log.dir=./logs
flume.log.dir=/opt/module/apache-flume-1.7.0/loggers
flume.log.file=flume.log
log4j.logger.org.apache.flume.lifecycle =
INFO
log4j.logger.org.jboss = WARN
log4j.logger.org.mortbay = INFO
log4j.logger.org.apache.avro.ipc.NettyTransceiver
= WARN
log4j.logger.org.apache.hadoop = INFO
log4j.logger.org.apache.hadoop.hive =
ERROR
# Define the root logger to the system
property "flume.root.logger".
log4j.rootLogger=${flume.root.logger}
# cevent:all log put on flume.log.dir. Stock log4j rolling file
appender
# Default log rotation configuration
log4j.appender.LOGFILE=org.apache.log4j.RollingFileAppender
log4j.appender.LOGFILE.MaxFileSize=100MB
log4j.appender.LOGFILE.MaxBackupIndex=10
log4j.appender.LOGFILE.File=${flume.log.dir}/${flume.log.file}
log4j.appender.LOGFILE.layout=org.apache.log4j.PatternLayout
log4j.appender.LOGFILE.layout.ConversionPattern=%d{dd
MMM yyyy HH:mm:ss,SSS} %-5p [%t] (%C.%M:%L) %x - %m%n
# Warning: If you enable the following
appender it will fill up your disk if you don't have a cleanup job!
# This uses the updated rolling file
appender from log4j-extras that supports a reliable time-based rolling
policy.
# See http://logging.apache.org/log4j/companions/extras/apidocs/org/apache/log4j/rolling/TimeBasedRollingPolicy.html
# cevent:log scall every day.
Add "DAILY" to flume.root.logger above if you want to use
this
log4j.appender.DAILY=org.apache.log4j.rolling.RollingFileAppender
log4j.appender.DAILY.rollingPolicy=org.apache.log4j.rolling.TimeBasedRollingPolicy
log4j.appender.DAILY.rollingPolicy.ActiveFileName=${flume.log.dir}/${flume.log.file}
log4j.appender.DAILY.rollingPolicy.FileNamePattern=${flume.log.dir}/${flume.log.file}.%d{yyyy-MM-dd}
log4j.appender.DAILY.layout=org.apache.log4j.PatternLayout
log4j.appender.DAILY.layout.ConversionPattern=%d{dd
MMM yyyy HH:mm:ss,SSS} %-5p [%t] (%C.%M:%L) %x - %m%n
# console
# Add "console" to
flume.root.logger above if you want to use this
log4j.appender.console=org.apache.log4j.ConsoleAppender
log4j.appender.console.target=System.err
"log4j.properties" 71L, 3281C 已写入
[cevent@hadoop207 conf]$ ll
总用量 16
-rw-r--r--. 1 cevent cevent 1661 9月 26 2016 flume-conf.properties.template
-rw-r--r--. 1 cevent cevent 1455 9月 26 2016 flume-env.ps1.template
-rw-r--r--. 1 cevent cevent 1563 6月 11 13:41 flume-env.sh
-rw-r--r--. 1 cevent cevent 3281 6月 11 17:18 log4j.properties
[cevent@hadoop207 conf]$ cd ..
[cevent@hadoop207 apache-flume-1.7.0]$
bin/flume-ng agent
--name a1 --conf conf/ --conf-file job/flume-netcat-logger.conf
-Dflume.root.logger=INFO,console
设置控制台输出logo
-Dflume.root.logger=INFO,console
Info: Sourcing environment configuration script
/opt/module/apache-flume-1.7.0/conf/flume-env.sh
Info: Including Hadoop libraries found via (/opt/module/hadoop-2.7.2/bin/hadoop)
for HDFS access
Info: Including Hive libraries found via (/opt/module/hive-1.2.1) for
Hive access
+ exec /opt/module/jdk1.7.0_79/bin/java -Xmx20m
-Dflume.root.logger=INFO,console -cp '/opt/module/apache-flume-1.7.0/conf:/opt/module/apache-flume-1.7.0/lib/*:/opt/module/hadoop-2.7.2/etc/hadoop:/opt/module/hadoop-2.7.2/share/hadoop/common/lib/*:/opt/module/hadoop-2.7.2/share/hadoop/common/*:/opt/module/hadoop-2.7.2/share/hadoop/hdfs:/opt/module/hadoop-2.7.2/share/hadoop/hdfs/lib/*:/opt/module/hadoop-2.7.2/share/hadoop/hdfs/*:/opt/module/hadoop-2.7.2/share/hadoop/yarn/lib/*:/opt/module/hadoop-2.7.2/share/hadoop/yarn/*:/opt/module/hadoop-2.7.2/share/hadoop/mapreduce/lib/*:/opt/module/hadoop-2.7.2/share/hadoop/mapreduce/*:/opt/module/hadoop-2.7.2/contrib/capacity-scheduler/*.jar:/opt/module/hive-1.2.1/lib/*'
-Djava.library.path=:/opt/module/hadoop-2.7.2/lib/native
org.apache.flume.node.Application --name a1 --conf-file
job/flume-netcat-logger.conf
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in
[jar:file:/opt/module/apache-flume-1.7.0/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/opt/module/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
explanation.
2020-06-11 17:20:04,619 (lifecycleSupervisor-1-0) [INFO -
org.apache.flume.node.PollingPropertiesFileConfigurationProvider.start(PollingPropertiesFileConfigurationProvider.java:62)]
Configuration provider starting
2020-06-11 17:20:04,623 (conf-file-poller-0) [INFO -
org.apache.flume.node.PollingPropertiesFileConfigurationProvider$FileWatcherRunnable.run(PollingPropertiesFileConfigurationProvider.java:134)]
Reloading configuration file:job/flume-netcat-logger.conf
2020-06-11 17:20:04,629 (conf-file-poller-0) [INFO -
org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:930)]
Added sinks: k1 Agent: a1
2020-06-11 17:20:04,629 (conf-file-poller-0) [INFO -
org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:1016)]
Processing:k1
2020-06-11 17:20:04,629 (conf-file-poller-0) [INFO -
org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:1016)]
Processing:k1
2020-06-11 17:20:04,640 (conf-file-poller-0) [INFO -
org.apache.flume.conf.FlumeConfiguration.validateConfiguration(FlumeConfiguration.java:140)]
Post-validation flume configuration contains configuration for agents: [a1]
2020-06-11 17:20:04,641 (conf-file-poller-0) [INFO -
org.apache.flume.node.AbstractConfigurationProvider.loadChannels(AbstractConfigurationProvider.java:147)]
Creating channels
2020-06-11 17:20:04,653 (conf-file-poller-0) [INFO -
org.apache.flume.channel.DefaultChannelFactory.create(DefaultChannelFactory.java:42)]
Creating instance of channel c1 type memory
2020-06-11 17:20:04,656 (conf-file-poller-0) [INFO -
org.apache.flume.node.AbstractConfigurationProvider.loadChannels(AbstractConfigurationProvider.java:201)]
Created channel c1
2020-06-11 17:20:04,657 (conf-file-poller-0) [INFO -
org.apache.flume.source.DefaultSourceFactory.create(DefaultSourceFactory.java:41)]
Creating instance of source r1, type netcat
2020-06-11 17:20:04,666 (conf-file-poller-0) [INFO -
org.apache.flume.sink.DefaultSinkFactory.create(DefaultSinkFactory.java:42)]
Creating instance of sink: k1, type: logger
2020-06-11 17:20:04,669 (conf-file-poller-0) [INFO - org.apache.flume.node.AbstractConfigurationProvider.getConfiguration(AbstractConfigurationProvider.java:116)]
Channel c1 connected to [r1, k1]
2020-06-11 17:20:04,675 (conf-file-poller-0) [INFO -
org.apache.flume.node.Application.startAllComponents(Application.java:137)] Starting
new configuration:{ sourceRunners:{r1=EventDrivenSourceRunner: {
source:org.apache.flume.source.NetcatSource{name:r1,state:IDLE} }}
sinkRunners:{k1=SinkRunner: {
policy:org.apache.flume.sink.DefaultSinkProcessor@6a8bda08 counterGroup:{
name:null counters:{} } }}
channels:{c1=org.apache.flume.channel.MemoryChannel{name: c1}} }
2020-06-11 17:20:04,682 (conf-file-poller-0) [INFO -
org.apache.flume.node.Application.startAllComponents(Application.java:144)]
Starting Channel c1
2020-06-11 17:20:04,724 (lifecycleSupervisor-1-0) [INFO -
org.apache.flume.instrumentation.MonitoredCounterGroup.register(MonitoredCounterGroup.java:119)]
Monitored counter group for type: CHANNEL, name: c1: Successfully registered
new MBean.
2020-06-11 17:20:04,725 (lifecycleSupervisor-1-0) [INFO -
org.apache.flume.instrumentation.MonitoredCounterGroup.start(MonitoredCounterGroup.java:95)]
Component type: CHANNEL, name: c1 started
2020-06-11 17:20:04,726 (conf-file-poller-0) [INFO -
org.apache.flume.node.Application.startAllComponents(Application.java:171)]
Starting Sink k1
2020-06-11 17:20:04,727 (conf-file-poller-0) [INFO -
org.apache.flume.node.Application.startAllComponents(Application.java:182)]
Starting Source r1
2020-06-11 17:20:04,728 (lifecycleSupervisor-1-4) [INFO - org.apache.flume.source.NetcatSource.start(NetcatSource.java:155)]
Source starting
2020-06-11 17:20:04,762 (lifecycleSupervisor-1-4) [INFO -
org.apache.flume.source.NetcatSource.start(NetcatSource.java:169)] Created
serverSocket:sun.nio.ch.ServerSocketChannelImpl[/127.0.0.1:44444]
2020-06-11 17:21:00,410
(SinkRunner-PollingRunner-DefaultSinkProcessor) [INFO -
org.apache.flume.sink.LoggerSink.process(LoggerSink.java:95)] Event: {
headers:{} body: 63 65 76 65 6E 74 cevent }
2020-06-11 17:21:30,419
(SinkRunner-PollingRunner-DefaultSinkProcessor) [INFO -
org.apache.flume.sink.LoggerSink.process(LoggerSink.java:95)] Event: {
headers:{} body: 65 63 68 6F echo }
^C2020-06-11 17:23:55,629 (agent-shutdown-hook) [INFO -
org.apache.flume.lifecycle.LifecycleSupervisor.stop(LifecycleSupervisor.java:78)]
Stopping lifecycle supervisor 10
2020-06-11 17:23:55,630 (agent-shutdown-hook) [INFO -
org.apache.flume.node.PollingPropertiesFileConfigurationProvider.stop(PollingPropertiesFileConfigurationProvider.java:84)]
Configuration provider stopping
2020-06-11 17:23:55,631 (agent-shutdown-hook) [INFO -
org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:149)]
Component type: CHANNEL, name: c1 stopped
2020-06-11 17:23:55,631 (agent-shutdown-hook) [INFO -
org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:155)]
Shutdown Metric for type: CHANNEL, name: c1. channel.start.time ==
1591867204725
2020-06-11 17:23:55,631 (agent-shutdown-hook) [INFO -
org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:161)]
Shutdown Metric for type: CHANNEL, name: c1. channel.stop.time ==
1591867435631
2020-06-11 17:23:55,632 (agent-shutdown-hook) [INFO -
org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:177)]
Shutdown Metric for type: CHANNEL, name: c1. channel.capacity == 1000
2020-06-11 17:23:55,632 (agent-shutdown-hook) [INFO -
org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:177)]
Shutdown Metric for type: CHANNEL, name: c1. channel.current.size == 0
2020-06-11 17:23:55,632 (agent-shutdown-hook) [INFO -
org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:177)]
Shutdown Metric for type: CHANNEL, name: c1. channel.event.put.attempt == 2
2020-06-11 17:23:55,632 (agent-shutdown-hook) [INFO -
org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:177)]
Shutdown Metric for type: CHANNEL, name: c1. channel.event.put.success == 2
2020-06-11 17:23:55,632 (agent-shutdown-hook) [INFO -
org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:177)]
Shutdown Metric for type: CHANNEL, name: c1. channel.event.take.attempt == 35
2020-06-11 17:23:55,632 (agent-shutdown-hook) [INFO -
org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:177)]
Shutdown Metric for type: CHANNEL, name: c1. channel.event.take.success == 2
2020-06-11 17:23:55,632 (agent-shutdown-hook) [INFO -
org.apache.flume.source.NetcatSource.stop(NetcatSource.java:196)] Source
stopping
【nc传输】
[cevent@hadoop207 ~]$ nc localhost 44444
cevent
OK
echo
OK
[cevent@hadoop210 soft]$ tar -zxvf kafka_2.11-0.11.0.0.tgz
[cevent@hadoop210 soft]$ mv kafka_2.11-0.11.0.0 /opt/module/
[cevent@hadoop210 soft]$ cd /opt/module/
[cevent@hadoop210 module]$ ll
总用量 24
drwxrwxr-x. 9 cevent cevent 4096 6月 28 18:04 flume-1.7.0
drwxr-xr-x. 12 cevent cevent 4096 6月 28 15:21 hadoop-2.7.2
drwxrwxr-x. 9 cevent cevent 4096 6月 28 16:35 hive-1.2.1
drwxr-xr-x. 8 cevent cevent 4096 3月 24 09:14 jdk1.7.0_79
drwxr-xr-x. 6 cevent cevent 4096 6月 23 2017 kafka_2.11-0.11.0.0
drwxr-xr-x. 11 cevent cevent 4096 6月 28 15:50 zookeeper-3.4.10
[cevent@hadoop210 module]$ cd
kafka_2.11-0.11.0.0/
[cevent@hadoop210 kafka_2.11-0.11.0.0]$
ll
总用量 52
drwxr-xr-x. 3 cevent cevent 4096 6月 23 2017 bin
drwxr-xr-x. 2 cevent cevent 4096 6月 23 2017 config
drwxr-xr-x. 2 cevent cevent 4096 6月 28 18:18 libs
-rw-r--r--. 1 cevent cevent 28824 6月 23 2017 LICENSE
-rw-r--r--. 1 cevent cevent 336 6月 23 2017 NOTICE
drwxr-xr-x. 2 cevent cevent 4096 6月 23 2017 site-docs
[cevent@hadoop210 kafka_2.11-0.11.0.0]$
vim config/server.properties
# Licensed to the Apache Software
Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding
copyright ownership.
# The ASF licenses this file to You under
the Apache License, Version 2.0
# (the "License"); you may not
use this file except in compliance with
# the License. You may obtain a copy of the License at
#
#
http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or
agreed to in writing, software
# distributed under the License is
distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY
KIND, either express or implied.
# See the License for the specific
language governing permissions and
# limitations under the License.
# see kafka.server.KafkaConfig for
additional details and defaults
############################# Server
Basics #############################
# The id of the broker. This must be set
to a unique integer for each broker.
broker.id=210
# Switch to enable topic deletion or not,
default value is false
#delete.topic.enable=true
############################# Socket
Server Settings #############################
# The address the socket server listens
on. It will get the value returned from
#
java.net.InetAddress.getCanonicalHostName() if not configured.
#
FORMAT:
#
listeners = listener_name://host_name:port
#
EXAMPLE:
#
listeners = PLAINTEXT://your.host.name:9092
#listeners=PLAINTEXT://:9092
# Hostname and port the broker will
advertise to producers and consumers. If not set,
# it uses the value for
"listeners" if configured.
Otherwise, it will use the value
# returned from java.net.InetAddress.getCanonicalHostName().
#advertised.listeners=PLAINTEXT://your.host.name:9092
# Maps listener names to security
protocols, the default is for them to be the same. See the config
documentation for more details
#listener.security.protocol.map=PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSL
# The number of threads that the server
uses for receiving requests from the network and sending responses to the
network
num.network.threads=3
# The number of threads that the server
uses for processing requests, which may include disk I/O
num.io.threads=8
# The send buffer (SO_SNDBUF) used by the
socket server
socket.send.buffer.bytes=102400
# The receive buffer (SO_RCVBUF) used by
the socket server
socket.receive.buffer.bytes=102400
# The maximum size of a request that the
socket server will accept (protection against OOM)
socket.request.max.bytes=104857600
############################# Log Basics #############################
# A comma seperated list of directories
under which to store log files
log.dirs=/opt/module/kafka_2.11-0.11.0.0/logs
# The default number of log partitions
per topic. More partitions allow greater
# parallelism for consumption, but this
will also result in more files across
# the brokers.
num.partitions=1
# The number of threads per data
directory to be used for log recovery at startup and flushing at shutdown.
# This value is recommended to be
increased for installations with data dirs located in RAID array.
num.recovery.threads.per.data.dir=1
############################# Internal
Topic Settings
#############################
# The replication factor for the group
metadata internal topics "__consumer_offsets" and "__transaction_state"
# For anything other than development
testing, a value greater than 1 is recommended for to ensure availability
such as 3.
offsets.topic.replication.factor=1
transaction.state.log.replication.factor=1
transaction.state.log.min.isr=1
############################# Log Flush
Policy #############################
# Messages are immediately written to the
filesystem but by default we only fsync() to sync
# the OS cache lazily. The following
configurations control the flush of data to disk.
# There are a few important trade-offs
here:
#
1. Durability: Unflushed data may be lost if you are not using
replication.
#
2. Latency: Very large flush intervals may lead to latency spikes when
the flush does occur as there will be a lot of data to flush.
#
3. Throughput: The flush is generally the most expensive operation,
and a small flush interval may lead to exceessive seeks.
# The settings below allow one to
configure the flush policy to flush data after a period of time or
# every N messages (or both). This can be
done globally and overridden on a per-topic basis.
# The number of messages to accept before
forcing a flush of data to disk
#log.flush.interval.messages=10000
# The maximum amount of time a message
can sit in a log before we force a flush
#log.flush.interval.ms=1000
############################# Log
Retention Policy #############################
# The following configurations control
the disposal of log segments. The policy can
# be set to delete segments after a
period of time, or after a given size has accumulated.
# A segment will be deleted whenever
*either* of these criteria are met. Deletion always happens
# from the end of the log.
# The minimum age of a log file to be
eligible for deletion due to age
log.retention.hours=168
# A size-based retention policy for logs.
Segments are pruned from the log as long as the remaining
# segments don't drop below
log.retention.bytes. Functions independently of log.retention.hours.
#log.retention.bytes=1073741824
# The maximum size of a log segment file.
When this size is reached a new log segment will be created.
log.segment.bytes=1073741824
# The interval at which log segments are
checked to see if they can be deleted according
# to the retention policies
log.retention.check.interval.ms=300000
############################# Zookeeper
#############################
# Zookeeper connection string (see
zookeeper docs for details).
# This is a comma separated host:port
pairs, each corresponding to a zk
# server. e.g.
"127.0.0.1:3000,127.0.0.1:3001,127.0.0.1:3002".
# You can also append an optional chroot
string to the urls to specify the
# root directory for all kafka znodes.
# zookeeper.connect=localhost:2181
zookeeper.connect=hadoop210.cevent.com:2181,hadoop211.cevent.com:2181,hadoop212.cevent.com:2181
# Timeout in ms for connecting to
zookeeper
zookeeper.connection.timeout.ms=6000
############################# Group
Coordinator Settings #############################
# The following configuration specifies the
time, in milliseconds, that the GroupCoordinator will delay the initial
consumer rebalance.
# The rebalance will be further delayed
by the value of group.initial.rebalance.delay.ms as new members join the
group, up to a maximum of max.poll.interval.ms.
# The default value for this is 3
seconds.
# We override this to 0 here as it makes
for a better out-of-the-box experience for development and testing.
[cevent@hadoop207 kafka_2.11-0.11.0.0]$ sudo vim /etc/profile
[sudo] password for cevent:
fi
# Path manipulation
if [ "$EUID" = "0" ];
then
pathmunge /sbin
pathmunge /usr/sbin
pathmunge /usr/local/sbin
else
pathmunge /usr/local/sbin after
pathmunge /usr/sbin after
pathmunge /sbin after
fi
HOSTNAME=`/bin/hostname 2>/dev/null`
HISTSIZE=1000
if [ "$HISTCONTROL" =
"ignorespace" ] ; then
export HISTCONTROL=ignoreboth
else
export HISTCONTROL=ignoredups
fi
export PATH USER LOGNAME MAIL HOSTNAME
HISTSIZE HISTCONTROL
# By default, we want umask to get set.
This sets it for login shell
# Current threshold for system reserved
uid/gids is 200
# You could check uidgid reservation
validity in
# /usr/share/doc/setup-*/uidgid file
if [ $UID -gt 199 ] && [
"`id -gn`" = "`id -un`" ]; then
umask 002
else
umask 022
fi
for i in /etc/profile.d/*.sh ; do
if [ -r "$i" ]; then
if [ "${-#*i}" != "$-" ]; then
. "$i"
else
. "$i" >/dev/null
2>&1
fi
fi
done
unset i
unset -f pathmunge
#JAVA_HOME
export JAVA_HOME=/opt/module/jdk1.7.0_79
export PATH=$PATH:$JAVA_HOME/bin
#HADOOP_HOME
export HADOOP_HOME=/opt/module/hadoop-2.7.2
export PATH=$PATH:$HADOOP_HOME/bin
export PATH=$PATH:$HADOOP_HOME/sbin
#HIVE_HOME
export HIVE_HOME=/opt/module/hive-1.2.1
export PATH=$PATH:$HIVE_HOME/bin
#FLUME_HOME
export FLUME_HOME=/opt/module/apache-flume-1.7.0
export PATH=$PATH:$FLUME_HOME/bin
#ZOOKEEPER_HOME
export ZOOKEEPER_HOME=/opt/module/zookeeper-3.4.10
export PATH=$PATH:$ZOOKEEPER_HOME/bin
#KAFKA_HOME
export KAFKA_HOME=/opt/module/kafka_2.11-0.11.0.0
export PATH=$PATH:$KAFKA_HOME/bin
执行同步
xsync
[cevent@hadoop207 kafka_2.11-0.11.0.0]$ source /etc/profile
[cevent@hadoop207 kafka_2.11-0.11.0.0]$ echo $KAFKA_HOME
/opt/module/kafka_2.11-0.11.0.0
[cevent@hadoop207 kafka_2.11-0.11.0.0]$ kafka-
kafka-acls.sh
kafka-reassign-partitions.sh
kafka-broker-api-versions.sh kafka-replay-log-producer.sh
kafka-configs.sh kafka-replica-verification.sh
kafka-console-consumer.sh kafka-run-class.sh
kafka-console-producer.sh kafka-server-start.sh
kafka-consumer-groups.sh kafka-server-stop.sh
kafka-consumer-offset-checker.sh kafka-simple-consumer-shell.sh
kafka-consumer-perf-test.sh kafka-streams-application-reset.sh
kafka-delete-records.sh kafka-topics.sh
kafka-mirror-maker.sh kafka-verifiable-consumer.sh
kafka-preferred-replica-election.sh kafka-verifiable-producer.sh
kafka-producer-perf-test.sh
[cevent@hadoop207 kafka_2.11-0.11.0.0]$ 启动kafka
kafka-server-start.sh
-daemon /opt/module/kafka_2.11-0.11.0.0/config/server.properties
[cevent@hadoop207 kafka_2.11-0.11.0.0]$
ll
总用量 56
drwxr-xr-x. 3 cevent cevent 4096 6月 23 2017 bin
drwxr-xr-x. 2 cevent cevent 4096 6月 17 16:08 config
drwxr-xr-x. 2 cevent cevent 4096 6月 15 13:22 libs
-rw-r--r--. 1 cevent cevent 28824 6月 23 2017 LICENSE
drwxrwxr-x. 2 cevent cevent 4096 6月 17 16:28 logs
-rw-r--r--. 1 cevent cevent 336 6月 23 2017 NOTICE
drwxr-xr-x. 2 cevent cevent 4096 6月 23 2017 site-docs
[cevent@hadoop207 kafka_2.11-0.11.0.0]$
cd logs/
[cevent@hadoop207 logs]$ ll
总用量 12
-rw-rw-r--. 1 cevent cevent 0 6月 17 16:28 controller.log
-rw-rw-r--. 1 cevent cevent 0 6月 17 16:28
kafka-authorizer.log
-rw-rw-r--. 1 cevent cevent 0 6月 17 16:28 kafka-request.log
-rw-rw-r--. 1 cevent cevent 1048 6月 17 16:28 kafkaServer-gc.log.0.current
-rw-rw-r--. 1 cevent cevent 813 6月 17 16:28 kafkaServer.out
-rw-rw-r--. 1 cevent cevent 0 6月 17 16:28 log-cleaner.log
-rw-rw-r--. 1 cevent cevent 813 6月 17 16:28 server.log
-rw-rw-r--. 1 cevent cevent 0 6月 17 16:28 state-change.log
[cevent@hadoop207 logs]$ cat server.log
[2020-06-17 16:28:55,630] FATAL (kafka.Kafka$)
org.apache.kafka.common.config.ConfigException:
Invalid value kafka207 for configuration broker.id: Not a number of type INT
at org.apache.kafka.common.config.ConfigDef.parseType(ConfigDef.java:713)
at
org.apache.kafka.common.config.ConfigDef.parseValue(ConfigDef.java:460)
at
org.apache.kafka.common.config.ConfigDef.parse(ConfigDef.java:453)
at org.apache.kafka.common.config.AbstractConfig.(AbstractConfig.java:62)
at
kafka.server.KafkaConfig.(KafkaConfig.scala:883)
at
kafka.server.KafkaConfig$.fromProps(KafkaConfig.scala:867)
at
kafka.server.KafkaConfig$.fromProps(KafkaConfig.scala:864)
at
kafka.server.KafkaServerStartable$.fromProps(KafkaServerStartable.scala:28)
at kafka.Kafka$.main(Kafka.scala:58)
at kafka.Kafka.main(Kafka.scala)
[cevent@hadoop207 module]$ cd kafka_2.11-0.11.0.0/
[cevent@hadoop207 kafka_2.11-0.11.0.0]$ cd config/
[cevent@hadoop207 config]$ ll
总用量 64
-rw-r--r--. 1 cevent cevent 906 6月 23 2017
connect-console-sink.properties
-rw-r--r--. 1 cevent cevent 909 6月 23 2017
connect-console-source.properties
-rw-r--r--. 1 cevent cevent 5807 6月 23 2017 connect-distributed.properties
-rw-r--r--. 1 cevent cevent 883 6月 23 2017
connect-file-sink.properties
-rw-r--r--. 1 cevent cevent 881 6月 23 2017
connect-file-source.properties
-rw-r--r--. 1 cevent cevent 1111 6月 23 2017 connect-log4j.properties
-rw-r--r--. 1 cevent cevent 2730 6月 23 2017 connect-standalone.properties
-rw-r--r--. 1 cevent cevent 1199 6月 23 2017 consumer.properties
-rw-r--r--. 1 cevent cevent 4696 6月 23 2017 log4j.properties
-rw-r--r--. 1 cevent cevent 1900 6月 23 2017 producer.properties
-rw-r--r--. 1 cevent cevent 7072 6月 17 16:08 server.properties
-rw-r--r--. 1 cevent cevent 1032 6月 23 2017 tools-log4j.properties
-rw-r--r--. 1 cevent cevent 1023 6月 23 2017 zookeeper.properties
[cevent@hadoop207 config]$ vim server.properties
# The id of the broker. This must be set
to a unique integer for each broker.
【hadoop207】
broker.id=7
【hadoop208】
broker.id=8
【hadoop209】
broker.id=9
先启动每个zookeeper
[cevent@hadoop208 zookeeper-3.4.10]$ bin/zkServer.sh
start
ZooKeeper JMX enabled by default
Using config:
/opt/module/zookeeper-3.4.10/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED
[cevent@hadoop208 zookeeper-3.4.10]$ jps
8227 Jps
7894 DataNode
8012 NodeManager
8188 QuorumPeerMain
[cevent@hadoop207 logs]$ kafka-server-start.sh -daemon
/opt/module/kafka_2.11-0.11.0.0/config/server.properties 启动服务
(这里$KAFKA_HOME/c 按tab可快捷呼出路径)
[2020-06-17 16:48:22,346] INFO Registered
broker 7 at path /brokers/ids/7 with addresses: EndPoint(hadoop207.cevent.com,9092,ListenerName(PLAINTEXT),PLAINTEXT)
(kafka.utils.ZkUtils)
[2020-06-17 16:48:22,348] WARN No
meta.properties file under dir
/opt/module/kafka_2.11-0.11.0.0/logs/meta.properties
(kafka.server.BrokerMetadataCheckpoint)
[2020-06-17 16:48:22,393] INFO Kafka
version : 0.11.0.0 (org.apache.kafka.common.utils.AppInfoParser)
[2020-06-17 16:48:22,393] INFO Kafka
commitId : cb8625948210849f (org.apache.kafka.common.utils.AppInfoParser)
[2020-06-17 16:48:22,393] INFO [Kafka
Server 7], started (kafka.server.KafkaServer)
[cevent@hadoop207 logs]$
kafka-server-start.sh -daemon
/opt/module/kafka_2.11-0.11.0.0/config/server.properties
[cevent@hadoop207 hadoop-2.7.2]$ sbin/start-dfs.sh
[cevent@hadoop207 hadoop-2.7.2]$ sbin/start-yarn.sh
[cevent@hadoop207 hadoop-2.7.2]$ jps
4167 Jps
3352 DataNode
3237 NameNode
3848 NodeManager
3731 ResourceManager
3562 SecondaryNameNode
[cevent@hadoop207 ~]$ cd /opt/module/zookeeper-3.4.10/
[cevent@hadoop207 zookeeper-3.4.10]$ bin/zkServer.sh start
ZooKeeper JMX enabled by default
Using config: /opt/module/zookeeper-3.4.10/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED
[cevent@hadoop207 zookeeper-3.4.10]$ jps
3352 DataNode
3237 NameNode
3848 NodeManager
3731 ResourceManager
4223 QuorumPeerMain
4246 Jps
3562 SecondaryNameNode
[cevent@hadoop207 ~]$ cd /opt/module/kafka_2.11-0.11.0.0/
[cevent@hadoop207 kafka_2.11-0.11.0.0]$ 启动kafka可$KAFKA_HOME/
kafka-server-start.sh
-daemon /opt/module/kafka_2.11-0.11.0.0/config/server.properties
[cevent@hadoop207 kafka_2.11-0.11.0.0]$ jps
4539 Kafka
3352 DataNode
3237 NameNode
3848 NodeManager
3731 ResourceManager
4223 QuorumPeerMain
4565 Jps
3562 SecondaryNameNode
【其他虚拟机】
[cevent@hadoop208 ~]$ cd /opt/module/zookeeper-3.4.10/
[cevent@hadoop208 zookeeper-3.4.10]$ bin/zkServer.sh start
ZooKeeper JMX enabled by default
Using config: /opt/module/zookeeper-3.4.10/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED
[cevent@hadoop208 zookeeper-3.4.10]$ jps
3406 QuorumPeerMain
3235 NodeManager
3437 Jps
[cevent@hadoop208 ~]$ cd /opt/module/kafka_2.11-0.11.0.0/
[cevent@hadoop208 kafka_2.11-0.11.0.0]$ kafka-server-start.sh -daemon config/server.properties
[cevent@hadoop208 kafka_2.11-0.11.0.0]$ jps
3732 Kafka
3406 QuorumPeerMain
3235 NodeManager
3797 Jps
HBase的解压
[cevent@hadoop207 module]$ tar -zxvf hbase-1.3.1-bin.tar.gz -C /opt/module/
[cevent@hadoop207 soft]$ cd /opt/module/
[cevent@hadoop207 module]$ ll
总用量 44
drwxrwxr-x. 12 cevent cevent 4096 6月
19 17:50 apache-flume-1.7.0
drwxrwxr-x.
8 cevent cevent 4096 6月 19 17:53 datas
drwxr-xr-x. 11 cevent cevent 4096 5月
22 2017 hadoop-2.7.2
drwxrwxr-x.
3 cevent cevent 4096 6月 5 13:27 hadoop-2.7.2-snappy
drwxrwxr-x. 7 cevent cevent 4096
6月 19 23:09 hbase-1.3.1
drwxrwxr-x. 10 cevent cevent 4096 5月
22 13:34 hive-1.2.1
drwxr-xr-x.
8 cevent cevent 4096 4月 11 2015 jdk1.7.0_79
drwxr-xr-x.
7 cevent cevent 4096 6月 17 18:23 kafka_2.11-0.11.0.0
drwxrwxr-x.
2 cevent cevent 4096 6月 19 18:22 kafka-monitor
-rw-rw-r--.
1 cevent cevent 23 6月
16 21:11 xsync.txt
drwxr-xr-x. 11 cevent cevent 4096 6月
17 11:54 zookeeper-3.4.10
[cevent@hadoop207 module]$ cd hbase-1.3.1/
[cevent@hadoop207 hbase-1.3.1]$ ll
总用量 348
drwxr-xr-x. 4 cevent cevent 4096 4月 5 2017 bin
-rw-r--r--. 1 cevent cevent 148959 4月 7 2017 CHANGES.txt
drwxr-xr-x. 2 cevent cevent 4096 4月 5 2017 conf
drwxr-xr-x. 12 cevent cevent 4096 4月 7 2017 docs
drwxr-xr-x. 7 cevent cevent 4096 4月 7 2017 hbase-webapps
-rw-r--r--. 1 cevent cevent 261 4月 7 2017 LEGAL
drwxrwxr-x. 3 cevent cevent 4096 6月 19 23:09 lib
-rw-r--r--. 1 cevent cevent 130696 4月 7 2017 LICENSE.txt
-rw-r--r--. 1 cevent cevent 43258 4月 7 2017 NOTICE.txt
-rw-r--r--. 1 cevent cevent 1477 9月 21 2016 README.txt
[cevent@hadoop207 hbase-1.3.1]$ cd conf/
[cevent@hadoop207 conf]$ ll
总用量 40
-rw-r--r--. 1 cevent cevent 1811 9月 21 2016 hadoop-metrics2-hbase.properties
-rw-r--r--. 1 cevent cevent 4537 11月 7 2016 hbase-env.cmd
-rw-r--r--. 1 cevent cevent 7468 11月 7 2016 hbase-env.sh
-rw-r--r--. 1 cevent cevent 2257 9月 21 2016 hbase-policy.xml
-rw-r--r--. 1 cevent cevent 934 9月 21 2016 hbase-site.xml
-rw-r--r--. 1 cevent cevent 4722 4月 5 2017 log4j.properties
-rw-r--r--. 1 cevent cevent 10 12月 1 2015 regionservers
[cevent@hadoop207 conf]$ vim hbase-env.sh
#
#/**
# * Licensed to the Apache Software
Foundation (ASF) under one
# * or more contributor license
agreements. See the NOTICE file
# * distributed with this work for
additional information
# * regarding copyright ownership. The ASF licenses this file
# * to you under the Apache License,
Version 2.0 (the
# * "License"); you may not use
this file except in compliance
# * with the License. You may obtain a copy of the License at
# *
# *
http://www.apache.org/licenses/LICENSE-2.0
# *
# * Unless required by applicable law or
agreed to in writing, software
# * distributed under the License is
distributed on an "AS IS" BASIS,
# * WITHOUT WARRANTIES OR CONDITIONS OF
ANY KIND, either express or implied.
# * See the License for the specific
language governing permissions and
# * limitations under the License.
# */
# Set environment variables here.
# This script sets variables multiple
times over the course of starting an hbase process,
# so try to keep things idempotent unless
you want to take an even deeper look
# into the startup scripts (bin/hbase,
etc.)
# The java implementation to use. Java 1.7+ required.
# export JAVA_HOME=/usr/java/jdk1.6.0/
#JAVA_HOME
export
JAVA_HOME=/opt/module/jdk1.7.0_79
export
PATH=$PATH:$JAVA_HOME/bin
# Extra Java CLASSPATH elements. Optional.
# export HBASE_CLASSPATH=
# The maximum amount of heap to use.
Default is left to JVM default.
# export HBASE_HEAPSIZE=1G
# Uncomment below if you intend to use
off heap cache. For example, to allocate 8G of
# offheap, set the value to
"8G".
# export HBASE_OFFHEAPSIZE=1G
# Extra Java runtime options.
# Below are what we set by default. May only work with SUN JVM.
# For more on why as well as other
possible settings,
# see
http://wiki.apache.org/hadoop/PerformanceTuning
export
HBASE_OPTS="-XX:+UseConcMarkSweepGC"
# Configure PermSize. Only needed in
JDK7. You can safely remove it for JDK8+
export HBASE_MASTER_OPTS="$HBASE_MASTER_OPTS
-XX:PermSize=128m -XX:MaxPermSize=128m"
export
HBASE_REGIONSERVER_OPTS="$HBASE_REGIONSERVER_OPTS -XX:PermSize=128m
-XX:MaxPermSize=128m"
# Uncomment one of the below three
options to enable java garbage collection logging for the server-side
processes.
# This enables basic gc logging to the
.out file.
# export SERVER_GC_OPTS="-verbose:gc
-XX:+PrintGCDetails -XX:+PrintGCDateStamps"
# This enables basic gc logging to its
own file.
# If FILE-PATH is not replaced, the log
file(.gc) would still be generated in the HBASE_LOG_DIR .
# export SERVER_GC_OPTS="-verbose:gc
-XX:+PrintGCDetails -XX:+PrintGCDateStamps -Xloggc:-PATH>"
# This enables basic GC logging to its
own file with automatic log rolling. Only applies to jdk 1.6.0_34+ and
1.7.0_2+.
# If FILE-PATH is not replaced, the log
file(.gc) would still be generated in the HBASE_LOG_DIR .
# export SERVER_GC_OPTS="-verbose:gc
-XX:+PrintGCDetails -XX:+PrintGCDateStamps -Xloggc:-PATH>
-XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=1
-XX:GCLogFileSize=512M"
# Uncomment one of the below three
options to enable java garbage collection logging for the client processes.
# This enables basic gc logging to the
.out file.
# export CLIENT_GC_OPTS="-verbose:gc
-XX:+PrintGCDetails -XX:+PrintGCDateStamps"
# This enables basic gc logging to its
own file.
# If FILE-PATH is not replaced, the log
file(.gc) would still be generated in the HBASE_LOG_DIR .
# export CLIENT_GC_OPTS="-verbose:gc
-XX:+PrintGCDetails -XX:+PrintGCDateStamps -Xloggc:-PATH>"
# This enables basic GC logging to its
own file with automatic log rolling. Only applies to jdk 1.6.0_34+ and
1.7.0_2+.
# If FILE-PATH is not replaced, the log
file(.gc) would still be generated in the HBASE_LOG_DIR .
# export CLIENT_GC_OPTS="-verbose:gc
-XX:+PrintGCDetails -XX:+PrintGCDateStamps -Xloggc:-PATH>
-XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=1
-XX:GCLogFileSize=512M"
# See the package documentation for
org.apache.hadoop.hbase.io.hfile for other configurations
# needed setting up off-heap block
caching.
# Uncomment and adjust to enable JMX
exporting
# See jmxremote.password and
jmxremote.access in $JRE_HOME/lib/management to configure remote password
access.
# More details at:
http://java.sun.com/javase/6/docs/technotes/guides/management/agent.html
# NOTE: HBase provides an alternative JMX
implementation to fix the random ports issue, please see JMX
# section in HBase Reference Guide for
instructions.
# export
HBASE_JMX_BASE="-Dcom.sun.management.jmxremote.ssl=false
-Dcom.sun.management.jmxremote.authenticate=false"
# export
HBASE_MASTER_OPTS="$HBASE_MASTER_OPTS $HBASE_JMX_BASE
-Dcom.sun.management.jmxremote.port=10101"
# export HBASE_REGIONSERVER_OPTS="$HBASE_REGIONSERVER_OPTS
$HBASE_JMX_BASE -Dcom.sun.management.jmxremote.port=10102"
# export
HBASE_THRIFT_OPTS="$HBASE_THRIFT_OPTS $HBASE_JMX_BASE
-Dcom.sun.management.jmxremote.port=10103"
# export
HBASE_ZOOKEEPER_OPTS="$HBASE_ZOOKEEPER_OPTS $HBASE_JMX_BASE
-Dcom.sun.management.jmxremote.port=10104"
# export
HBASE_REST_OPTS="$HBASE_REST_OPTS $HBASE_JMX_BASE
-Dcom.sun.management.jmxremote.port=10105"
# File naming hosts on which
HRegionServers will run.
$HBASE_HOME/conf/regionservers by default.
# export
HBASE_REGIONSERVERS=${HBASE_HOME}/conf/regionservers
# Uncomment and adjust to keep all the
Region Server pages mapped to be memory resident
#HBASE_REGIONSERVER_MLOCK=true
#HBASE_REGIONSERVER_UID="hbase"
# File naming hosts on which backup
HMaster will run.
$HBASE_HOME/conf/backup-masters by default.
# export
HBASE_BACKUP_MASTERS=${HBASE_HOME}/conf/backup-masters
# Extra ssh options. Empty by default.
# export HBASE_SSH_OPTS="-o
ConnectTimeout=1 -o SendEnv=HBASE_CONF_DIR"
# Where log files are stored. $HBASE_HOME/logs by default.
# export HBASE_LOG_DIR=${HBASE_HOME}/logs
# Enable remote JDWP debugging of major
HBase processes. Meant for Core Developers
# export
HBASE_MASTER_OPTS="$HBASE_MASTER_OPTS -Xdebug -Xrunjdwp:transport=dt_socket,server=y,suspend=n,address=8070"
# export
HBASE_REGIONSERVER_OPTS="$HBASE_REGIONSERVER_OPTS -Xdebug
-Xrunjdwp:transport=dt_socket,server=y,suspend=n,address=8071"
# export
HBASE_THRIFT_OPTS="$HBASE_THRIFT_OPTS -Xdebug -Xrunjdwp:transport=dt_socket,server=y,suspend=n,address=8072"
# export
HBASE_ZOOKEEPER_OPTS="$HBASE_ZOOKEEPER_OPTS -Xdebug
-Xrunjdwp:transport=dt_socket,server=y,suspend=n,address=8073"
# A string representing this instance of
hbase. $USER by default.
# export HBASE_IDENT_STRING=$USER
# The scheduling priority for daemon
processes. See 'man nice'.
# export HBASE_NICENESS=10
# The directory where pid files are
stored. /tmp by default.
# export HBASE_PID_DIR=/var/hadoop/pids
# Seconds to sleep between slave commands. Unset by default. This
# can be useful in large clusters, where,
e.g., slave rsyncs can
# otherwise arrive faster than the master
can service them.
# export HBASE_SLAVE_SLEEP=0.1
# Tell HBase whether it should manage
it's own instance of Zookeeper or not.
# export HBASE_MANAGES_ZK=true 取消HBase默认开启自己的zookeeper
export
HBASE_MANAGES_ZK=false
# The default log rolling policy is RFA,
where the log file is rolled as per the size defined for the
# RFA appender. Please refer to the
log4j.properties file to see more details on this appender.
# In case one needs to do log rolling on
a date change, one should set the environment property
# HBASE_ROOT_LOGGER to
",DRFA".
# For example:
# HBASE_ROOT_LOGGER=INFO,DRFA
[cevent@hadoop207 conf]$ ll
总用量 40
-rw-r--r--. 1 cevent cevent 1811 9月 21 2016 hadoop-metrics2-hbase.properties
-rw-r--r--. 1 cevent cevent 4537 11月 7 2016 hbase-env.cmd
-rw-r--r--. 1 cevent cevent 7497 6月 19 23:17 hbase-env.sh
-rw-r--r--. 1 cevent cevent 2257 9月 21 2016 hbase-policy.xml
-rw-r--r--. 1 cevent cevent 934 9月 21 2016 hbase-site.xml
-rw-r--r--. 1 cevent cevent 4722 4月 5 2017 log4j.properties
-rw-r--r--. 1 cevent cevent 10 12月 1 2015 regionservers
[cevent@hadoop207 conf]$ vim hbase-site.xml
"1.0"?>
-stylesheet type="text/xsl"
href="configuration.xsl"?>
<!--
/**
*
*
Licensed to the Apache Software Foundation (ASF) under one
*
or more contributor license agreements.
See the NOTICE file
*
distributed with this work for additional information
*
regarding copyright ownership. The ASF
licenses this file
*
to you under the Apache License, Version 2.0 (the
*
"License"); you may not use this file except in compliance
*
with the License. You may obtain a
copy of the License at
*
*
http://www.apache.org/licenses/LICENSE-2.0
*
*
Unless required by applicable law or agreed to in writing, software
*
distributed under the License is distributed on an "AS IS" BASIS,
*
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
*
See the License for the specific language governing permissions and
*
limitations under the License.
*/
-->
hbase.root.dir</name>
hdfs://hadoop207.cevent.com:8020/HBase</value>
</property>
hbase.cluster.distributed</name>
true</value>
</property>
<!-- 0.98后的新变动,之前版本没有.port,默认端口为60000 -->
hbase.master.port</name>
16000</value>
</property>
hbase.zookeeper.quorum</name>
hadoop207.cevent.com,hadoop208.cevent.com,hadoop209.cevent.com</value>
</property>
hbase.zookeeper.property.dataDir</name>
/opt/module/zookeeper-3.4.10/data/zkData</value>
</property>
</configuration>
regionservers:
[cevent@hadoop207 conf]$ vim regionservers
hadoop207.cevent.com
hadoop208.cevent.com
hadoop209.cevent.com
[cevent@hadoop207 conf]$ ll
总用量 40
-rw-r--r--. 1 cevent cevent 1811 9月 21 2016 hadoop-metrics2-hbase.properties
-rw-r--r--. 1 cevent cevent 4537 11月 7 2016 hbase-env.cmd
-rw-r--r--. 1 cevent cevent 7497 6月 19 23:17 hbase-env.sh
-rw-r--r--. 1 cevent cevent 2257 9月 21 2016 hbase-policy.xml
-rw-r--r--. 1 cevent cevent 1586 6月 19 23:23 hbase-site.xml
-rw-r--r--. 1 cevent cevent 4722 4月 5 2017 log4j.properties
-rw-r--r--. 1 cevent cevent 63 6月 19 23:26 regionservers
[cevent@hadoop207 conf]$ ln -s /opt/module/hadoop-2.7.2/etc/hadoop/core-site.xml
[cevent@hadoop207 conf]$ ln -s /opt/module/hadoop-2.7.2/etc/hadoop/hdfs-site.xml
[cevent@hadoop207 conf]$ ll
总用量 40
lrwxrwxrwx. 1 cevent cevent 49 6月 19 23:29
core-site.xml -> /opt/module/hadoop-2.7.2/etc/hadoop/core-site.xml
-rw-r--r--. 1 cevent cevent 1811 9月 21 2016 hadoop-metrics2-hbase.properties
-rw-r--r--. 1 cevent cevent 4537 11月 7 2016 hbase-env.cmd
-rw-r--r--. 1 cevent cevent 7497 6月 19 23:17 hbase-env.sh
-rw-r--r--. 1 cevent cevent 2257 9月 21 2016 hbase-policy.xml
-rw-r--r--. 1 cevent cevent 1586 6月 19 23:23 hbase-site.xml
lrwxrwxrwx. 1 cevent cevent 49 6月 19 23:29
hdfs-site.xml -> /opt/module/hadoop-2.7.2/etc/hadoop/hdfs-site.xml
-rw-r--r--. 1 cevent cevent 4722 4月 5 2017 log4j.properties
-rw-r--r--. 1 cevent cevent 63 6月 19 23:26 regionservers
[cevent@hadoop207 module]$ xsync hbase-1.3.1/
[cevent@hadoop208 module]$ ll
总用量 24
drwxr-xr-x. 12 cevent cevent 4096 6月 16 21:35 hadoop-2.7.2
drwxrwxr-x. 7 cevent cevent
4096 6月 19 23:31
hbase-1.3.1
drwxr-xr-x. 8 cevent cevent 4096 3月 24 09:14 jdk1.7.0_79
drwxr-xr-x. 7 cevent cevent 4096 6月 18 09:50 kafka_2.11-0.11.0.0
-rw-rw-r--. 1 cevent cevent 23 6月 16 21:11 xsync.txt
drwxr-xr-x. 11 cevent cevent 4096 6月 17 13:36 zookeeper-3.4.10
[cevent@hadoop209 ~]$ cd /opt/module/
[cevent@hadoop209 module]$ ll
总用量 24
drwxr-xr-x. 12 cevent cevent 4096 6月 16 21:37 hadoop-2.7.2
drwxrwxr-x. 7 cevent cevent
4096 6月 19 23:33
hbase-1.3.1
drwxr-xr-x. 8 cevent cevent 4096 3月 24 09:14 jdk1.7.0_79
drwxr-xr-x. 7 cevent cevent 4096 6月 18 09:51 kafka_2.11-0.11.0.0
-rw-rw-r--. 1 cevent cevent 23 6月 16 21:11 xsync.txt
drwxr-xr-x. 11 cevent cevent 4096 6月 17 13:36 zookeeper-3.4.10
[cevent@hadoop207 hbase-1.3.1]$ bin/hbase-daemon.sh start master 启动master
starting master, logging to /opt/module/hbase-1.3.1/bin/../logs/hbase-cevent-master-hadoop207.cevent.com.out
[cevent@hadoop207 hbase-1.3.1]$ bin/hbase-daemons.sh start regionserver 启动regionserver
hadoop207.cevent.com: regionserver
running as process 4993. Stop it first.
hadoop208.cevent.com: starting
regionserver, logging to
/opt/module/hbase-1.3.1/bin/../logs/hbase-cevent-regionserver-hadoop208.cevent.com.out
hadoop209.cevent.com: starting
regionserver, logging to
/opt/module/hbase-1.3.1/bin/../logs/hbase-cevent-regionserver-hadoop209.cevent.com.out
[cevent@hadoop207 hbase-1.3.1]$ jps
5471 HMaster
4197 NodeManager
4542 QuorumPeerMain
3580 NameNode
5544 Jps
3878 SecondaryNameNode
3696 DataNode
4079 ResourceManager
4993 HRegionServer
(1)查看hadoop配置的端口
[cevent@hadoop207 hadoop-2.7.2]$ cd
etc/hadoop/
[cevent@hadoop207 hadoop]$ vim
core-site.xml
"1.0"
encoding="UTF-8"?>
-stylesheet
type="text/xsl" href="configuration.xsl"?>
<!--
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS"
BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
implied.
See the License for the specific language governing permissions and
limitations under the License. See accompanying LICENSE file.
-->
<!-- Put site-specific property
overrides in this file. -->
<!-- 指定HDFS中NameNode地址 ,设置的hadoop207.cevent.com=hostname
-->
fs.defaultFS</name>
hdfs://hadoop207.cevent.com:8020</value>
</property>
<!-- 指定tmp数据存储位置 -->
hadoop.tmp.dir</name>
/opt/module/hadoop-2.7.2/data/tmp</value>
</property>
</configuration>
(2)更改HBase配置的端口
[cevent@hadoop207 hbase-1.3.1]$ vim
conf/hbase-site.xml
"1.0"?>
-stylesheet
type="text/xsl" href="configuration.xsl"?>
<!--
/**
*
* Licensed to the Apache Software
Foundation (ASF) under one
* or more contributor license
agreements. See the NOTICE file
* distributed with this work for
additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version
2.0 (the
* "License"); you may not use
this file except in compliance
* with the License. You may obtain a copy of the License at
*
*
http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or
agreed to in writing, software
* distributed under the License is
distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY
KIND, either express or implied.
* See the License for the specific language
governing permissions and
* limitations under the License.
*/
-->
hbase.rootdir</name>
hdfs://hadoop207.cevent.com:8020/HBase</value>
</property>
hbase.cluster.distributed</name>
true</value>
</property>
<!-- 0.98后的新变动,之前版本没有.port,默认端口为60000 -->
hbase.master.port</name>
16000</value>
</property>
hbase.zookeeper.quorum</name>
hadoop207.cevent.com,hadoop208.cevent.com,hadoop209.cevent.com</value>
</property>
hbase.zookeeper.property.dataDir</name>
/opt/module/zookeeper-3.4.10/data/zkData</value>
</property>
</configuration>
[cevent@hadoop207 hbase-1.3.1]$ jps
10056 Jps
9430 HMaster
4197 NodeManager
8151 ZooKeeperMain
4542 QuorumPeerMain
3580 NameNode
3878 SecondaryNameNode
3696 DataNode
4079 ResourceManager
访问页面:http://hadoop207.cevent.com:16010/master-status
ln -s
$HBASE_HOME/lib/hbase-common-1.3.1.jar $HIVE_HOME/lib/hbase-common-1.3.1.jar
ln -s $HBASE_HOME/lib/hbase-server-1.3.1.jar
$HIVE_HOME/lib/hbase-server-1.3.1.jar
ln -s $HBASE_HOME/lib/hbase-client-1.3.1.jar
$HIVE_HOME/lib/hbase-client-1.3.1.jar
ln -s $HBASE_HOME/lib/hbase-protocol-1.3.1.jar
$HIVE_HOME/lib/hbase-protocol-1.3.1.jar
ln -s $HBASE_HOME/lib/hbase-it-1.3.1.jar
$HIVE_HOME/lib/hbase-it-1.3.1.jar
ln -s
$HBASE_HOME/lib/htrace-core-3.1.0-incubating.jar $HIVE_HOME/lib/htrace-core-3.1.0-incubating.jar
ln -s
$HBASE_HOME/lib/hbase-hadoop2-compat-1.3.1.jar
$HIVE_HOME/lib/hbase-hadoop2-compat-1.3.1.jar
ln -s $HBASE_HOME/lib/hbase-hadoop-compat-1.3.1.jar
$HIVE_HOME/lib/hbase-hadoop-compat-1.3.1.jar
hive.zookeeper.quorum</name>
hadoop207.cevent.com,hadoop208.cevent.com,hadoop209.cevent.com</value>
The list of
ZooKeeper servers to talk to. This is only needed for read/write
locks.</description>
</property>
hive.zookeeper.client.port</name>
2181</value>
The port of
ZooKeeper servers to talk to. This is only needed for read/write
locks.</description>
</property>
[cevent@hadoop207 conf]$ vim
/opt/module/hive-1.2.1/conf/hive-site.xml
<!--修改默认的reduce生成数量-->
"1.0"?>
-stylesheet
type="text/xsl" href="configuration.xsl"?>
<!--mySQL数据库地址-->
javax.jdo.option.ConnectionURL</name>
jdbc:mysql://hadoop207.cevent.com:3306/metastore?createDatabaseIfNotExist=true</value>
JDBC connect string for a JDBC
metastore</description>
</property>
javax.jdo.option.ConnectionDriverName</name>
com.mysql.jdbc.Driver</value>
Driver class name for a JDBC
metastore</description>
</property>
<!--mySQL数据库访问用户名及密码-->
javax.jdo.option.ConnectionUserName</name>
root</value>
username to use against metastore database</description>
</property>
javax.jdo.option.ConnectionPassword</name>
cevent</value>
password to use against metastore
database</description>
</property>
<!-- 自定义hive查询显示的信息 -->
hive.cli.print.header</name>
true</value>
</property>
hive.cli.print.current.db</name>
true</value>
</property>
hive.zookeeper.quorum</name>
hadoop207.cevent.com,hadoop208.cevent.com,hadoop209.cevent.com</value>
The list of ZooKeeper servers to talk to. This is
only needed for read/write locks.</description>
</property>
hive.zookeeper.client.port</name>
2181</value>
The port of ZooKeeper servers to talk to. This is
only needed for read/write locks.</description>
</property>
</configuration>
1. 将准备好的 hive-hbase-handler-1.2.1.jar放入lib
[cevent@hadoop207 conf]$ cd /opt/module/hive-1.2.1/lib
[cevent@hadoop207 lib]$ ll hive-hbase-handler-1.2.1.jar
-rw-rw-r--. 1 cevent cevent 115935 6月
19 2015 hive-hbase-handler-1.2.1.jar
[cevent@hadoop207 lib]$ rm hive-hbase-handler-1.2.1.jar 删除自带的handler
[cevent@hadoop207 lib]$ mv /opt/soft/hive-hbase-handler-1.2.1.jar
/opt/module/hive-1.2.1/lib/ 传入编译好的handler
1 row selected (0.366 seconds)
0:
jdbc:hive2://hadoop210.cevent.com:10000> CREATE TABLE
hive_hbase_emp_table(
0:
jdbc:hive2://hadoop210.cevent.com:10000> empno int,
0:
jdbc:hive2://hadoop210.cevent.com:10000> ename string,
0: jdbc:hive2://hadoop210.cevent.com:10000>
job string,
0:
jdbc:hive2://hadoop210.cevent.com:10000> mgr int,
0:
jdbc:hive2://hadoop210.cevent.com:10000> hiredate string,
0:
jdbc:hive2://hadoop210.cevent.com:10000> sal double,
0:
jdbc:hive2://hadoop210.cevent.com:10000> comm double,
0: jdbc:hive2://hadoop210.cevent.com:10000>
deptno int)
0:
jdbc:hive2://hadoop210.cevent.com:10000> STORED BY
'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
0:
jdbc:hive2://hadoop210.cevent.com:10000> WITH SERDEPROPERTIES
("hbase.columns.mapping" = ":key,info:ename,info:job,info:mgr,info:hiredate,info:sal,info:comm,info:deptno")
0:
jdbc:hive2://hadoop210.cevent.com:10000> TBLPROPERTIES
("hbase.table.name" = "hbase_emp_table");
Error: Error while processing statement:
FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask.
org.apache.hadoop.hbase.HTableDescriptor.addFamily(Lorg/apache/hadoop/hbase/HColumnDescriptor;)V
(state=08S01,code=1)
[cevent@hadoop207 hbase-1.3.1]$ hdfs dfs
-rmr /HBase/WALs/hadoop207.cevent.com,16020,1592703009107-splitting
推荐使用:rmr: DEPRECATED: Please use 'rm -r' instead.
CREATE TABLE hive_hbase_emp_table(
empno int,
ename string,
job string,
mgr int,
hiredate string,
sal double,
comm double,
deptno int)
STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
WITH SERDEPROPERTIES
("hbase.columns.mapping" =
":key,info:ename,info:job,info:mgr,info:hiredate,info:sal,info:comm,info:deptno")
TBLPROPERTIES ("hbase.table.name" =
"hbase_emp_table");
[cevent@hadoop210 hive-1.2.1]$ start-dfs.sh
[cevent@hadoop210 hive-1.2.1]$ start-yarn.sh
[cevent@hadoop210 hive-1.2.1]$ zkServer.sh start
[cevent@hadoop210 hbase-1.3.1]$ bin/start-hbase.sh
starting master, logging to
/opt/module/hbase-1.3.1/bin/../logs/hbase-cevent-master-hadoop210.cevent.com.out
hadoop210.cevent.com: starting
regionserver, logging to
/opt/module/hbase-1.3.1/bin/../logs/hbase-cevent-regionserver-hadoop210.cevent.com.out
hadoop211.cevent.com: starting
regionserver, logging to /opt/module/hbase-1.3.1/bin/../logs/hbase-cevent-regionserver-hadoop211.cevent.com.out
hadoop212.cevent.com: starting
regionserver, logging to
/opt/module/hbase-1.3.1/bin/../logs/hbase-cevent-regionserver-hadoop212.cevent.com.out
[cevent@hadoop210 hbase-1.3.1]$ jps
10394 Jps
9192 DataNode
9999 QuorumPeerMain
10320 HRegionServer
10174 HMaster
9540 ResourceManager
9051 NameNode
9660 NodeManager
9373 SecondaryNameNode
[cevent@hadoop210 hive-1.2.1]$ bin/hiveserver2
OK
[cevent@hadoop210 hive-1.2.1]$ bin/beeline
Beeline version 1.2.1 by Apache Hive
beeline> !connect
jdbc:hive2://hadoop210.cevent.com:10000
Connecting to
jdbc:hive2://hadoop210.cevent.com:10000
Enter username for
jdbc:hive2://hadoop210.cevent.com:10000: cevent
Enter password for
jdbc:hive2://hadoop210.cevent.com:10000: ******
Connected to: Apache Hive (version 1.2.1)
Driver: Hive JDBC (version 1.2.1)
org/apache/hadoop/hive/hbase/HBaseStorageHandler : Unsupported major.minor
version 52.0 (state=,code=0)
0:
jdbc:hive2://hadoop210.cevent.com:10000> CREATE TABLE
hive_hbase_emp_table(
0:
jdbc:hive2://hadoop210.cevent.com:10000> empno int,
0: jdbc:hive2://hadoop210.cevent.com:10000>
ename string,
0:
jdbc:hive2://hadoop210.cevent.com:10000> job string,
0:
jdbc:hive2://hadoop210.cevent.com:10000> mgr int,
0: jdbc:hive2://hadoop210.cevent.com:10000>
hiredate string,
0:
jdbc:hive2://hadoop210.cevent.com:10000> sal double,
0:
jdbc:hive2://hadoop210.cevent.com:10000> comm double,
0:
jdbc:hive2://hadoop210.cevent.com:10000> deptno int)
0: jdbc:hive2://hadoop210.cevent.com:10000>
STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
0: jdbc:hive2://hadoop210.cevent.com:10000>
WITH SERDEPROPERTIES ("hbase.columns.mapping" =
":key,info:ename,info:job,info:mgr,info:hiredate,info:sal,info:comm,info:deptno")
0:
jdbc:hive2://hadoop210.cevent.com:10000> TBLPROPERTIES
("hbase.table.name" = "hbase_emp_table");
Error: java.lang.UnsupportedClassVersionError:
org/apache/hadoop/hive/hbase/HBaseStorageHandler : Unsupported major.minor
version 52.0 (state=,code=0)
1)官网下载源码
链接:http://archive.apache.org/dist/hive/hive-1.2.1/
[cevent@hadoop210 hadoop-2.7.2]$ sbin/start-dfs.sh
[cevent@hadoop210 hadoop-2.7.2]$ sbin/start-yarn.sh
[cevent@hadoop210 hadoop-2.7.2]$ zkServer.sh start
[cevent@hadoop210 hadoop-2.7.2]$ jps
4370 ResourceManager
4488 NodeManager
4841 QuorumPeerMain
4965 Jps
3448 NameNode
3745 SecondaryNameNode
3565 DataNode
[cevent@hadoop210 hbase-1.3.1]$ bin/start-hbase.sh
starting master, logging to
/opt/module/hbase-1.3.1/logs/hbase-cevent-master-hadoop210.cevent.com.out
hadoop212.cevent.com: starting
regionserver, logging to /opt/module/hbase-1.3.1/bin/../logs/hbase-cevent-regionserver-hadoop212.cevent.com.out
hadoop211.cevent.com: starting
regionserver, logging to
/opt/module/hbase-1.3.1/bin/../logs/hbase-cevent-regionserver-hadoop211.cevent.com.out
hadoop210.cevent.com: starting regionserver,
logging to
/opt/module/hbase-1.3.1/bin/../logs/hbase-cevent-regionserver-hadoop210.cevent.com.out
[cevent@hadoop210 hbase-1.3.1]$ jps
4370 ResourceManager
5115 HMaster
4488 NodeManager
4841 QuorumPeerMain
3448 NameNode
3745 SecondaryNameNode
3565 DataNode
5393 Jps
5262 HRegionServer
[cevent@hadoop210 hive-1.2.1]$ ll
总用量 532
drwxrwxr-x. 3 cevent cevent 4096 6月 28 16:23 bin
drwxrwxr-x. 2 cevent cevent 4096 6月 28 21:08 conf
-rw-rw-r--. 1 cevent cevent 21061 6月 28 16:35 derby.log
drwxrwxr-x. 4 cevent cevent 4096 6月 28 16:23 examples
drwxrwxr-x. 7 cevent cevent 4096 6月 28 16:23 hcatalog
drwxrwxr-x. 4 cevent cevent 4096 6月 29 09:24 lib
-rw-rw-r--. 1 cevent cevent 24754 4月 30 2015 LICENSE
drwxrwxr-x. 5 cevent cevent 4096 6月 28 16:35 metastore_db
-rw-rw-r--. 1 cevent cevent 397 6月 19 2015 NOTICE
-rw-rw-r--. 1 cevent cevent 4366 6月 19 2015 README.txt
-rw-rw-r--. 1 cevent cevent 421129 6月 19 2015 RELEASE_NOTES.txt
drwxrwxr-x. 3 cevent cevent 4096 6月 28 16:23 scripts
-rw-rw-r--. 1 cevent cevent 24896 6月 28 22:32 zookeeper.out
[cevent@hadoop210 hive-1.2.1]$ bin/hiveserver2
[cevent@hadoop210 hive-1.2.1]$ bin/beeline
Beeline version 1.2.1 by Apache Hive
beeline> !connect
jdbc:hive2://hadoop210.cevent.com:10000
Connecting to
jdbc:hive2://hadoop210.cevent.com:10000
Enter username for
jdbc:hive2://hadoop210.cevent.com:10000: cevent
Enter password for
jdbc:hive2://hadoop210.cevent.com:10000: ******
Connected to: Apache Hive (version 1.2.1)
Driver: Hive JDBC (version 1.2.1)
Transaction isolation:
TRANSACTION_REPEATABLE_READ
0:
jdbc:hive2://hadoop210.cevent.com:10000> show tables;
+-----------+--+
| tab_name |
+-----------+--+
| student |
+-----------+--+
1 row selected (1.458 seconds) 创建hive表和hbase表
0:
jdbc:hive2://hadoop210.cevent.com:10000> CREATE
TABLE hive_hbase_emp_table(
0:
jdbc:hive2://hadoop210.cevent.com:10000> empno
int,
0:
jdbc:hive2://hadoop210.cevent.com:10000> ename
string,
0: jdbc:hive2://hadoop210.cevent.com:10000>
job string,
0:
jdbc:hive2://hadoop210.cevent.com:10000> mgr int,
0:
jdbc:hive2://hadoop210.cevent.com:10000> hiredate
string,
0: jdbc:hive2://hadoop210.cevent.com:10000> sal double,
0: jdbc:hive2://hadoop210.cevent.com:10000>
comm double,
0:
jdbc:hive2://hadoop210.cevent.com:10000> deptno
int)
0:
jdbc:hive2://hadoop210.cevent.com:10000> STORED
BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
0: jdbc:hive2://hadoop210.cevent.com:10000>
WITH
SERDEPROPERTIES
("hbase.columns.mapping"
=
":key,info:ename,info:job,info:mgr,info:hiredate,info:sal,info:comm,info:deptno")
0:
jdbc:hive2://hadoop210.cevent.com:10000> TBLPROPERTIES
("hbase.table.name" = "hbase_emp_table");
No rows affected (5.286 seconds)
0: jdbc:hive2://hadoop210.cevent.com:10000>
show tables;
+-----------------------+--+
|
tab_name |
+-----------------------+--+
| hive_hbase_emp_table |
| student |
+-----------------------+--+
2 rows selected (0.057 seconds)
0: jdbc:hive2://hadoop210.cevent.com:10000>
插入数据
insert into table
hive_hbase_emp_table values(2020,'cevent','developer',201,'2020-06-29',20000,2000,619);
INFO
: Number of reduce tasks is set to 0 since there's no reduce operator
INFO
: number of splits:1
INFO
: Submitting tokens for job: job_1593393499225_0001
INFO
: The url to track the job:
http://hadoop210.cevent.com:8088/proxy/application_1593393499225_0001/
INFO
: Starting Job = job_1593393499225_0001, Tracking URL =
http://hadoop210.cevent.com:8088/proxy/application_1593393499225_0001/
INFO
: Kill Command = /opt/module/hadoop-2.7.2/bin/hadoop job -kill job_1593393499225_0001
INFO
: Hadoop job information for Stage-0: number of mappers: 1; number of
reducers: 0
INFO
: 2020-06-29 09:33:36,570 Stage-0 map = 0%, reduce = 0%
INFO
: 2020-06-29 09:33:44,934 Stage-0 map = 100%, reduce = 0%, Cumulative CPU 2.79 sec
INFO
: MapReduce Total cumulative CPU time: 2 seconds 790 msec
INFO
: Ended Job = job_1593393499225_0001
No rows affected (37.702 seconds)
0: jdbc:hive2://hadoop210.cevent.com:10000>
select * from hive_hbase_emp_table; 查看hbase中清除数据
+-----------------------------+-----------------------------+---------------------------+---------------------------+--------------------------------+---------------------------+----------------------------+------------------------------+--+
| hive_hbase_emp_table.empno |
hive_hbase_emp_table.ename |
hive_hbase_emp_table.job |
hive_hbase_emp_table.mgr |
hive_hbase_emp_table.hiredate |
hive_hbase_emp_table.sal |
hive_hbase_emp_table.comm |
hive_hbase_emp_table.deptno |
+-----------------------------+-----------------------------+---------------------------+---------------------------+--------------------------------+---------------------------+----------------------------+------------------------------+--+
+-----------------------------+-----------------------------+---------------------------+---------------------------+--------------------------------+---------------------------+----------------------------+------------------------------+--+
No rows selected (1.009 seconds)
[cevent@hadoop210 hbase-1.3.1]$ hbase shell
SLF4J: Class path contains multiple SLF4J
bindings.
SLF4J: Found binding in [jar:file:/opt/module/hbase-1.3.1/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/opt/module/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See
http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type
[org.slf4j.impl.Log4jLoggerFactory]
HBase Shell; enter 'help'
for list of supported commands.
Type "exit" to
leave the HBase Shell
Version 1.3.1,
r930b9a55528fe45d8edce7af42fef2d35e77677a, Thu Apr 6 19:36:54 PDT 2017
hbase(main):001:0> list
TABLE
hbase_emp_table
student
2 row(s) in 0.2990 seconds
=> ["hbase_emp_table",
"student"]
hbase(main):002:0> scan
'hbase_emp_table'
ROW COLUMN+CELL
0 row(s) in 0.2450 seconds
ROW COLUMN+CELL
0 row(s) in 0.0070 seconds
hbase(main):004:0> scan 'hbase_emp_table'
ROW COLUMN+CELL
2020 column=info:comm,
timestamp=1593394424573, value=2000.0
2020 column=info:deptno,
timestamp=1593394424573, value=619
2020 column=info:ename,
timestamp=1593394424573, value=cevent
2020 column=info:hiredate,
timestamp=1593394424573, value=2020-06-29
2020 column=info:job,
timestamp=1593394424573, value=developer
2020 column=info:mgr,
timestamp=1593394424573, value=201
2020 column=info:sal,
timestamp=1593394424573, value=20000.0
1 row(s) in 0.0540 seconds
hbase(main):005:0> truncate 'hbase_emp_table'
清洗数据
Truncating 'hbase_emp_table' table (it
may take a while):
-
Disabling table...
-
Truncating table...
0 row(s) in 7.7160 seconds