硬件:
Dell PowerEdge SC420 Pentium 4GB RAM 80GBHDD
系统:
CentOS 6.2
下载地址:
ftp://mirror.nandomedia.com/pub/CentOS/6.2/isos/i386/CentOS-6.2-i386-LiveCD.iso
烧制CD
从CD引导启动
安装CentOS 6.2
从桌面单击Install onto Hard Disk Drive
重启
以root登录
软件:
下载java地址:
http://download.oracle.com/otn-pub/java/jdk/7u4-b20/jdk-7u4-linux-i586.rpm
# rpm -ivh {downloadfolder}/jdk-7u4-linux-i586.rpm
安装java 到 /usr/java/jdk1.7.0_04
# java -version
java version "1.7.0_04"
Java(TM) SE Runtime Environment (build1.7.0_04-b20)
Java HotSpot(TM) Client VM (build 23.0-b21,mixed mode, sharing)
# cd
# pwd
/root
# vi .bashrc
添加以下两行到文本最后
export JAVA_HOME=/usr/java/jdk1.7.0_04
export PATH=${JAVA_HOME}/bin:${PATH}
按键 Esc
输入 : wq
# . ~/.bashrc
下载Hadoop:
http://apache.mirrorcatalogs.com/hadoop/common/hadoop-2.0.0-alpha/hadoop-2.0.0-alpha.tar.gz
# cd /usr
# tar -xvzf {downloadfolder}/hadoop-2.0.0-alpha.tar.gz
# ls /usr/hadoop-2.0.0-alpha
bin include lib LICENSE.txt output sbin
etc input libexec NOTICE.txt README.txtshare
# service sshd status
openssh-daemon (pid xxxx) is running...
如果不能运行则# service sshd restart
# groupadd hadoop
# useradd -g hadoop -m hadoop
# cd /usr/hadoop-2.0.0-alpha
# chown -R hadoop:hadoop
# su - hadoop
# ssh-keygen -t rsa -P ''
当提示需要一个文件夹时,点击回车。
# cat .ssh/id_rsa.pub >>.ssh/authorized_keys
# ssh localhost
登入时将不会提示需要密码
# exit
# whoami
hadoop
# bin/hadoop version
Hadoop 2.0.0-alpha
Subversionhttp://svn.apache.org/repos/asf/hadoop/common/branches/branch-2.0.0-alpha/hadoop-common-project/hadoop-common-r 1338348
Compiled by hortonmu on Wed May 16 01:28:50UTC 2012
From source with checksum954e3f6c91d058b06b1e81a02813303f
# mkdir input
# cpshare/hadoop/common/templates/conf/*.xml input
# bin/hadoop jar share/hadoop/mapreduce/hadoop-mapreduce-examples-2.0.0-alpha.jargrep input output 'dfs[a-z.]+'
...
INFO input.FileInputFormat: Total inputpaths to process : 6
...
...
12/05/27 03:38:30 INFO mapreduce.Job:Counters: 27
File System Counters
FILE: Number of bytes read=91368
FILE: Number of bytes written=505456
FILE: Number of read operations=0
FILE: Number of large read operations=0
FILE: Number of write operations=0
Map-Reduce Framework
Map input records=32
Map output records=32
Map output bytes=1006
Map output materialized bytes=1076
Input split bytes=133
Combine input records=0
Combine output records=0
Reduce input groups=3
Reduce shuffle bytes=0
Reduce input records=32
Reduce output records=32
Spilled Records=64
Shuffled Maps =0
Failed Shuffles=0
Merged Map outputs=0
GC time elapsed (ms)=50
CPU time spent (ms)=0
Physical memory (bytes) snapshot=0
Virtual memory (bytes) snapshot=0
Total committed heap usage(bytes)=242368512
File Input Format Counters
Bytes Read=1368
File Output Format Counters
Bytes Written=830
# ls input/
capacity-scheduler.xml core-site.xmlhadoop-policy.xml hdfs-site.xml mapred-queue-acls.xml mapred-site.xml
{matching the, INFO input.FileInputFormat:Total input paths to process : 6}
# cat output/part-r-00000
3 dfs.datanode.data.dir
2 dfs.namenode.http
2 dfs.namenode.https
1 dfsqa
1 dfsadmin
1 dfs.webhdfs.enabled
1 dfs.umaskmode
1 dfs.support.append
1 dfs.secondary.namenode.keytab.file
1 dfs.secondary.namenode.kerberos.principal
1 dfs.secondary.namenode.kerberos.https.principal
1 dfs.permissions.superusergroup
1 dfs.namenode.secondary.http
1 dfs.namenode.safemode.threshold
1 dfs.namenode.replication.min.
1 dfs.namenode.name.dir
1 dfs.namenode.keytab.file
1 dfs.namenode.kerberos.principal
1 dfs.namenode.kerberos.https.principal
1 dfs.include
1 dfs.https.port
1 dfs.hosts.exclude
1 dfs.hosts
1 dfs.exclude
1 dfs.datanode.keytab.file
1 dfs.datanode.kerberos.principal
1 dfs.datanode.http.address
1 dfs.datanode.data.dir.perm
1 dfs.datanode.address
1 dfs.cluster.administrators
1 dfs.block.access.token.enable
1 dfs
{matching the, Reduce output records=32}
完成,快速安装并核实hadoop 2.0.0 alpha独立版
引用
http://hadoop.apache.org/common/docs/stable/single_node_setup.html