core-site.xml
fs.defaultFS
hdfs://earth
true
hadoop.tmp.dir
/data1/tmp-security
true
ha.zookeeper.quorum
hadoop-btzk0001.eniot.io:2181,hadoop-btzk0002.eniot.io:2181,hadoop-btzk0003.eniot.io:2181
ha.failover-controller.active-standby-elector.zk.op.retries
120
fs.du.interval
1200000
fs.df.interval
60000
fs.trash.interval
1440
true
io.file.buffer.size
65536
io.compression.codecs
org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec,com.hadoop.compression.lzo.LzoCodec,com.hadoop.compression.lzo.LzopCodec,org.apache.hadoop.io.compress.BZip2Codec,org.apache.hadoop.io.compress.SnappyCodec,org.apache.hadoop.io.compress.Lz4Codec
io.compression.codec.lzo.class
com.hadoop.compression.lzo.LzoCodec
lzo.text.input.format.ignore.nonlzo
false
hadoop.ssl.hostname.verifier
ALLOW_ALL
hdfs-site.xml
dfs.webhdfs.enabled
false
dfs.replication
3
dfs.nameservices
earth
dfs.ha.namenodes.earth
nn1,nn2
dfs.ha.automatic-failover.enabled
true
dfs.ha.fencing.methods
sshfence
shell(/bin/true)
dfs.ha.fencing.ssh.private-key-files
/home/hdfs/.ssh/id_rsa
dfs.ha.fencing.ssh.connect-timeout
30000
dfs.permissions.enabled
true
dfs.permissions.superusergroup
hadoop
dfs.cluster.administrators
hadoop
dfs.data.transfer.protection
integrity
dfs.blockreport.initialDelay
180
dfs.blockreport.intervalMsec
21600000
dfs.blockreport.split.threshold
1000000
dfs.block.scanner.volume.bytes.per.second
1048576
dfs.namenode.resource.du.reserved
100G
dfs.namenode.name.dir.restore
false
dfs.namenode.checkpoint.txns
10000000
dfs.namenode.rpc-address.earth.nn1
hadoop-btnn0001.eniot.io:8020
dfs.namenode.https-address.earth.nn1
hadoop-btnn0001.eniot.io:50470
dfs.namenode.rpc-address.earth.nn2
hadoop-btnn0002.eniot.io:8020
dfs.namenode.https-address.earth.nn2
hadoop-btnn0002.eniot.io:50470
dfs.namenode.acls.enabled
True
dfs.namenode.name.dir
/data1/hdfs/hadoopNNData
true
dfs.namenode.shared.edits.dir
qjournal://hadoop-btnn0001.eniot.io:8485;hadoop-btnn0002.eniot.io:8485;hadoop-btnn0003.eniot.io:8485/earth
dfs.namenode.delegation.token.max-lifetime
31536000000
dfs.namenode.delegation.token.renew-interval
31536000000
dfs.namenode.checkpoint.period
7200
dfs.datanode.scan.period.hours
504
dfs.datanode.handler.count
10
dfs.datanode.available-space-volume-choosing-policy.balanced-space-preference-fraction
0.75f
dfs.datanode.available-space-volume-choosing-policy.balanced-space-threshold
10737418240
dfs.datanode.address
0.0.0.0:50010
dfs.datanode.http.address
0.0.0.0:50075
dfs.datanode.data.dir
/data1/dfs/data
true
balancer时,hdfs移动数据的速度,默认值为1M/S的速度。一般情况下设置为10M
50485760
dfs.datanode.balance.max.concurrent.moves
25
dfs.datanode.du.reserved
0
dfs.journalnode.edits.dir
/data1/hdfs/journal
dfs.http.policy
HTTPS_ONLY
dfs.https.port
50470
dfs.image.transfer.timeout
1800000
dfs.image.transfer.bandwidthPerSec
5242880
dfs.hosts.exclude
/usr/local/hadoop/hadoop-release/etc/hadoop/hosts-dfs.exclude
dfs.block.access.token.enable
true
dfs.client.read.shortcircuit
true
dfs.domain.socket.path
/var/lib/hadoop-hdfs/dn_socket
dfs.client.failover.proxy.provider.earth
org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider
dfs.namenode.keytab.file
/etc/security/keytab/hdfs.keytab
dfs.namenode.kerberos.principal
hdfs/[email protected]
dfs.datanode.keytab.file
/etc/security/keytab/hdfs.keytab
dfs.datanode.kerberos.principal
hdfs/[email protected]
dfs.journalnode.keytab.file
/etc/security/keytab/hdfs.keytab
dfs.journalnode.kerberos.principal
hdfs/[email protected]
dfs.journalnode.kerberos.internal.spnego.principal
HTTP/[email protected]
dfs.web.authentication.kerberos.keytab
/etc/security/keytab/hdfs.keytab
dfs.web.authentication.kerberos.principal
HTTP/[email protected]
dfs.permissions
true
dfs.namenode.inode.attributes.provider.class
org.apache.ranger.authorization.hadoop.RangerHdfsAuthorizer
dfs.qjournal.start-segment.timeout.ms
20000
dfs.qjournal.prepare-recovery.timeout.ms
120000
dfs.qjournal.accept-recovery.timeout.ms
120000
dfs.qjournal.finalize-segment.timeout.ms
120000
dfs.qjournal.select-input-streams.timeout.ms
20000
dfs.qjournal.get-journal-state.timeout.ms
120000
dfs.qjournal.new-epoch.timeout.ms
120000
dfs.qjournal.write-txns.timeout.ms
20000
yarn-site.xml
yarn.http.policy
HTTPS_ONLY
yarn.acl.enable
true
yarn.admin.acl
yarn,hadoop
yarn.web-proxy.address
hadoop-btnn0003.eniot.io:8089
yarn.resourcemanager.zk-state-store.parent-path
/rmstore_batch
yarn.resourcemanager.cluster-id
yarn-cluster-batch
yarn.resourcemanager.recovery.enabled
true
yarn.resourcemanager.ha.enabled
true
yarn.resourcemanager.ha.rm-ids
rm1,rm2
yarn.resourcemanager.connect.retry-interval.ms
2000
yarn.resourcemanager.am.max-retries
5
yarn.resourcemanager.zk-acl
world:anyone:rwcda
yarn.resourcemanager.zk-address
hadoop-btzk0001.eniot.io:2181,hadoop-btzk0002.eniot.io:2181,hadoop-btzk0003.eniot.io:2181
yarn.resourcemanager.store.class
org.apache.hadoop.yarn.server.resourcemanager.recovery.ZKRMStateStore
yarn.resourcemanager.hostname.rm1
hadoop-btnn0001.eniot.io
yarn.resourcemanager.scheduler.address.rm1
hadoop-btnn0001.eniot.io:8030
yarn.resourcemanager.resource-tracker.address.rm1
hadoop-btnn0001.eniot.io:8031
yarn.resourcemanager.address.rm1
hadoop-btnn0001.eniot.io:8032
yarn.resourcemanager.admin.address.rm1
hadoop-btnn0001.eniot.io:8033
yarn.resourcemanager.webapp.https.address.rm1
hadoop-btnn0001.eniot.io:8090
yarn.resourcemanager.hostname.rm2
hadoop-btnn0002.eniot.io
yarn.resourcemanager.scheduler.address.rm2
hadoop-btnn0002.eniot.io:8030
yarn.resourcemanager.resource-tracker.address.rm2
hadoop-btnn0002.eniot.io:8031
yarn.resourcemanager.address.rm2
hadoop-btnn0002.eniot.io:8032
yarn.resourcemanager.admin.address.rm2
hadoop-btnn0002.eniot.io:8033
yarn.resourcemanager.webapp.https.address.rm2
hadoop-btnn0002.eniot.io:8090
yarn.log.server.url
https://hadoop-btnn0003.eniot.io:19888/jobhistory/logs
yarn.resourcemanager.nodes.exclude-path
/usr/local/hadoop/hadoop-release/etc/hadoop/hosts-yarn.exclude
yarn.resourcemanager.proxy-user-privileges.enabled
true
yarn.nodemanager.linux-container-executor.group
hadoop
yarn.nodemanager.disk-health-checker.min-healthy-disks
0.25
yarn.nodemanager.disk-health-checker.max-disk-utilization-per-disk-percentage
95.0
yarn.nodemanager.disk-health-checker.min-free-space-per-disk-mb
0
yarn.nodemanager.vmem-check-enabled
false
yarn.nodemanager.recovery.enabled
true
yarn.nodemanager.linux-container-executor.path
/usr/local/hadoop/hadoop-release/bin/container-executor
yarn.nodemanager.address
0.0.0.0:45454
yarn.nodemanager.local-dirs
/data1/yarn/local
yarn.nodemanager.log-dirs
/data1/yarn/userlogs
yarn.nodemanager.log.retain-seconds
10800
yarn.nodemanager.aux-services
mapreduce_shuffle
yarn.nodemanager.resource.cpu-vcores
4
yarn.nodemanager.resource.memory-mb
28942
yarn.resourcemanager.scheduler.class
org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairScheduler
yarn.scheduler.maximum-allocation-mb
20480
yarn.scheduler.minimum-allocation-mb
128
yarn.scheduler.increment-allocation-mb
128
yarn.scheduler.fair.allocation.file
/usr/local/hadoop/hadoop-release/etc/hadoop/fair-scheduler.xml
yarn.scheduler.fair.max.assign
10
yarn.scheduler.fair.user-as-default-queue
false
yarn.scheduler.fair.allow-undeclared-pools
false
yarn.log-aggregation-enable
true
yarn.log-aggregation.retain-seconds
604800
yarn.log-aggregation.retain-check-interval-seconds
86400
yarn.nodemanager.remote-app-log-dir
/mr-history
yarn.nodemanager.remote-app-log-dir-suffix
logs
yarn.nodemanager.localizer.cache.cleanup.interval-ms
600000
yarn.nodemanager.log-aggregator.on-fail.remain-log-in-sec
259200
yarn.nodemanager.principal
yarn/[email protected]
yarn.nodemanager.keytab
/etc/security/keytab/yarn.keytab
yarn.resourcemanager.keytab
/etc/security/keytab/yarn.keytab
yarn.resourcemanager.principal
yarn/[email protected]
yarn.web-proxy.keytab
/etc/security/keytab/yarn.keytab
yarn.web-proxy.principal
HTTP/[email protected]
yarn.nodemanager.container-executor.class
org.apache.hadoop.yarn.server.nodemanager.LinuxContainerExecutor
mapred-site.xm
yarn.app.mapreduce.am.resource.mb
2048
mapreduce.framework.name
yarn
mapreduce.cluster.local.dir
/data1/yarn/mapred/data
mapreduce.task.io.sort.mb
300
mapred.child.env
LD_LIBRARY_PATH=/usr/local/hadoop/hadoop-release/lib/libexec
mapred.child.java.opts
-Xmx3072m -XX:-UseGCOverheadLimit
mapreduce.tasktracker.http.address
0.0.0.0:50060
mapreduce.task.userlog.limit.kb
51200
mapreduce.reduce.shuffle.memory.limit.percent
0.25
yarn.app.mapreduce.task.container.log.backups
1
yarn.app.mapreduce.am.staging-dir
/tmp/hadoop-yarn/staging
mapreduce.jobhistory.done-dir
${yarn.app.mapreduce.am.staging-dir}/history/done
mapreduce.jobhistory.address
hadoop-btnn0003.eniot.io:10020
mapreduce.jobhistory.cleaner.enable
true
mapreduce.jobhistory.max-age-ms
604800000
mapreduce.jobhistory.cleaner.interval-ms
86400000
mapreduce.jobhistory.move.interval-ms
180000
mapreduce.jobhistory.http.policy
HTTPS_ONLY
mapreduce.jobhistory.webapp.https.address
hadoop-btnn0003.eniot.io:19888
mapreduce.map.speculative
false
mapreduce.reduce.speculative
false
mapreduce.map.output.compress
true
mapreduce.map.output.compress.codec
com.hadoop.compression.lzo.LzoCodec
mapreduce.map.memory.mb
2048
mapreduce.reduce.memory.mb
4096
mapreduce.map.java.opts
-Duser.home=. -Djava.net.preferIPv4Stack=true -Xmx1800m -XX:+UseParNewGC -XX:+UseConcMarkSweepGC -XX:+UseCMSCompactAtFullCollection -XX:CMSFullGCsBeforeCompaction=15 -XX:CMSInitiatingOccupancyFraction=70
mapreduce.reduce.java.opts
-Duser.home=. -Djava.net.preferIPv4Stack=true -Xmx3600m -XX:+UseParNewGC -XX:+UseConcMarkSweepGC -XX:+UseCMSCompactAtFullCollection -XX:CMSFullGCsBeforeCompaction=15 -XX:CMSInitiatingOccupancyFraction=70
mapreduce.jobhistory.keytab
/etc/security/keytab/mapred.keytab
mapreduce.jobhistory.principal
mapred/[email protected]
fail-schduler.xml
0 mb, 0 vcores
0 mb, 0 vcores
50
fifo
1.0
600
32768 mb, 5 vcores
163840 mb, 24 vcores
50
fifo
1.0
600
0 mb, 0 vcores
0 mb, 0 vcores
50
fifo
1.0
600
32768 mb, 16 vcores
32768 mb, 16 vcores
50
fifo
1.0
600
0 mb, 0 vcores
8192 mb, 4 vcores
10
fifo
1.0
600
other
dfs.datanode.max.transfer.threads: 数据传输最大线程,默认:16384
dfs.datanode.ipc.address:
DN的IPC监听端口,写0的话监听在随机端口通过心跳传输给NN,默认:8010
dfs.namenode.checkpoint.dir:
standy NN节点存储 checkpoint
文件目录,默认:/hadoop/hdfs/namesecondary
dfs.namenode.stale.datanode.interval:
标记一个dn为“down机”时间间隔,即:如果
nn没有接到一个dn心跳超过这个时间间隔,dn将标记为“过期”,过期的间隔不能太小 (默认: 3000)
dfs.namenode.safemode.threshold-pct
指定应有多少比例的数据块满足最小副本数要求,默认0.999f 说明:
(NN在启动的时候首先进入安全模式,如果DN丢失的block达到此参数设置的比例,则系统会一直处于安全模式状态即只读状态。如果设为1则HDFS永远是处于SafeMode)