java.net.ConnectException: Connection timed out: no further information

首先,本地读取云集群hdfs文件报连接超时,这个问题困的我。。。。。。。
百度,谷歌很多都是防火墙,端口,路径的错误,这些都不是这个问题的根源,防火墙开着也可以读取hdfs中的文件,所以有必要记下这个问题,错误如下:

19/06/03 15:25:09 WARN BlockReaderFactory: I/O error constructing remote block reader.
	java.net.ConnectException: Connection timed out: no further information
		at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
		at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
		at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
		at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
		at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
		at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
		at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
		at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
		at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
		at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
		at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
		at java.io.DataInputStream.read(DataInputStream.java:100)
		at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
		at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
		at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
		at org.apache.hadoop.mapred.LineRecordReader.(LineRecordReader.java:134)
		at org.apache.hadoop.mapred.TextInputFormat.getRecordReader(TextInputFormat.java:67)
		at org.apache.spark.rdd.HadoopRDD$$anon$1.liftedTree1$1(HadoopRDD.scala:257)
		at org.apache.spark.rdd.HadoopRDD$$anon$1.(HadoopRDD.scala:256)
		at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:214)
		at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:94)
		at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:324)
		at org.apache.spark.rdd.RDD.iterator(RDD.scala:288)
		at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:49)
		at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:324)
		at org.apache.spark.rdd.RDD.iterator(RDD.scala:288)
		at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
		at org.apache.spark.scheduler.Task.run(Task.scala:109)
		at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
		at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
		at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
		at java.lang.Thread.run(Thread.java:748)
	19/06/03 15:25:09 WARN BlockReaderFactory: I/O error constructing remote block reader.
	java.net.ConnectException: Connection timed out: no further information
		at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
		at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
		at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
		at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
		at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
		at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
		at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
		at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
		at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
		at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
		at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
		at java.io.DataInputStream.read(DataInputStream.java:100)
		at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
		at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
		at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
		at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
		at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
		at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
		at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:277)
		at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:214)
		at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
		at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:37)
		at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
		at scala.collection.Iterator$class.foreach(Iterator.scala:893)
		at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
		at org.apache.spark.rdd.RDD$$anonfun$foreach$1$$anonfun$apply$28.apply(RDD.scala:927)
		at org.apache.spark.rdd.RDD$$anonfun$foreach$1$$anonfun$apply$28.apply(RDD.scala:927)
		at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2074)
		at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2074)
		at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
		at org.apache.spark.scheduler.Task.run(Task.scala:109)
		at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
		at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
		at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
		at java.lang.Thread.run(Thread.java:748)
	19/06/03 15:25:09 WARN DFSClient: Failed to connect to /IP:50010 for block, add to deadNodes and continue. java.net.ConnectException: Connection timed out: no further information
	java.net.ConnectException: Connection timed out: no further information
		at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
		at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
		at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
		at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
		at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
		at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
		at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
		at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
		at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
		at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
		at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
		at java.io.DataInputStream.read(DataInputStream.java:100)
		at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
		at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
		at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
		at org.apache.hadoop.mapred.LineRecordReader.(LineRecordReader.java:134)
		at org.apache.hadoop.mapred.TextInputFormat.getRecordReader(TextInputFormat.java:67)
		at org.apache.spark.rdd.HadoopRDD$$anon$1.liftedTree1$1(HadoopRDD.scala:257)
		at org.apache.spark.rdd.HadoopRDD$$anon$1.(HadoopRDD.scala:256)
		at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:214)
		at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:94)
		at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:324)
		at org.apache.spark.rdd.RDD.iterator(RDD.scala:288)
		at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:49)
		at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:324)
		at org.apache.spark.rdd.RDD.iterator(RDD.scala:288)
		at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
		at org.apache.spark.scheduler.Task.run(Task.scala:109)
		at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
		at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
		at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
		at java.lang.Thread.run(Thread.java:748)
	19/06/03 15:25:09 WARN DFSClient: Failed to connect to /IP:50010 for block, add to deadNodes and continue. java.net.ConnectException: Connection timed out: no further information
	java.net.ConnectException: Connection timed out: no further information
		at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
		at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
		at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
		at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
		at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer(DFSClient.java:3101)
		at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer(BlockReaderFactory.java:755)
		at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:670)
		at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:337)
		at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:576)
		at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:800)
		at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:847)
		at java.io.DataInputStream.read(DataInputStream.java:100)
		at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)
		at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)
		at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)
		at org.apache.hadoop.mapred.LineRecordReader.skipUtfByteOrderMark(LineRecordReader.java:206)
		at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:244)
		at org.apache.hadoop.mapred.LineRecordReader.next(LineRecordReader.java:47)
		at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:277)
		at org.apache.spark.rdd.HadoopRDD$$anon$1.getNext(HadoopRDD.scala:214)
		at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
		at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:37)
		at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:408)
		at scala.collection.Iterator$class.foreach(Iterator.scala:893)
		at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
		at org.apache.spark.rdd.RDD$$anonfun$foreach$1$$anonfun$apply$28.apply(RDD.scala:927)
		at org.apache.spark.rdd.RDD$$anonfun$foreach$1$$anonfun$apply$28.apply(RDD.scala:927)
		at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2074)
		at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2074)
		at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
		at org.apache.spark.scheduler.Task.run(Task.scala:109)
		at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
		at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
		at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
		at java.lang.Thread.run(Thread.java:748)	

解决方案:
问题在于但我们使用本地读取hdfs文件时,云集群datanode使用的是内网进行通信,所以我们把hdfs-site.xml和core-site.xml文件导入到Idea中时,需要修改hdfs-site.xml中的以下配置:


        dfs.datanode.use.datanode.hostname
        true

通过查阅资料,建议在hdfs-site.xml中增加dfs.datanode.
use.datanode.hostname属性,表示datanode之间的通信也通过域名方式

你可能感兴趣的:(bugbug)