spark on yarn提交job错误

java.io.IOException: Cannot run program "/etc/hadoop/conf/topology_script.py" (in directory "I:\软件开发\zmap"): CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1048)
 at org.apache.hadoop.util.Shell.runCommand(Shell.java:889)
 at org.apache.hadoop.util.Shell.run(Shell.java:855)
 at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1163)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.runResolveCommand(ScriptBasedMapping.java:251)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.resolve(ScriptBasedMapping.java:188)
 at org.apache.hadoop.net.CachedDNSToSwitchMapping.resolve(CachedDNSToSwitchMapping.java:119)
 at org.apache.hadoop.yarn.util.RackResolver.coreResolve(RackResolver.java:101)
 at org.apache.hadoop.yarn.util.RackResolver.resolve(RackResolver.java:81)
 at org.apache.spark.scheduler.cluster.YarnScheduler.getRackForHost(YarnScheduler.scala:37)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$1.apply(TaskSchedulerImpl.scala:337)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$1.apply(TaskSchedulerImpl.scala:326)
 at scala.collection.Iterator$class.foreach(Iterator.scala:893)
 at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
 at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
 at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
 at org.apache.spark.scheduler.TaskSchedulerImpl.resourceOffers(TaskSchedulerImpl.scala:326)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint.org$apache$spark$scheduler$cluster$CoarseGrainedSchedulerBackend$DriverEndpoint$$makeOffers(CoarseGrainedSchedulerBackend.scala:237)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint$$anonfun$receiveAndReply$1.applyOrElse(CoarseGrainedSchedulerBackend.scala:200)
 at org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:105)
 at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:205)
 at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:101)
 at org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:216)
 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
 at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessImpl.create(Native Method)
 at java.lang.ProcessImpl.(ProcessImpl.java:386)
 at java.lang.ProcessImpl.start(ProcessImpl.java:137)
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1029)
 ... 25 more
18/05/12 22:49:33 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 0.0 (TID 0, slave1, executor 1, partition 0, PROCESS_LOCAL, 4854 bytes)
18/05/12 22:49:33 INFO storage.BlockManagerMasterEndpoint: Registering block manager slave1:44181 with 366.3 MB RAM, BlockManagerId(1, slave1, 44181, None)
18/05/12 22:49:33 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint: Registered executor NettyRpcEndpointRef(spark-client://Executor) (192.168.9.54:33600) with ID 2
18/05/12 22:49:33 WARN net.ScriptBasedMapping: Exception running /etc/hadoop/conf/topology_script.py 192.168.9.54
java.io.IOException: Cannot run program "/etc/hadoop/conf/topology_script.py" (in directory "I:\软件开发\zmap"): CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1048)
 at org.apache.hadoop.util.Shell.runCommand(Shell.java:889)
 at org.apache.hadoop.util.Shell.run(Shell.java:855)
 at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1163)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.runResolveCommand(ScriptBasedMapping.java:251)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.resolve(ScriptBasedMapping.java:188)
 at org.apache.hadoop.net.CachedDNSToSwitchMapping.resolve(CachedDNSToSwitchMapping.java:119)
 at org.apache.hadoop.yarn.util.RackResolver.coreResolve(RackResolver.java:101)
 at org.apache.hadoop.yarn.util.RackResolver.resolve(RackResolver.java:81)
 at org.apache.spark.scheduler.cluster.YarnScheduler.getRackForHost(YarnScheduler.scala:37)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$1.apply(TaskSchedulerImpl.scala:337)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$1.apply(TaskSchedulerImpl.scala:326)
 at scala.collection.Iterator$class.foreach(Iterator.scala:893)
 at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
 at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
 at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
 at org.apache.spark.scheduler.TaskSchedulerImpl.resourceOffers(TaskSchedulerImpl.scala:326)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint.org$apache$spark$scheduler$cluster$CoarseGrainedSchedulerBackend$DriverEndpoint$$makeOffers(CoarseGrainedSchedulerBackend.scala:237)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint$$anonfun$receiveAndReply$1.applyOrElse(CoarseGrainedSchedulerBackend.scala:200)
 at org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:105)
 at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:205)
 at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:101)
 at org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:216)
 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
 at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessImpl.create(Native Method)
 at java.lang.ProcessImpl.(ProcessImpl.java:386)
 at java.lang.ProcessImpl.start(ProcessImpl.java:137)
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1029)
 ... 25 more
18/05/12 22:49:33 WARN net.ScriptBasedMapping: Exception running /etc/hadoop/conf/topology_script.py 192.168.9.59
java.io.IOException: Cannot run program "/etc/hadoop/conf/topology_script.py" (in directory "I:\软件开发\zmap"): CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1048)
 at org.apache.hadoop.util.Shell.runCommand(Shell.java:889)
 at org.apache.hadoop.util.Shell.run(Shell.java:855)
 at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1163)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.runResolveCommand(ScriptBasedMapping.java:251)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.resolve(ScriptBasedMapping.java:188)
 at org.apache.hadoop.net.CachedDNSToSwitchMapping.resolve(CachedDNSToSwitchMapping.java:119)
 at org.apache.hadoop.yarn.util.RackResolver.coreResolve(RackResolver.java:101)
 at org.apache.hadoop.yarn.util.RackResolver.resolve(RackResolver.java:81)
 at org.apache.spark.scheduler.cluster.YarnScheduler.getRackForHost(YarnScheduler.scala:37)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$1.apply(TaskSchedulerImpl.scala:337)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$1.apply(TaskSchedulerImpl.scala:326)
 at scala.collection.Iterator$class.foreach(Iterator.scala:893)
 at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
 at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
 at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
 at org.apache.spark.scheduler.TaskSchedulerImpl.resourceOffers(TaskSchedulerImpl.scala:326)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint.org$apache$spark$scheduler$cluster$CoarseGrainedSchedulerBackend$DriverEndpoint$$makeOffers(CoarseGrainedSchedulerBackend.scala:237)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint$$anonfun$receiveAndReply$1.applyOrElse(CoarseGrainedSchedulerBackend.scala:200)
 at org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:105)
 at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:205)
 at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:101)
 at org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:216)
 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
 at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessImpl.create(Native Method)
 at java.lang.ProcessImpl.(ProcessImpl.java:386)
 at java.lang.ProcessImpl.start(ProcessImpl.java:137)
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1029)
 ... 25 more
18/05/12 22:49:33 INFO scheduler.TaskSetManager: Starting task 1.0 in stage 0.0 (TID 1, slave3, executor 2, partition 1, PROCESS_LOCAL, 4855 bytes)
18/05/12 22:49:33 INFO storage.BlockManagerMasterEndpoint: Registering block manager slave3:38138 with 366.3 MB RAM, BlockManagerId(2, slave3, 38138, None)
18/05/12 22:49:33 INFO storage.BlockManagerInfo: Added broadcast_2_piece0 in memory on slave3:38138 (size: 1714.0 B, free: 366.3 MB)
18/05/12 22:49:33 WARN net.ScriptBasedMapping: Exception running /etc/hadoop/conf/topology_script.py 192.168.9.54
java.io.IOException: Cannot run program "/etc/hadoop/conf/topology_script.py" (in directory "I:\软件开发\zmap"): CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1048)
 at org.apache.hadoop.util.Shell.runCommand(Shell.java:889)
 at org.apache.hadoop.util.Shell.run(Shell.java:855)
 at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1163)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.runResolveCommand(ScriptBasedMapping.java:251)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.resolve(ScriptBasedMapping.java:188)
 at org.apache.hadoop.net.CachedDNSToSwitchMapping.resolve(CachedDNSToSwitchMapping.java:119)
 at org.apache.hadoop.yarn.util.RackResolver.coreResolve(RackResolver.java:101)
 at org.apache.hadoop.yarn.util.RackResolver.resolve(RackResolver.java:81)
 at org.apache.spark.scheduler.cluster.YarnScheduler.getRackForHost(YarnScheduler.scala:37)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$1.apply(TaskSchedulerImpl.scala:337)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$1.apply(TaskSchedulerImpl.scala:326)
 at scala.collection.Iterator$class.foreach(Iterator.scala:893)
 at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
 at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
 at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
 at org.apache.spark.scheduler.TaskSchedulerImpl.resourceOffers(TaskSchedulerImpl.scala:326)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint.org$apache$spark$scheduler$cluster$CoarseGrainedSchedulerBackend$DriverEndpoint$$makeOffers(CoarseGrainedSchedulerBackend.scala:237)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint$$anonfun$receive$1.applyOrElse(CoarseGrainedSchedulerBackend.scala:137)
 at org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:117)
 at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:205)
 at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:101)
 at org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:216)
 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
 at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessImpl.create(Native Method)
 at java.lang.ProcessImpl.(ProcessImpl.java:386)
 at java.lang.ProcessImpl.start(ProcessImpl.java:137)
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1029)
 ... 25 more
18/05/12 22:49:33 WARN net.ScriptBasedMapping: Exception running /etc/hadoop/conf/topology_script.py 192.168.9.59
java.io.IOException: Cannot run program "/etc/hadoop/conf/topology_script.py" (in directory "I:\软件开发\zmap"): CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1048)
 at org.apache.hadoop.util.Shell.runCommand(Shell.java:889)
 at org.apache.hadoop.util.Shell.run(Shell.java:855)
 at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1163)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.runResolveCommand(ScriptBasedMapping.java:251)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.resolve(ScriptBasedMapping.java:188)
 at org.apache.hadoop.net.CachedDNSToSwitchMapping.resolve(CachedDNSToSwitchMapping.java:119)
 at org.apache.hadoop.yarn.util.RackResolver.coreResolve(RackResolver.java:101)
 at org.apache.hadoop.yarn.util.RackResolver.resolve(RackResolver.java:81)
 at org.apache.spark.scheduler.cluster.YarnScheduler.getRackForHost(YarnScheduler.scala:37)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$1.apply(TaskSchedulerImpl.scala:337)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$1.apply(TaskSchedulerImpl.scala:326)
 at scala.collection.Iterator$class.foreach(Iterator.scala:893)
 at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
 at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
 at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
 at org.apache.spark.scheduler.TaskSchedulerImpl.resourceOffers(TaskSchedulerImpl.scala:326)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint.org$apache$spark$scheduler$cluster$CoarseGrainedSchedulerBackend$DriverEndpoint$$makeOffers(CoarseGrainedSchedulerBackend.scala:237)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint$$anonfun$receive$1.applyOrElse(CoarseGrainedSchedulerBackend.scala:137)
 at org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:117)
 at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:205)
 at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:101)
 at org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:216)
 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
 at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessImpl.create(Native Method)
 at java.lang.ProcessImpl.(ProcessImpl.java:386)
 at java.lang.ProcessImpl.start(ProcessImpl.java:137)
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1029)
 ... 25 more
18/05/12 22:49:33 INFO storage.BlockManagerInfo: Added broadcast_2_piece0 in memory on slave1:44181 (size: 1714.0 B, free: 366.3 MB)
18/05/12 22:49:34 WARN net.ScriptBasedMapping: Exception running /etc/hadoop/conf/topology_script.py 192.168.9.54
java.io.IOException: Cannot run program "/etc/hadoop/conf/topology_script.py" (in directory "I:\软件开发\zmap"): CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1048)
 at org.apache.hadoop.util.Shell.runCommand(Shell.java:889)
 at org.apache.hadoop.util.Shell.run(Shell.java:855)
 at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1163)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.runResolveCommand(ScriptBasedMapping.java:251)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.resolve(ScriptBasedMapping.java:188)
 at org.apache.hadoop.net.CachedDNSToSwitchMapping.resolve(CachedDNSToSwitchMapping.java:119)
 at org.apache.hadoop.yarn.util.RackResolver.coreResolve(RackResolver.java:101)
 at org.apache.hadoop.yarn.util.RackResolver.resolve(RackResolver.java:81)
 at org.apache.spark.scheduler.cluster.YarnScheduler.getRackForHost(YarnScheduler.scala:37)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$1.apply(TaskSchedulerImpl.scala:337)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$1.apply(TaskSchedulerImpl.scala:326)
 at scala.collection.Iterator$class.foreach(Iterator.scala:893)
 at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
 at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
 at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
 at org.apache.spark.scheduler.TaskSchedulerImpl.resourceOffers(TaskSchedulerImpl.scala:326)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint.org$apache$spark$scheduler$cluster$CoarseGrainedSchedulerBackend$DriverEndpoint$$makeOffers(CoarseGrainedSchedulerBackend.scala:261)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint$$anonfun$receive$1.applyOrElse(CoarseGrainedSchedulerBackend.scala:128)
 at org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:117)
 at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:205)
 at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:101)
 at org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:216)
 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
 at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessImpl.create(Native Method)
 at java.lang.ProcessImpl.(ProcessImpl.java:386)
 at java.lang.ProcessImpl.start(ProcessImpl.java:137)
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1029)
 ... 25 more
18/05/12 22:49:34 WARN net.ScriptBasedMapping: Exception running /etc/hadoop/conf/topology_script.py 192.168.9.54
java.io.IOException: Cannot run program "/etc/hadoop/conf/topology_script.py" (in directory "I:\软件开发\zmap"): CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1048)
 at org.apache.hadoop.util.Shell.runCommand(Shell.java:889)
 at org.apache.hadoop.util.Shell.run(Shell.java:855)
 at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1163)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.runResolveCommand(ScriptBasedMapping.java:251)
 at org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.resolve(ScriptBasedMapping.java:188)
 at org.apache.hadoop.net.CachedDNSToSwitchMapping.resolve(CachedDNSToSwitchMapping.java:119)
 at org.apache.hadoop.yarn.util.RackResolver.coreResolve(RackResolver.java:101)
 at org.apache.hadoop.yarn.util.RackResolver.resolve(RackResolver.java:81)
 at org.apache.spark.scheduler.cluster.YarnScheduler.getRackForHost(YarnScheduler.scala:37)
 at org.apache.spark.scheduler.TaskSetManager.dequeueTask(TaskSetManager.scala:404)
 at org.apache.spark.scheduler.TaskSetManager.resourceOffer(TaskSetManager.scala:457)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$org$apache$spark$scheduler$TaskSchedulerImpl$$resourceOfferSingleTaskSet$1.apply$mcVI$sp(TaskSchedulerImpl.scala:295)
 at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160)
 at org.apache.spark.scheduler.TaskSchedulerImpl.org$apache$spark$scheduler$TaskSchedulerImpl$$resourceOfferSingleTaskSet(TaskSchedulerImpl.scala:290)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$4$$anonfun$apply$9.apply(TaskSchedulerImpl.scala:375)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$4$$anonfun$apply$9.apply(TaskSchedulerImpl.scala:373)
 at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
 at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$4.apply(TaskSchedulerImpl.scala:373)
 at org.apache.spark.scheduler.TaskSchedulerImpl$$anonfun$resourceOffers$4.apply(TaskSchedulerImpl.scala:370)
 at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
 at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
 at org.apache.spark.scheduler.TaskSchedulerImpl.resourceOffers(TaskSchedulerImpl.scala:370)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint.org$apache$spark$scheduler$cluster$CoarseGrainedSchedulerBackend$DriverEndpoint$$makeOffers(CoarseGrainedSchedulerBackend.scala:261)
 at org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint$$anonfun$receive$1.applyOrElse(CoarseGrainedSchedulerBackend.scala:128)
 at org.apache.spark.rpc.netty.Inbox$$anonfun$process$1.apply$mcV$sp(Inbox.scala:117)
 at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:205)
 at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:101)
 at org.apache.spark.rpc.netty.Dispatcher$MessageLoop.run(Dispatcher.scala:216)
 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
 at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: CreateProcess error=2, 系统找不到指定的文件。
 at java.lang.ProcessImpl.create(Native Method)
 at java.lang.ProcessImpl.(ProcessImpl.java:386)
 at java.lang.ProcessImpl.start(ProcessImpl.java:137)
 at java.lang.ProcessBuilder.start(ProcessBuilder.java:1029)
 ... 32 more
18/05/12 22:49:34 WARN scheduler.TaskSetManager: Lost task 1.0 in stage 0.0 (TID 1, slave3, executor 2): java.lang.ClassNotFoundException: org.apache.hadoop.hbase.spark.HBaseContext$$anonfun$foreachPartition$1
 at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
 at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
 at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
 at java.lang.Class.forName0(Native Method)
 at java.lang.Class.forName(Class.java:348)
 at org.apache.spark.serializer.JavaDeserializationStream$$anon$1.resolveClass(JavaSerializer.scala:67)
 at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1859)
 at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1745)
 at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2033)
 at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1567)
 at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2278)
 at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2202)
 at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2060)
 at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1567)
 at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2278)
 at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2202)
 at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2060)
 at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1567)
 at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2278)
 at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2202)
 at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2060)
 at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1567)
 at java.io.ObjectInputStream.readObject(ObjectInputStream.java:427)
 at org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:75)
 at org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:114)
 at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:80)
 at org.apache.spark.scheduler.Task.run(Task.scala:108)
 at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338)
 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
 at java.lang.Thread.run(Thread.java:748)

你可能感兴趣的:(spark)