作业由于Spark中的阶段故障而中止

时间:2019-02-26 22:25:14

标签: scala apache-spark yarn

当我提交我的spark作业时,它失败并显示以下错误,我试图理解为什么有时会说fileNotFound Exception。

Job aborted due to stage failure: Task 15 in stage 8.0 failed 4 times, most recent failure: Lost task 15.4 in stage 8.0 (TID 1769, dn37.prdc.local): java.io.FileNotFoundException: /data03/hadoop/yarn/local/usercache/user/appcache/application_123XXXXXXXXX_XXXX/blockmgr-42c15000-e62p-4c73-8mac-f6ec3d3da0f1/13/shuffle_10_15_0.index.cb9f4626-2765-4864-acb3-58f55aadce1e (No such file or directory)
        at java.io.FileOutputStream.open0(Native Method)
        at java.io.FileOutputStream.open(FileOutputStream.java:270)
        at java.io.FileOutputStream.<init>(FileOutputStream.java:213)
        at java.io.FileOutputStream.<init>(FileOutputStream.java:162)
        at org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:141)
        at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)
        at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
        at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)

任何人都可以表达为什么我们会看到此类错误。

谢谢, 宝贝

0 个答案:

没有答案