当我提交我的spark作业时,它失败并显示以下错误,我试图理解为什么有时会说fileNotFound Exception。
Job aborted due to stage failure: Task 15 in stage 8.0 failed 4 times, most recent failure: Lost task 15.4 in stage 8.0 (TID 1769, dn37.prdc.local): java.io.FileNotFoundException: /data03/hadoop/yarn/local/usercache/user/appcache/application_123XXXXXXXXX_XXXX/blockmgr-42c15000-e62p-4c73-8mac-f6ec3d3da0f1/13/shuffle_10_15_0.index.cb9f4626-2765-4864-acb3-58f55aadce1e (No such file or directory)
at java.io.FileOutputStream.open0(Native Method)
at java.io.FileOutputStream.open(FileOutputStream.java:270)
at java.io.FileOutputStream.<init>(FileOutputStream.java:213)
at java.io.FileOutputStream.<init>(FileOutputStream.java:162)
at org.apache.spark.shuffle.IndexShuffleBlockResolver.writeIndexFileAndCommit(IndexShuffleBlockResolver.scala:141)
at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:128)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
任何人都可以表达为什么我们会看到此类错误。
谢谢, 宝贝