我正在尝试在群集上启动Spark作业(Spark 1.4.0)。无论是命令行还是Eclipse,我都会收到有关Spark Utils类中缺少withDummyCallSite
函数的错误。在maven依赖项中,我可以看到spark-core_2.10-1.4.0.jar被加载,它应该包含这个函数。我正在运行Java 1.7,与之前编译代码的Java版本相同。我可以在Spark Master监视器上看到该作业已启动,因此它似乎不是防火墙问题。这是我在控制台中看到的错误(来自命令行和Eclipse):
ERROR 09:53:06,314 Logging.scala:75 -- Task 0 in stage 1.0 failed 4 times; aborting job
SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
SLF4J: Defaulting to no-operation (NOP) logger implementation
SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details.
java.lang.NoSuchMethodError: org.apache.spark.util.Utils$.withDummyCallSite(Lorg/apache/spark/SparkContext;Lscala/Function0;)Ljava/lang/Object;
at org.apache.spark.sql.parquet.ParquetRelation2.buildScan(newParquet.scala:269)
at org.apache.spark.sql.sources.HadoopFsRelation.buildScan(interfaces.scala:530)
at org.apache.spark.sql.sources.DataSourceStrategy$$anonfun$8.apply(DataSourceStrategy.scala:98)
at org.apache.spark.sql.sources.DataSourceStrategy$$anonfun$8.apply(DataSourceStrategy.scala:98)
at org.apache.spark.sql.sources.DataSourceStrategy$$anonfun$pruneFilterProject$1.apply(DataSourceStrategy.scala:266)
at org.apache.spark.sql.sources.DataSourceStrategy$$anonfun$pruneFilterProject$1.apply(DataSourceStrategy.scala:265)
at org.apache.spark.sql.sources.DataSourceStrategy$.pruneFilterProjectRaw(DataSourceStrategy.scala:296)
at org.apache.spark.sql.sources.DataSourceStrategy$.pruneFilterProject(DataSourceStrategy.scala:261)
at org.apache.spark.sql.sources.DataSourceStrategy$.apply(DataSourceStrategy.scala:94)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59)
at org.apache.spark.sql.catalyst.planning.QueryPlanner.planLater(QueryPlanner.scala:54)
at org.apache.spark.sql.execution.SparkStrategies$HashAggregation$.apply(SparkStrategies.scala:162)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59)
at org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan$lzycompute(SQLContext.scala:932)
at org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan(SQLContext.scala:930)
at org.apache.spark.sql.SQLContext$QueryExecution.executedPlan$lzycompute(SQLContext.scala:936)
at org.apache.spark.sql.SQLContext$QueryExecution.executedPlan(SQLContext.scala:936)
at org.apache.spark.sql.DataFrame.collect(DataFrame.scala:1255)
at org.apache.spark.sql.DataFrame.count(DataFrame.scala:1269)
(为简洁而截断日志)
提前感谢任何指针!
答案 0 :(得分:1)
请检查maven使用密钥(CNTR + Shift + T)如何解决您的课程。确保它没有从类路径中的两个不同的jar解析。
如果您的类是从任何传递依赖项引用的,请使用您需要的版本将所需的jar添加为直接依赖项。
您可以参考这些链接以获取更多参考。