我试图在IntelliJ IDE中配置Scala
我的Scala&我机器中的Spark版本
Welcome to Scala 2.12.5 (Java HotSpot(TM) 64-Bit Server VM, Java 1.8.0_121).
apache-spark/2.2.1
SBT档案
scalaVersion := "2.12.5"
resolvers += "MavenRepository" at "http://central.maven.org/maven2"
libraryDependencies ++= {
val sparkVersion = "2.2.1"
Seq( "org.apache.spark" %% "spark-core" % sparkVersion)
}
错误我得到了
Error:Error while importing SBT project:<br/>...<br/><pre>[info] Resolving jline#jline;2.14.5 ...
[error] (*:ssExtractDependencies) sbt.ResolveException: unresolved dependency: org.apache.spark#spark-core_2.12;2.2.1: not found
[error] unresolved dependency: org.apache.spark#spark-core_2.12;1.4.0: not found
[error] (*:update) sbt.ResolveException: unresolved dependency: org.apache.spark#spark-core_2.12;2.2.1: not found
[error] unresolved dependency: org.apache.spark#spark-core_2.12;1.4.0: not found
答案 0 :(得分:3)
您在sbt项目中定义的spark核心版本无法下载。您可以查看maven dependency以获取有关可用版本的更多信息
正如您所看到的,对于spark-core 2.2.1版,要下载的最新版本是在Scala 2.11中编译的 info here
所以
您将sbt构建文件更改为
scalaVersion := "2.11.8"
resolvers += "MavenRepository" at "http://central.maven.org/maven2"
libraryDependencies ++= {
val sparkVersion = "2.2.1"
Seq( "org.apache.spark" %% "spark-core" % sparkVersion)
}
或将依赖项中的构建版本定义为
libraryDependencies ++= {
val sparkVersion = "2.2.1"
Seq("org.apache.spark" % "spark-core_2.11" % sparkVersion)
}
我希望答案很有帮助
答案 1 :(得分:0)
Spark-2.2.1
不支持scalaVersion-2.12
。你必须这样做:
scalaVersion:=“2.11.8”
libraryDependencies + =“org.apache.spark”%“spark-core”%“$ sparkVersion”
由于