1
我正在使用spark 1.3.0。我的build.sbt看起来如下:从sbt文件中的火花库中排除hadoop依赖项
libraryDependencies ++= Seq(
"org.apache.spark" %% "spark-core" % "1.3.0" % "provided",
"org.apache.spark" %% "spark-sql" % "1.3.0" % "provided",
"org.apache.spark" %% "spark-streaming" % "1.3.0" % "provided",
"org.apache.spark" %% "spark-mllib" % "1.3.0" % "provided",
"org.springframework.security" % "spring-security-web" % "3.0.7.RELEASE",
"com.databricks" % "spark-csv_2.10" % "1.4.0"
)
// META-INF discarding
mergeStrategy in assembly <<= (mergeStrategy in assembly) { (old) =>
{
case PathList("META-INF", xs @ _*) => MergeStrategy.discard
case x => MergeStrategy.first
}
}
在编译期间使用此sbt文件hadoop 2.2.0。但我的运行环境包含hadoop 2.6.0。任何人都可以帮助我如何从spark库中排除hadoop依赖项,并在sbt文件中提到hadoop 2.6.0?
谢谢