Cod*_*joy 1 scala sbt apache-spark
我必须建立一个jar
从我scala
使用的代码sbt
:
sudo sbt package
Run Code Online (Sandbox Code Playgroud)
它使用了我的构建文件:
name := "PSG CCD"
version := "1.0"
scalaVersion := "2.11.8"
resolvers += "Spark Packages Repo" at "http://dl.bintray.com/spark-packages/maven"
libraryDependencies ++= Seq(
"org.apache.spark" %% "spark-core" % "2.2.0",
"org.apache.spark" %% "spark-sql" % "2.2.0",
"org.apache.spark" %% "spark-streaming" % "2.2.0",
"neo4j-contrib" % "neo4j-spark-connector" % "2.0.0-M2"
)
Run Code Online (Sandbox Code Playgroud)
我也建好了jar
.然后,我通过FTP将其传输到我的spark
服务器并执行以下操作spark-submit
:
spark-submit --class "PSGApp" --master local[4] psg_ccd.jar
Run Code Online (Sandbox Code Playgroud)
我收到此错误:
Exception in thread "main" java.lang.NoClassDefFoundError: org/neo4j/spark/Neo4j
at java.lang.Class.getDeclaredMethods0(Native Method)
at java.lang.Class.privateGetDeclaredMethods(Class.java:2701)
at java.lang.Class.privateGetMethodRecursive(Class.java:3048)
at java.lang.Class.getMethod0(Class.java:3018)
at java.lang.Class.getMethod(Class.java:1784)
at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:739)
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:119)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.ClassNotFoundException: org.neo4j.spark.Neo4j
at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
[truncated, 10 more messages]
Run Code Online (Sandbox Code Playgroud)
所以即使我neo4j-spark-connector
在依赖文件中有它,但是当我运行它时似乎找不到它spark
.现在,这spark
是跑了不同的机器上我建立了jar
从scala
上.这有关系吗?是否有一个lib文件夹我需要复制并删除某处?
我猜测neo4j
spark
连接依赖的库信息没有内置到jar
我试图运行.
也许我错过了一个强制转换的开关?
您需要使用sbt assembly插件来生成具有依赖项的胖jar.Stackoverflow中有很多示例.我向您展示一个:如何在IntelliJ IDEA中使用SBT构建Uber JAR(Fat JAR)?