小编Sum*_*mit的帖子

在yarn集群模式下运行spark应用程序时找不到HADOOP_HOME

我正在尝试以纱线集群模式运行应用程序。shell脚本的设置如下:

spark-submit --class "com.Myclass"  \
--num-executors 2 \
 --executor-cores 2 \
 --master yarn \
 --supervise \
 --deploy-mode cluster \
../target/ \
Run Code Online (Sandbox Code Playgroud)

此外,我收到以下错误。以下是来自 YARN LOGS APPLICATIONID 的错误详细信息

INFO : org.apache.spark.deploy.yarn.ApplicationMaster - Registered signal handlers for [TERM, HUP, INT]
DEBUG: org.apache.hadoop.util.Shell - Failed to detect a valid hadoop home directory
java.io.IOException: HADOOP_HOME or hadoop.home.dir are not set.
    at org.apache.hadoop.util.Shell.checkHadoopHome(Shell.java:307)
    at org.apache.hadoop.util.Shell.<clinit>(Shell.java:332)
    at org.apache.hadoop.util.StringUtils.<clinit>(StringUtils.java:79)
    at org.apache.hadoop.yarn.conf.YarnConfiguration.<clinit>(YarnConfiguration.java:590)
    at org.apache.spark.deploy.yarn.YarnSparkHadoopUtil.newConfiguration(YarnSparkHadoopUtil.scala:62)
    at org.apache.spark.deploy.SparkHadoopUtil.<init>(SparkHadoopUtil.scala:52)
    at org.apache.spark.deploy.yarn.YarnSparkHadoopUtil.<init>(YarnSparkHadoopUtil.scala:47)
Run Code Online (Sandbox Code Playgroud)

我尝试像下面这样修改spark-env.sh,我看到 Hadoop_Home 已记录,但仍然出现上述错误。修改并添加以下条目到spark-env.sh

export HADOOP_HOME="/usr/lib/hadoop"
echo "&&&&&&&&&&&&&&&&&&&&&& HADOOP HOME " 
echo …
Run Code Online (Sandbox Code Playgroud)

hadoop-yarn apache-spark

5
推荐指数
1
解决办法
2万
查看次数

标签 统计

apache-spark ×1

hadoop-yarn ×1