相关疑难解决方法(0)

为什么要删除Spark执行程序(在日志中使用"ExecutorAllocationManager:请求删除executorIds")?

我试图在6个c4.2xlarge节点的AWS集群中执行spark工作,我不知道为什么Spark会杀死执行程序...任何帮助将不胜感激

这里的spark submit命令:

. /usr/bin/spark-submit --packages="com.databricks:spark-avro_2.11:3.2.0" --jars RedshiftJDBC42-1.2.1.1001.jar --deploy-mode client --master yarn --num-executors 12 --executor-cores 3 --executor-memory 7G --driver-memory 7g --py-files dependencies.zip iface_extractions.py 2016-10-01 > output.log
Run Code Online (Sandbox Code Playgroud)

在这一行,这一行开始删除执行程序

17/05/25 14:42:50 INFO ExecutorAllocationManager: Request to remove executorIds: 5, 3
Run Code Online (Sandbox Code Playgroud)

输出spark-submit日志:

Ivy Default Cache set to: /home/hadoop/.ivy2/cache
The jars for the packages stored in: /home/hadoop/.ivy2/jars
:: loading settings :: url = jar:file:/usr/lib/spark/jars/ivy-2.4.0.jar!/org/apache/ivy/core/settings/ivysettings.xml
com.databricks#spark-avro_2.11 added as a dependency
:: resolving dependencies :: org.apache.spark#spark-submit-parent;1.0
    confs: [default]
    found com.databricks#spark-avro_2.11;3.2.0 in central
    found org.slf4j#slf4j-api;1.7.5 in central
    found …
Run Code Online (Sandbox Code Playgroud)

hadoop-yarn apache-spark pyspark

4
推荐指数
1
解决办法
5602
查看次数

标签 统计

apache-spark ×1

hadoop-yarn ×1

pyspark ×1