环境:
尝试执行以下脚本:
from pyspark.sql import SparkSession
from pyspark_llap import HiveWarehouseSession
# Create spark session
spark = SparkSession.builder.appName("LLAP Test - CLI").enableHiveSupport().getOrCreate()
# Create HWC session
hive = HiveWarehouseSession.session(spark).userPassword('hive','hive').build()
# Execute a query to read from Spark using HWC
hive.executeQuery("select * from wifi_table where partit='2019-12-02'").show(20)
Run Code Online (Sandbox Code Playgroud)
问题: 当使用spark-submit提交应用程序或使用pyspark shell与上述脚本(或任何使用HiveWarehouseSession执行查询的脚本)时,spark作业卡住,抛出异常:java.lang.RuntimeException:java .io.IOException:shadecurator.org.apache.curator.CuratorConnectionLossException:KeeperErrorCode = ConnectionLoss
执行的命令如下:
$ …Run Code Online (Sandbox Code Playgroud)