Lil*_*ian 1 apache-spark pyspark jupyter
我使用了 jupyter notebook,pyspark,然后,我的第一个命令是:
rdd = sc.parallelize([2, 3, 4])
Run Code Online (Sandbox Code Playgroud)
然后,它表明
NameError Traceback (most recent call last)
<ipython-input-1-c540c4a1d203> in <module>()
----> 1 rdd = sc.parallelize([2, 3, 4])
NameError: name 'sc' is not defined.
Run Code Online (Sandbox Code Playgroud)
未定义如何修复此错误“sc”。
Have you initialized the SparkContext?
You could try this:
#Initializing PySpark
from pyspark import SparkContext, SparkConf
# #Spark Config
conf = SparkConf().setAppName("sample_app")
sc = SparkContext(conf=conf)
Run Code Online (Sandbox Code Playgroud)
| 归档时间: |
|
| 查看次数: |
11197 次 |
| 最近记录: |