它可以在没有聚合或计数的情况下迭代 Pyspark groupBy 数据帧吗?
例如 Pandas 中的代码:
for i, d in df2:
mycode ....
^^ if using pandas ^^
Is there a difference in how to iterate groupby in Pyspark or have to use aggregation and count?
Run Code Online (Sandbox Code Playgroud)