from pyspark.sql.functions import pandas_udfimport numpy as npC:\opt\spark\spark-2.3.0-bin-hadoop2.7\pythonreturnType_placeholder)
C:\opt\spark\spark-2.3.0-
我试图按多个列进行分组,并将它们聚合,使它们在分组后成为一个列表。_python_agg_general(arg, *args, **kwargs) 2871
/usr/local/lib/python2.7/dist-packages/pandas/core/groupby.pyc in _python_agg_general(self, fun
使用Pandas数据帧按特性分组,我希望按列c_b分组,并计算列c_a和列c_c的唯一计数。输入文件,hello,python,numpy,0.0ho,c++,vector,0.0go/site-packages/pandas/core/frame.py", line 1997, in __getitem__