./.virtualenv/etl-orderlines-generic-pivot/lib/python3.7/site-packages/pyspark/context.py:566: in parallelizedata, tempFile)
../../../.virtualenv/etl-orderlines-generic-pivot/lib/python3.7/site-packages&
我已经为pyspark dataframe中的两个特性应用了groupby和计算标准差。from pyspark.sql import functions as f
val1 = [('a',20,100),('a',100,100),('a',50,100),('b',0,100),('b',0,100),('c',0,0),('c',0,50),('c',0,100),('c