下面是我到目前为止掌握的代码://read data from Azure blobvar df = spark.read.parquet(some_path)
df.createOrReplaceTempView("data_sample")
//have some sqlqueries, the one below is justan example date,
c
1Group1计薪>0,<= 5Group2计薪>5,<=10Group3计薪>10,<=20计薪df.groupBy('STATE').agg(count('*') as group1).where('SALARY' >0 and 'SALARY' <=5)
.agg(count('*') as <em