: val result = CurrentCost(df).aggregate 这将根据调用的实现生成运行时异常。对于实施1: Caused by: java.io.NotSerializableException: org.apache.spark.sql.Column
Serialization stack:- object not serializable (class: org.apache.spark.sql.Column, value: CASE WHEN ((NOT de
当我试图显示Dataset值时,它也会抛出一个错误。我做错什么了?Task not serializable at org.apache.spark.util.ClosureCleaner$.ensureSerializable(ClosureCleaner.scala:403)
at org.apache.spark.util.ClosureC
at org.apache.spark.rdd.PairRDDFunctions.:112) at org.apache.spark.rdd.PairRDDFunctions.aggregateByKeyat big_data.job_one.App$.main(App.scala:116)
at big_data.job_one.App.main(App.s
现在只剩下两个步骤:需要使用columns.Convert的相同子集对每个X_i执行特定的操作,每个X-i列都要进入FloatType.最低代码:import org.apache.spark.sql.types.FloatType
val foo = (FloatType))df.show()Exception in thread "main" java.la