import org.apache.accumulo.core.client.ClientConfiguration
//conf.registerKryoClasses(Array(classOf[org.apache.accumulo.core.data.Key:51)
at org.apache.spark
:114) at org.apache.spark.scheduler.Task.run$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1499)
at org.apache.spark.
$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1273) ~[spark-core(DAGScheduler.scala:1264) ~[spark-core_2.10-1.4.1.3.jar:1.4.1.3]
at org.apache.spark.scheduler.DAGScheduler(DAGSchedu
它们在命令行中使用以下命令非常健谈(显然忽略了-Dtest= --所有核心测试都在运行中):7:03:30.251 INFO org.apache.spark.scheduler.TaskSetManager: Finished TID 4417 in 23ms on localhost (progress: 4/4)
17:
:63) at org.apache.spark.scheduler.DAGScheduler.submitMissingTasks(DAGScheduler.scala:1006)
at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler<
(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:88) at org.apache.spark.executor.Executor$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1283) at org.apache.spark.scheduler.DAG
(QueryExecutorImpl.java:2178) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:88)
at org.apache.spark.scheduler.Task$$anonfun$run$2.app