2

在 DStream 中的每个 RDD 上运行 SparkSQL 时出现 ClassCastException。

java.lang.ClassCastException: org.apache.spark.rdd.ShuffledRDDPartition cannot be cast to org.apache.spark.rdd.ParallelCollectionPartition
    at org.apache.spark.rdd.ParallelCollectionRDD.compute(ParallelCollectionRDD.scala:102)
    at org.apache.spark.rdd.ParallelCollectionRDD.compute(ParallelCollectionRDD.scala:85)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
    at org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
    at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:111)
    at org.apache.spark.scheduler.Task.run(Task.scala:51)
    at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:187)
4

0 回答 0