在 DStream 中的每个 RDD 上运行 SparkSQL 时出现 ClassCastException。
java.lang.ClassCastException: org.apache.spark.rdd.ShuffledRDDPartition cannot be cast to org.apache.spark.rdd.ParallelCollectionPartition
at org.apache.spark.rdd.ParallelCollectionRDD.compute(ParallelCollectionRDD.scala:102)
at org.apache.spark.rdd.ParallelCollectionRDD.compute(ParallelCollectionRDD.scala:85)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:111)
at org.apache.spark.scheduler.Task.run(Task.scala:51)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:187)