File tree Expand file tree Collapse file tree
spark/src/main/scala/org/apache/spark/sql/comet Expand file tree Collapse file tree Original file line number Diff line number Diff line change @@ -65,7 +65,7 @@ private[spark] class CometExecRDD(
6565 subqueries : Seq [ScalarSubquery ],
6666 broadcastedHadoopConfForEncryption : Option [Broadcast [SerializableConfiguration ]] = None ,
6767 encryptedFilePaths : Seq [String ] = Seq .empty)
68- extends RDD [ColumnarBatch ](sc, Nil ) {
68+ extends RDD [ColumnarBatch ](sc, inputRDDs.map(rdd => new OneToOneDependency (rdd)) ) {
6969
7070 // Determine partition count: from inputs if available, otherwise from parameter
7171 private val numPartitions : Int = if (inputRDDs.nonEmpty) {
@@ -133,9 +133,6 @@ private[spark] class CometExecRDD(
133133 it
134134 }
135135
136- override def getDependencies : Seq [Dependency [_]] =
137- inputRDDs.map(rdd => new OneToOneDependency (rdd))
138-
139136 // Duplicates logic from Spark's ZippedPartitionsBaseRDD.getPreferredLocations
140137 override def getPreferredLocations (split : Partition ): Seq [String ] = {
141138 if (inputRDDs.isEmpty) return Nil
You can’t perform that action at this time.
0 commit comments