Skip to content

Commit ec0f453

Browse files
committed
fix: enable ParquetFilterSuite tests for native_datafusion (apache#3320)
Remove IgnoreCometNativeDataFusion tags from two ParquetFilterSuite tests that were disabled due to a CometNativeExec crash when executing without a serialized plan. The crash was fixed by PR apache#3511 which added a dedicated doExecuteColumnar() to CometNativeScanExec. The test assertions are already guarded by isCometEnabled/isCometScanOnly checks.
1 parent 8cfbca2 commit ec0f453

1 file changed

Lines changed: 11 additions & 29 deletions

File tree

dev/diffs/3.5.8.diff

Lines changed: 11 additions & 29 deletions
Original file line numberDiff line numberDiff line change
@@ -965,7 +965,7 @@ index 3cf2bfd17ab..49728c35c42 100644
965965
SQLConf.ANSI_ENABLED.key -> "true") {
966966
withTable("t") {
967967
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/StringFunctionsSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/StringFunctionsSuite.scala
968-
index fa1a64460fc..1d2e215d6a3 100644
968+
index fa1a64460fc..134f0db1fb8 100644
969969
--- a/sql/core/src/test/scala/org/apache/spark/sql/StringFunctionsSuite.scala
970970
+++ b/sql/core/src/test/scala/org/apache/spark/sql/StringFunctionsSuite.scala
971971
@@ -17,6 +17,8 @@
@@ -2038,7 +2038,7 @@ index 07e2849ce6f..3e73645b638 100644
20382038
ParquetOutputFormat.WRITER_VERSION -> ParquetProperties.WriterVersion.PARQUET_2_0.toString
20392039
)
20402040
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala
2041-
index 8e88049f51e..49f2001dc6b 100644
2041+
index 8e88049f51e..99561d9e7c7 100644
20422042
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala
20432043
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala
20442044
@@ -1095,7 +1095,11 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
@@ -2064,17 +2064,7 @@ index 8e88049f51e..49f2001dc6b 100644
20642064
import testImplicits._
20652065

20662066
withSQLConf(SQLConf.PARQUET_VECTORIZED_READER_ENABLED.key -> "true",
2067-
@@ -1548,7 +1553,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
2068-
}
2069-
}
2070-
2071-
- test("SPARK-31026: Parquet predicate pushdown for fields having dots in the names") {
2072-
+ test("SPARK-31026: Parquet predicate pushdown for fields having dots in the names",
2073-
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3320")) {
2074-
import testImplicits._
2075-
2076-
withAllParquetReaders {
2077-
@@ -1580,13 +1586,18 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
2067+
@@ -1580,7 +1585,11 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
20782068
// than the total length but should not be a single record.
20792069
// Note that, if record level filtering is enabled, it should be a single record.
20802070
// If no filter is pushed down to Parquet, it should be the total length of data.
@@ -2087,15 +2077,7 @@ index 8e88049f51e..49f2001dc6b 100644
20872077
}
20882078
}
20892079
}
2090-
}
2091-
2092-
- test("Filters should be pushed down for Parquet readers at row group level") {
2093-
+ test("Filters should be pushed down for Parquet readers at row group level",
2094-
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3320")) {
2095-
import testImplicits._
2096-
2097-
withSQLConf(
2098-
@@ -1607,7 +1618,11 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
2080+
@@ -1607,7 +1616,11 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
20992081
// than the total length but should not be a single record.
21002082
// Note that, if record level filtering is enabled, it should be a single record.
21012083
// If no filter is pushed down to Parquet, it should be the total length of data.
@@ -2108,7 +2090,7 @@ index 8e88049f51e..49f2001dc6b 100644
21082090
}
21092091
}
21102092
}
2111-
@@ -1699,7 +1714,7 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
2093+
@@ -1699,7 +1712,7 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
21122094
(attr, value) => sources.StringContains(attr, value))
21132095
}
21142096

@@ -2117,7 +2099,7 @@ index 8e88049f51e..49f2001dc6b 100644
21172099
import testImplicits._
21182100
// keep() should take effect on StartsWith/EndsWith/Contains
21192101
Seq(
2120-
@@ -1743,7 +1758,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
2102+
@@ -1743,7 +1756,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
21212103
}
21222104
}
21232105

@@ -2127,7 +2109,7 @@ index 8e88049f51e..49f2001dc6b 100644
21272109
val schema = StructType(Seq(
21282110
StructField("a", IntegerType, nullable = false)
21292111
))
2130-
@@ -1933,7 +1949,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
2112+
@@ -1933,7 +1947,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
21312113
}
21322114
}
21332115

@@ -2137,7 +2119,7 @@ index 8e88049f51e..49f2001dc6b 100644
21372119
withTempPath { dir =>
21382120
val count = 10
21392121
val tableName = "spark_25207"
2140-
@@ -1984,7 +2001,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
2122+
@@ -1984,7 +1999,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
21412123
}
21422124
}
21432125

@@ -2147,7 +2129,7 @@ index 8e88049f51e..49f2001dc6b 100644
21472129
// block 1:
21482130
// null count min max
21492131
// page-0 0 0 99
2150-
@@ -2044,7 +2062,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
2132+
@@ -2044,7 +2060,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared
21512133
}
21522134
}
21532135

@@ -2157,7 +2139,7 @@ index 8e88049f51e..49f2001dc6b 100644
21572139
withTempPath { dir =>
21582140
val path = dir.getCanonicalPath
21592141
spark.range(100).selectExpr("id * 2 AS id")
2160-
@@ -2276,7 +2295,11 @@ class ParquetV1FilterSuite extends ParquetFilterSuite {
2142+
@@ -2276,7 +2293,11 @@ class ParquetV1FilterSuite extends ParquetFilterSuite {
21612143
assert(pushedParquetFilters.exists(_.getClass === filterClass),
21622144
s"${pushedParquetFilters.map(_.getClass).toList} did not contain ${filterClass}.")
21632145

@@ -2170,7 +2152,7 @@ index 8e88049f51e..49f2001dc6b 100644
21702152
} else {
21712153
assert(selectedFilters.isEmpty, "There is filter pushed down")
21722154
}
2173-
@@ -2336,7 +2359,11 @@ class ParquetV2FilterSuite extends ParquetFilterSuite {
2155+
@@ -2336,7 +2357,11 @@ class ParquetV2FilterSuite extends ParquetFilterSuite {
21742156
assert(pushedParquetFilters.exists(_.getClass === filterClass),
21752157
s"${pushedParquetFilters.map(_.getClass).toList} did not contain ${filterClass}.")
21762158

0 commit comments

Comments
 (0)