From 352bc6d3cb540f9697460c06b4e21437901de3c6 Mon Sep 17 00:00:00 2001 From: huaxingao Date: Tue, 23 Aug 2022 22:55:58 -0700 Subject: [PATCH 1/2] [SQL][SPARK-39528][FOLLOWUP] Make DynamicPartitionPruningV2FilterSuite extend DynamicPartitionPruningV2Suite --- .../org/apache/spark/sql/DynamicPartitionPruningSuite.scala | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala index fd213d120b6a4..6067cdecf2933 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala @@ -1807,8 +1807,7 @@ class DynamicPartitionPruningV2SuiteAEOn extends DynamicPartitionPruningV2Suite with EnableAdaptiveExecutionSuite abstract class DynamicPartitionPruningV2FilterSuite - extends DynamicPartitionPruningDataSourceSuiteBase { - override protected def runAnalyzeColumnCommands: Boolean = false + extends DynamicPartitionPruningV2Suite { override protected def initState(): Unit = { spark.conf.set("spark.sql.catalog.testcat", classOf[InMemoryTableWithV2FilterCatalog].getName) From 4d8e54604e51910af83e8d603108737f632c9ccb Mon Sep 17 00:00:00 2001 From: huaxingao Date: Wed, 24 Aug 2022 07:59:49 -0700 Subject: [PATCH 2/2] call super.initState --- .../org/apache/spark/sql/DynamicPartitionPruningSuite.scala | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala index 6067cdecf2933..f7b51db1c9020 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala @@ -1810,8 +1810,8 @@ abstract class DynamicPartitionPruningV2FilterSuite extends DynamicPartitionPruningV2Suite { override protected def initState(): Unit = { + super.initState() spark.conf.set("spark.sql.catalog.testcat", classOf[InMemoryTableWithV2FilterCatalog].getName) - spark.conf.set("spark.sql.defaultCatalog", "testcat") } }