diff --git a/sql/core/benchmarks/WideTableBenchmark-results.txt b/sql/core/benchmarks/WideTableBenchmark-results.txt index 3b41a3e036c4d..7bc388aaa549f 100644 --- a/sql/core/benchmarks/WideTableBenchmark-results.txt +++ b/sql/core/benchmarks/WideTableBenchmark-results.txt @@ -6,12 +6,12 @@ OpenJDK 64-Bit Server VM 1.8.0_191-b12 on Linux 3.10.0-862.3.2.el7.x86_64 Intel(R) Xeon(R) CPU E5-2670 v2 @ 2.50GHz projection on wide table: Best/Avg Time(ms) Rate(M/s) Per Row(ns) Relative ------------------------------------------------------------------------------------------------ -split threshold 10 38932 / 39307 0.0 37128.1 1.0X -split threshold 100 31991 / 32556 0.0 30508.8 1.2X -split threshold 1024 10993 / 11041 0.1 10483.5 3.5X -split threshold 2048 8959 / 8998 0.1 8543.8 4.3X -split threshold 4096 8116 / 8134 0.1 7739.8 4.8X -split threshold 8196 8069 / 8098 0.1 7695.5 4.8X -split threshold 65536 57068 / 57339 0.0 54424.3 0.7X +split threshold 10 39634 / 39829 0.0 37798.3 1.0X +split threshold 100 30121 / 30571 0.0 28725.8 1.3X +split threshold 1024 9678 / 9725 0.1 9229.9 4.1X +split threshold 2048 8634 / 8662 0.1 8233.6 4.6X +split threshold 4096 8561 / 8576 0.1 8164.6 4.6X +split threshold 8192 8393 / 8408 0.1 8003.8 4.7X +split threshold 65536 57063 / 57273 0.0 54419.1 0.7X diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/WideTableBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/WideTableBenchmark.scala index ffefef1d4fce3..c61db3ce4b949 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/WideTableBenchmark.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/WideTableBenchmark.scala @@ -18,6 +18,7 @@ package org.apache.spark.sql.execution.benchmark import org.apache.spark.benchmark.Benchmark +import org.apache.spark.sql.Row import org.apache.spark.sql.internal.SQLConf /** @@ -42,7 +43,7 @@ object WideTableBenchmark extends SqlBasedBenchmark { Seq("10", "100", "1024", "2048", "4096", "8192", "65536").foreach { n => benchmark.addCase(s"split threshold $n", numIters = 5) { iter => withSQLConf(SQLConf.CODEGEN_METHOD_SPLIT_THRESHOLD.key -> n) { - df.selectExpr(columns: _*).foreach(identity(_)) + df.selectExpr(columns: _*).foreach((x => x): Row => Unit) } } }