Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ abstract class BenchmarkBase {
* Implementations of this method are supposed to use the wrapper method `runBenchmark`
* for each benchmark scenario.
*/
def runBenchmarkSuite(): Unit
def runBenchmarkSuite(mainArgs: Array[String]): Unit

final def runBenchmark(benchmarkName: String)(func: => Any): Unit = {
val separator = "=" * 96
Expand All @@ -51,7 +51,7 @@ abstract class BenchmarkBase {
output = Some(new FileOutputStream(file))
}

runBenchmarkSuite()
runBenchmarkSuite(args)

output.foreach { o =>
if (o != null) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ import org.apache.spark.serializer.KryoTest._
object KryoBenchmark extends BenchmarkBase {

val N = 1000000
override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
val name = "Benchmark Kryo Unsafe vs safe Serialization"
runBenchmark(name) {
val benchmark = new Benchmark(name, N, 10, output = output)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@ import org.apache.spark.sql.catalyst.encoders.ExpressionEncoder
*/
object UDTSerializationBenchmark extends BenchmarkBase {

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {

runBenchmark("VectorUDT de/serialization") {
val iters = 1e2.toInt
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -102,7 +102,7 @@ object HashBenchmark extends BenchmarkBase {
}
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
val singleInt = new StructType().add("i", IntegerType)
test("single ints", singleInt, 1 << 15, 1 << 14)

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@ object HashByteArrayBenchmark extends BenchmarkBase {
benchmark.run()
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("Benchmark for MurMurHash 3 and xxHash64") {
test(8, 42L, 1 << 10, 1 << 11)
test(16, 42L, 1 << 10, 1 << 11)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ object UnsafeProjectionBenchmark extends BenchmarkBase {
(1 to numRows).map(_ => encoder.toRow(generator().asInstanceOf[Row]).copy()).toArray
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("unsafe projection") {
val iters = 1024 * 16
val numRows = 1024 * 16
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -256,7 +256,7 @@ object DatasetBenchmark extends SqlBasedBenchmark {
.getOrCreate()
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
val numRows = 100000000
val numChains = 10
runBenchmark("Dataset Benchmark") {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -44,7 +44,7 @@ import org.apache.spark.unsafe.map.BytesToBytesMap
*/
object AggregateBenchmark extends SqlBasedBenchmark {

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("aggregate without grouping") {
val N = 500L << 22
codegenBenchmark("agg w/o group", N) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -80,7 +80,7 @@ object BloomFilterBenchmark extends SqlBasedBenchmark {
}
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
writeBenchmark()
readBenchmark()
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -585,7 +585,7 @@ object DataSourceReadBenchmark extends BenchmarkBase with SQLHelper {
}
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("SQL Single Numeric Column Scan") {
Seq(ByteType, ShortType, IntegerType, LongType, FloatType, DoubleType).foreach {
dataType => numericScanBenchmark(1024 * 1024 * 15, dataType)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -198,7 +198,7 @@ object FilterPushdownBenchmark extends BenchmarkBase with SQLHelper {
}
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("Pushdown for many distinct value case") {
withTempPath { dir =>
withTempTable("orcTable", "parquetTable") {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -164,7 +164,7 @@ object JoinBenchmark extends SqlBasedBenchmark {
}
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("Join Benchmark") {
broadcastHashJoinLongKey()
broadcastHashJoinLongKeyWithDuplicates()
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -148,7 +148,7 @@ object MiscBenchmark extends SqlBasedBenchmark {
}
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
filterAndAggregateWithoutGroup(500L << 22)
limitAndAggregateWithoutGroup(500L << 20)
sample(500 << 18)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ object PrimitiveArrayBenchmark extends BenchmarkBase {
.config("spark.sql.autoBroadcastJoinThreshold", 1)
.getOrCreate()

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("Write primitive arrays in dataset") {
writeDatasetArray(4)
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@ import org.apache.spark.benchmark.Benchmark
*/
object RangeBenchmark extends SqlBasedBenchmark {

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
import spark.implicits._

runBenchmark("range") {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -119,7 +119,7 @@ object SortBenchmark extends BenchmarkBase {
benchmark.run()
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("radix sort") {
sortBenchmark()
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -194,7 +194,7 @@ object UnsafeArrayDataBenchmark extends BenchmarkBase {
benchmark.run
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("Benchmark UnsafeArrayData") {
readUnsafeArray(10)
writeUnsafeArray(10)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -208,7 +208,7 @@ object WideSchemaBenchmark extends SqlBasedBenchmark {
deleteTmpFiles()
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {

runBenchmarkWithDeleteTmpFiles("parsing large select expressions") {
parsingLargeSelectExpressions()
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -233,7 +233,7 @@ object CompressionSchemeBenchmark extends BenchmarkBase with AllCompressionSchem
runDecodeBenchmark("STRING Decode", iters, count, STRING, testData)
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("Compression Scheme Benchmark") {
bitEncodingBenchmark(1024)
shortEncodingBenchmark(1024)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -443,7 +443,7 @@ object ColumnarBatchBenchmark extends BenchmarkBase {
benchmark.run
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("Int Read/Write") {
intAccess(1024 * 40)
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -212,7 +212,7 @@ object ObjectHashAggregateExecBenchmark extends BenchmarkBase with SQLHelper {
Column(approxPercentile.toAggregateExpression(isDistinct))
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("Hive UDAF vs Spark AF") {
hiveUDAFvsSparkAF(2 << 15)
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -336,7 +336,7 @@ object OrcReadBenchmark extends BenchmarkBase with SQLHelper {
}
}

override def runBenchmarkSuite(): Unit = {
override def runBenchmarkSuite(mainArgs: Array[String]): Unit = {
runBenchmark("SQL Single Numeric Column Scan") {
Seq(ByteType, ShortType, IntegerType, LongType, FloatType, DoubleType).foreach { dataType =>
numericScanBenchmark(1024 * 1024 * 15, dataType)
Expand Down