From eded731fba1b943fc062a516c26b4db1a461576e Mon Sep 17 00:00:00 2001 From: Max Gekk Date: Tue, 21 Dec 2021 21:36:54 +0300 Subject: [PATCH] Fix the typo in function names: crete --- .../org/apache/spark/sql/avro/AvroDeserializer.scala | 4 ++-- .../scala/org/apache/spark/sql/avro/AvroSerializer.scala | 4 ++-- .../spark/sql/execution/datasources/DataSourceUtils.scala | 8 ++++---- .../datasources/parquet/ParquetRowConverter.scala | 6 +++--- .../datasources/parquet/ParquetWriteSupport.scala | 6 +++--- 5 files changed, 14 insertions(+), 14 deletions(-) diff --git a/external/avro/src/main/scala/org/apache/spark/sql/avro/AvroDeserializer.scala b/external/avro/src/main/scala/org/apache/spark/sql/avro/AvroDeserializer.scala index d7f2fa83e30b8..54cd3ba3464e9 100644 --- a/external/avro/src/main/scala/org/apache/spark/sql/avro/AvroDeserializer.scala +++ b/external/avro/src/main/scala/org/apache/spark/sql/avro/AvroDeserializer.scala @@ -62,10 +62,10 @@ private[sql] class AvroDeserializer( private lazy val decimalConversions = new DecimalConversion() - private val dateRebaseFunc = DataSourceUtils.creteDateRebaseFuncInRead( + private val dateRebaseFunc = DataSourceUtils.createDateRebaseFuncInRead( datetimeRebaseMode, "Avro") - private val timestampRebaseFunc = DataSourceUtils.creteTimestampRebaseFuncInRead( + private val timestampRebaseFunc = DataSourceUtils.createTimestampRebaseFuncInRead( datetimeRebaseMode, "Avro") private val converter: Any => Option[Any] = try { diff --git a/external/avro/src/main/scala/org/apache/spark/sql/avro/AvroSerializer.scala b/external/avro/src/main/scala/org/apache/spark/sql/avro/AvroSerializer.scala index 32a84d07ac3a0..f2f754aabd3ed 100644 --- a/external/avro/src/main/scala/org/apache/spark/sql/avro/AvroSerializer.scala +++ b/external/avro/src/main/scala/org/apache/spark/sql/avro/AvroSerializer.scala @@ -60,10 +60,10 @@ private[sql] class AvroSerializer( converter.apply(catalystData) } - private val dateRebaseFunc = DataSourceUtils.creteDateRebaseFuncInWrite( + private val dateRebaseFunc = DataSourceUtils.createDateRebaseFuncInWrite( datetimeRebaseMode, "Avro") - private val timestampRebaseFunc = DataSourceUtils.creteTimestampRebaseFuncInWrite( + private val timestampRebaseFunc = DataSourceUtils.createTimestampRebaseFuncInWrite( datetimeRebaseMode, "Avro") private val converter: Any => Any = { diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSourceUtils.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSourceUtils.scala index 67d03998a2a24..76bc03ad5e946 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSourceUtils.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSourceUtils.scala @@ -172,7 +172,7 @@ object DataSourceUtils extends PredicateHelper { QueryExecutionErrors.sparkUpgradeInWritingDatesError(format, config) } - def creteDateRebaseFuncInRead( + def createDateRebaseFuncInRead( rebaseMode: LegacyBehaviorPolicy.Value, format: String): Int => Int = rebaseMode match { case LegacyBehaviorPolicy.EXCEPTION => days: Int => @@ -184,7 +184,7 @@ object DataSourceUtils extends PredicateHelper { case LegacyBehaviorPolicy.CORRECTED => identity[Int] } - def creteDateRebaseFuncInWrite( + def createDateRebaseFuncInWrite( rebaseMode: LegacyBehaviorPolicy.Value, format: String): Int => Int = rebaseMode match { case LegacyBehaviorPolicy.EXCEPTION => days: Int => @@ -196,7 +196,7 @@ object DataSourceUtils extends PredicateHelper { case LegacyBehaviorPolicy.CORRECTED => identity[Int] } - def creteTimestampRebaseFuncInRead( + def createTimestampRebaseFuncInRead( rebaseMode: LegacyBehaviorPolicy.Value, format: String): Long => Long = rebaseMode match { case LegacyBehaviorPolicy.EXCEPTION => micros: Long => @@ -208,7 +208,7 @@ object DataSourceUtils extends PredicateHelper { case LegacyBehaviorPolicy.CORRECTED => identity[Long] } - def creteTimestampRebaseFuncInWrite( + def createTimestampRebaseFuncInWrite( rebaseMode: LegacyBehaviorPolicy.Value, format: String): Long => Long = rebaseMode match { case LegacyBehaviorPolicy.EXCEPTION => micros: Long => diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetRowConverter.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetRowConverter.scala index 8635510f5a1ec..243bc164183d4 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetRowConverter.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetRowConverter.scala @@ -190,13 +190,13 @@ private[parquet] class ParquetRowConverter( */ def currentRecord: InternalRow = currentRow - private val dateRebaseFunc = DataSourceUtils.creteDateRebaseFuncInRead( + private val dateRebaseFunc = DataSourceUtils.createDateRebaseFuncInRead( datetimeRebaseMode, "Parquet") - private val timestampRebaseFunc = DataSourceUtils.creteTimestampRebaseFuncInRead( + private val timestampRebaseFunc = DataSourceUtils.createTimestampRebaseFuncInRead( datetimeRebaseMode, "Parquet") - private val int96RebaseFunc = DataSourceUtils.creteTimestampRebaseFuncInRead( + private val int96RebaseFunc = DataSourceUtils.createTimestampRebaseFuncInRead( int96RebaseMode, "Parquet INT96") // Converters for each field. diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetWriteSupport.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetWriteSupport.scala index 933f79db4684e..183ea394d8b15 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetWriteSupport.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetWriteSupport.scala @@ -82,16 +82,16 @@ class ParquetWriteSupport extends WriteSupport[InternalRow] with Logging { private val datetimeRebaseMode = LegacyBehaviorPolicy.withName( SQLConf.get.getConf(SQLConf.PARQUET_REBASE_MODE_IN_WRITE)) - private val dateRebaseFunc = DataSourceUtils.creteDateRebaseFuncInWrite( + private val dateRebaseFunc = DataSourceUtils.createDateRebaseFuncInWrite( datetimeRebaseMode, "Parquet") - private val timestampRebaseFunc = DataSourceUtils.creteTimestampRebaseFuncInWrite( + private val timestampRebaseFunc = DataSourceUtils.createTimestampRebaseFuncInWrite( datetimeRebaseMode, "Parquet") private val int96RebaseMode = LegacyBehaviorPolicy.withName( SQLConf.get.getConf(SQLConf.PARQUET_INT96_REBASE_MODE_IN_WRITE)) - private val int96RebaseFunc = DataSourceUtils.creteTimestampRebaseFuncInWrite( + private val int96RebaseFunc = DataSourceUtils.createTimestampRebaseFuncInWrite( int96RebaseMode, "Parquet INT96") override def init(configuration: Configuration): WriteContext = {