diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/SpecializedGettersReader.java b/sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/SpecializedGettersReader.java index ea0648a6cb90..7576eee98fe1 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/SpecializedGettersReader.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/catalyst/expressions/SpecializedGettersReader.java @@ -83,6 +83,9 @@ public static Object read( if (handleUserDefinedType && dataType instanceof UserDefinedType) { return obj.get(ordinal, ((UserDefinedType)dataType).sqlType()); } + if (dataType instanceof TimeType) { + return obj.getLong(ordinal); + } throw new UnsupportedOperationException("Unsupported data type " + dataType.simpleString()); } diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/Encoders.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/Encoders.scala index 5d31b5bbf12a..74e9c063606f 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/Encoders.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/Encoders.scala @@ -128,6 +128,14 @@ object Encoders { */ def INSTANT: Encoder[java.time.Instant] = ExpressionEncoder() + /** + * Creates an encoder that serializes instances of the `java.time.LocalTime` class + * to the internal representation of nullable Catalyst's TimeType. + * + * @since 3.0.0 + */ + def LOCALTIME: Encoder[java.time.LocalTime] = ExpressionEncoder() + /** * An encoder for arrays of bytes. * diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/CatalystTypeConverters.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/CatalystTypeConverters.scala index 488252aa0c7b..d2c7df11e4db 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/CatalystTypeConverters.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/CatalystTypeConverters.scala @@ -21,7 +21,7 @@ import java.lang.{Iterable => JavaIterable} import java.math.{BigDecimal => JavaBigDecimal} import java.math.{BigInteger => JavaBigInteger} import java.sql.{Date, Timestamp} -import java.time.{Instant, LocalDate} +import java.time.{Instant, LocalDate, LocalTime} import java.util.{Map => JavaMap} import javax.annotation.Nullable @@ -66,6 +66,7 @@ object CatalystTypeConverters { case DateType => DateConverter case TimestampType if SQLConf.get.datetimeJava8ApiEnabled => InstantConverter case TimestampType => TimestampConverter + case TimeType => LocalTimeConverter case dt: DecimalType => new DecimalConverter(dt) case BooleanType => BooleanConverter case ByteType => ByteConverter @@ -341,6 +342,18 @@ object CatalystTypeConverters { DateTimeUtils.microsToInstant(row.getLong(column)) } + private object LocalTimeConverter extends CatalystTypeConverter[LocalTime, LocalTime, Any] { + override def toCatalystImpl(scalaValue: LocalTime): Long = { + DateTimeUtils.localTimeToMicros(scalaValue) + } + override def toScala(catalystValue: Any): LocalTime = { + if (catalystValue == null) null + else DateTimeUtils.microsToLocalTime(catalystValue.asInstanceOf[Long]) + } + override def toScalaImpl(row: InternalRow, column: Int): LocalTime = + DateTimeUtils.microsToLocalTime(row.getLong(column)) + } + private class DecimalConverter(dataType: DecimalType) extends CatalystTypeConverter[Any, JavaBigDecimal, Decimal] { @@ -452,6 +465,7 @@ object CatalystTypeConverters { case ld: LocalDate => LocalDateConverter.toCatalyst(ld) case t: Timestamp => TimestampConverter.toCatalyst(t) case i: Instant => InstantConverter.toCatalyst(i) + case t: LocalTime => LocalTimeConverter.toCatalyst(t) case d: BigDecimal => new DecimalConverter(DecimalType(d.precision, d.scale)).toCatalyst(d) case d: JavaBigDecimal => new DecimalConverter(DecimalType(d.precision, d.scale)).toCatalyst(d) case seq: Seq[Any] => new GenericArrayData(seq.map(convertToCatalyst).toArray) diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/DeserializerBuildHelper.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/DeserializerBuildHelper.scala index e55c25c4b0c5..8d67fe0c23eb 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/DeserializerBuildHelper.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/DeserializerBuildHelper.scala @@ -118,6 +118,15 @@ object DeserializerBuildHelper { returnNullable = false) } + def createDeserializerForLocalTime(path: Expression): Expression = { + StaticInvoke( + DateTimeUtils.getClass, + ObjectType(classOf[java.time.LocalTime]), + "microsToLocalTime", + path :: Nil, + returnNullable = false) + } + def createDeserializerForJavaBigDecimal( path: Expression, returnNullable: Boolean): Expression = { diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/InternalRow.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/InternalRow.scala index bdab407688a6..1da87dc7a2af 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/InternalRow.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/InternalRow.scala @@ -131,7 +131,7 @@ object InternalRow { case ByteType => (input, ordinal) => input.getByte(ordinal) case ShortType => (input, ordinal) => input.getShort(ordinal) case IntegerType | DateType => (input, ordinal) => input.getInt(ordinal) - case LongType | TimestampType => (input, ordinal) => input.getLong(ordinal) + case LongType | TimestampType | TimeType => (input, ordinal) => input.getLong(ordinal) case FloatType => (input, ordinal) => input.getFloat(ordinal) case DoubleType => (input, ordinal) => input.getDouble(ordinal) case StringType => (input, ordinal) => input.getUTF8String(ordinal) @@ -166,7 +166,8 @@ object InternalRow { case ByteType => (input, v) => input.setByte(ordinal, v.asInstanceOf[Byte]) case ShortType => (input, v) => input.setShort(ordinal, v.asInstanceOf[Short]) case IntegerType | DateType => (input, v) => input.setInt(ordinal, v.asInstanceOf[Int]) - case LongType | TimestampType => (input, v) => input.setLong(ordinal, v.asInstanceOf[Long]) + case LongType | TimestampType | TimeType => + (input, v) => input.setLong(ordinal, v.asInstanceOf[Long]) case FloatType => (input, v) => input.setFloat(ordinal, v.asInstanceOf[Float]) case DoubleType => (input, v) => input.setDouble(ordinal, v.asInstanceOf[Double]) case DecimalType.Fixed(precision, _) => diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/JavaTypeInference.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/JavaTypeInference.scala index c5be3efc6371..060b63d37b18 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/JavaTypeInference.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/JavaTypeInference.scala @@ -106,6 +106,7 @@ object JavaTypeInference { case c: Class[_] if c == classOf[java.sql.Date] => (DateType, true) case c: Class[_] if c == classOf[java.time.Instant] => (TimestampType, true) case c: Class[_] if c == classOf[java.sql.Timestamp] => (TimestampType, true) + case c: Class[_] if c == classOf[java.time.LocalTime] => (TimeType, true) case _ if typeToken.isArray => val (dataType, nullable) = inferDataType(typeToken.getComponentType, seenTypeSet) @@ -235,6 +236,9 @@ object JavaTypeInference { case c if c == classOf[java.sql.Timestamp] => createDeserializerForSqlTimestamp(path) + case c if c == classOf[java.time.LocalTime] => + createDeserializerForLocalTime(path) + case c if c == classOf[java.lang.String] => createDeserializerForString(path, returnNullable = true) @@ -390,6 +394,8 @@ object JavaTypeInference { case c if c == classOf[java.sql.Date] => createSerializerForSqlDate(inputObject) + case c if c == classOf[java.time.LocalTime] => createSerializerForJavaLocalTime(inputObject) + case c if c == classOf[java.math.BigDecimal] => createSerializerForJavaBigDecimal(inputObject) diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/ScalaReflection.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/ScalaReflection.scala index 2a4e84241a94..9a9354475035 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/ScalaReflection.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/ScalaReflection.scala @@ -230,6 +230,9 @@ object ScalaReflection extends ScalaReflection { case t if isSubtype(t, localTypeOf[java.sql.Timestamp]) => createDeserializerForSqlTimestamp(path) + case t if isSubtype(t, localTypeOf[java.time.LocalTime]) => + createDeserializerForLocalTime(path) + case t if isSubtype(t, localTypeOf[java.lang.String]) => createDeserializerForString(path, returnNullable = false) @@ -496,6 +499,9 @@ object ScalaReflection extends ScalaReflection { case t if isSubtype(t, localTypeOf[java.sql.Date]) => createSerializerForSqlDate(inputObject) + case t if isSubtype(t, localTypeOf[java.time.LocalTime]) => + createSerializerForJavaLocalTime(inputObject) + case t if isSubtype(t, localTypeOf[BigDecimal]) => createSerializerForScalaBigDecimal(inputObject) @@ -671,6 +677,7 @@ object ScalaReflection extends ScalaReflection { Schema(TimestampType, nullable = true) case t if isSubtype(t, localTypeOf[java.time.LocalDate]) => Schema(DateType, nullable = true) case t if isSubtype(t, localTypeOf[java.sql.Date]) => Schema(DateType, nullable = true) + case t if isSubtype(t, localTypeOf[java.time.LocalTime]) => Schema(TimeType, nullable = true) case t if isSubtype(t, localTypeOf[BigDecimal]) => Schema(DecimalType.SYSTEM_DEFAULT, nullable = true) case t if isSubtype(t, localTypeOf[java.math.BigDecimal]) => @@ -771,6 +778,7 @@ object ScalaReflection extends ScalaReflection { StringType -> classOf[UTF8String], DateType -> classOf[DateType.InternalType], TimestampType -> classOf[TimestampType.InternalType], + TimeType -> classOf[TimeType.InternalType], BinaryType -> classOf[BinaryType.InternalType], CalendarIntervalType -> classOf[CalendarInterval] ) @@ -784,7 +792,8 @@ object ScalaReflection extends ScalaReflection { FloatType -> classOf[java.lang.Float], DoubleType -> classOf[java.lang.Double], DateType -> classOf[java.lang.Integer], - TimestampType -> classOf[java.lang.Long] + TimestampType -> classOf[java.lang.Long], + TimeType -> classOf[java.lang.Long] ) def dataTypeJavaClass(dt: DataType): Class[_] = { diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SerializerBuildHelper.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SerializerBuildHelper.scala index 75c278e78114..870dd9efb63e 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SerializerBuildHelper.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SerializerBuildHelper.scala @@ -101,6 +101,15 @@ object SerializerBuildHelper { returnNullable = false) } + def createSerializerForJavaLocalTime(inputObject: Expression): Expression = { + StaticInvoke( + DateTimeUtils.getClass, + TimeType, + "localTimeToMicros", + inputObject :: Nil, + returnNullable = false) + } + def createSerializerForJavaBigDecimal(inputObject: Expression): Expression = { CheckOverflow(StaticInvoke( Decimal.getClass, diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/dsl/package.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/dsl/package.scala index 796043fff665..f0c1dbe634e0 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/dsl/package.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/dsl/package.scala @@ -252,6 +252,9 @@ package object dsl { /** Creates a new AttributeReference of type timestamp */ def timestamp: AttributeReference = AttributeReference(s, TimestampType, nullable = true)() + /** Creates a new AttributeReference of type date */ + def time: AttributeReference = AttributeReference(s, TimeType, nullable = true)() + /** Creates a new AttributeReference of type binary */ def binary: AttributeReference = AttributeReference(s, BinaryType, nullable = true)() diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/encoders/RowEncoder.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/encoders/RowEncoder.scala index afe8a23f8f15..9ebcb4820134 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/encoders/RowEncoder.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/encoders/RowEncoder.scala @@ -53,6 +53,7 @@ import org.apache.spark.sql.types._ * TimestampType -> java.sql.Timestamp if spark.sql.datetime.java8API.enabled is false * TimestampType -> java.time.Instant if spark.sql.datetime.java8API.enabled is true * + * TimeType -> java.time.LocalTime * BinaryType -> byte array * ArrayType -> scala.collection.Seq or Array * MapType -> scala.collection.Map @@ -108,6 +109,8 @@ object RowEncoder { createSerializerForSqlDate(inputObject) } + case TimeType => createSerializerForJavaLocalTime(inputObject) + case d: DecimalType => CheckOverflow(StaticInvoke( Decimal.getClass, @@ -226,6 +229,7 @@ object RowEncoder { } else { ObjectType(classOf[java.sql.Date]) } + case TimeType => ObjectType(classOf[java.time.LocalTime]) case _: DecimalType => ObjectType(classOf[java.math.BigDecimal]) case StringType => ObjectType(classOf[java.lang.String]) case _: ArrayType => ObjectType(classOf[scala.collection.Seq[_]]) @@ -281,6 +285,8 @@ object RowEncoder { createDeserializerForSqlDate(input) } + case TimeType => createDeserializerForLocalTime(input) + case _: DecimalType => createDeserializerForJavaBigDecimal(input, returnNullable = false) case StringType => createDeserializerForString(input, returnNullable = false) diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/InterpretedUnsafeProjection.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/InterpretedUnsafeProjection.scala index 55a5bd380859..8fca20b9a2d5 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/InterpretedUnsafeProjection.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/InterpretedUnsafeProjection.scala @@ -146,7 +146,7 @@ object InterpretedUnsafeProjection { case IntegerType | DateType => (v, i) => writer.write(i, v.getInt(i)) - case LongType | TimestampType => + case LongType | TimestampType | TimeType => (v, i) => writer.write(i, v.getLong(i)) case FloatType => diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/SpecificInternalRow.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/SpecificInternalRow.scala index 75feaf670c84..7b7d62d5a975 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/SpecificInternalRow.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/SpecificInternalRow.scala @@ -200,8 +200,8 @@ final class SpecificInternalRow(val values: Array[MutableValue]) extends BaseGen case ShortType => new MutableShort // We use INT for DATE internally case IntegerType | DateType => new MutableInt - // We use Long for Timestamp internally - case LongType | TimestampType => new MutableLong + // We use Long for Timestamp and Time internally + case LongType | TimestampType | TimeType => new MutableLong case FloatType => new MutableFloat case DoubleType => new MutableDouble case _ => new MutableAny diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/CodeGenerator.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/CodeGenerator.scala index 95fad412002e..aad3de010600 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/CodeGenerator.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/CodeGenerator.scala @@ -1630,7 +1630,7 @@ object CodeGenerator extends Logging { case ByteType => JAVA_BYTE case ShortType => JAVA_SHORT case IntegerType | DateType => JAVA_INT - case LongType | TimestampType => JAVA_LONG + case LongType | TimestampType | TimeType => JAVA_LONG case FloatType => JAVA_FLOAT case DoubleType => JAVA_DOUBLE case _: DecimalType => "Decimal" @@ -1651,7 +1651,7 @@ object CodeGenerator extends Logging { case ByteType => java.lang.Byte.TYPE case ShortType => java.lang.Short.TYPE case IntegerType | DateType => java.lang.Integer.TYPE - case LongType | TimestampType => java.lang.Long.TYPE + case LongType | TimestampType | TimeType => java.lang.Long.TYPE case FloatType => java.lang.Float.TYPE case DoubleType => java.lang.Double.TYPE case _: DecimalType => classOf[Decimal] diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/literals.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/literals.scala index 9cef3ecadc54..5c72a50ddca4 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/literals.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/literals.scala @@ -27,7 +27,7 @@ import java.lang.{Short => JavaShort} import java.math.{BigDecimal => JavaBigDecimal} import java.nio.charset.StandardCharsets import java.sql.{Date, Timestamp} -import java.time.{Instant, LocalDate} +import java.time.{Instant, LocalDate, LocalTime} import java.util import java.util.Objects import javax.xml.bind.DatatypeConverter @@ -42,7 +42,7 @@ import org.apache.spark.sql.AnalysisException import org.apache.spark.sql.catalyst.{CatalystTypeConverters, InternalRow, ScalaReflection} import org.apache.spark.sql.catalyst.expressions.codegen._ import org.apache.spark.sql.catalyst.util._ -import org.apache.spark.sql.catalyst.util.DateTimeUtils.instantToMicros +import org.apache.spark.sql.catalyst.util.DateTimeUtils.{instantToMicros, localTimeToMicros} import org.apache.spark.sql.internal.SQLConf import org.apache.spark.sql.types._ import org.apache.spark.unsafe.types._ @@ -71,6 +71,7 @@ object Literal { case t: Timestamp => Literal(DateTimeUtils.fromJavaTimestamp(t), TimestampType) case ld: LocalDate => Literal(ld.toEpochDay.toInt, DateType) case d: Date => Literal(DateTimeUtils.fromJavaDate(d), DateType) + case lt: LocalTime => Literal(localTimeToMicros(lt), TimeType) case a: Array[Byte] => Literal(a, BinaryType) case a: collection.mutable.WrappedArray[_] => apply(a.array) case a: Array[_] => @@ -105,6 +106,7 @@ object Literal { case _ if clz == classOf[Date] => DateType case _ if clz == classOf[Instant] => TimestampType case _ if clz == classOf[Timestamp] => TimestampType + case _ if clz == classOf[LocalTime] => TimeType case _ if clz == classOf[JavaBigDecimal] => DecimalType.SYSTEM_DEFAULT case _ if clz == classOf[Array[Byte]] => BinaryType case _ if clz == classOf[JavaShort] => ShortType @@ -160,6 +162,7 @@ object Literal { case dt: DecimalType => Literal(Decimal(0, dt.precision, dt.scale)) case DateType => create(0, DateType) case TimestampType => create(0L, TimestampType) + case TimeType => create(0L, TimeType) case StringType => Literal("") case BinaryType => Literal("".getBytes(StandardCharsets.UTF_8)) case CalendarIntervalType => Literal(new CalendarInterval(0, 0)) @@ -179,7 +182,7 @@ object Literal { case ByteType => v.isInstanceOf[Byte] case ShortType => v.isInstanceOf[Short] case IntegerType | DateType => v.isInstanceOf[Int] - case LongType | TimestampType => v.isInstanceOf[Long] + case LongType | TimestampType | TimeType => v.isInstanceOf[Long] case FloatType => v.isInstanceOf[Float] case DoubleType => v.isInstanceOf[Double] case _: DecimalType => v.isInstanceOf[Decimal] @@ -336,7 +339,7 @@ case class Literal (value: Any, dataType: DataType) extends LeafExpression { } case ByteType | ShortType => ExprCode.forNonNullValue(JavaCode.expression(s"($javaType)$value", dataType)) - case TimestampType | LongType => + case TimestampType | LongType | TimeType => toExprCode(s"${value}L") case _ => val constRef = ctx.addReferenceObj("literal", value, javaType) diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/DateTimeUtils.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/DateTimeUtils.scala index 65a9bee5eaed..da7acc80bfc8 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/DateTimeUtils.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/DateTimeUtils.scala @@ -20,6 +20,7 @@ package org.apache.spark.sql.catalyst.util import java.sql.{Date, Timestamp} import java.time._ import java.time.temporal.{ChronoField, ChronoUnit, IsoFields} +import java.time.temporal.ChronoField.MICRO_OF_DAY import java.util.{Locale, TimeZone} import java.util.concurrent.TimeUnit._ @@ -363,6 +364,10 @@ object DateTimeUtils { def daysToLocalDate(days: Int): LocalDate = LocalDate.ofEpochDay(days) + def localTimeToMicros(localTime: LocalTime): Long = localTime.getLong(MICRO_OF_DAY) + + def microsToLocalTime(us: Long): LocalTime = LocalTime.ofNanoOfDay(us * NANOS_PER_MICROS) + /** * Trim and parse a given UTF8 date string to a corresponding [[Int]] value. * The return type is [[Option]] in order to distinguish between 0 and null. The following diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DataType.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DataType.scala index 3a10a56f6937..8ae8069e4a7a 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DataType.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DataType.scala @@ -130,7 +130,7 @@ object DataType { private val nonDecimalNameToType = { Seq(NullType, DateType, TimestampType, BinaryType, IntegerType, BooleanType, LongType, - DoubleType, FloatType, ShortType, ByteType, StringType, CalendarIntervalType) + DoubleType, FloatType, ShortType, ByteType, StringType, CalendarIntervalType, TimeType) .map(t => t.typeName -> t).toMap } diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/types/TimeType.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/types/TimeType.scala new file mode 100644 index 000000000000..e6d1e2a0a828 --- /dev/null +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/types/TimeType.scala @@ -0,0 +1,61 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.spark.sql.types + +import scala.math.Ordering +import scala.reflect.runtime.universe.typeTag + +import org.apache.spark.annotation.Unstable + +/** + * The time type represents local time in microsecond precision. + * Valid range is [00:00:00.000000, 23:59:59.999999]. + * + * Please use the singleton `DataTypes.TimeType` to refer the type. + * @since 3.0.0 + */ +@Unstable +class TimeType private () extends AtomicType { + + /** + * Internally, time is stored as the number of microseconds since 00:00:00.000000. + */ + private[sql] type InternalType = Long + + @transient private[sql] lazy val tag = typeTag[InternalType] + + private[sql] val ordering = implicitly[Ordering[InternalType]] + + /** + * The default size of a value of the TimeType is 8 bytes. + */ + override def defaultSize: Int = 8 + + private[spark] override def asNullable: TimeType = this +} + +/** + * The companion case object and its class is separated so the companion object also subclasses + * the TimeType class. Otherwise, the companion object would be of type "TimeType$" + * in byte code. Defined with a private constructor so the companion object is the only possible + * instantiation. + * + * @since 3.0.0 + */ +@Unstable +case object TimeType extends TimeType diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/CatalystTypeConvertersSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/CatalystTypeConvertersSuite.scala index b9e7cf304989..08b50229df95 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/CatalystTypeConvertersSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/CatalystTypeConvertersSuite.scala @@ -17,7 +17,7 @@ package org.apache.spark.sql.catalyst -import java.time.{Instant, LocalDate} +import java.time.{Instant, LocalDate, LocalTime} import org.apache.spark.SparkFunSuite import org.apache.spark.sql.Row @@ -216,4 +216,33 @@ class CatalystTypeConvertersSuite extends SparkFunSuite with SQLHelper { } } } + + test("converting java.time.LocalTime to TimeType") { + Seq( + "00:00:00", + "01:02:03.999", + "02:59:01", + "12:30:02.0", + "22:00:00.000001", + "23:59:59.999999").foreach { time => + val input = LocalTime.parse(time) + val result = CatalystTypeConverters.convertToCatalyst(input) + val expected = DateTimeUtils.localTimeToMicros(input) + assert(result === expected) + } + } + + test("converting TimeType to java.time.LocalTime") { + Seq( + 0, + 1, + 59000000, + 3600000001L, + 43200999999L, + 86399000000L, + 86399999999L).foreach { us => + val localTime = DateTimeUtils.microsToLocalTime(us) + assert(CatalystTypeConverters.createToScalaConverter(TimeType)(us) === localTime) + } + } } diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/encoders/RowEncoderSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/encoders/RowEncoderSuite.scala index 5d21e4a2a83c..577594c5b34c 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/encoders/RowEncoderSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/encoders/RowEncoderSuite.scala @@ -332,6 +332,16 @@ class RowEncoderSuite extends CodegenInterpretedPlanTest { } } + test("encoding/decoding TimeType to/from java.time.LocalTime") { + val schema = new StructType().add("t", TimeType) + val encoder = RowEncoder(schema).resolveAndBind() + val localTime = java.time.LocalTime.parse("20:38:45.123456") + val row = encoder.toRow(Row(localTime)) + assert(row.getLong(0) === DateTimeUtils.localTimeToMicros(localTime)) + val readback = encoder.fromRow(row) + assert(readback.get(0).equals(localTime)) + } + for { elementType <- Seq(IntegerType, StringType) containsNull <- Seq(true, false) diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/LiteralExpressionSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/LiteralExpressionSuite.scala index 269f1a09ac53..51b0802631ca 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/LiteralExpressionSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/LiteralExpressionSuite.scala @@ -18,7 +18,7 @@ package org.apache.spark.sql.catalyst.expressions import java.nio.charset.StandardCharsets -import java.time.{Instant, LocalDate, LocalDateTime, ZoneOffset} +import java.time.{Instant, LocalDate, LocalDateTime, LocalTime, ZoneOffset} import java.util.TimeZone import scala.reflect.runtime.universe.TypeTag @@ -81,6 +81,7 @@ class LiteralExpressionSuite extends SparkFunSuite with ExpressionEvalHelper { checkEvaluation(Literal.default(StructType(StructField("a", StringType) :: Nil)), Row("")) // ExamplePointUDT.sqlType is ArrayType(DoubleType, false). checkEvaluation(Literal.default(new ExamplePointUDT), Array()) + checkEvaluation(Literal.default(TimeType), LocalTime.ofNanoOfDay(0)) } test("boolean literals") { @@ -316,4 +317,24 @@ class LiteralExpressionSuite extends SparkFunSuite with ExpressionEvalHelper { assert(literalStr === expected) } } + + test("construct literals from java.time.LocalTime") { + Seq( + LocalTime.of(0, 0, 0), + LocalTime.of(0, 0, 0, 1000), + LocalTime.of(1, 1, 1), + LocalTime.of(23, 59, 59), + LocalTime.of(23, 59, 59, 999999000), + LocalTime.of(10, 11, 12, 131415000), + LocalTime.of(21, 0, 0)).foreach { localTime => + checkEvaluation(Literal(localTime), localTime) + } + } + + test("construct literals from arrays of java.time.LocalTime") { + val localTime0 = LocalTime.of(10, 10, 10) + checkEvaluation(Literal(Array(localTime0)), Array(localTime0)) + val localTime1 = LocalTime.of(21, 21, 21) + checkEvaluation(Literal(Array(localTime0, localTime1)), Array(localTime0, localTime1)) + } } diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SQLImplicits.scala b/sql/core/src/main/scala/org/apache/spark/sql/SQLImplicits.scala index 71cbc3ab14d9..895a782923d4 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/SQLImplicits.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/SQLImplicits.scala @@ -88,6 +88,9 @@ abstract class SQLImplicits extends LowPrioritySQLImplicits { /** @since 3.0.0 */ implicit def newInstantEncoder: Encoder[java.time.Instant] = Encoders.INSTANT + /** @since 3.0.0 */ + implicit def newLocalTimeEncoder: Encoder[java.time.LocalTime] = Encoders.LOCALTIME + // Boxed primitives /** @since 2.0.0 */ diff --git a/sql/core/src/test/java/test/org/apache/spark/sql/JavaDatasetSuite.java b/sql/core/src/test/java/test/org/apache/spark/sql/JavaDatasetSuite.java index 1e5f55e494b7..976059411d6c 100644 --- a/sql/core/src/test/java/test/org/apache/spark/sql/JavaDatasetSuite.java +++ b/sql/core/src/test/java/test/org/apache/spark/sql/JavaDatasetSuite.java @@ -23,6 +23,7 @@ import java.sql.Timestamp; import java.time.Instant; import java.time.LocalDate; +import java.time.LocalTime; import java.util.*; import org.apache.spark.sql.streaming.GroupStateTimeout; @@ -411,6 +412,14 @@ public void testLocalDateAndInstantEncoders() { Assert.assertEquals(data, ds.collectAsList()); } + @Test + public void testLocalTimeEncoder() { + Encoder encoder = Encoders.LOCALTIME(); + List data = Arrays.asList(LocalTime.ofNanoOfDay(0)); + Dataset ds = spark.createDataset(data, encoder); + Assert.assertEquals(data, ds.collectAsList()); + } + public static class KryoSerializable { String value; diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala index 42e5ee58954e..ced9c46ab1de 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DatasetSuite.scala @@ -1841,6 +1841,11 @@ class DatasetSuite extends QueryTest with SharedSparkSession { val instant = java.time.Instant.parse("2019-03-30T09:54:00Z") assert(spark.range(1).map { _ => instant }.head === instant) } + + test("implicit encoder for LocalTime") { + val localTime = java.time.LocalTime.of(19, 30, 30) + assert(spark.range(1).map { _ => localTime }.head === localTime) + } } object AssertExecutionId {