diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/DeleteMarkerProcedure.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/DeleteMarkerProcedure.scala new file mode 100644 index 0000000000000..8804d9fb5fcb0 --- /dev/null +++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/DeleteMarkerProcedure.scala @@ -0,0 +1,78 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.spark.sql.hudi.command.procedures + +import org.apache.hudi.table.HoodieSparkTable +import org.apache.hudi.table.marker.WriteMarkersFactory +import org.apache.spark.internal.Logging +import org.apache.spark.sql.Row +import org.apache.spark.sql.types.{DataTypes, Metadata, StructField, StructType} + +import java.util.function.Supplier +import scala.util.{Failure, Success, Try} + +class DeleteMarkerProcedure extends BaseProcedure with ProcedureBuilder with Logging { + private val PARAMETERS = Array[ProcedureParameter]( + ProcedureParameter.required(0, "table", DataTypes.StringType, None), + ProcedureParameter.required(1, "instant_Time", DataTypes.StringType, None) + ) + + private val OUTPUT_TYPE = new StructType(Array[StructField]( + StructField("delete_marker_result", DataTypes.BooleanType, nullable = true, Metadata.empty)) + ) + + def parameters: Array[ProcedureParameter] = PARAMETERS + + def outputType: StructType = OUTPUT_TYPE + + override def call(args: ProcedureArgs): Seq[Row] = { + super.checkArgs(PARAMETERS, args) + + val tableName = getArgValueOrDefault(args, PARAMETERS(0)) + val instantTime = getArgValueOrDefault(args, PARAMETERS(1)).get.asInstanceOf[String] + val basePath = getBasePath(tableName) + + val result = Try { + val client = createHoodieClient(jsc, basePath) + val config = client.getConfig + val context = client.getEngineContext + val table = HoodieSparkTable.create(config, context) + WriteMarkersFactory.get(config.getMarkersType, table, instantTime) + .quietDeleteMarkerDir(context, config.getMarkersDeleteParallelism) + } match { + case Success(_) => + logInfo(s"Marker $instantTime deleted.") + true + case Failure(e) => + logWarning(s"Failed: Could not clean marker instantTime: $instantTime.", e) + false + } + + Seq(Row(result)) + } + + override def build: Procedure = new DeleteMarkerProcedure() +} + +object DeleteMarkerProcedure { + val NAME: String = "delete_marker" + + def builder: Supplier[ProcedureBuilder] = new Supplier[ProcedureBuilder] { + override def get(): DeleteMarkerProcedure = new DeleteMarkerProcedure() + } +} diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/HoodieProcedures.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/HoodieProcedures.scala index e7de3e784a2fd..2b720bb94d2d8 100644 --- a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/HoodieProcedures.scala +++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/HoodieProcedures.scala @@ -44,6 +44,7 @@ object HoodieProcedures { mapBuilder.put(ShowCommitsProcedure.NAME, ShowCommitsProcedure.builder) mapBuilder.put(ShowCommitsMetadataProcedure.NAME, ShowCommitsMetadataProcedure.builder) mapBuilder.put(ShowSavepointsProcedure.NAME, ShowSavepointsProcedure.builder) + mapBuilder.put(DeleteMarkerProcedure.NAME, DeleteMarkerProcedure.builder) mapBuilder.build } } diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/TestCallProcedure.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/TestCallProcedure.scala index f75569a1171f5..848d09ab62bd0 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/TestCallProcedure.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/TestCallProcedure.scala @@ -17,6 +17,8 @@ package org.apache.spark.sql.hudi.procedure +import org.apache.hudi.common.model.IOType +import org.apache.hudi.common.testutils.FileCreateUtils import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase class TestCallProcedure extends HoodieSparkSqlTestBase { @@ -131,4 +133,42 @@ class TestCallProcedure extends HoodieSparkSqlTestBase { assertResult(1){commits.length} } } + + test("Test Call delete_marker Procedure") { + withTempDir { tmp => + val tableName = generateTableName + val tablePath = s"${tmp.getCanonicalPath}/$tableName" + // create table + spark.sql( + s""" + |create table $tableName ( + | id int, + | name string, + | price double, + | ts long + |) using hudi + | location '$tablePath' + | tblproperties ( + | primaryKey = 'id', + | preCombineField = 'ts' + | ) + """.stripMargin) + + // Check required fields + checkExceptionContain(s"""call delete_marker(table => '$tableName')""")( + s"Argument: instant_Time is required") + + val instantTime = "101" + FileCreateUtils.createMarkerFile(tablePath, "", instantTime, "f0", IOType.APPEND) + assertResult(1) { + FileCreateUtils.getTotalMarkerFileCount(tablePath, "", instantTime, IOType.APPEND) + } + + checkAnswer(s"""call delete_marker(table => '$tableName', instant_Time => '$instantTime')""")(Seq(true)) + + assertResult(0) { + FileCreateUtils.getTotalMarkerFileCount(tablePath, "", instantTime, IOType.APPEND) + } + } + } }