From 0b79b91c1fe3d23eebd939bdf13d48bbafc002b3 Mon Sep 17 00:00:00 2001 From: lamber-ken Date: Mon, 30 Dec 2019 09:29:17 +0800 Subject: [PATCH 1/3] [HUDI-482] Fix missing @Override annotation on method --- .../apache/hudi/cli/HoodieHistoryFileNameProvider.java | 1 + .../main/java/org/apache/hudi/cli/HoodieSplashScreen.java | 3 +++ .../main/java/org/apache/hudi/AbstractHoodieClient.java | 1 + .../src/main/java/org/apache/hudi/HoodieWriteClient.java | 1 + .../apache/hudi/func/SparkBoundedInMemoryExecutor.java | 1 + .../org/apache/hudi/index/bloom/BloomIndexFileInfo.java | 1 + .../main/java/org/apache/hudi/index/hbase/HBaseIndex.java | 2 ++ .../main/java/org/apache/hudi/io/HoodieCreateHandle.java | 1 + .../main/java/org/apache/hudi/io/HoodieMergeHandle.java | 6 ++++++ .../org/apache/hudi/io/storage/HoodieParquetWriter.java | 1 + .../org/apache/hudi/table/HoodieCopyOnWriteTable.java | 1 + .../apache/hudi/common/table/log/HoodieLogFileReader.java | 1 + .../hudi/common/table/log/HoodieLogFormatWriter.java | 2 ++ .../hudi/common/table/timeline/HoodieActiveTimeline.java | 1 + .../common/table/view/AbstractTableFileSystemView.java | 3 +++ .../hudi/common/table/view/HoodieTableFileSystemView.java | 2 ++ .../table/view/SpillableMapBasedFileSystemView.java | 2 ++ .../org/apache/hudi/common/util/ObjectSizeCalculator.java | 1 + .../apache/hudi/common/util/collection/DiskBasedMap.java | 1 + .../hudi/common/util/collection/LazyFileIterable.java | 1 + .../org/apache/hudi/hadoop/HoodieParquetInputFormat.java | 2 ++ .../hudi/hadoop/hive/HoodieCombineHiveInputFormat.java | 8 +++++++- .../hudi/utilities/deltastreamer/HoodieDeltaStreamer.java | 1 + 23 files changed, 43 insertions(+), 1 deletion(-) diff --git a/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieHistoryFileNameProvider.java b/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieHistoryFileNameProvider.java index af01f6675d620..95f983416a50d 100644 --- a/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieHistoryFileNameProvider.java +++ b/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieHistoryFileNameProvider.java @@ -30,6 +30,7 @@ @Order(Ordered.HIGHEST_PRECEDENCE) public class HoodieHistoryFileNameProvider extends DefaultHistoryFileNameProvider { + @Override public String getHistoryFileName() { return "hoodie-cmd.log"; } diff --git a/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieSplashScreen.java b/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieSplashScreen.java index b7f016b501ee5..f2a458c196c94 100644 --- a/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieSplashScreen.java +++ b/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieSplashScreen.java @@ -50,14 +50,17 @@ public class HoodieSplashScreen extends DefaultBannerProvider { + "* *" + OsUtils.LINE_SEPARATOR + "===================================================================" + OsUtils.LINE_SEPARATOR; + @Override public String getBanner() { return screen; } + @Override public String getVersion() { return "1.0"; } + @Override public String getWelcomeMessage() { return "Welcome to Apache Hudi CLI. Please type help if you are looking for help. "; } diff --git a/hudi-client/src/main/java/org/apache/hudi/AbstractHoodieClient.java b/hudi-client/src/main/java/org/apache/hudi/AbstractHoodieClient.java index 1d60840ae5421..dd108be023b08 100644 --- a/hudi-client/src/main/java/org/apache/hudi/AbstractHoodieClient.java +++ b/hudi-client/src/main/java/org/apache/hudi/AbstractHoodieClient.java @@ -72,6 +72,7 @@ protected AbstractHoodieClient(JavaSparkContext jsc, HoodieWriteConfig clientCon /** * Releases any resources used by the client. */ + @Override public void close() { stopEmbeddedServerView(true); } diff --git a/hudi-client/src/main/java/org/apache/hudi/HoodieWriteClient.java b/hudi-client/src/main/java/org/apache/hudi/HoodieWriteClient.java index 0f5411648f852..9d0332d62ab27 100644 --- a/hudi-client/src/main/java/org/apache/hudi/HoodieWriteClient.java +++ b/hudi-client/src/main/java/org/apache/hudi/HoodieWriteClient.java @@ -973,6 +973,7 @@ private void rollbackInternal(String commitToRollback) { /** * Releases any resources used by the client. */ + @Override public void close() { // Stop timeline-server if running super.close(); diff --git a/hudi-client/src/main/java/org/apache/hudi/func/SparkBoundedInMemoryExecutor.java b/hudi-client/src/main/java/org/apache/hudi/func/SparkBoundedInMemoryExecutor.java index 9741f7ce9c3ad..a576e3cc36534 100644 --- a/hudi-client/src/main/java/org/apache/hudi/func/SparkBoundedInMemoryExecutor.java +++ b/hudi-client/src/main/java/org/apache/hudi/func/SparkBoundedInMemoryExecutor.java @@ -48,6 +48,7 @@ public SparkBoundedInMemoryExecutor(final HoodieWriteConfig hoodieConfig, Bounde this.sparkThreadTaskContext = TaskContext.get(); } + @Override public void preExecute() { // Passing parent thread's TaskContext to newly launched thread for it to access original TaskContext properties. TaskContext$.MODULE$.setTaskContext(sparkThreadTaskContext); diff --git a/hudi-client/src/main/java/org/apache/hudi/index/bloom/BloomIndexFileInfo.java b/hudi-client/src/main/java/org/apache/hudi/index/bloom/BloomIndexFileInfo.java index 51e1b76520a25..655d99bb7cd17 100644 --- a/hudi-client/src/main/java/org/apache/hudi/index/bloom/BloomIndexFileInfo.java +++ b/hudi-client/src/main/java/org/apache/hudi/index/bloom/BloomIndexFileInfo.java @@ -88,6 +88,7 @@ public int hashCode() { return Objects.hashCode(fileId, minRecordKey, maxRecordKey); } + @Override public String toString() { final StringBuilder sb = new StringBuilder("BloomIndexFileInfo {"); sb.append(" fileId=").append(fileId); diff --git a/hudi-client/src/main/java/org/apache/hudi/index/hbase/HBaseIndex.java b/hudi-client/src/main/java/org/apache/hudi/index/hbase/HBaseIndex.java index dfe9dab96d3d4..677408b52d4ed 100644 --- a/hudi-client/src/main/java/org/apache/hudi/index/hbase/HBaseIndex.java +++ b/hudi-client/src/main/java/org/apache/hudi/index/hbase/HBaseIndex.java @@ -154,6 +154,7 @@ private Connection getHBaseConnection() { */ private void addShutDownHook() { Runtime.getRuntime().addShutdownHook(new Thread() { + @Override public void run() { try { hbaseConnection.close(); @@ -167,6 +168,7 @@ public void run() { /** * Ensure that any resources used for indexing are released here. */ + @Override public void close() { this.hBaseIndexQPSResourceAllocator.releaseQPSResources(); } diff --git a/hudi-client/src/main/java/org/apache/hudi/io/HoodieCreateHandle.java b/hudi-client/src/main/java/org/apache/hudi/io/HoodieCreateHandle.java index 6916bad788bb4..095e0a0e35331 100644 --- a/hudi-client/src/main/java/org/apache/hudi/io/HoodieCreateHandle.java +++ b/hudi-client/src/main/java/org/apache/hudi/io/HoodieCreateHandle.java @@ -94,6 +94,7 @@ public boolean canWrite(HoodieRecord record) { /** * Perform the actual writing of the given record into the backing file. */ + @Override public void write(HoodieRecord record, Option avroRecord) { Option recordMetadata = record.getData().getMetadata(); try { diff --git a/hudi-client/src/main/java/org/apache/hudi/io/HoodieMergeHandle.java b/hudi-client/src/main/java/org/apache/hudi/io/HoodieMergeHandle.java index 45519acca806f..518b883340043 100644 --- a/hudi-client/src/main/java/org/apache/hudi/io/HoodieMergeHandle.java +++ b/hudi-client/src/main/java/org/apache/hudi/io/HoodieMergeHandle.java @@ -92,6 +92,7 @@ public static Schema createHoodieWriteSchema(Schema originalSchema) { return HoodieAvroUtils.addMetadataFields(originalSchema); } + @Override public Path makeNewPath(String partitionPath) { Path path = FSUtils.getPartitionPath(config.getBasePath(), partitionPath); try { @@ -103,6 +104,7 @@ public Path makeNewPath(String partitionPath) { return new Path(path.toString(), FSUtils.makeDataFileName(instantTime, writeToken, fileId)); } + @Override public Schema getWriterSchema() { return writerSchema; } @@ -113,6 +115,7 @@ public Schema getWriterSchema() { * - Whether it belongs to the same partitionPath as existing records - Whether the current file written bytes lt max * file size */ + @Override public boolean canWrite(HoodieRecord record) { return false; } @@ -120,6 +123,7 @@ public boolean canWrite(HoodieRecord record) { /** * Perform the actual writing of the given record into the backing file. */ + @Override public void write(HoodieRecord record, Option insertValue) { // NO_OP } @@ -127,6 +131,7 @@ public void write(HoodieRecord record, Option insertValue) { /** * Perform the actual writing of the given record into the backing file. */ + @Override public void write(HoodieRecord record, Option avroRecord, Option exception) { Option recordMetadata = record.getData().getMetadata(); if (exception.isPresent() && exception.get() instanceof Throwable) { @@ -141,6 +146,7 @@ public void write(HoodieRecord record, Option avroRecord, Option< /** * Rewrite the GenericRecord with the Schema containing the Hoodie Metadata fields. */ + @Override protected GenericRecord rewriteRecord(GenericRecord record) { return HoodieAvroUtils.rewriteRecord(record, writerSchema); } diff --git a/hudi-client/src/main/java/org/apache/hudi/io/storage/HoodieParquetWriter.java b/hudi-client/src/main/java/org/apache/hudi/io/storage/HoodieParquetWriter.java index fa7b165286fd8..5f1fb8fc56984 100644 --- a/hudi-client/src/main/java/org/apache/hudi/io/storage/HoodieParquetWriter.java +++ b/hudi-client/src/main/java/org/apache/hudi/io/storage/HoodieParquetWriter.java @@ -93,6 +93,7 @@ public void writeAvroWithMetadata(R avroRecord, HoodieRecord record) throws IOEx writeSupport.add(record.getRecordKey()); } + @Override public boolean canWrite() { return fs.getBytesWritten(file) < maxFileSize; } diff --git a/hudi-client/src/main/java/org/apache/hudi/table/HoodieCopyOnWriteTable.java b/hudi-client/src/main/java/org/apache/hudi/table/HoodieCopyOnWriteTable.java index ed9c65dccee14..f9447d34db1c3 100644 --- a/hudi-client/src/main/java/org/apache/hudi/table/HoodieCopyOnWriteTable.java +++ b/hudi-client/src/main/java/org/apache/hudi/table/HoodieCopyOnWriteTable.java @@ -279,6 +279,7 @@ public Iterator> handleInsertPartition(String commitTime, Inte * @param jsc JavaSparkContext * @return Cleaner Plan */ + @Override public HoodieCleanerPlan scheduleClean(JavaSparkContext jsc) { try { HoodieCleanHelper cleaner = new HoodieCleanHelper(this, config); diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFileReader.java b/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFileReader.java index 985ac99c101ee..354f8090d476f 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFileReader.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFileReader.java @@ -108,6 +108,7 @@ public HoodieLogFile getLogFile() { */ private void addShutDownHook() { Runtime.getRuntime().addShutdownHook(new Thread() { + @Override public void run() { try { close(); diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFormatWriter.java b/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFormatWriter.java index c6adbc3ca9818..b8d5f8933c1dc 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFormatWriter.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFormatWriter.java @@ -107,6 +107,7 @@ public FileSystem getFs() { return fs; } + @Override public HoodieLogFile getLogFile() { return logFile; } @@ -212,6 +213,7 @@ private void flush() throws IOException { output.hsync(); } + @Override public long getCurrentSize() throws IOException { if (output == null) { throw new IllegalStateException("Cannot get current size as the underlying stream has been closed already"); diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieActiveTimeline.java b/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieActiveTimeline.java index a20353b086d70..01d5ea23f1063 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieActiveTimeline.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieActiveTimeline.java @@ -145,6 +145,7 @@ public HoodieTimeline getCommitsTimeline() { * timeline * With Async compaction a requested/inflight compaction-instant is a valid baseInstant for a file-slice as * there could be delta-commits with that baseInstant. */ + @Override public HoodieTimeline getCommitsAndCompactionTimeline() { return getTimelineOfActions(Sets.newHashSet(COMMIT_ACTION, DELTA_COMMIT_ACTION, COMPACTION_ACTION)); } diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java b/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java index 9beeea0c55e6f..bd7462d83f2d3 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java @@ -182,6 +182,7 @@ protected List buildFileGroups(Stream dataFileS /** * Clears the partition Map and reset view states. */ + @Override public final void reset() { try { writeLock.lock(); @@ -380,6 +381,7 @@ public final Option getDataFileOn(String partitionStr, String in /** * Get Latest data file for a partition and file-Id. */ + @Override public final Option getLatestDataFile(String partitionStr, String fileId) { try { readLock.lock(); @@ -434,6 +436,7 @@ public final Stream getLatestFileSlices(String partitionStr) { /** * Get Latest File Slice for a given fileId in a given partition. */ + @Override public final Option getLatestFileSlice(String partitionStr, String fileId) { try { readLock.lock(); diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/view/HoodieTableFileSystemView.java b/hudi-common/src/main/java/org/apache/hudi/common/table/view/HoodieTableFileSystemView.java index 34dadb0fec3e8..dd711244a064c 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/view/HoodieTableFileSystemView.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/view/HoodieTableFileSystemView.java @@ -205,6 +205,7 @@ public Stream fetchAllStoredFileGroups() { }); } + @Override public void close() { closed = true; super.reset(); @@ -212,6 +213,7 @@ public void close() { fgIdToPendingCompaction = null; } + @Override public boolean isClosed() { return closed; } diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/view/SpillableMapBasedFileSystemView.java b/hudi-common/src/main/java/org/apache/hudi/common/table/view/SpillableMapBasedFileSystemView.java index 928e060bb7cb9..3ada17e95d1eb 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/view/SpillableMapBasedFileSystemView.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/view/SpillableMapBasedFileSystemView.java @@ -76,6 +76,7 @@ protected Map> createPartitionToFileGroups() { } } + @Override protected Map> createFileIdToPendingCompactionMap( Map> fgIdToPendingCompaction) { try { @@ -91,6 +92,7 @@ protected Map> createFileId } } + @Override public Stream getAllFileGroups() { return ((ExternalSpillableMap) partitionToFileGroupsMap).valueStream() .flatMap(fg -> ((List) fg).stream()); diff --git a/hudi-common/src/main/java/org/apache/hudi/common/util/ObjectSizeCalculator.java b/hudi-common/src/main/java/org/apache/hudi/common/util/ObjectSizeCalculator.java index dc50c567152d1..b11ac6c1d1491 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/util/ObjectSizeCalculator.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/util/ObjectSizeCalculator.java @@ -124,6 +124,7 @@ public static long getObjectSize(Object obj) throws UnsupportedOperationExceptio private final LoadingCache, ClassSizeInfo> classSizeInfos = CacheBuilder.newBuilder().build(new CacheLoader, ClassSizeInfo>() { + @Override public ClassSizeInfo load(Class clazz) { return new ClassSizeInfo(clazz); } diff --git a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java index e26f1f23923aa..e764a170d2753 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java @@ -125,6 +125,7 @@ private void initFile(File writeOnlyFile) throws IOException { */ private void addShutDownHook() { Runtime.getRuntime().addShutdownHook(new Thread() { + @Override public void run() { try { if (writeOnlyFileHandle != null) { diff --git a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/LazyFileIterable.java b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/LazyFileIterable.java index aaaf32e8ebff7..4b8e71684c1a5 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/LazyFileIterable.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/LazyFileIterable.java @@ -115,6 +115,7 @@ private void close() { private void addShutdownHook() { Runtime.getRuntime().addShutdownHook(new Thread() { + @Override public void run() { close(); } diff --git a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieParquetInputFormat.java b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieParquetInputFormat.java index 5d6e52c41abee..e8f7de0f88ce0 100644 --- a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieParquetInputFormat.java +++ b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieParquetInputFormat.java @@ -173,10 +173,12 @@ private Map> groupFileStatus(FileStatus[ return grouped; } + @Override public void setConf(Configuration conf) { this.conf = conf; } + @Override public Configuration getConf() { return conf; } diff --git a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java index 1bf0da1e2e1a5..5d39bfc972be0 100644 --- a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java +++ b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java @@ -805,6 +805,7 @@ public static class HoodieCombineFileInputFormatShim extends CombineFileIn public HoodieCombineFileInputFormatShim() {} + @Override public Path[] getInputPathsShim(JobConf conf) { try { return FileInputFormat.getInputPaths(conf); @@ -813,6 +814,7 @@ public Path[] getInputPathsShim(JobConf conf) { } } + @Override public void createPool(JobConf conf, PathFilter... filters) { super.createPool(conf, filters); } @@ -822,6 +824,7 @@ public RecordReader getRecordReader(InputSplit split, JobConf job, Reporte throw new IOException("CombineFileInputFormat.getRecordReader not needed."); } + @Override protected List listStatus(JobContext job) throws IOException { LOG.info("Listing status in HoodieCombineHiveInputFormat.HoodieCombineFileInputFormatShim"); List result; @@ -851,6 +854,7 @@ protected List listStatus(JobContext job) throws IOException { return result; } + @Override public CombineFileSplit[] getSplits(JobConf job, int numSplits) throws IOException { long minSize = job.getLong(org.apache.hadoop.mapreduce.lib.input.FileInputFormat.SPLIT_MINSIZE, 0L); if (job.getLong("mapreduce.input.fileinputformat.split.minsize.per.node", 0L) == 0L) { @@ -879,12 +883,14 @@ public CombineFileSplit[] getSplits(JobConf job, int numSplits) throws IOExcepti return (CombineFileSplit[]) inputSplitShims.toArray(new HadoopShimsSecure.InputSplitShim[inputSplitShims.size()]); } + @Override public HadoopShimsSecure.InputSplitShim getInputSplitShim() throws IOException { return new HadoopShimsSecure.InputSplitShim(); } + @Override public RecordReader getRecordReader(JobConf job, CombineFileSplit split, Reporter reporter, - Class> rrClass) throws IOException { + Class> rrClass) throws IOException { return new HadoopShimsSecure.CombineFileRecordReader(job, split, reporter, rrClass); } diff --git a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java index a8eab1da09352..f8ddadbe5cd86 100644 --- a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java +++ b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java @@ -554,6 +554,7 @@ private HoodieInstant fetchNextCompactionInstant() throws InterruptedException { /** * Start Compaction Service. */ + @Override protected Pair startService() { ExecutorService executor = Executors.newFixedThreadPool(maxConcurrentCompaction); List> compactionFutures = From aaff453114017f5a3602a25e108c1431c14c269d Mon Sep 17 00:00:00 2001 From: lamber-ken Date: Mon, 30 Dec 2019 09:59:45 +0800 Subject: [PATCH 2/3] trigger rebuild From 851414e168d487f630f6069e55fb324ac5135f4a Mon Sep 17 00:00:00 2001 From: lamber-ken Date: Mon, 30 Dec 2019 10:45:54 +0800 Subject: [PATCH 3/3] revert code padding style --- .../apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java index 5d39bfc972be0..cd1cea32eef88 100644 --- a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java +++ b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java @@ -890,7 +890,7 @@ public HadoopShimsSecure.InputSplitShim getInputSplitShim() throws IOException { @Override public RecordReader getRecordReader(JobConf job, CombineFileSplit split, Reporter reporter, - Class> rrClass) throws IOException { + Class> rrClass) throws IOException { return new HadoopShimsSecure.CombineFileRecordReader(job, split, reporter, rrClass); }