diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/impl/BlockManagerImpl.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/impl/BlockManagerImpl.java index 722a76391f74..1cf3421c0bdd 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/impl/BlockManagerImpl.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/impl/BlockManagerImpl.java @@ -351,7 +351,7 @@ public List listBlock(Container container, long startLocalID, int result = new ArrayList<>(); String startKey = (startLocalID == -1) ? cData.startKeyEmpty() : cData.getBlockKey(startLocalID); - List> range = + List> range = db.getStore().getBlockDataTable() .getSequentialRangeKVs(startKey, count, cData.containerPrefix(), cData.getUnprefixedKeyFilter()); diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingTask.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingTask.java index c23e3b4ebcdb..37d3b9175dfd 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingTask.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingTask.java @@ -181,8 +181,7 @@ public ContainerBackgroundTaskResult deleteViaSchema1( // # of blocks to delete is throttled KeyPrefixFilter filter = containerData.getDeletingBlockKeyFilter(); - List> toDeleteBlocks = - blockDataTable + List> toDeleteBlocks = blockDataTable .getSequentialRangeKVs(containerData.startKeyEmpty(), (int) blocksToDelete, containerData.containerPrefix(), filter); diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/DatanodeTable.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/DatanodeTable.java index 7f1aba66aad4..5c4055086767 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/DatanodeTable.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/DatanodeTable.java @@ -111,7 +111,7 @@ public VALUE getReadCopy(KEY key) throws IOException { } @Override - public List> getRangeKVs( + public List> getRangeKVs( KEY startKey, int count, KEY prefix, MetadataKeyFilters.MetadataKeyFilter... filters) throws IOException, IllegalArgumentException { @@ -119,7 +119,7 @@ public List> getRangeKVs( } @Override - public List> getSequentialRangeKVs( + public List> getSequentialRangeKVs( KEY startKey, int count, KEY prefix, MetadataKeyFilters.MetadataKeyFilter... filters) throws IOException, IllegalArgumentException { diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/SchemaOneDeletedBlocksTable.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/SchemaOneDeletedBlocksTable.java index c4f8b66570b0..fd0b6f856403 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/SchemaOneDeletedBlocksTable.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/SchemaOneDeletedBlocksTable.java @@ -99,7 +99,7 @@ public ChunkInfoList getReadCopy(String key) throws IOException { } @Override - public List> getRangeKVs( + public List> getRangeKVs( String startKey, int count, String prefix, MetadataKeyFilters.MetadataKeyFilter... filters) throws IOException, IllegalArgumentException { @@ -112,7 +112,7 @@ public List> getRangeKVs( } @Override - public List> getSequentialRangeKVs( + public List> getSequentialRangeKVs( String startKey, int count, String prefix, MetadataKeyFilters.MetadataKeyFilter... filters) throws IOException, IllegalArgumentException { @@ -143,7 +143,7 @@ private static String unprefix(String key) { } private static List> unprefix( - List> kvs) { + List> kvs) { return kvs.stream() .map(kv -> Table.newKeyValue(unprefix(kv.getKey()), kv.getValue())) diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/TestSchemaOneBackwardsCompatibility.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/TestSchemaOneBackwardsCompatibility.java index f92c89480a70..54133bd750a8 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/TestSchemaOneBackwardsCompatibility.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/TestSchemaOneBackwardsCompatibility.java @@ -192,7 +192,7 @@ public void testBlockIteration(String schemaVersion) throws Exception { refCountedDB.getStore().getDeletedBlocksTable(); // Test rangeKVs. - List> deletedBlocks = + List> deletedBlocks = deletedBlocksTable.getRangeKVs(cData.startKeyEmpty(), 100, cData.containerPrefix()); @@ -357,7 +357,7 @@ public void testReadDeletedBlockChunkInfo(String schemaVersion) KeyValueContainerData cData = newKvData(); try (DBHandle refCountedDB = BlockUtils.getDB(cData, conf)) { // Read blocks that were already deleted before the upgrade. - List> deletedBlocks = + List> deletedBlocks = refCountedDB.getStore().getDeletedBlocksTable() .getRangeKVs(cData.startKeyEmpty(), 100, cData.containerPrefix()); @@ -416,7 +416,7 @@ public void testReadBlockData(String schemaVersion) throws Exception { } // Test decoding keys from the database. - List> blockKeyValues = + List> blockKeyValues = blockDataTable.getRangeKVs(cData.startKeyEmpty(), 100, cData.containerPrefix(), cData.getUnprefixedKeyFilter()); @@ -462,7 +462,7 @@ public void testReadDeletingBlockData(String schemaVersion) throws Exception { } // Test decoding keys from the database. - List> blockKeyValues = + List> blockKeyValues = blockDataTable.getRangeKVs(cData.startKeyEmpty(), 100, cData.containerPrefix(), cData.getDeletingBlockKeyFilter()); @@ -535,7 +535,7 @@ public void testReadDeletedBlocks(String schemaVersion) throws Exception { } // Test decoding keys from the database. - List> chunkInfoKeyValues = + List> chunkInfoKeyValues = deletedBlocksTable.getRangeKVs(cData.startKeyEmpty(), 100, cData.containerPrefix()); diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/Table.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/Table.java index 9add35f5d5c3..475e1bf0bba8 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/Table.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/Table.java @@ -273,7 +273,7 @@ default TableCacheMetrics createCacheMetrics() throws IOException { * @throws IOException if there are I/O errors. * @throws IllegalArgumentException if count is less than 0. */ - List> getRangeKVs(KEY startKey, + List> getRangeKVs(KEY startKey, int count, KEY prefix, MetadataKeyFilters.MetadataKeyFilter... filters) throws IOException, IllegalArgumentException; @@ -295,7 +295,7 @@ List> getRangeKVs(KEY startKey, * @throws IOException * @throws IllegalArgumentException */ - List> getSequentialRangeKVs(KEY startKey, + List> getSequentialRangeKVs(KEY startKey, int count, KEY prefix, MetadataKeyFilters.MetadataKeyFilter... filters) throws IOException, IllegalArgumentException; diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/TypedTable.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/TypedTable.java index abb5edf81d49..325d938a168d 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/TypedTable.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/TypedTable.java @@ -395,7 +395,7 @@ public void deleteRange(KEY beginKey, KEY endKey) throws IOException { } @Override - public Table.KeyValueIterator iterator(KEY prefix, KeyValueIterator.Type type) throws IOException { + public KeyValueIterator iterator(KEY prefix, KeyValueIterator.Type type) throws IOException { if (supportCodecBuffer) { final CodecBuffer prefixBuffer = encodeKeyCodecBuffer(prefix); try { @@ -469,7 +469,7 @@ public List> getRangeKVs( final byte[] startKeyBytes = encodeKey(startKey); final byte[] prefixBytes = encodeKey(prefix); - List> rangeKVBytes = + List> rangeKVBytes = rawTable.getRangeKVs(startKeyBytes, count, prefixBytes, filters); return convert(rangeKVBytes); } @@ -485,14 +485,13 @@ public List> getSequentialRangeKVs( final byte[] startKeyBytes = encodeKey(startKey); final byte[] prefixBytes = encodeKey(prefix); - List> rangeKVBytes = + List> rangeKVBytes = rawTable.getSequentialRangeKVs(startKeyBytes, count, prefixBytes, filters); return convert(rangeKVBytes); } - private List> convert(List> rangeKVBytes) - throws CodecException { + private List> convert(List> rangeKVBytes) throws CodecException { final List> rangeKVs = new ArrayList<>(); for (KeyValue kv : rangeKVBytes) { rangeKVs.add(Table.newKeyValue(decodeKey(kv.getKey()), decodeValue(kv.getValue()))); diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/MapBackedTableIterator.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/MapBackedTableIterator.java index 0ca7b452512e..7dc22b6d80fa 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/MapBackedTableIterator.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/MapBackedTableIterator.java @@ -22,12 +22,11 @@ import java.util.Map; import java.util.TreeMap; import org.apache.hadoop.hdds.utils.db.Table; -import org.apache.hadoop.hdds.utils.db.TableIterator; /** * Generic Table Iterator implementation that can be used for unit tests to reduce redundant mocking in tests. */ -public class MapBackedTableIterator implements TableIterator> { +public class MapBackedTableIterator implements Table.KeyValueIterator { private Iterator> itr; private final String prefix; diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/TestRDBSnapshotProvider.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/TestRDBSnapshotProvider.java index d012b6bcbb8e..61652dbd7ca2 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/TestRDBSnapshotProvider.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/TestRDBSnapshotProvider.java @@ -47,7 +47,6 @@ import org.apache.hadoop.hdds.utils.db.Table; import org.apache.hadoop.hdds.utils.db.Table.KeyValue; import org.apache.hadoop.hdds.utils.db.TableConfig; -import org.apache.hadoop.hdds.utils.db.TableIterator; import org.apache.hadoop.hdds.utils.db.managed.ManagedColumnFamilyOptions; import org.apache.hadoop.hdds.utils.db.managed.ManagedDBOptions; import org.junit.jupiter.api.AfterEach; @@ -190,7 +189,7 @@ public void compareDB(File db1, File db2, int columnFamilyUsed) final String name = families.get(i); final Table table1 = rdbStore1.getTable(name); final Table table2 = rdbStore2.getTable(name); - try (TableIterator> iterator + try (Table.KeyValueIterator iterator = table1.iterator()) { while (iterator.hasNext()) { KeyValue keyValue = iterator.next(); diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/InMemoryTestTable.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/InMemoryTestTable.java index bf960afe3e3e..6caf7336f4c2 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/InMemoryTestTable.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/InMemoryTestTable.java @@ -105,14 +105,14 @@ public long getEstimatedKeyCount() { } @Override - public List> getRangeKVs(KEY startKey, int count, KEY prefix, + public List> getRangeKVs(KEY startKey, int count, KEY prefix, MetadataKeyFilters.MetadataKeyFilter... filters) throws IOException, IllegalArgumentException { throw new UnsupportedOperationException(); } @Override - public List> getSequentialRangeKVs(KEY startKey, int count, KEY prefix, + public List> getSequentialRangeKVs(KEY startKey, int count, KEY prefix, MetadataKeyFilters.MetadataKeyFilter... filters) { throw new UnsupportedOperationException(); } diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestRDBTableStore.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestRDBTableStore.java index e421527b2e6e..e9a729f0ff17 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestRDBTableStore.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestRDBTableStore.java @@ -323,7 +323,7 @@ public void putGetTypedTableCodec() throws Exception { @Test public void forEachAndIterator() throws Exception { final int iterCount = 100; - try (Table testTable = rdbStore.getTable("Sixth")) { + try (Table testTable = rdbStore.getTable("Sixth")) { for (int x = 0; x < iterCount; x++) { byte[] key = RandomStringUtils.secure().next(10).getBytes(StandardCharsets.UTF_8); @@ -332,9 +332,9 @@ public void forEachAndIterator() throws Exception { testTable.put(key, value); } int localCount = 0; - try (TableIterator iter = testTable.iterator()) { + try (Table.KeyValueIterator iter = testTable.iterator()) { while (iter.hasNext()) { - Table.KeyValue keyValue = iter.next(); + iter.next(); localCount++; } @@ -491,7 +491,7 @@ public void testIteratorRemoveFromDB() throws Exception { // Remove without next removes first entry. try (Table testTable = rdbStore.getTable("Fifth")) { writeToTable(testTable, 3); - try (TableIterator> iterator = + try (Table.KeyValueIterator iterator = testTable.iterator()) { iterator.removeFromDB(); } @@ -503,7 +503,7 @@ public void testIteratorRemoveFromDB() throws Exception { // Remove after seekToLast removes lastEntry try (Table testTable = rdbStore.getTable("Sixth")) { writeToTable(testTable, 3); - try (TableIterator> iterator = + try (Table.KeyValueIterator iterator = testTable.iterator()) { iterator.seekToLast(); iterator.removeFromDB(); @@ -516,7 +516,7 @@ public void testIteratorRemoveFromDB() throws Exception { // Remove after seek deletes that entry. try (Table testTable = rdbStore.getTable("Sixth")) { writeToTable(testTable, 3); - try (TableIterator> iterator = + try (Table.KeyValueIterator iterator = testTable.iterator()) { iterator.seek(bytesOf[3]); iterator.removeFromDB(); @@ -529,7 +529,7 @@ public void testIteratorRemoveFromDB() throws Exception { // Remove after next() deletes entry that was returned by next. try (Table testTable = rdbStore.getTable("Sixth")) { writeToTable(testTable, 3); - try (TableIterator> iterator = + try (Table.KeyValueIterator iterator = testTable.iterator()) { iterator.seek(bytesOf[2]); iterator.next(); @@ -564,8 +564,7 @@ public void testPrefixedIterator() throws Exception { // iterator should seek to right pos in the middle byte[] samplePrefix = testPrefixes.get(2).getBytes( StandardCharsets.UTF_8); - try (TableIterator> iter = testTable.iterator( + try (Table.KeyValueIterator iter = testTable.iterator( samplePrefix)) { int keyCount = 0; while (iter.hasNext()) { @@ -656,7 +655,7 @@ public void testPrefixedRangeKVs() throws Exception { // test start at first byte[] startKey = samplePrefix; - List> rangeKVs = testTable + List> rangeKVs = testTable .getRangeKVs(startKey, 3, samplePrefix); assertEquals(3, rangeKVs.size()); @@ -710,8 +709,7 @@ public void testDumpAndLoadBasic() throws Exception { testTable2.loadFromFile(dumpFile); // check loaded keys - try (TableIterator> iter = testTable2.iterator( + try (Table.KeyValueIterator iter = testTable2.iterator( samplePrefix)) { int keyCount = 0; while (iter.hasNext()) { @@ -751,8 +749,7 @@ public void testDumpAndLoadEmpty() throws Exception { testTable2.loadFromFile(dumpFile); // check loaded keys - try (TableIterator> iter = testTable2.iterator( + try (Table.KeyValueIterator iter = testTable2.iterator( samplePrefix)) { int keyCount = 0; while (iter.hasNext()) { diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestTypedRDBTableStore.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestTypedRDBTableStore.java index 318bfdbc5def..70ac80848d16 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestTypedRDBTableStore.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestTypedRDBTableStore.java @@ -41,7 +41,6 @@ import java.util.Set; import org.apache.commons.lang3.RandomStringUtils; import org.apache.hadoop.hdds.StringUtils; -import org.apache.hadoop.hdds.utils.db.Table.KeyValue; import org.apache.hadoop.hdds.utils.db.cache.CacheKey; import org.apache.hadoop.hdds.utils.db.cache.CacheValue; import org.apache.hadoop.hdds.utils.db.cache.TableCache.CacheType; @@ -227,7 +226,7 @@ public void forEachAndIterator() throws Exception { } int localCount = 0; - try (TableIterator> iter = + try (Table.KeyValueIterator iter = testTable.iterator()) { while (iter.hasNext()) { Table.KeyValue keyValue = iter.next(); diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMCertStore.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMCertStore.java index b3751ab35585..a54cd1e4fa89 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMCertStore.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMCertStore.java @@ -172,7 +172,7 @@ public List listCertificate(NodeType role, .collect(Collectors.toList()); } - private List> + private List> getValidCertTableList(NodeType role, BigInteger startSerialID, int count) throws IOException { // Implemented for role SCM and CertType VALID_CERTS.