diff --git a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/ContainerClientMetrics.java b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/ContainerClientMetrics.java index 2634ffb726b5..f30b869d866a 100644 --- a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/ContainerClientMetrics.java +++ b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/ContainerClientMetrics.java @@ -66,7 +66,6 @@ public final class ContainerClientMetrics { @Metric private MutableCounterLong flushesDuringWrite; - private MutableQuantiles[] listBlockLatency; private MutableQuantiles[] getBlockLatency; private MutableQuantiles[] getCommittedBlockLengthLatency; diff --git a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/XceiverClientManager.java b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/XceiverClientManager.java index ef2e3573aae9..7e38e7d9c2ce 100644 --- a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/XceiverClientManager.java +++ b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/XceiverClientManager.java @@ -275,7 +275,6 @@ public long getStaleThreshold(TimeUnit unit) { return unit.convert(staleThreshold, MILLISECONDS); } - public int getMaxSize() { return maxSize; } diff --git a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/XceiverClientRatis.java b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/XceiverClientRatis.java index f185c3e85576..4accbdbaec44 100644 --- a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/XceiverClientRatis.java +++ b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/XceiverClientRatis.java @@ -239,7 +239,6 @@ private RaftClient getClient() { return Objects.requireNonNull(client.get(), "client is null"); } - @VisibleForTesting public ConcurrentMap getCommitInfoMap() { return commitInfoMap; diff --git a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/client/HddsClientUtils.java b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/client/HddsClientUtils.java index ac2503339794..a09ab64eae5a 100644 --- a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/client/HddsClientUtils.java +++ b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/client/HddsClientUtils.java @@ -231,7 +231,6 @@ public static int getMaxOutstandingRequests(ConfigurationSource config) { .getMaxOutstandingRequests(); } - // This will return the underlying exception after unwrapping // the exception to see if it matches with expected exception // list otherwise will return the exception back. diff --git a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/BlockDataStreamOutput.java b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/BlockDataStreamOutput.java index d9e3adca139b..7eccdc2fafff 100644 --- a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/BlockDataStreamOutput.java +++ b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/BlockDataStreamOutput.java @@ -139,6 +139,7 @@ public class BlockDataStreamOutput implements ByteBufferStreamOutput { // buffers for which putBlock is yet to be executed private List buffersForPutBlock; private boolean isDatastreamPipelineMode; + /** * Creates a new BlockDataStreamOutput. * @@ -604,7 +605,6 @@ private void validateResponse( } } - private void setIoException(Throwable e) { IOException ioe = getIoException(); if (ioe == null) { diff --git a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/BlockOutputStream.java b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/BlockOutputStream.java index 1e950554d7c5..a03fd037bda0 100644 --- a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/BlockOutputStream.java +++ b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/BlockOutputStream.java @@ -819,7 +819,6 @@ void validateResponse( } } - public IOException setIoException(Throwable e) { IOException ioe = getIoException(); if (ioe == null) { @@ -1245,6 +1244,7 @@ ContainerCommandResponseProto getResponse() { */ private static class FlushRuntimeException extends RuntimeException { private final IOException cause; + FlushRuntimeException(IOException cause) { this.cause = cause; } diff --git a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/BufferPool.java b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/BufferPool.java index fedc311148db..03a0f931ab3c 100644 --- a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/BufferPool.java +++ b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/BufferPool.java @@ -55,7 +55,6 @@ public class BufferPool { private final Lock lock = new ReentrantLock(); private final Condition notFull = lock.newCondition(); - public static BufferPool empty() { return EMPTY; } diff --git a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/ChunkInputStream.java b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/ChunkInputStream.java index 23c96fc7d6a7..74ba51de8884 100644 --- a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/ChunkInputStream.java +++ b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/ChunkInputStream.java @@ -675,7 +675,6 @@ private boolean chunkStreamEOF() { } } - /** * Release the buffers upto the given index. * @param releaseUptoBufferIndex bufferIndex (inclusive) upto which the diff --git a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/ECBlockOutputStream.java b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/ECBlockOutputStream.java index b928f74f06de..d798b3a9385a 100644 --- a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/ECBlockOutputStream.java +++ b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/ECBlockOutputStream.java @@ -64,6 +64,7 @@ public class ECBlockOutputStream extends BlockOutputStream { private CompletableFuture putBlkRspFuture = null; + /** * Creates a new ECBlockOutputStream. * diff --git a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/StreamBuffer.java b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/StreamBuffer.java index 4c348e74c259..3e2861803686 100644 --- a/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/StreamBuffer.java +++ b/hadoop-hdds/client/src/main/java/org/apache/hadoop/hdds/scm/storage/StreamBuffer.java @@ -46,7 +46,6 @@ public int position() { return buffer.position(); } - public void put(StreamBuffer sb) { buffer.put(sb.buffer); } diff --git a/hadoop-hdds/client/src/main/java/org/apache/hadoop/ozone/client/io/ECBlockInputStream.java b/hadoop-hdds/client/src/main/java/org/apache/hadoop/ozone/client/io/ECBlockInputStream.java index 25f2f46d9a53..e3c67a99a0c4 100644 --- a/hadoop-hdds/client/src/main/java/org/apache/hadoop/ozone/client/io/ECBlockInputStream.java +++ b/hadoop-hdds/client/src/main/java/org/apache/hadoop/ozone/client/io/ECBlockInputStream.java @@ -75,6 +75,7 @@ public class ECBlockInputStream extends BlockExtendedInputStream { private boolean closed = false; private boolean seeked = false; private OzoneClientConfig config; + protected ECReplicationConfig getRepConfig() { return repConfig; } diff --git a/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/ECStreamTestUtil.java b/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/ECStreamTestUtil.java index db7d7a111e44..639b4e92a147 100644 --- a/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/ECStreamTestUtil.java +++ b/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/ECStreamTestUtil.java @@ -212,7 +212,6 @@ public static Map createIndexMap(int... idxs) { return map; } - /** * A stream factory which can be used in tests to provide TestBlockStream * instances. @@ -357,7 +356,7 @@ public int read(ByteBuffer buf) throws IOException { buf.put(data.get()); } return toRead; - }; + } private void throwError() throws IOException { if (errorToThrow != null) { diff --git a/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/TestECBlockInputStream.java b/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/TestECBlockInputStream.java index 00bf2e915742..68fd37221f0c 100644 --- a/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/TestECBlockInputStream.java +++ b/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/TestECBlockInputStream.java @@ -642,7 +642,7 @@ public int read(ByteBuffer buf) throws IOException { } position += toRead; return toRead; - }; + } @Override protected int readWithStrategy(ByteReaderStrategy strategy) throws diff --git a/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/TestECBlockReconstructedInputStream.java b/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/TestECBlockReconstructedInputStream.java index 71f0a0a07952..1c88d94d3d8c 100644 --- a/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/TestECBlockReconstructedInputStream.java +++ b/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/TestECBlockReconstructedInputStream.java @@ -157,7 +157,6 @@ public void testReadDataByteBufferMultipleStripes() throws IOException { } } - @Test public void testReadDataWithUnbuffer() throws IOException { // Read buffer is 16kb + 5 bytes so it does not align with stripes exactly @@ -349,8 +348,6 @@ private void resetAndAdvanceDataGenerator(long position) { } } - - /** * Return a list of num ByteBuffers of the given size. * @param num Number of buffers to create diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/HddsUtils.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/HddsUtils.java index 0fe5d9a69749..c450e46e8ffb 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/HddsUtils.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/HddsUtils.java @@ -99,7 +99,6 @@ @InterfaceStability.Stable public final class HddsUtils { - private static final Logger LOG = LoggerFactory.getLogger(HddsUtils.class); public static final ByteString REDACTED = diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/JavaUtils.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/JavaUtils.java index bf59fc553d8f..4d1ec4fc0696 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/JavaUtils.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/JavaUtils.java @@ -47,7 +47,6 @@ public static boolean isJavaVersionAtMost(int version) { return JAVA_SPEC_VER <= version; } - /** * Private constructor. */ diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/client/DecommissionUtils.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/client/DecommissionUtils.java index d3130de729ee..bad4973237ee 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/client/DecommissionUtils.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/client/DecommissionUtils.java @@ -45,7 +45,6 @@ @InterfaceStability.Stable public final class DecommissionUtils { - private static final Logger LOG = LoggerFactory.getLogger(DecommissionUtils.class); private DecommissionUtils() { diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/protocol/DatanodeDetails.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/protocol/DatanodeDetails.java index a9bcbad385f0..a780a96637da 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/protocol/DatanodeDetails.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/protocol/DatanodeDetails.java @@ -412,7 +412,6 @@ public Port getStandalonePort() { return getPort(Name.STANDALONE); } - /** * Starts building a new DatanodeDetails from the protobuf input. * @@ -679,7 +678,6 @@ public boolean equals(Object obj) { id.equals(((DatanodeDetails) obj).id); } - /** * Checks hostname, ipAddress and port of the 2 nodes are the same. * @param datanodeDetails dnDetails object to compare with. diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/ratis/RatisHelper.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/ratis/RatisHelper.java index 4d89d4fd2cda..1b77b31275c1 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/ratis/RatisHelper.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/ratis/RatisHelper.java @@ -413,7 +413,6 @@ public static void createRaftServerProperties(ConfigurationSource ozoneConf, }); } - private static Map getDatanodeRatisPrefixProps( ConfigurationSource configuration) { return configuration.getPropsMatchPrefixAndTrimPrefix( @@ -531,7 +530,6 @@ static RaftPeer newRaftPeer(RaftPeer peer, RaftPeerId target) { return RaftPeer.newBuilder(peer).setPriority(priority).build(); } - /** * Use raft client to send admin request, transfer the leadership. * 1. Set priority and send setConfiguration request @@ -647,7 +645,6 @@ public static int calculateAttempts(Duration pollInterval, Duration maxDuration) return (int) (max / interval); } - public static void setFirstElectionTimeoutDuration( ConfigurationSource conf, RaftProperties properties, String configKey) { long firstElectionTimeout = conf.getTimeDuration(configKey, -1, TimeUnit.MILLISECONDS); diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/ratis/retrypolicy/RequestTypeDependentRetryPolicyCreator.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/ratis/retrypolicy/RequestTypeDependentRetryPolicyCreator.java index ffa0efb84b67..02d4118807bf 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/ratis/retrypolicy/RequestTypeDependentRetryPolicyCreator.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/ratis/retrypolicy/RequestTypeDependentRetryPolicyCreator.java @@ -114,6 +114,7 @@ private static ExceptionDependentRetry createExceptionDependentPolicy( .setDefaultPolicy(multipleLinearRandomRetry) .build(); } + private static TimeDuration toTimeDuration(Duration duration) { return toTimeDuration(duration.toMillis()); } diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/AddSCMRequest.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/AddSCMRequest.java index aa585964dc69..fbdfa1767db5 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/AddSCMRequest.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/AddSCMRequest.java @@ -45,6 +45,7 @@ public HddsProtos.AddScmRequestProto getProtobuf() { return HddsProtos.AddScmRequestProto.newBuilder().setClusterId(clusterId) .setScmId(scmId).setRatisAddr(ratisAddr).build(); } + /** * Builder for AddSCMRequest. */ @@ -53,7 +54,6 @@ public static class Builder { private String scmId; private String ratisAddr; - /** * sets the cluster id. * @param cid clusterId to be set diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfig.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfig.java index 7b2e920b764c..db5fde5efe65 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfig.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfig.java @@ -151,7 +151,6 @@ public void setKerberosPrincipal(String kerberosPrincipal) { this.principal = kerberosPrincipal; } - public void setKerberosKeytab(String kerberosKeytab) { this.keytab = kerberosKeytab; } diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfigKeys.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfigKeys.java index c6abf7062a3a..44229338ec64 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfigKeys.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfigKeys.java @@ -445,7 +445,6 @@ public final class ScmConfigKeys { public static final String OZONE_SCM_PIPELINE_SCRUB_INTERVAL_DEFAULT = "150s"; - // Allow SCM to auto create factor ONE ratis pipeline. public static final String OZONE_SCM_PIPELINE_AUTO_CREATE_FACTOR_ONE = "ozone.scm.pipeline.creation.auto.factor.one"; @@ -497,7 +496,6 @@ public final class ScmConfigKeys { "ozone.scm.ca.list.retry.interval"; public static final long OZONE_SCM_CA_LIST_RETRY_INTERVAL_DEFAULT = 10; - public static final String OZONE_SCM_EVENT_PREFIX = "ozone.scm.event."; public static final String OZONE_SCM_EVENT_CONTAINER_REPORT_THREAD_POOL_SIZE = diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/XceiverClientSpi.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/XceiverClientSpi.java index 7a43bf6f6d83..35e65271bb1c 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/XceiverClientSpi.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/XceiverClientSpi.java @@ -149,6 +149,7 @@ public static IOException getIOExceptionForSendCommand( return new IOException("Failed to execute command " + HddsUtils.processForDebug(request), e); } + /** * Sends a given command to server gets a waitable future back. * diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/container/common/helpers/ContainerWithPipeline.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/container/common/helpers/ContainerWithPipeline.java index 2e1a8f531779..cfd2d93a5df7 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/container/common/helpers/ContainerWithPipeline.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/container/common/helpers/ContainerWithPipeline.java @@ -66,7 +66,6 @@ public HddsProtos.ContainerWithPipeline getProtobuf(int clientVersion) return builder.build(); } - @Override public String toString() { return containerInfo.toString() + " | " + pipeline.toString(); diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/container/common/helpers/ExcludeList.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/container/common/helpers/ExcludeList.java index 58cfdd411fb1..bfe242eeb8b2 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/container/common/helpers/ExcludeList.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/container/common/helpers/ExcludeList.java @@ -44,7 +44,6 @@ public class ExcludeList { private long expiryTime = 0; private final Clock clock; - public ExcludeList() { clock = Clock.system(ZoneOffset.UTC); } diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/InnerNodeImpl.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/InnerNodeImpl.java index d594613021ba..682026fdb1d7 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/InnerNodeImpl.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/InnerNodeImpl.java @@ -677,6 +677,7 @@ private Map getExcludedScopeNodeCount( } return nodeCounts; } + protected static class Factory implements InnerNode.Factory { protected Factory() { } diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NetUtils.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NetUtils.java index 183a2c47a548..7280b78a3557 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NetUtils.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NetUtils.java @@ -36,6 +36,7 @@ public final class NetUtils { private NetUtils() { // Prevent instantiation } + /** * Normalize a path by stripping off any trailing. * {@link NetConstants#PATH_SEPARATOR} @@ -74,7 +75,6 @@ public static int locationToDepth(String location) { newLocation.split(NetConstants.PATH_SEPARATOR_STR).length; } - /** * Remove node from mutableExcludedNodes if it's covered by excludedScope. * Please noted that mutableExcludedNodes content might be changed after the diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NetworkTopology.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NetworkTopology.java index 34f66e674646..680b46ecba99 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NetworkTopology.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NetworkTopology.java @@ -27,10 +27,12 @@ public interface NetworkTopology { /** Exception for invalid network topology detection. */ class InvalidTopologyException extends RuntimeException { private static final long serialVersionUID = 1L; + public InvalidTopologyException(String msg) { super(msg); } } + /** * Add a leaf node. This will be called when a new datanode is added. * @param node node to be added; can be null diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeImpl.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeImpl.java index 4c70ad7b9621..10e3bf0fce46 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeImpl.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeImpl.java @@ -248,6 +248,7 @@ public boolean isAncestor(String nodePath) { NetUtils.addSuffix(nodePath).startsWith( NetUtils.addSuffix(this.getNetworkFullPath())); } + @Override public boolean isDescendant(Node node) { if (node == null) { diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchema.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchema.java index fb3b4e2c346a..c1cf4cb7e199 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchema.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchema.java @@ -133,6 +133,7 @@ public void setDefaultName(String name) { public int getCost() { return this.cost; } + public void setCost(int cost) { this.cost = cost; } @@ -170,6 +171,7 @@ public String toString() { public int getCost() { return cost; } + public static LayerType getType(String typeStr) { for (LayerType type: LayerType.values()) { if (typeStr.equalsIgnoreCase(type.toString())) { diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchemaLoader.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchemaLoader.java index 7493909ee176..832a2d47d68b 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchemaLoader.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/net/NodeSchemaLoader.java @@ -66,6 +66,7 @@ public final class NodeSchemaLoader { private static final int LAYOUT_VERSION = 1; private static volatile NodeSchemaLoader instance = null; + private NodeSchemaLoader() { } public static NodeSchemaLoader getInstance() { diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java index 722c60f3d3bc..605c3bc254c4 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/pipeline/Pipeline.java @@ -223,7 +223,6 @@ public boolean sameDatanodes(Pipeline pipeline) { return getNodeSet().equals(pipeline.getNodeSet()); } - /** * Return the replica index of the specific datanode in the datanode set. *

@@ -604,6 +603,7 @@ public Builder setLeaderId(UUID leaderId1) { this.leaderId = leaderId1; return this; } + public Builder setNodes(List nodes) { this.nodeStatus = new LinkedHashMap<>(); nodes.forEach(node -> nodeStatus.put(node, -1L)); @@ -642,7 +642,6 @@ public Builder setSuggestedLeaderId(UUID uuid) { return this; } - public Builder setReplicaIndexes(Map indexes) { this.replicaIndexes = indexes == null ? ImmutableMap.of() : ImmutableMap.copyOf(indexes); return this; diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/protocolPB/ContainerCommandResponseBuilders.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/protocolPB/ContainerCommandResponseBuilders.java index 9ce971c08139..6d32decf54a5 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/protocolPB/ContainerCommandResponseBuilders.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/protocolPB/ContainerCommandResponseBuilders.java @@ -168,6 +168,7 @@ public static ContainerCommandResponseProto getListBlockResponse( .setListBlock(builder) .build(); } + /** * Returns successful getCommittedBlockLength Response. * @param msg - Request. @@ -281,6 +282,7 @@ public static ContainerCommandResponseProto getGetSmallFileResponseSuccess( .setGetSmallFile(getSmallFile) .build(); } + /** * Returns a ReadContainer Response. * diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/storage/ContainerProtocolCalls.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/storage/ContainerProtocolCalls.java index 3937adc2523b..317b1da83935 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/storage/ContainerProtocolCalls.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/storage/ContainerProtocolCalls.java @@ -563,6 +563,7 @@ public static void createContainer(XceiverClientSpi client, long containerID, String encodedToken) throws IOException { createContainer(client, containerID, encodedToken, null, 0); } + /** * createContainer call that creates a container on the datanode. * @param client - client diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/security/token/ShortLivedTokenIdentifier.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/security/token/ShortLivedTokenIdentifier.java index 11f2aebf2bf5..112034391cef 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/security/token/ShortLivedTokenIdentifier.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/security/token/ShortLivedTokenIdentifier.java @@ -74,7 +74,6 @@ public Instant getExpiry() { return expiry; } - public String getOwnerId() { return ownerId; } diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/utils/CertificateCodec.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/utils/CertificateCodec.java index 788bb14f1b04..8a50c21a4dfa 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/utils/CertificateCodec.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/utils/CertificateCodec.java @@ -70,6 +70,7 @@ public class CertificateCodec { private final Set permissionSet = Stream.of(OWNER_READ, OWNER_WRITE, OWNER_EXECUTE) .collect(Collectors.toSet()); + /** * Creates a CertificateCodec with component name. * diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/security/x509/exception/CertificateException.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/security/x509/exception/CertificateException.java index 8f23310f5d03..181f8888380a 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/security/x509/exception/CertificateException.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/security/x509/exception/CertificateException.java @@ -25,6 +25,7 @@ public class CertificateException extends SCMSecurityException { private ErrorCode errorCode; + /** * Ctor. * @param message - Error Message. diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/tracing/TracingUtil.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/tracing/TracingUtil.java index f49932e43100..867802120ef2 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/tracing/TracingUtil.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/tracing/TracingUtil.java @@ -197,7 +197,6 @@ public static void executeAsChildSpan(String spanName, executeInSpan(span, runnable); } - /** * Create an active span with auto-close at finish. *

diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/BackgroundService.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/BackgroundService.java index 692571e83e64..959bee8d8c5f 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/BackgroundService.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/BackgroundService.java @@ -102,7 +102,6 @@ public void runPeriodicalTaskNow() throws Exception { } } - // start service public void start() { exec.scheduleWithFixedDelay(service, 0, interval, unit); diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/LeakTracker.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/LeakTracker.java index ced3b3caab86..4b34656de30c 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/LeakTracker.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/LeakTracker.java @@ -30,6 +30,7 @@ final class LeakTracker extends WeakReference implements UncheckedAutoCloseable { private final Set allLeaks; private final Runnable leakReporter; + LeakTracker(Object referent, ReferenceQueue referenceQueue, Set allLeaks, Runnable leakReporter) { super(referent, referenceQueue); diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/db/CodecBuffer.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/db/CodecBuffer.java index 89e3e477d19a..d575c4255e50 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/db/CodecBuffer.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/db/CodecBuffer.java @@ -79,6 +79,7 @@ public class CodecBuffer implements UncheckedAutoCloseable { private static class Factory { private static volatile BiFunction constructor = CodecBuffer::new; + static void set(BiFunction f, String name) { constructor = f; LOG.info("Successfully set constructor to {}: {}", name, f); diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/db/ShortCodec.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/db/ShortCodec.java index ed46cb5c5510..e2075dc9ee33 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/db/ShortCodec.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/utils/db/ShortCodec.java @@ -56,7 +56,6 @@ public Short fromCodecBuffer(@Nonnull CodecBuffer buffer) { return buffer.asReadOnlyByteBuffer().getShort(); } - @Override public byte[] toPersistedFormat(Short object) { return ByteBuffer.wrap(new byte[Short.BYTES]).putShort(object).array(); diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/OzoneConfigKeys.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/OzoneConfigKeys.java index 99f39f2e0a4b..e1a1df0a9629 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/OzoneConfigKeys.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/OzoneConfigKeys.java @@ -43,7 +43,6 @@ public final class OzoneConfigKeys { public static final String OZONE_OM_DB_DIRS_PERMISSIONS = "ozone.om.db.dirs.permissions"; - public static final String OZONE_OM_DB_DIRS = "ozone.om.db.dirs"; /** @@ -140,7 +139,6 @@ public final class OzoneConfigKeys { public static final String OZONE_OM_LEASE_SOFT_LIMIT_DEFAULT = "60s"; - /** * When set to true, allocate a random free port for ozone container, so that * a mini cluster is able to launch multiple containers on a node. @@ -197,7 +195,6 @@ public final class OzoneConfigKeys { "ozone.client.ec.grpc.write.timeout"; public static final String OZONE_CLIENT_EC_GRPC_WRITE_TIMEOUT_DEFAULT = "30s"; - /** * Ozone administrator users delimited by comma. * If not set, only the user who launches an ozone service will be the @@ -318,7 +315,6 @@ public final class OzoneConfigKeys { public static final String OZONE_RECOVERING_CONTAINER_TIMEOUT_DEFAULT = "20m"; - public static final String OZONE_KEY_PREALLOCATION_BLOCKS_MAX = "ozone.key.preallocation.max.blocks"; public static final int OZONE_KEY_PREALLOCATION_BLOCKS_MAX_DEFAULT @@ -449,7 +445,6 @@ public final class OzoneConfigKeys { public static final String OZONE_CONTAINER_COPY_WORKDIR = "hdds.datanode.replication.work.dir"; - public static final int OZONE_CLIENT_BYTES_PER_CHECKSUM_MIN_SIZE = 8 * 1024; public static final String OZONE_CLIENT_READ_TIMEOUT @@ -509,7 +504,6 @@ public final class OzoneConfigKeys { public static final String OZONE_FREON_HTTP_AUTH_CONFIG_PREFIX = "ozone.freon.http.auth."; - public static final String OZONE_NETWORK_TOPOLOGY_AWARE_READ_KEY = "ozone.network.topology.aware.read"; public static final boolean OZONE_NETWORK_TOPOLOGY_AWARE_READ_DEFAULT = true; @@ -625,7 +619,6 @@ public final class OzoneConfigKeys { public static final String OZONE_FS_LISTING_PAGE_SIZE_MAX = "ozone.fs.listing.page.size.max"; - public static final String FS_TRASH_CLASSNAME = "fs.trash.classname"; public static final String FS_TRASH_CLASSNAME_DEFAULT = "org.apache.hadoop.ozone.om.TrashPolicyOzone"; @@ -664,7 +657,6 @@ public final class OzoneConfigKeys { public static final String HDDS_SCM_CLIENT_FAILOVER_MAX_RETRY = "hdds.scmclient.failover.max.retry"; - public static final String OZONE_XCEIVER_CLIENT_METRICS_PERCENTILES_INTERVALS_SECONDS_KEY = "ozone.xceiver.client.metrics.percentiles.intervals.seconds"; @@ -678,7 +670,6 @@ public final class OzoneConfigKeys { "ozone.security.crypto.compliance.mode"; public static final String OZONE_SECURITY_CRYPTO_COMPLIANCE_MODE_UNRESTRICTED = "unrestricted"; - /** * There is no need to instantiate this class. */ diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/common/ChecksumByteBufferFactory.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/common/ChecksumByteBufferFactory.java index 505b057a42a4..410dfb328cad 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/common/ChecksumByteBufferFactory.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/common/ChecksumByteBufferFactory.java @@ -68,7 +68,7 @@ public static java.util.zip.Checksum createChecksum() { : new RuntimeException(t); } } - }; + } public static ChecksumByteBuffer crc32Impl() { return new ChecksumByteBufferImpl(new CRC32()); diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/lock/BootstrapStateHandler.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/lock/BootstrapStateHandler.java index 49befe2c8faf..d6de873b842e 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/lock/BootstrapStateHandler.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/lock/BootstrapStateHandler.java @@ -31,6 +31,7 @@ public Lock lock() throws InterruptedException { semaphore.acquire(); return this; } + public void unlock() { semaphore.release(); } diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/upgrade/UpgradeException.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/upgrade/UpgradeException.java index ec1b3abaf4d4..20a7d02d178f 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/upgrade/UpgradeException.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/upgrade/UpgradeException.java @@ -99,6 +99,7 @@ public UpgradeException.ResultCodes getResult() { public String toString() { return result + " " + super.toString(); } + /** * Error codes to make it easy to decode these exceptions. */ diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/upgrade/UpgradeFinalization.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/upgrade/UpgradeFinalization.java index 0afcc2c7b230..218a09b6edc5 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/upgrade/UpgradeFinalization.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/upgrade/UpgradeFinalization.java @@ -167,6 +167,7 @@ public static void emitCancellationMsg(String component) { + "will not cancel finalization in " + component + ". Progress can be" + "monitored in the Ozone Manager's log."); } + private UpgradeFinalization() { // no instances } diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/util/OzoneNetUtils.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/util/OzoneNetUtils.java index cb962b8058be..d3f6e3f72014 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/util/OzoneNetUtils.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/ozone/util/OzoneNetUtils.java @@ -88,7 +88,6 @@ private static String getHostNameWithoutDomain(final String fqdn) { return fqdn.split("\\.")[0]; } - /** * Match input address to local address. * Return true if it matches, false otherwsie. @@ -98,7 +97,6 @@ public static boolean isAddressLocal(InetSocketAddress addr) { return inetAddress != null && NetUtils.isLocalAddress(inetAddress); } - public static boolean isUnresolved(boolean flexibleFqdnResolutionEnabled, InetSocketAddress addr) { return !flexibleFqdnResolutionEnabled && addr.isUnresolved() diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/TestHddsUtils.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/TestHddsUtils.java index 71fb2eae2333..5eb1b5e31fee 100644 --- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/TestHddsUtils.java +++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/TestHddsUtils.java @@ -192,7 +192,6 @@ void testGetSCMAddresses() { "An invalid value should cause an IllegalArgumentException"); } - @Test void testGetSCMAddressesWithHAConfig() { OzoneConfiguration conf = new OzoneConfiguration(); diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/protocol/TestDatanodeDetails.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/protocol/TestDatanodeDetails.java index 68cacb6e3dff..dcbf9553dd03 100644 --- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/protocol/TestDatanodeDetails.java +++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/protocol/TestDatanodeDetails.java @@ -51,6 +51,7 @@ void protoIncludesNewPortsOnlyForV1() { subject.toProto(VERSION_HANDLES_UNKNOWN_DN_PORTS.toProtoValue()); assertPorts(protoV1, ALL_PORTS); } + @Test void testRequiredPortsProto() { DatanodeDetails subject = MockDatanodeDetails.randomDatanodeDetails(); diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestRatisHelper.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestRatisHelper.java index bb8a0e9ef681..2f10f550bb5a 100644 --- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestRatisHelper.java +++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestRatisHelper.java @@ -76,7 +76,6 @@ public void testCreateRaftGrpcPropertiesForClient() { } - @Test public void testCreateRaftGrpcPropertiesForServer() { diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestReplicationManagerReport.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestReplicationManagerReport.java index ee0f920d8af9..e669b8cdc59b 100644 --- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestReplicationManagerReport.java +++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestReplicationManagerReport.java @@ -71,7 +71,6 @@ void testMetricCanBeIncremented() { report.getStat(HddsProtos.LifeCycleState.QUASI_CLOSED)); } - @Test void testJsonOutput() throws IOException { report.increment(HddsProtos.LifeCycleState.OPEN); diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java index 13b87ebbacbf..8706f0f7f035 100644 --- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java +++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java @@ -162,6 +162,4 @@ public void testNonHAWithRestDefaults() { assertEquals("localhost:" + OZONE_SCM_DATANODE_PORT_DEFAULT, scmNodeInfos.get(0).getScmDatanodeAddress()); } - - } diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/security/x509/keys/TestHDDSKeyGenerator.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/security/x509/keys/TestHDDSKeyGenerator.java index 2718763feca3..f4f5cd147c72 100644 --- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/security/x509/keys/TestHDDSKeyGenerator.java +++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/security/x509/keys/TestHDDSKeyGenerator.java @@ -47,6 +47,7 @@ public void init() { conf.set(OZONE_METADATA_DIRS, tempPath.getPath()); config = new SecurityConfig(conf); } + /** * In this test we verify that we are able to create a key pair, then get * bytes of that and use ASN1. parser to parse it back to a private key. diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/container/ContainerTestHelper.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/container/ContainerTestHelper.java index 6277a3584c6b..67c71f9618ab 100644 --- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/container/ContainerTestHelper.java +++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/container/ContainerTestHelper.java @@ -230,7 +230,6 @@ public static ContainerCommandRequestProto getWriteSmallFileRequest( return request.build(); } - public static ContainerCommandRequestProto getReadSmallFileRequest( Pipeline pipeline, ContainerProtos.PutBlockRequestProto putKey) throws Exception { @@ -293,7 +292,6 @@ public static ContainerCommandRequestProto getCreateContainerRequest( return getCreateContainerRequest(containerID, pipeline, ContainerProtos.ContainerDataProto.State.OPEN); } - /** * Returns a create container command for test purposes. There are a bunch of * tests where we need to just send a request and get a reply. diff --git a/hadoop-hdds/config/src/main/java/org/apache/hadoop/hdds/conf/ConfigFileGenerator.java b/hadoop-hdds/config/src/main/java/org/apache/hadoop/hdds/conf/ConfigFileGenerator.java index 91ff6029af1d..372d75753b63 100644 --- a/hadoop-hdds/config/src/main/java/org/apache/hadoop/hdds/conf/ConfigFileGenerator.java +++ b/hadoop-hdds/config/src/main/java/org/apache/hadoop/hdds/conf/ConfigFileGenerator.java @@ -59,7 +59,6 @@ public Element visitDeclared(DeclaredType t, Void aVoid) { } }; - @Override public boolean process(Set annotations, RoundEnvironment roundEnv) { diff --git a/hadoop-hdds/config/src/main/java/org/apache/hadoop/hdds/conf/ConfigurationSource.java b/hadoop-hdds/config/src/main/java/org/apache/hadoop/hdds/conf/ConfigurationSource.java index a213482a5c89..ff659ddba35d 100644 --- a/hadoop-hdds/config/src/main/java/org/apache/hadoop/hdds/conf/ConfigurationSource.java +++ b/hadoop-hdds/config/src/main/java/org/apache/hadoop/hdds/conf/ConfigurationSource.java @@ -151,6 +151,7 @@ default Map getPropsMatchPrefix(String keyPrefix) { default boolean isConfigured(String key) { return get(key) != null; } + /** * Create a Configuration object and inject the required configuration values. * diff --git a/hadoop-hdds/config/src/test/java/org/apache/hadoop/hdds/conf/TestConfigurationSource.java b/hadoop-hdds/config/src/test/java/org/apache/hadoop/hdds/conf/TestConfigurationSource.java index de0718e9f603..f38658af2012 100644 --- a/hadoop-hdds/config/src/test/java/org/apache/hadoop/hdds/conf/TestConfigurationSource.java +++ b/hadoop-hdds/config/src/test/java/org/apache/hadoop/hdds/conf/TestConfigurationSource.java @@ -42,6 +42,7 @@ void getPropsMatchPrefix() { assertEquals(ImmutableMap.of("somePrefix.key", "value"), c.getPropsMatchPrefix("somePrefix.")); } + @Test void reconfigurableProperties() { String prefix = "ozone.test.config"; diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/hdds/scm/VersionInfo.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/hdds/scm/VersionInfo.java index afce723d1483..ea0261dc7f0f 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/hdds/scm/VersionInfo.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/hdds/scm/VersionInfo.java @@ -26,7 +26,6 @@ public final class VersionInfo { private static final VersionInfo[] VERSION_INFOS = {new VersionInfo("First version of SCM", 1)}; - public static final String DESCRIPTION_KEY = "Description"; private final String description; private final int version; diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/HddsPolicyProvider.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/HddsPolicyProvider.java index e4893800b4d8..9944f52e9af7 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/HddsPolicyProvider.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/HddsPolicyProvider.java @@ -55,7 +55,6 @@ public static HddsPolicyProvider getInstance() { return SUPPLIER.get(); } - @Override public Service[] getServices() { return DN_SERVICES.toArray(new Service[0]); diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/helpers/ContainerMetrics.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/helpers/ContainerMetrics.java index d37a8626cfd0..fcf3edb631ac 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/helpers/ContainerMetrics.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/helpers/ContainerMetrics.java @@ -57,7 +57,6 @@ public class ContainerMetrics implements Closeable { @Metric private MutableCounterLong numReadStateMachine; @Metric private MutableCounterLong bytesReadStateMachine; - private final EnumMap numOpsArray; private final EnumMap opsBytesArray; private final EnumMap opsLatency; @@ -129,6 +128,7 @@ public void incContainerBytesStats(ContainerProtos.Type type, long bytes) { public void incContainerDeleteFailedBlockCountNotZero() { containerDeleteFailedBlockCountNotZero.incr(); } + public void incContainerDeleteFailedNonEmpty() { containerDeleteFailedNonEmpty.incr(); } diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/impl/ContainerData.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/impl/ContainerData.java index 47ca8f1cc0f1..d431b494d783 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/impl/ContainerData.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/impl/ContainerData.java @@ -186,7 +186,6 @@ public ContainerType getContainerType() { return containerType; } - /** * Returns the state of the container. * @return ContainerLifeCycleState diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/impl/HddsDispatcher.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/impl/HddsDispatcher.java index 0ab5094ac869..c4e03e453349 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/impl/HddsDispatcher.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/impl/HddsDispatcher.java @@ -458,6 +458,7 @@ private void updateBCSID(Container container, container2BCSIDMap.computeIfPresent(containerId, (u, v) -> v = bcsID); } } + /** * Create a container using the input container request. * @param containerRequest - the container request which requires container diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/ReportManager.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/ReportManager.java index 11ddde998de9..05588a8cfd78 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/ReportManager.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/ReportManager.java @@ -104,7 +104,6 @@ public static final class Builder { private ReportPublisherFactory publisherFactory; private String threadNamePrefix = ""; - private Builder(ConfigurationSource conf) { this.reportPublishers = new ArrayList<>(); this.publisherFactory = new ReportPublisherFactory(conf); diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/DatanodeStateMachine.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/DatanodeStateMachine.java index e2860f0fd80a..c9412c5be3d0 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/DatanodeStateMachine.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/DatanodeStateMachine.java @@ -132,6 +132,7 @@ public class DatanodeStateMachine implements Closeable { private final DatanodeQueueMetrics queueMetrics; private final ReconfigurationHandler reconfigurationHandler; + /** * Constructs a datanode state machine. * @param datanodeDetails - DatanodeDetails used to identify a datanode @@ -309,7 +310,6 @@ public DatanodeDetails getDatanodeDetails() { return datanodeDetails; } - /** * Returns the Connection manager for this state machine. * @@ -734,6 +734,7 @@ public StatusAndMessages queryUpgradeStatus() return upgradeFinalizer.reportStatus(datanodeDetails.getUuidString(), true); } + public UpgradeFinalizer getUpgradeFinalizer() { return upgradeFinalizer; } diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/EndpointStateMachine.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/EndpointStateMachine.java index eec64a04f1b1..94bc0549e66a 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/EndpointStateMachine.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/EndpointStateMachine.java @@ -265,7 +265,6 @@ public void setPassive(boolean passive) { isPassive = passive; } - /** * States that an Endpoint can be in. *

diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/SCMConnectionManager.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/SCMConnectionManager.java index d2497d458186..01da280300a3 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/SCMConnectionManager.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/SCMConnectionManager.java @@ -78,7 +78,6 @@ public SCMConnectionManager(ConfigurationSource conf) { this); } - /** * Returns Config. * @@ -97,7 +96,6 @@ public int getRpcTimeout() { return rpcTimeout; } - /** * Takes a read lock. */ diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/DeleteBlocksCommandHandler.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/DeleteBlocksCommandHandler.java index fc8118ee520a..eb2f21a0be14 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/DeleteBlocksCommandHandler.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/DeleteBlocksCommandHandler.java @@ -188,15 +188,19 @@ public DeleteCmdInfo(DeleteBlocksCommand command, OzoneContainer container, this.container = container; this.connectionManager = connectionManager; } + public DeleteBlocksCommand getCmd() { return this.cmd; } + public StateContext getContext() { return this.context; } + public OzoneContainer getContainer() { return this.container; } + public SCMConnectionManager getConnectionManager() { return this.connectionManager; } @@ -208,6 +212,7 @@ public SCMConnectionManager getConnectionManager() { public static final class DeleteBlockTransactionExecutionResult { private final DeleteBlockTransactionResult result; private final boolean lockAcquisitionFailed; + public DeleteBlockTransactionExecutionResult( DeleteBlockTransactionResult result, boolean lockAcquisitionFailed) { this.result = result; diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/DeleteContainerCommandHandler.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/DeleteContainerCommandHandler.java index ae036a1c8f82..d63bb359f732 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/DeleteContainerCommandHandler.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/DeleteContainerCommandHandler.java @@ -75,6 +75,7 @@ protected DeleteContainerCommandHandler(Clock clock, DeleteContainerCommandHandler.class.getSimpleName()); this.opsLatencyMs = registry.newRate(SCMCommandProto.Type.deleteContainerCommand + "Ms"); } + @Override public void handle(final SCMCommand command, final OzoneContainer ozoneContainer, diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/XceiverServerSpi.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/XceiverServerSpi.java index 84813d1ef189..687845b80337 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/XceiverServerSpi.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/XceiverServerSpi.java @@ -57,7 +57,6 @@ void submitRequest(ContainerCommandRequestProto request, */ boolean isExist(HddsProtos.PipelineID pipelineId); - /** * Join a new pipeline. */ diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/CSMMetrics.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/CSMMetrics.java index 54426d40fbf9..6248e09fb90c 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/CSMMetrics.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/CSMMetrics.java @@ -223,7 +223,6 @@ public void recordWriteStateMachineCompletionNs(long latencyNanos) { writeStateMachineDataNs.add(latencyNanos); } - public void recordWriteStateMachineQueueingLatencyNs(long latencyNanos) { writeStateMachineQueueingLatencyNs.add(latencyNanos); } @@ -243,6 +242,7 @@ public void incNumDataCacheMiss() { public void incNumDataCacheHit() { numDataCacheHit.incr(); } + public void incNumEvictedCacheCount() { numEvictedCacheCount.incr(); } diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java index e38b485641ef..602506b20c38 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java @@ -377,6 +377,7 @@ public void buildMissingContainerSet(File snapshotFile) throws IOException { dispatcher.buildMissingContainerSetAndValidate(container2BCSIDMap); } } + /** * As a part of taking snapshot with Ratis StateMachine, it will persist * the existing container set in the snapshotFile. diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/XceiverServerRatis.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/XceiverServerRatis.java index 88cb0c78fcde..78962f6dfbd1 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/XceiverServerRatis.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/XceiverServerRatis.java @@ -879,6 +879,7 @@ void handleApplyTransactionFailure(RaftGroupId groupId, triggerPipelineClose(groupId, msg, ClosePipelineInfo.Reason.STATEMACHINE_TRANSACTION_FAILED); } + /** * The fact that the snapshot contents cannot be used to actually catch up * the follower, it is the reason to initiate close pipeline and diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/ContainerCache.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/ContainerCache.java index 41bda6dcbdf6..bf4eba91d131 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/ContainerCache.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/ContainerCache.java @@ -44,6 +44,7 @@ public final class ContainerCache extends LRUMap { private static final float LOAD_FACTOR = 0.75f; private final Striped rocksDBLock; private static ContainerCacheMetrics metrics; + /** * Constructs a cache that holds DBHandle references. */ diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/DatanodeStoreCache.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/DatanodeStoreCache.java index 7e0c8c415b37..3b8d78a494ce 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/DatanodeStoreCache.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/DatanodeStoreCache.java @@ -65,7 +65,6 @@ public static synchronized void setMiniClusterMode(boolean isMiniCluster) { getInstance().miniClusterMode = isMiniCluster; } - public void addDB(String containerDBPath, RawDB db) { datanodeStoreMap.putIfAbsent(containerDBPath, db); LOG.info("Added db {} to cache", containerDBPath); diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/DiskCheckUtil.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/DiskCheckUtil.java index 113062e0b9fa..73e69eddc15b 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/DiskCheckUtil.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/utils/DiskCheckUtil.java @@ -81,9 +81,11 @@ public interface DiskChecks { default boolean checkExistence(File storageDir) { return true; } + default boolean checkPermissions(File storageDir) { return true; } + default boolean checkReadWrite(File storageDir, File testFileDir, int numBytesToWrite) { return true; diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/KeyValueContainer.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/KeyValueContainer.java index fd0604906b8f..09cadd5d13fc 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/KeyValueContainer.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/KeyValueContainer.java @@ -334,7 +334,6 @@ private void updateContainerFile(File containerFile) writeToContainerFile(containerFile, false); } - @Override public void delete() throws StorageContainerException { try { diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/BlockUtils.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/BlockUtils.java index 730689539f94..7dc02dffbb83 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/BlockUtils.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/BlockUtils.java @@ -143,6 +143,7 @@ public static DBHandle getDB(KeyValueContainerData containerData, throw new StorageContainerException(message, UNABLE_TO_READ_METADATA_DB); } } + /** * Remove a DB handler from cache. * diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/ChunkUtils.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/ChunkUtils.java index e176252a8076..68bf7a2ff7a0 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/ChunkUtils.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/ChunkUtils.java @@ -365,7 +365,6 @@ public static boolean validateChunkForOverwrite(File chunkFile, return false; } - /** * Validates chunk data and returns a boolean value that indicates if the * chunk data should be overwritten. diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/KeyValueContainerLocationUtil.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/KeyValueContainerLocationUtil.java index d2882780a794..70026f5f8e0c 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/KeyValueContainerLocationUtil.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/KeyValueContainerLocationUtil.java @@ -32,6 +32,7 @@ public final class KeyValueContainerLocationUtil { private KeyValueContainerLocationUtil() { } + /** * Returns Container Metadata Location. * @param hddsVolumeDir base dir of the hdds volume where scm directories diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingTask.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingTask.java index 38fe872f30e4..156657e462e4 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingTask.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingTask.java @@ -113,7 +113,6 @@ public int getSize() { } } - @Override public BackgroundTaskResult call() throws Exception { ContainerBackgroundTaskResult result = diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/DatanodeSchemaThreeDBDefinition.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/DatanodeSchemaThreeDBDefinition.java index d81c36022f74..df0f0f9e0dcd 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/DatanodeSchemaThreeDBDefinition.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/DatanodeSchemaThreeDBDefinition.java @@ -176,6 +176,7 @@ public static byte[] getContainerKeyPrefixBytes(long containerID) { public static String getKeyWithoutPrefix(String keyWithPrefix) { return keyWithPrefix.substring(keyWithPrefix.indexOf(separator) + 1); } + /** * * @param key rocksDB original key diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/DatanodeStoreWithIncrementalChunkList.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/DatanodeStoreWithIncrementalChunkList.java index 570e3b077fc6..6115c0cd36b6 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/DatanodeStoreWithIncrementalChunkList.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/metadata/DatanodeStoreWithIncrementalChunkList.java @@ -53,7 +53,6 @@ public DatanodeStoreWithIncrementalChunkList(ConfigurationSource config, super(config, dbDef, openReadOnly); } - @Override public BlockData getCompleteBlockData(BlockData blockData, BlockID blockID, String blockKey) throws IOException { diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/replication/AbstractReplicationTask.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/replication/AbstractReplicationTask.java index d080f92c99af..2ba2b676a913 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/replication/AbstractReplicationTask.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/replication/AbstractReplicationTask.java @@ -65,6 +65,7 @@ protected AbstractReplicationTask(long containerID, public long getContainerId() { return containerId; } + public Status getStatus() { return status; } @@ -138,7 +139,7 @@ protected void setShouldOnlyRunOnInServiceDatanodes( */ protected Object getCommandForDebug() { return ""; - }; + } @Override public String toString() { diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/replication/GrpcContainerUploader.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/replication/GrpcContainerUploader.java index 8b51f88e5fc1..bf381e3715be 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/replication/GrpcContainerUploader.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/replication/GrpcContainerUploader.java @@ -156,6 +156,7 @@ public WrappedRequestStreamObserver( this.observer = observer; this.responseObserver = responseObserver; } + @Override public boolean isReady() { if (responseObserver.isError()) { diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/stream/DirstreamClientHandler.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/stream/DirstreamClientHandler.java index d8f8b416bad2..783e885b6498 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/stream/DirstreamClientHandler.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/stream/DirstreamClientHandler.java @@ -118,6 +118,7 @@ public void doRead(ChannelHandlerContext ctx, ByteBuf buffer) public boolean isAtTheEnd() { return getCurrentFileName().equals(END_MARKER); } + @Override public void channelUnregistered(ChannelHandlerContext ctx) { try { diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/stream/StreamingClient.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/stream/StreamingClient.java index 85af66341adc..55233e749bcc 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/stream/StreamingClient.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/stream/StreamingClient.java @@ -103,7 +103,6 @@ public void stream(String id, long timeout, TimeUnit unit) { } } - @Override public void close() { group.shutdownGracefully(); diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/StorageContainerNodeProtocol.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/StorageContainerNodeProtocol.java index ffe5a40fb498..210c87e99bf3 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/StorageContainerNodeProtocol.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/StorageContainerNodeProtocol.java @@ -71,7 +71,7 @@ RegisteredCommand register(DatanodeDetails datanodeDetails, */ default List> processHeartbeat(DatanodeDetails datanodeDetails) { return processHeartbeat(datanodeDetails, null); - }; + } /** * Send heartbeat to indicate the datanode is alive and doing well. diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/CommandStatus.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/CommandStatus.java index 344b6c4150fd..e482acbbfbb4 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/CommandStatus.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/CommandStatus.java @@ -93,6 +93,7 @@ public CommandStatus getFromProtoBuf( .setMsg(cmdStatusProto.getMsg()) .build(); } + /** * Returns a CommandStatus from the protocol buffers. * diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/DeleteBlocksCommand.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/DeleteBlocksCommand.java index 93b444d555b2..e955cabdfa47 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/DeleteBlocksCommand.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/DeleteBlocksCommand.java @@ -30,7 +30,6 @@ public class DeleteBlocksCommand extends private List blocksTobeDeleted; - public DeleteBlocksCommand(List blocks) { super(); this.blocksTobeDeleted = blocks; diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/ReconstructECContainersCommand.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/ReconstructECContainersCommand.java index 40cb03821aff..93bbfd8202a4 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/ReconstructECContainersCommand.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/ReconstructECContainersCommand.java @@ -145,6 +145,7 @@ public String toString() { Arrays.toString(missingContainerIndexes.toByteArray())); return sb.toString(); } + /** * To store the datanode details with replica index. */ diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/SCMCommand.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/SCMCommand.java index fba9b95057ca..6c3406a14357 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/SCMCommand.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/protocol/commands/SCMCommand.java @@ -46,6 +46,7 @@ public abstract class SCMCommand implements SCMCommand(long id) { this.id = id; } + /** * Returns the type of this command. * @return Type diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/SCMTestUtils.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/SCMTestUtils.java index 8a520da3cdbb..8fcd3f157883 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/SCMTestUtils.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/SCMTestUtils.java @@ -83,7 +83,6 @@ private static RPC.Server startRpcServer(Configuration conf, return rpcServer; } - /** * Start Datanode RPC server. */ diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/ScmTestMock.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/ScmTestMock.java index 708cd3240d55..28a0d6cf6f48 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/ScmTestMock.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/ScmTestMock.java @@ -73,6 +73,7 @@ public ScmTestMock(String clusterId) { this.clusterId = clusterId; this.scmId = UUID.randomUUID().toString(); } + /** * Returns the number of heartbeats made to this class. * @@ -291,7 +292,6 @@ public void updateContainerReport( } } - /** * Return the number of StorageReports of a datanode. * @param datanodeDetails diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/TestContainerLayoutVersion.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/TestContainerLayoutVersion.java index 52bc1f956818..ddf20b51fa54 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/TestContainerLayoutVersion.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/TestContainerLayoutVersion.java @@ -39,7 +39,6 @@ public void testV1() { assertEquals(1, FILE_PER_CHUNK.getVersion()); } - @Test public void testV2() { assertEquals(2, FILE_PER_BLOCK.getVersion()); diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/helpers/TestBlockData.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/helpers/TestBlockData.java index 81c9e48c4b20..b585fb1a2b4f 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/helpers/TestBlockData.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/helpers/TestBlockData.java @@ -84,7 +84,6 @@ static long assertAddChunk(List expected, return info.getLen(); } - static void removeChunk(List expected, BlockData computed) { final int i = ThreadLocalRandom.current().nextInt(expected.size()); diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/impl/TestContainerDataYaml.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/impl/TestContainerDataYaml.java index 33a838c99801..ce27d7a0140c 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/impl/TestContainerDataYaml.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/impl/TestContainerDataYaml.java @@ -200,7 +200,6 @@ public void testIncorrectContainerFile(ContainerLayoutVersion layout) { assertThat(exception).hasMessageContaining("No enum constant"); } - @ContainerLayoutTestInfo.ContainerTest void testCheckBackWardCompatibilityOfContainerFile( ContainerLayoutVersion layout) throws Exception { diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/TestDeleteContainerCommandHandler.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/TestDeleteContainerCommandHandler.java index f9c346889345..e4f35691544f 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/TestDeleteContainerCommandHandler.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/statemachine/commandhandler/TestDeleteContainerCommandHandler.java @@ -200,6 +200,7 @@ private static DeleteContainerCommandHandler createSubjectWithPoolSize( static class ThreadPoolWithLockExecutor extends ThreadPoolExecutor { private CountDownLatch countDownLatch; + ThreadPoolWithLockExecutor(ThreadFactory threadFactory, CountDownLatch latch) { super(1, 1, 0, TimeUnit.MILLISECONDS, new LinkedBlockingQueue(), threadFactory); diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/TestContainerStateMachine.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/TestContainerStateMachine.java index f021ca3ffd6d..d0b2dc5358e3 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/TestContainerStateMachine.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/TestContainerStateMachine.java @@ -108,13 +108,11 @@ public void setup() throws IOException { RaftGroupId.randomId(), dispatcher, controller, executor, ratisServer, conf, "containerOp"); } - @AfterEach public void teardown() { stateMachine.close(); } - @AfterAll public void shutdown() { executor.forEach(ThreadPoolExecutor::shutdown); diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestReservedVolumeSpace.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestReservedVolumeSpace.java index 5638f5841a11..ef5279a16880 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestReservedVolumeSpace.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestReservedVolumeSpace.java @@ -210,7 +210,6 @@ public void testMinFreeSpaceCalculator() throws Exception { assertEquals(10, conf.getObject(DatanodeConfiguration.class).getMinFreeSpace(capacity)); } - private long getExpectedDefaultReserved(HddsVolume volume) { long totalCapacity = volume.getVolumeUsage().get().realUsage().getCapacity(); return (long) Math.ceil(totalCapacity * HDDS_DATANODE_DIR_DU_RESERVED_PERCENT_DEFAULT); diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestStorageVolumeHealthChecks.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestStorageVolumeHealthChecks.java index a3e12a96ee39..25c6f05585a1 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestStorageVolumeHealthChecks.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/volume/TestStorageVolumeHealthChecks.java @@ -116,7 +116,6 @@ public void testVolumeFullHealth(StorageVolume.Builder builder) throws Except verifyFullVolumeHealthWithDiskReadWriteStatus(builder, true, false); } - public void verifyFullVolumeHealthWithDiskReadWriteStatus(StorageVolume.Builder builder, boolean... checkResult) throws Exception { diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueBlockIterator.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueBlockIterator.java index 577f65c2e583..034bc9640656 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueBlockIterator.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueBlockIterator.java @@ -122,7 +122,6 @@ public void setup() throws Exception { db = BlockUtils.getDB(containerData, conf); } - @AfterEach public void tearDown() throws Exception { db.close(); diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainer.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainer.java index 407a6d8261f0..c700b235faf9 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainer.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainer.java @@ -492,7 +492,6 @@ private void populateWithoutBlock(KeyValueContainer container, container.update(metadata, true); } - /** * Set container state to CLOSED. */ @@ -626,7 +625,6 @@ public void testReportOfUnhealthyContainer( assertNotNull(keyValueContainer.getContainerReport()); } - @ContainerTestVersionInfo.ContainerTest public void testUpdateContainer(ContainerTestVersionInfo versionInfo) throws Exception { @@ -670,7 +668,6 @@ public void testUpdateContainerUnsupportedRequest( assertThat(exception).hasMessageContaining(keyValueContainerData.toString()); } - @ContainerTestVersionInfo.ContainerTest public void testContainerRocksDB(ContainerTestVersionInfo versionInfo) throws Exception { diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainerIntegrityChecks.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainerIntegrityChecks.java index e29e2031e259..08c44372b928 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainerIntegrityChecks.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainerIntegrityChecks.java @@ -109,7 +109,6 @@ protected OzoneConfiguration getConf() { return conf; } - /** * Creates a container with normal and deleted blocks. * First it will insert normal blocks, and then it will insert diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainerMetadataInspector.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainerMetadataInspector.java index d782db01e3d5..09c92cda6a53 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainerMetadataInspector.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainerMetadataInspector.java @@ -412,7 +412,6 @@ private void checkJsonErrorsReport(JsonNode jsonReport, assertTrue(matchFound); } - public void setDBBlockAndByteCounts(KeyValueContainerData containerData, long blockCount, long byteCount) throws Exception { setDB(containerData, blockCount, byteCount, diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandler.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandler.java index 6b0e84295cad..76e696e91dd5 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandler.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandler.java @@ -432,7 +432,6 @@ public void testDeleteContainer() throws IOException { } } - @Test public void testDeleteContainerTimeout() throws IOException { final String testDir = tempDir.toString(); diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandlerWithUnhealthyContainer.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandlerWithUnhealthyContainer.java index 1b7d388edc9c..34a1d68a1f6a 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandlerWithUnhealthyContainer.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueHandlerWithUnhealthyContainer.java @@ -106,7 +106,6 @@ private static Stream getAllClientVersions() { .mapToObj(rid -> Arguments.of(client, rid))); } - @ParameterizedTest @MethodSource("getAllClientVersions") public void testGetBlockWithReplicaIndexMismatch(ClientVersion clientVersion, int replicaIndex) { diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/impl/TestKeyValueStreamDataChannel.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/impl/TestKeyValueStreamDataChannel.java index 093e2007a257..e4c7decefd16 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/impl/TestKeyValueStreamDataChannel.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/impl/TestKeyValueStreamDataChannel.java @@ -82,6 +82,7 @@ public class TestKeyValueStreamDataChannel { .setVersion(ClientVersion.CURRENT.toProtoValue()) .build(); static final int PUT_BLOCK_PROTO_SIZE = PUT_BLOCK_PROTO.toByteString().size(); + static { LOG.info("PUT_BLOCK_PROTO_SIZE = {}", PUT_BLOCK_PROTO_SIZE); } diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestBackgroundContainerDataScanner.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestBackgroundContainerDataScanner.java index 3da90a294aea..93f8a957e781 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestBackgroundContainerDataScanner.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestBackgroundContainerDataScanner.java @@ -217,7 +217,6 @@ public void testWithVolumeFailure() throws Exception { verify(openCorruptMetadata, never()).scanData(any(), any()); } - @Test @Override public void testShutdownDuringScan() throws Exception { diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestContainerReader.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestContainerReader.java index 4688c46214fa..dbe6df0c7b28 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestContainerReader.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestContainerReader.java @@ -84,7 +84,6 @@ public class TestContainerReader { private ContainerSet containerSet; private OzoneConfiguration conf; - private RoundRobinVolumeChoosingPolicy volumeChoosingPolicy; private UUID datanodeId; private String clusterId = UUID.randomUUID().toString(); diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestContainerScannersAbstract.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestContainerScannersAbstract.java index d65e747b472a..90be5b24c788 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestContainerScannersAbstract.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/ozoneimpl/TestContainerScannersAbstract.java @@ -81,7 +81,6 @@ public abstract class TestContainerScannersAbstract { protected ContainerScannerConfiguration conf; protected ContainerController controller; - private Collection> containers; public void setup() { diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/replication/TestPushReplicator.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/replication/TestPushReplicator.java index fce32b874cea..6b6a511ce0db 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/replication/TestPushReplicator.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/replication/TestPushReplicator.java @@ -81,7 +81,6 @@ void uploadCompletesNormally(CopyContainerCompression compression) output.assertClosedExactlyOnce(); } - @Test void uploadFailsWithException() throws IOException { // GIVEN diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/replication/TestReplicationSupervisor.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/replication/TestReplicationSupervisor.java index 0884d7301089..9acb73486a06 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/replication/TestReplicationSupervisor.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/replication/TestReplicationSupervisor.java @@ -450,7 +450,6 @@ public void testReplicationImportReserveSpace(ContainerLayoutVersion layout) } - private File containerTarFile( long containerId, ContainerData containerData) throws IOException { File yamlFile = new File(tempDir, "container.yaml"); diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/stream/TestDirstreamClientHandler.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/stream/TestDirstreamClientHandler.java index ed250fd1a547..7b996616fa5e 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/stream/TestDirstreamClientHandler.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/stream/TestDirstreamClientHandler.java @@ -52,7 +52,6 @@ public void oneFileStream() throws IOException { } - @Test public void splitAtHeader() throws IOException { @@ -85,7 +84,6 @@ public void splitInHeader() throws IOException { } - @Test public void splitSecondHeader() throws IOException { @@ -101,7 +99,6 @@ public void splitSecondHeader() throws IOException { assertTrue(handler.isAtTheEnd()); } - @Test public void splitContent() throws IOException { diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/stream/TestStreamingServer.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/stream/TestStreamingServer.java index 2c95fccac6bb..47e280e97ae4 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/stream/TestStreamingServer.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/stream/TestStreamingServer.java @@ -108,6 +108,7 @@ public void ssl() throws Exception { assertArrayEquals(CONTENT, targetContent); } + @Test public void failedStream() throws Exception { Files.createDirectories(sourceDir.resolve(SUBDIR)); diff --git a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/upgrade/UpgradeTestHelper.java b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/upgrade/UpgradeTestHelper.java index e45a5243a2de..2b6112bd0d91 100644 --- a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/upgrade/UpgradeTestHelper.java +++ b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/upgrade/UpgradeTestHelper.java @@ -179,7 +179,6 @@ public static File addDbVolume(OzoneConfiguration conf, Path tempFolder) throws return vol; } - public static void dispatchRequest( ContainerDispatcher dispatcher, ContainerProtos.ContainerCommandRequestProto request) { diff --git a/hadoop-hdds/dev-support/checkstyle/checkstyle.xml b/hadoop-hdds/dev-support/checkstyle/checkstyle.xml index 47010b89e9aa..1382bed67f08 100644 --- a/hadoop-hdds/dev-support/checkstyle/checkstyle.xml +++ b/hadoop-hdds/dev-support/checkstyle/checkstyle.xml @@ -134,6 +134,14 @@ + + + + + diff --git a/hadoop-hdds/erasurecode/src/main/java/org/apache/ozone/erasurecode/rawcoder/CoderUtil.java b/hadoop-hdds/erasurecode/src/main/java/org/apache/ozone/erasurecode/rawcoder/CoderUtil.java index fa214c64a305..92737377630a 100644 --- a/hadoop-hdds/erasurecode/src/main/java/org/apache/ozone/erasurecode/rawcoder/CoderUtil.java +++ b/hadoop-hdds/erasurecode/src/main/java/org/apache/ozone/erasurecode/rawcoder/CoderUtil.java @@ -96,7 +96,6 @@ static void resetOutputBuffers(byte[][] buffers, int[] offsets, } } - /** * Convert an array of this chunks to an array of ByteBuffers. * @@ -136,8 +135,6 @@ static ByteBuffer cloneAsDirectByteBuffer(byte[] input, int offset, int len) { return directBuffer; } - - /** * Find the valid input from all the inputs. * @param inputs input buffers to look for valid input diff --git a/hadoop-hdds/erasurecode/src/main/java/org/apache/ozone/erasurecode/rawcoder/XORRawErasureCoderFactory.java b/hadoop-hdds/erasurecode/src/main/java/org/apache/ozone/erasurecode/rawcoder/XORRawErasureCoderFactory.java index 01dd11cc4343..820b5713651f 100644 --- a/hadoop-hdds/erasurecode/src/main/java/org/apache/ozone/erasurecode/rawcoder/XORRawErasureCoderFactory.java +++ b/hadoop-hdds/erasurecode/src/main/java/org/apache/ozone/erasurecode/rawcoder/XORRawErasureCoderFactory.java @@ -28,7 +28,6 @@ public class XORRawErasureCoderFactory implements RawErasureCoderFactory { public static final String CODER_NAME = "xor_java"; - @Override public RawErasureEncoder createEncoder( ECReplicationConfig replicationConfig) { diff --git a/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/TestCoderBase.java b/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/TestCoderBase.java index 02711b0d89cf..3f5d66bc1468 100644 --- a/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/TestCoderBase.java +++ b/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/TestCoderBase.java @@ -24,6 +24,7 @@ import org.apache.commons.lang3.RandomUtils; import org.apache.hadoop.hdds.conf.ConfigurationSource; import org.apache.hadoop.hdds.conf.OzoneConfiguration; + /** * Test base of common utilities for tests not only raw coders but also block * coders. diff --git a/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/rawcoder/TestXORRawCoder.java b/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/rawcoder/TestXORRawCoder.java index ddcc42507620..f882e32536cc 100644 --- a/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/rawcoder/TestXORRawCoder.java +++ b/hadoop-hdds/erasurecode/src/test/java/org/apache/ozone/erasurecode/rawcoder/TestXORRawCoder.java @@ -25,6 +25,4 @@ public class TestXORRawCoder extends TestXORRawCoderBase { public TestXORRawCoder() { super(XORRawErasureCoderFactory.class, XORRawErasureCoderFactory.class); } - - } diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocolPB/SCMSecurityProtocolClientSideTranslatorPB.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocolPB/SCMSecurityProtocolClientSideTranslatorPB.java index 775aecd7ee05..a9c71162618b 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocolPB/SCMSecurityProtocolClientSideTranslatorPB.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/protocolPB/SCMSecurityProtocolClientSideTranslatorPB.java @@ -114,6 +114,7 @@ private SCMSecurityResponse handleError(SCMSecurityResponse resp) } return resp; } + /** * Closes this stream and releases any system resources associated * with it. If the stream is already closed then invoking this diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/container/common/helpers/DeletedBlocksTransactionInfoWrapper.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/container/common/helpers/DeletedBlocksTransactionInfoWrapper.java index 238c067d602e..e728c1d7e470 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/container/common/helpers/DeletedBlocksTransactionInfoWrapper.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/container/common/helpers/DeletedBlocksTransactionInfoWrapper.java @@ -86,7 +86,6 @@ public static DeletedBlocksTransaction toTxn( .build(); } - public long getTxID() { return txID; } diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/metadata/Replicate.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/metadata/Replicate.java index 557673cd530d..c5297ce11e64 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/metadata/Replicate.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/metadata/Replicate.java @@ -46,5 +46,6 @@ * For CLIENT, the call will be submitted to Ratis Server via Ratis Client. */ enum InvocationType { DIRECT, CLIENT } + InvocationType invocationType() default InvocationType.DIRECT; } diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/protocolPB/ScmBlockLocationProtocolClientSideTranslatorPB.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/protocolPB/ScmBlockLocationProtocolClientSideTranslatorPB.java index 13a8846f0423..a7324d481c48 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/protocolPB/ScmBlockLocationProtocolClientSideTranslatorPB.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/protocolPB/ScmBlockLocationProtocolClientSideTranslatorPB.java @@ -302,6 +302,7 @@ public boolean addSCM(AddSCMRequest request) throws IOException { resp = wrappedResponse.getAddScmResponse(); return resp.getSuccess(); } + /** * Sort the datanodes based on distance from client. * @return list of datanodes; diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/proxy/SCMFailoverProxyProviderBase.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/proxy/SCMFailoverProxyProviderBase.java index 504504e597cb..b616f0f621a5 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/proxy/SCMFailoverProxyProviderBase.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/scm/proxy/SCMFailoverProxyProviderBase.java @@ -166,7 +166,6 @@ protected synchronized void loadConfigs() { } } - @VisibleForTesting public synchronized void changeCurrentProxy(String nodeId) { currentProxyIndex = scmNodeIds.indexOf(nodeId); diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/OzoneSecretManager.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/OzoneSecretManager.java index 96a89509a1a8..3ba16ce99829 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/OzoneSecretManager.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/OzoneSecretManager.java @@ -86,7 +86,6 @@ public OzoneSecretManager(SecurityConfig secureConf, long tokenMaxLifetime, this.currentKey = new AtomicReference<>(); } - /** * Compute HMAC of the identifier using the private key and return the output * as password. @@ -139,6 +138,7 @@ public byte[] createPassword(T identifier) { */ public abstract long renewToken(Token token, String renewer) throws IOException; + /** * Cancel a token by removing it from store and cache. * diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/symmetric/DefaultSecretKeyClient.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/symmetric/DefaultSecretKeyClient.java index 05712ae80a6b..c2ac712c3028 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/symmetric/DefaultSecretKeyClient.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/symmetric/DefaultSecretKeyClient.java @@ -31,14 +31,12 @@ public class DefaultSecretKeyClient implements SecretKeyClient { private final SecretKeySignerClient signerClientDelegate; private final SecretKeyVerifierClient verifierClientDelegate; - DefaultSecretKeyClient(SecretKeySignerClient signerClientDelegate, SecretKeyVerifierClient verifierClientDelegate) { this.signerClientDelegate = signerClientDelegate; this.verifierClientDelegate = verifierClientDelegate; } - @Override public ManagedSecretKey getCurrentSecretKey() { return signerClientDelegate.getCurrentSecretKey(); diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/authority/CertificateApprover.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/authority/CertificateApprover.java index c78b2eedbfce..55ba2b81555a 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/authority/CertificateApprover.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/authority/CertificateApprover.java @@ -66,7 +66,6 @@ X509Certificate sign( String certSerialId) throws IOException, CertificateException; - /** * Approval Types for a certificate request. */ diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/client/DefaultCertificateClient.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/client/DefaultCertificateClient.java index a204fc9a5ee8..a4ae1a794134 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/client/DefaultCertificateClient.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/client/DefaultCertificateClient.java @@ -788,7 +788,6 @@ protected InitResponse handleCase(InitCase init) } } - /** * Recover the state if needed. * */ diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/client/SCMCertificateClient.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/client/SCMCertificateClient.java index feef456d152c..64c16ba28237 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/client/SCMCertificateClient.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/security/x509/certificate/client/SCMCertificateClient.java @@ -337,7 +337,6 @@ private void getRootCASignedSCMCert() { } } - /** * For primary SCM get sub-ca signed certificate and root CA certificate by * root CA certificate server and store it using certificate client. diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/ServerUtils.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/ServerUtils.java index 30c31eee6abf..9dffa3e1f18a 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/ServerUtils.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/ServerUtils.java @@ -44,7 +44,6 @@ public final class ServerUtils { private static final Logger LOG = LoggerFactory.getLogger( ServerUtils.class); - private ServerUtils() { } @@ -84,7 +83,6 @@ public static long sanitizeUserArgs(String key, long valueTocheck, return valueTocheck; } - /** * After starting an RPC server, updates configuration with the actual * listening address of that server. The listening address may be different @@ -103,7 +101,6 @@ public static InetSocketAddress updateRPCListenAddress( rpcServer.getListenerAddress()); } - /** * After starting an server, updates configuration with the actual * listening address of that server. The listening address may be different @@ -221,7 +218,6 @@ private static boolean isSymbolic(String permissionValue) { return permissionValue.matches(".*[rwx].*"); } - /** * Retrieves the permissions' configuration value for a given config key. * @@ -255,7 +251,6 @@ public static String getPermissions(String key, ConfigurationSource conf) { "Invalid configuration value for key: " + key); } - /** * Checks and creates Ozone Metadir Path if it does not exist. * @@ -312,6 +307,4 @@ public static String getDefaultRatisDirectory(ConfigurationSource conf) { File metaDirPath = ServerUtils.getOzoneMetaDirPath(conf); return (new File(metaDirPath, "ratis")).getPath(); } - - } diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/ServiceRuntimeInfo.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/ServiceRuntimeInfo.java index 8f8592e7cf1a..955443bdba88 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/ServiceRuntimeInfo.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/ServiceRuntimeInfo.java @@ -40,7 +40,7 @@ public interface ServiceRuntimeInfo { */ default String getNamespace() { return ""; - }; + } /** * Gets the version of Hadoop. diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/events/EventQueue.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/events/EventQueue.java index 1d1a36cffd72..02778d5e89f1 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/events/EventQueue.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/events/EventQueue.java @@ -309,7 +309,6 @@ public void setSilent(boolean silent) { isSilent = silent; } - @VisibleForTesting public Map> getExecutorAndHandler( Event event) { diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/events/EventWatcher.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/events/EventWatcher.java index 6f48181dccea..17d8ec3e942d 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/events/EventWatcher.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/events/EventWatcher.java @@ -155,7 +155,6 @@ private synchronized Void handleTimeout(EventPublisher publisher, return null; } - /** * Check if a specific payload is in-progress. */ diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/events/IEventInfo.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/events/IEventInfo.java index 89f4482a8219..f69a1b99926d 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/events/IEventInfo.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/events/IEventInfo.java @@ -22,6 +22,7 @@ */ public interface IEventInfo { long getCreateTime(); + default String getEventId() { return ""; } diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/http/BaseHttpServer.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/http/BaseHttpServer.java index 3144fd63ac58..26457d9d0c68 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/http/BaseHttpServer.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/http/BaseHttpServer.java @@ -244,7 +244,6 @@ public static HttpServer2.Builder newHttpServer2BuilderForOzone( return builder; } - /** * Add a servlet to BaseHttpServer. * @@ -368,7 +367,6 @@ public void updateConnectorAddress() { } } - public static HttpServer2.Builder loadSslConfToHttpServerBuilder( HttpServer2.Builder builder, ConfigurationSource sslConf) { return builder @@ -412,6 +410,7 @@ static String getPassword(ConfigurationSource conf, String alias) { } return password; } + /** * Load HTTPS-related configuration. */ diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/http/HttpServer2.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/http/HttpServer2.java index 06008dd4f102..aa7eb407ae50 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/http/HttpServer2.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/server/http/HttpServer2.java @@ -214,6 +214,7 @@ public final class HttpServer2 implements FilterContainer { private static final String X_FRAME_OPTIONS = "X-FRAME-OPTIONS"; private static final Pattern PATTERN_HTTP_HEADER_REGEX = Pattern.compile(HTTP_HEADER_REGEX); + /** * Class to construct instances of HTTP server with specific options. */ @@ -1509,7 +1510,6 @@ public static boolean userHasAdministratorAccess( return adminsAcl != null && adminsAcl.isUserAllowed(remoteUserUGI); } - /** * A very simple servlet to serve up a text representation of the current * stack traces. It both returns the stacks to the caller and logs them. diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/DecayRpcSchedulerUtil.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/DecayRpcSchedulerUtil.java index 63d742965d94..56e38913b7fe 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/DecayRpcSchedulerUtil.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/DecayRpcSchedulerUtil.java @@ -88,7 +88,6 @@ public static String checkMetricNameForUsername(String recordName, return null; } - /** * Create a username metrics tag. * @param username caller username diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/HddsServerUtil.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/HddsServerUtil.java index 07a8da2c90e8..b3e21db17f34 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/HddsServerUtil.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/HddsServerUtil.java @@ -222,7 +222,6 @@ public static InetSocketAddress getScmDataNodeBindAddress( ScmConfigKeys.OZONE_SCM_DATANODE_PORT_DEFAULT))); } - /** * Retrieve the socket address that should be used by DataNodes to connect * to Recon. @@ -268,7 +267,6 @@ public static long getScmHeartbeatInterval(ConfigurationSource conf) { HDDS_HEARTBEAT_INTERVAL_DEFAULT, TimeUnit.MILLISECONDS); } - /** * Heartbeat Interval - Defines the initial heartbeat frequency from a datanode to * SCM. diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/MetadataKeyFilters.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/MetadataKeyFilters.java index 187954e90074..551560a31f6e 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/MetadataKeyFilters.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/MetadataKeyFilters.java @@ -46,6 +46,7 @@ public static KeyPrefixFilter getUnprefixedKeyFilter() { return new MetadataKeyFilters.KeyPrefixFilter() .addFilter("#", true); } + /** * Interface for RocksDB key filters. */ diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/DBDefinition.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/DBDefinition.java index 7913f3b3e51d..01061969fb7a 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/DBDefinition.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/DBDefinition.java @@ -77,7 +77,6 @@ default List getColumnFamilyNames() { */ List> getColumnFamilies(String name); - /** * @return The unique column family for the given name. */ diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/DBStoreBuilder.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/DBStoreBuilder.java index c6f9e1c7cf1b..abb806bb1abf 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/DBStoreBuilder.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/DBStoreBuilder.java @@ -302,6 +302,7 @@ public DBStoreBuilder setEnableRocksDbMetrics(boolean enableRocksDbMetrics) { this.enableRocksDbMetrics = enableRocksDbMetrics; return this; } + /** * Set the {@link ManagedDBOptions} and default * {@link ManagedColumnFamilyOptions} based on {@code prof}. diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/RDBMetrics.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/RDBMetrics.java index d84ee57b5216..624d1bb761ca 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/RDBMetrics.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/RDBMetrics.java @@ -97,7 +97,6 @@ public void incNumDBKeyMayExistMisses() { numDBKeyMayExistMisses.incr(); } - @VisibleForTesting public long getNumDBKeyMayExistChecks() { return numDBKeyMayExistChecks.value(); diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/RDBStore.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/RDBStore.java index 0e7ee39db0a6..b2ecc2faed71 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/RDBStore.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/RDBStore.java @@ -209,7 +209,6 @@ public RocksDBCheckpointDiffer getRocksDBCheckpointDiffer() { return rocksDBCheckpointDiffer; } - /** * Returns the RocksDB's DBOptions. */ diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/RDBTable.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/RDBTable.java index 0f172027f0cf..f72b0084bdc5 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/RDBTable.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/RDBTable.java @@ -39,7 +39,6 @@ @InterfaceAudience.Private class RDBTable implements Table { - private static final Logger LOG = LoggerFactory.getLogger(RDBTable.class); @@ -93,7 +92,6 @@ public void putWithBatch(BatchOperation batch, byte[] key, byte[] value) } } - @Override public boolean isEmpty() throws IOException { try (TableIterator> keyIter = iterator()) { @@ -195,7 +193,6 @@ public void delete(ByteBuffer key) throws IOException { db.delete(family, key); } - @Override public void deleteRange(byte[] beginKey, byte[] endKey) throws IOException { db.deleteRange(family, beginKey, endKey); diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/Table.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/Table.java index 5a07558cd546..18f98a08ed63 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/Table.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/Table.java @@ -83,7 +83,6 @@ void putWithBatch(BatchOperation batch, KEY key, VALUE value) */ VALUE get(KEY key) throws IOException; - /** * Skip checking cache and get the value mapped to the given key in byte * array or returns null if the key is not found. @@ -96,7 +95,6 @@ default VALUE getSkipCache(KEY key) throws IOException { throw new NotImplementedException("getSkipCache is not implemented"); } - /** * Returns the value mapped to the given key in byte array or returns null * if the key is not found. @@ -112,7 +110,6 @@ default VALUE getReadCopy(KEY key) throws IOException { throw new NotImplementedException("getReadCopy is not implemented"); } - /** * Returns the value mapped to the given key in byte array or returns null * if the key is not found. @@ -419,7 +416,6 @@ public int hashCode() { }; } - /** A {@link TableIterator} to iterate {@link KeyValue}s. */ interface KeyValueIterator extends TableIterator> { diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/cache/FullTableCache.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/cache/FullTableCache.java index 524eca9fd9b1..2f60d0cf4d24 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/cache/FullTableCache.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/cache/FullTableCache.java @@ -62,7 +62,6 @@ public class FullTableCache implements TableCache { private final CacheStatsRecorder statsRecorder; - public FullTableCache(String threadNamePrefix) { // As for full table cache only we need elements to be inserted in sorted // manner, so that list will be easy. But look ups have log(N) time diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/cache/PartialTableCache.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/cache/PartialTableCache.java index 8f179fbd7aef..9b7240364947 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/cache/PartialTableCache.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/cache/PartialTableCache.java @@ -54,7 +54,6 @@ public class PartialTableCache implements TableCache { private final ExecutorService executorService; private final CacheStatsRecorder statsRecorder; - public PartialTableCache(String threadNamePrefix) { // We use concurrent Hash map for O(1) lookup for get API. // During list operation for partial cache we anyway merge between DB and diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/cache/TableNoCache.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/cache/TableNoCache.java index 17bb961ac3d0..b3ec98619ce6 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/cache/TableNoCache.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/hdds/utils/db/cache/TableNoCache.java @@ -38,6 +38,7 @@ public final class TableNoCache implements TableCache { public static final CacheStats EMPTY_STAT = new CacheStats(0, 0, 0); private static final TableCache NO_CACHE_INSTANCE = new TableNoCache<>(); + public static TableCache instance() { return (TableCache) NO_CACHE_INSTANCE; } diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/audit/AuditLogger.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/audit/AuditLogger.java index 28ab88440bdd..026970361973 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/audit/AuditLogger.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/audit/AuditLogger.java @@ -148,6 +148,7 @@ private String getLowerCaseOp(String op) { */ public static class PerformanceStringBuilder { private final StringBuilder builder = new StringBuilder(128).append('{'); + /** * Appends metadata operation latency in milliseconds. * @param nanos Latency in nanoseconds. diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/BlockGroup.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/BlockGroup.java index 7fd00bb9341c..5e8b0e1724be 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/BlockGroup.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/BlockGroup.java @@ -30,6 +30,7 @@ public final class BlockGroup { private String groupID; private List blockIDs; + private BlockGroup(String groupID, List blockIDs) { this.groupID = groupID; this.blockIDs = blockIDs; diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/DeleteBlockGroupResult.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/DeleteBlockGroupResult.java index 6c7444dfe7b2..ea9bc4a4caf7 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/DeleteBlockGroupResult.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/DeleteBlockGroupResult.java @@ -31,6 +31,7 @@ public class DeleteBlockGroupResult { private String objectKey; private List blockResultList; + public DeleteBlockGroupResult(String objectKey, List blockResultList) { this.objectKey = objectKey; diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/Storage.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/Storage.java index 15047510d151..ae357d8d7519 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/Storage.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/Storage.java @@ -58,7 +58,6 @@ public abstract class Storage { private final StorageState state; private final StorageInfo storageInfo; - /** * Determines the state of the Version file. */ @@ -192,7 +191,6 @@ public File getVersionFile() { return new File(getCurrentDir(), STORAGE_FILE_VERSION); } - /** * Check to see if current/ directory is empty. This method is used * before determining to format the directory. diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/StorageInfo.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/StorageInfo.java index 77bc23a4537b..ebb3eab5c45b 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/StorageInfo.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/common/StorageInfo.java @@ -187,7 +187,6 @@ private void verifyCreationTime() { Preconditions.checkNotNull(creationTime); } - public void writeTo(File to) throws IOException { IOUtils.writePropertiesToFile(to, properties); diff --git a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/lease/Lease.java b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/lease/Lease.java index d98cfb9df223..7310cd9b2766 100644 --- a/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/lease/Lease.java +++ b/hadoop-hdds/framework/src/main/java/org/apache/hadoop/ozone/lease/Lease.java @@ -49,7 +49,6 @@ public class Lease { */ private Callable callback; - /** * Creates a lease on the specified resource with given timeout. * diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/ssl/TestSSLConnectionWithReload.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/ssl/TestSSLConnectionWithReload.java index 8b1739c2b2c9..b1be337b2b9a 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/ssl/TestSSLConnectionWithReload.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/ssl/TestSSLConnectionWithReload.java @@ -132,6 +132,7 @@ private ContainerCommandResponseProto sendRequest( public void onNext(ContainerCommandResponseProto value) { replyFuture.complete(value); } + @Override public void onError(Throwable t) { } diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/authority/TestDefaultProfile.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/authority/TestDefaultProfile.java index c2952a4c66f7..5f09b347b937 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/authority/TestDefaultProfile.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/authority/TestDefaultProfile.java @@ -281,8 +281,6 @@ public void testInValidExtendedKeyUsage() throws IOException, assertFalse(approver.verfiyExtensions(csr)); } - - /** * Generates an CSR with the extension specified. * This function is used to get an Invalid CSR and test that PKI profile diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/client/TestDnCertificateClientInit.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/client/TestDnCertificateClientInit.java index 3ddbc318c856..cf7de7b11f34 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/client/TestDnCertificateClientInit.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/x509/certificate/client/TestDnCertificateClientInit.java @@ -103,7 +103,6 @@ public void tearDown() throws IOException { dnCertificateClient = null; } - @ParameterizedTest @MethodSource("parameters") public void testInitDatanode(boolean pvtKeyPresent, boolean pubKeyPresent, diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/server/events/TestEventQueueChain.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/server/events/TestEventQueueChain.java index 1d9dba2c4071..c2ab7d805c20 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/server/events/TestEventQueueChain.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/server/events/TestEventQueueChain.java @@ -42,7 +42,6 @@ public void simpleEvent() { queue.processAll(5000); } - static class FailedNode { private final String nodeId; diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/server/http/TestRatisNameRewrite.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/server/http/TestRatisNameRewrite.java index ac26bc97a5b9..8a1a2ecc3090 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/server/http/TestRatisNameRewrite.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/server/http/TestRatisNameRewrite.java @@ -33,7 +33,6 @@ */ public class TestRatisNameRewrite { - private static Stream parameters() { return Stream.of( arguments( diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestRDBTableStore.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestRDBTableStore.java index a5cf3b6e6c58..fd593016d19d 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestRDBTableStore.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/TestRDBTableStore.java @@ -468,7 +468,6 @@ public void testGetIfExist() throws Exception { } } - @Test public void testCountEstimatedRowsInTable() throws Exception { try (Table testTable = rdbStore.getTable("Eighth")) { @@ -552,7 +551,6 @@ private void writeToTable(Table testTable, int num) throws IOException { } } - @Test public void testPrefixedIterator() throws Exception { int containerCount = 3; @@ -642,7 +640,6 @@ public void testStringPrefixedIteratorCloseDb() throws Exception { } } - @Test public void testPrefixedRangeKVs() throws Exception { int containerCount = 3; diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/cache/TestTableCache.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/cache/TestTableCache.java index 6acd908f0a78..3692b7432901 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/cache/TestTableCache.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/utils/db/cache/TestTableCache.java @@ -59,6 +59,7 @@ private void createTableCache(TableCache.CacheType cacheType) { private static Stream cacheTypeList() { return Stream.of(TableCache.CacheType.FULL_CACHE, TableCache.CacheType.PARTIAL_CACHE); } + @ParameterizedTest @MethodSource("cacheTypeList") public void testPartialTableCache(TableCache.CacheType cacheType) { @@ -494,7 +495,6 @@ public void testTableCache(TableCache.CacheType cacheType) { verifyStats(tableCache, 0, 0, 0); } - @ParameterizedTest @MethodSource("cacheTypeList") public void testTableCacheWithNonConsecutiveEpochList( diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/audit/DummyEntity.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/audit/DummyEntity.java index 057323af0d64..5f8791a2aec6 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/audit/DummyEntity.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/audit/DummyEntity.java @@ -32,6 +32,7 @@ public DummyEntity() { this.key1 = "value1"; this.key2 = "value2"; } + public String getKey1() { return key1; } diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/upgrade/TestBasicUpgradeFinalizer.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/upgrade/TestBasicUpgradeFinalizer.java index a37665dcdb80..2be89cacfe44 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/upgrade/TestBasicUpgradeFinalizer.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/upgrade/TestBasicUpgradeFinalizer.java @@ -85,7 +85,6 @@ public void testFinalizerPhasesAreInvokedInOrder() throws IOException { finalizer.postCalled); } - @Test public void testAlreadyFinalizedDoesNotTriggerNewFinalization() throws IOException { diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/upgrade/TestLayoutVersionInstanceFactory.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/upgrade/TestLayoutVersionInstanceFactory.java index 7a872e202361..f759b26dbf83 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/upgrade/TestLayoutVersionInstanceFactory.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/upgrade/TestLayoutVersionInstanceFactory.java @@ -36,7 +36,6 @@ public class TestLayoutVersionInstanceFactory { private final MockInterface m1 = new MockClassV1(); private final MockInterface m2 = new MockClassV2(); - @Test void testRegister() { LayoutVersionManager lvm = getMockLvm(1, 2); @@ -107,7 +106,6 @@ void testMethodBasedVersionFactory() { assertEquals("v1", method.get()); } - private VersionFactoryKey getKey(String key, Integer version) { VersionFactoryKey.Builder vfKey = new VersionFactoryKey.Builder().key(key); if (version != null) { @@ -116,8 +114,6 @@ private VersionFactoryKey getKey(String key, Integer version) { return vfKey.build(); } - - @Test void testOnFinalize() { LayoutVersionManager lvm = getMockLvm(1, 3); diff --git a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/upgrade/TestUpgradeFinalizerActions.java b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/upgrade/TestUpgradeFinalizerActions.java index 46681cbba6a3..9ff33db914e9 100644 --- a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/upgrade/TestUpgradeFinalizerActions.java +++ b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/ozone/upgrade/TestUpgradeFinalizerActions.java @@ -189,7 +189,6 @@ public Optional action(UpgradeActionType phase) { } } - /** * Mock DN Upgrade Action that fails. */ diff --git a/hadoop-hdds/rocks-native/src/main/java/org/apache/hadoop/hdds/utils/NativeConstants.java b/hadoop-hdds/rocks-native/src/main/java/org/apache/hadoop/hdds/utils/NativeConstants.java index 7d49eb20d768..379ea32b8537 100644 --- a/hadoop-hdds/rocks-native/src/main/java/org/apache/hadoop/hdds/utils/NativeConstants.java +++ b/hadoop-hdds/rocks-native/src/main/java/org/apache/hadoop/hdds/utils/NativeConstants.java @@ -25,6 +25,7 @@ public final class NativeConstants { private NativeConstants() { } + public static final String ROCKS_TOOLS_NATIVE_LIBRARY_NAME = "ozone_rocksdb_tools"; public static final String ROCKS_TOOLS_NATIVE_PROPERTY = "rocks_tools_native"; } diff --git a/hadoop-hdds/rocks-native/src/main/java/org/apache/hadoop/hdds/utils/db/managed/ManagedRawSSTFileIterator.java b/hadoop-hdds/rocks-native/src/main/java/org/apache/hadoop/hdds/utils/db/managed/ManagedRawSSTFileIterator.java index 14e6b9374c05..391c04a89a8f 100644 --- a/hadoop-hdds/rocks-native/src/main/java/org/apache/hadoop/hdds/utils/db/managed/ManagedRawSSTFileIterator.java +++ b/hadoop-hdds/rocks-native/src/main/java/org/apache/hadoop/hdds/utils/db/managed/ManagedRawSSTFileIterator.java @@ -38,10 +38,15 @@ public class ManagedRawSSTFileIterator implements ClosableIterator { } private native boolean hasNext(long handle); + private native void next(long handle); + private native byte[] getKey(long handle); + private native byte[] getValue(long handle); + private native long getSequenceNumber(long handle); + private native int getType(long handle); @Override diff --git a/hadoop-hdds/rocks-native/src/main/java/org/apache/hadoop/hdds/utils/db/managed/ManagedRawSSTFileReader.java b/hadoop-hdds/rocks-native/src/main/java/org/apache/hadoop/hdds/utils/db/managed/ManagedRawSSTFileReader.java index eadf1afcb879..97f7a7a49e79 100644 --- a/hadoop-hdds/rocks-native/src/main/java/org/apache/hadoop/hdds/utils/db/managed/ManagedRawSSTFileReader.java +++ b/hadoop-hdds/rocks-native/src/main/java/org/apache/hadoop/hdds/utils/db/managed/ManagedRawSSTFileReader.java @@ -75,7 +75,6 @@ public ManagedRawSSTFileIterator newIterator( private native long newRawSSTFileReader(long optionsHandle, String filePath, int readSize); - private native long newIterator(long handle, boolean hasFrom, long fromSliceHandle, boolean hasTo, long toSliceHandle); diff --git a/hadoop-hdds/rocks-native/src/test/java/org/apache/hadoop/hdds/utils/db/managed/TestManagedRawSSTFileIterator.java b/hadoop-hdds/rocks-native/src/test/java/org/apache/hadoop/hdds/utils/db/managed/TestManagedRawSSTFileIterator.java index 98f8708e211a..319ef651753d 100644 --- a/hadoop-hdds/rocks-native/src/test/java/org/apache/hadoop/hdds/utils/db/managed/TestManagedRawSSTFileIterator.java +++ b/hadoop-hdds/rocks-native/src/test/java/org/apache/hadoop/hdds/utils/db/managed/TestManagedRawSSTFileIterator.java @@ -95,7 +95,6 @@ public static void init() throws NativeLibraryNotLoadedException { ManagedRawSSTFileReader.loadLibrary(); } - @ParameterizedTest @MethodSource("keyValueFormatArgs") public void testSSTDumpIteratorWithKeyFormat(String keyFormat, String valueFormat) throws Exception { diff --git a/hadoop-hdds/rocksdb-checkpoint-differ/src/main/java/org/apache/ozone/rocksdiff/RocksDBCheckpointDiffer.java b/hadoop-hdds/rocksdb-checkpoint-differ/src/main/java/org/apache/ozone/rocksdiff/RocksDBCheckpointDiffer.java index e57c501640d6..0444e4244c36 100644 --- a/hadoop-hdds/rocksdb-checkpoint-differ/src/main/java/org/apache/ozone/rocksdiff/RocksDBCheckpointDiffer.java +++ b/hadoop-hdds/rocksdb-checkpoint-differ/src/main/java/org/apache/ozone/rocksdiff/RocksDBCheckpointDiffer.java @@ -786,6 +786,7 @@ private void preconditionChecksForLoadAllCompactionLogs() { Preconditions.checkNotNull(activeRocksDB, "activeRocksDB must be set before calling loadAllCompactionLogs."); } + /** * Helper function that prepends SST file name with SST backup directory path * (or DB checkpoint path if compaction hasn't happened yet as SST files won't @@ -1174,7 +1175,6 @@ public void pruneOlderSnapshotsWithCompactionHistory() { } } - /** * Returns the list of input files from the compaction entries which are * older than the maximum allowed in the compaction DAG. diff --git a/hadoop-hdds/rocksdb-checkpoint-differ/src/test/java/org/apache/ozone/compaction/log/TestCompactionLogEntry.java b/hadoop-hdds/rocksdb-checkpoint-differ/src/test/java/org/apache/ozone/compaction/log/TestCompactionLogEntry.java index b65a16d446f0..d19e8295e335 100644 --- a/hadoop-hdds/rocksdb-checkpoint-differ/src/test/java/org/apache/ozone/compaction/log/TestCompactionLogEntry.java +++ b/hadoop-hdds/rocksdb-checkpoint-differ/src/test/java/org/apache/ozone/compaction/log/TestCompactionLogEntry.java @@ -103,7 +103,6 @@ public void testGetProtobuf( } } - @ParameterizedTest(name = "{0}") @MethodSource("compactionLogEntryValidScenarios") public void testFromProtobuf( diff --git a/hadoop-hdds/rocksdb-checkpoint-differ/src/test/java/org/apache/ozone/rocksdiff/TestRocksDBCheckpointDiffer.java b/hadoop-hdds/rocksdb-checkpoint-differ/src/test/java/org/apache/ozone/rocksdiff/TestRocksDBCheckpointDiffer.java index 96dc9305f018..b975ea26dc34 100644 --- a/hadoop-hdds/rocksdb-checkpoint-differ/src/test/java/org/apache/ozone/rocksdiff/TestRocksDBCheckpointDiffer.java +++ b/hadoop-hdds/rocksdb-checkpoint-differ/src/test/java/org/apache/ozone/rocksdiff/TestRocksDBCheckpointDiffer.java @@ -1168,7 +1168,6 @@ public void testPruneBackwardDag(String description, assertEquals(actualFileNodesRemoved, expectedFileNodesRemoved); } - /** * Test cases for pruneBackwardDag. */ diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/PlacementPolicy.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/PlacementPolicy.java index b48cf6794e90..917a29b2a8e6 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/PlacementPolicy.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/PlacementPolicy.java @@ -39,6 +39,7 @@ default List chooseDatanodes( favoredNodes, nodesRequired, metadataSizeRequired, dataSizeRequired); } + /** * Given an initial set of datanodes and the size required, * return set of datanodes that satisfy the nodes and size requirement. diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/block/BlockManagerImpl.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/block/BlockManagerImpl.java index d766e3b05d4a..9b46968424cc 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/block/BlockManagerImpl.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/block/BlockManagerImpl.java @@ -70,6 +70,7 @@ public class BlockManagerImpl implements BlockManager, BlockmanagerMXBean { private ObjectName mxBean; private final SequenceIdGenerator sequenceIdGen; private ScmBlockDeletingServiceMetrics metrics; + /** * Constructor. * diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/block/ScmBlockDeletingServiceMetrics.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/block/ScmBlockDeletingServiceMetrics.java index e8c0178612e5..2f298a5ebc5c 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/block/ScmBlockDeletingServiceMetrics.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/block/ScmBlockDeletingServiceMetrics.java @@ -168,10 +168,12 @@ public void incrDNCommandsSent(UUID id, long delta) { numCommandsDatanode.computeIfAbsent(id, k -> new DatanodeCommandCounts()) .incrCommandsSent(delta); } + public void incrDNCommandsSuccess(UUID id, long delta) { numCommandsDatanode.computeIfAbsent(id, k -> new DatanodeCommandCounts()) .incrCommandsSuccess(delta); } + public void incrDNCommandsFailure(UUID id, long delta) { numCommandsDatanode.computeIfAbsent(id, k -> new DatanodeCommandCounts()) .incrCommandsFailure(delta); @@ -312,6 +314,7 @@ public long getNumCommandsDatanodeSent() { } return sent; } + public long getNumCommandsDatanodeSuccess() { long successCount = 0; for (DatanodeCommandCounts v : numCommandsDatanode.values()) { @@ -319,6 +322,7 @@ public long getNumCommandsDatanodeSuccess() { } return successCount; } + public long getNumCommandsDatanodeFailed() { long failCount = 0; for (DatanodeCommandCounts v : numCommandsDatanode.values()) { diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/AbstractContainerReportHandler.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/AbstractContainerReportHandler.java index 52f42594429a..23d5ccf8f799 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/AbstractContainerReportHandler.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/AbstractContainerReportHandler.java @@ -94,7 +94,6 @@ public String toString() { }; } - /** * Process the given ContainerReplica received from specified datanode. * diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerManager.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerManager.java index b1cbd129b702..104b3433c91e 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerManager.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerManager.java @@ -52,10 +52,10 @@ void reinitialize(Table containerStore) ContainerInfo getContainer(ContainerID containerID) throws ContainerNotFoundException; - default List getContainers() { return getContainers(ContainerID.valueOf(0), Integer.MAX_VALUE); } + /** * Returns containers under certain conditions. * Search container IDs from start ID(exclusive), @@ -72,7 +72,6 @@ default List getContainers() { */ List getContainers(ContainerID startID, int count); - List getContainers(ReplicationType type); /** diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerManagerImpl.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerManagerImpl.java index 813c91c2398c..2e6bfa6f3dd3 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerManagerImpl.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerManagerImpl.java @@ -136,7 +136,6 @@ public ContainerInfo getContainer(final ContainerID id) id + " not found.")); } - @Override public List getContainers(ReplicationType type) { return containerStateManager.getContainerInfos(type); diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerStateManagerImpl.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerStateManagerImpl.java index 5ada3ecbbebc..34864b705191 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerStateManagerImpl.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/ContainerStateManagerImpl.java @@ -517,7 +517,6 @@ private ContainerInfo findContainerWithSpace(final long size, return null; } - @Override public void removeContainer(final HddsProtos.ContainerID id) throws IOException { @@ -588,6 +587,7 @@ public Builder setSCMDBTransactionBuffer(DBTransactionBuffer buffer) { this.transactionBuffer = buffer; return this; } + public Builder setConfiguration(final Configuration config) { conf = config; return this; diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/ContainerBalancer.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/ContainerBalancer.java index 5677c0373dcf..872747f27e38 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/ContainerBalancer.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/ContainerBalancer.java @@ -195,6 +195,7 @@ public ContainerBalancerStatusInfo getBalancerStatusInfo() throws IOException { lock.unlock(); } } + /** * Checks if ContainerBalancer is in valid state to call stop. * diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/ContainerBalancerMetrics.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/ContainerBalancerMetrics.java index 969940b4d891..b4acf2a2fe75 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/ContainerBalancerMetrics.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/ContainerBalancerMetrics.java @@ -358,6 +358,7 @@ public void incrementNumContainerMovesFailedInLatestIteration( long valueToAdd) { numContainerMovesFailedInLatestIteration.incr(valueToAdd); } + public void resetNumContainerMovesFailedInLatestIteration() { numContainerMovesFailedInLatestIteration.incr( -getNumContainerMovesFailedInLatestIteration()); diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/ContainerBalancerSelectionCriteria.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/ContainerBalancerSelectionCriteria.java index 800465bee27b..2c2395af6ba1 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/ContainerBalancerSelectionCriteria.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/ContainerBalancerSelectionCriteria.java @@ -231,7 +231,6 @@ public void addToExcludeDueToFailContainers(ContainerID container) { this.excludeContainersDueToFailure.add(container); } - private NavigableSet getCandidateContainers(DatanodeDetails node) { NavigableSet newSet = new TreeSet<>(orderContainersByUsedBytes().reversed()); diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/DataMoveInfo.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/DataMoveInfo.java index d7a7d6c5df8d..455e1310a89d 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/DataMoveInfo.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/balancer/DataMoveInfo.java @@ -29,7 +29,6 @@ public class DataMoveInfo { private final Map sizeEnteringNodes; private final Map sizeLeavingNodes; - public DataMoveInfo( long sizeScheduledForMove, long dataSizeMoved, diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/metrics/SCMContainerManagerMetrics.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/metrics/SCMContainerManagerMetrics.java index 014aa5a7c0bd..ed3670ea6645 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/metrics/SCMContainerManagerMetrics.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/metrics/SCMContainerManagerMetrics.java @@ -43,7 +43,6 @@ public final class SCMContainerManagerMetrics { private @Metric MutableCounterLong numFailureDeleteContainers; private @Metric MutableCounterLong numListContainerOps; - private @Metric MutableCounterLong numContainerReportsProcessedSuccessful; private @Metric MutableCounterLong numContainerReportsProcessedFailed; private @Metric MutableCounterLong numICRReportsProcessedSuccessful; diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/ContainerPlacementPolicyFactory.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/ContainerPlacementPolicyFactory.java index d1ff5c5fa401..aef416358882 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/ContainerPlacementPolicyFactory.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/ContainerPlacementPolicyFactory.java @@ -46,7 +46,6 @@ public final class ContainerPlacementPolicyFactory { private ContainerPlacementPolicyFactory() { } - public static PlacementPolicy getPolicy( ConfigurationSource conf, final NodeManager nodeManager, NetworkTopology clusterMap, final boolean fallback, diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/ContainerPlacementStatusDefault.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/ContainerPlacementStatusDefault.java index 2d9c80542fcf..9328025419e3 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/ContainerPlacementStatusDefault.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/ContainerPlacementStatusDefault.java @@ -35,7 +35,6 @@ public class ContainerPlacementStatusDefault private final int maxReplicasPerRack; private final List rackReplicaCnts; - public ContainerPlacementStatusDefault(int currentRacks, int requiredRacks, int totalRacks, int maxReplicasPerRack, List rackReplicaCnts) { this.requiredRacks = requiredRacks; diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ContainerHealthResult.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ContainerHealthResult.java index bf4f0b92fdd6..34e8b9d323fc 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ContainerHealthResult.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ContainerHealthResult.java @@ -222,6 +222,7 @@ public boolean isUnrecoverable() { public void setHasUnReplicatedOfflineIndexes(boolean val) { hasUnReplicatedOfflineIndexes = val; } + /** * Indicates whether a container has some indexes which are only on nodes * which are DECOMMISSIONING or ENTERING_MAINTENANCE. These containers may diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ContainerReplicaPendingOps.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ContainerReplicaPendingOps.java index 6ee168cd9e1f..d2d4389d7372 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ContainerReplicaPendingOps.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ContainerReplicaPendingOps.java @@ -163,7 +163,6 @@ public boolean completeAddReplica(ContainerID containerID, return completed; } - /** * Remove a stored ContainerReplicaOp from the given ContainerID as it has * been deleted successfully. diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ECContainerReplicaCount.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ECContainerReplicaCount.java index 07e1dc0f44eb..36be55c41b07 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ECContainerReplicaCount.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ECContainerReplicaCount.java @@ -330,6 +330,7 @@ private Set healthyReplicas() { distinct.addAll(maintenanceIndexes.keySet()); return distinct; } + /** * Returns an unsorted list of indexes which need additional copies to * ensure the container is sufficiently replicated. These missing indexes will @@ -545,8 +546,6 @@ public boolean isSufficientlyReplicated() { return isSufficientlyReplicated(false); } - - @Override public String toString() { StringBuilder sb = new StringBuilder(); diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationManager.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationManager.java index 889a8d1ef5c3..f8312a62db5a 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationManager.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationManager.java @@ -111,13 +111,11 @@ public class ReplicationManager implements SCMService, ContainerReplicaPendingOp */ private final ContainerManager containerManager; - /** * SCMContext from StorageContainerManager. */ private SCMContext scmContext; - /** * ReplicationManager specific configuration. */ @@ -632,6 +630,7 @@ public void sendLowPriorityReplicateContainerCommand( command.setPriority(ReplicationCommandPriority.LOW); sendDatanodeCommand(command, container, source, scmDeadlineEpochMs); } + /** * Sends a command to a datanode with the command deadline set to the default * in ReplicationManager config. @@ -1132,6 +1131,7 @@ public static class ReplicationManagerConfiguration + "sent to datanodes. After this timeout the command will be " + "retried.") private long eventTimeout = Duration.ofMinutes(12).toMillis(); + public void setInterval(Duration interval) { this.interval = interval; } @@ -1156,6 +1156,7 @@ public void setEventTimeout(Duration timeout) { + "the datanodes will not process a command after SCM believes it " + "should have expired.") private long datanodeTimeoutOffset = Duration.ofMinutes(6).toMillis(); + public long getDatanodeTimeoutOffset() { return datanodeTimeoutOffset; } diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationManagerMetrics.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationManagerMetrics.java index 27c69b9702d7..13206e585f5d 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationManagerMetrics.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationManagerMetrics.java @@ -67,7 +67,6 @@ public final class ReplicationManagerMetrics implements MetricsSource { "Tracked inflight container deletion requests skipped" + " due to the configured limit."); - private static final MetricsInfo INFLIGHT_MOVE = Interns.info( "InflightMove", "Tracked inflight container move requests."); @@ -225,7 +224,6 @@ public final class ReplicationManagerMetrics implements MetricsSource { + "to the pending commands on all source datanodes") private MutableCounterLong replicateContainerCmdsDeferredTotal; - public ReplicationManagerMetrics(ReplicationManager manager) { this.registry = new MetricsRegistry(METRICS_SOURCE_NAME); this.replicationManager = manager; diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationManagerUtil.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationManagerUtil.java index 1502a9c4cf09..e4e5a6938d28 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationManagerUtil.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationManagerUtil.java @@ -199,7 +199,6 @@ An UNHEALTHY replica with unique origin node id of a QUASI_CLOSED container shou return new ExcludedAndUsedNodes(excludedNodes, usedNodes); } - /** * Simple class to hold the excluded and used nodes lists. */ diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/report/ContainerReportValidator.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/report/ContainerReportValidator.java index e7b7a086361b..683301b6fb30 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/report/ContainerReportValidator.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/report/ContainerReportValidator.java @@ -35,6 +35,7 @@ public final class ContainerReportValidator { private Map replicaValidators; private static final ContainerReportValidator CONTAINER_REPORT_VALIDATOR; + private ContainerReportValidator() { this.replicaValidators = ImmutableMap.of(ReplicationType.EC, new ECReplicaValidator()); @@ -50,6 +51,7 @@ private static class ECReplicaValidator implements ReplicaValidator { private static final Logger LOG = LoggerFactory.getLogger(ECReplicaValidator.class); + @Override public boolean validate(ReplicationConfig replicationConfig, DatanodeDetails datanodeDetails, @@ -81,6 +83,7 @@ public boolean validate(ReplicationConfig replicationConfig, static { CONTAINER_REPORT_VALIDATOR = new ContainerReportValidator(); } + private static boolean validateReplica(ReplicationConfig replicationConfig, DatanodeDetails datanodeDetails, ContainerReplicaProto replicaProto) { @@ -91,12 +94,11 @@ private static boolean validateReplica(ReplicationConfig replicationConfig, datanodeDetails, replicaProto)) .orElse(true); } + public static boolean validate(ContainerInfo containerInfo, DatanodeDetails datanodeDetails, ContainerReplicaProto replicaProto) { return validateReplica(containerInfo.getReplicationConfig(), datanodeDetails, replicaProto); } - - } diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/states/ContainerStateMap.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/states/ContainerStateMap.java index 05196f846c1a..593d965a6d16 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/states/ContainerStateMap.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/container/states/ContainerStateMap.java @@ -140,6 +140,7 @@ ContainerReplica removeReplica(ContainerID containerID, DatanodeID datanodeID) { return entry == null ? null : entry.removeReplica(datanodeID); } } + /** * Map {@link LifeCycleState} to {@link ContainerInfo}. * Note that a {@link ContainerInfo} can only exists in at most one of the {@link LifeCycleState}s. diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/InterSCMGrpcClient.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/InterSCMGrpcClient.java index aa6f4f7f0860..204962463371 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/InterSCMGrpcClient.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/InterSCMGrpcClient.java @@ -84,7 +84,6 @@ public InterSCMGrpcClient(final String host, withDeadlineAfter(timeout, TimeUnit.SECONDS); } - @Override public CompletableFuture download(final Path outputPath) { // By default, on every checkpoint, the rocks db will be flushed diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/RatisUtil.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/RatisUtil.java index 456d020f9fb8..a482ae34f239 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/RatisUtil.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/RatisUtil.java @@ -57,7 +57,6 @@ public final class RatisUtil { private RatisUtil() { } - /** * Constructs new Raft Properties instance using {@link ConfigurationSource}. * diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMHAInvocationHandler.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMHAInvocationHandler.java index f33945175c7e..74e4f54e024c 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMHAInvocationHandler.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMHAInvocationHandler.java @@ -38,7 +38,6 @@ */ public class SCMHAInvocationHandler implements InvocationHandler { - private static final Logger LOG = LoggerFactory .getLogger(SCMHAInvocationHandler.class); diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMHAManagerImpl.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMHAManagerImpl.java index 2a1b9ff65eed..00915406a4ce 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMHAManagerImpl.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMHAManagerImpl.java @@ -428,7 +428,6 @@ public boolean removeSCM(RemoveSCMRequest request) throws IOException { return ratisServer.removeSCM(request); } - void stopServices() throws Exception { // just stop the SCMMetaData store. All other background diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisRequest.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisRequest.java index 3150bdff12d4..67db93dd1a26 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisRequest.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisRequest.java @@ -84,6 +84,7 @@ public Object[] getArguments() { public Class[] getParameterTypes() { return parameterTypes.clone(); } + /** * Encodes the request into Ratis Message. */ diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisServerImpl.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisServerImpl.java index b00358701856..ff23c0b8fdef 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisServerImpl.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMRatisServerImpl.java @@ -274,7 +274,6 @@ public boolean isStopped() { return isStopped; } - @Override public List getRatisRoles() { Collection peers = division.getGroup().getPeers(); @@ -299,6 +298,7 @@ public List getRatisRoles() { } return ratisRoles; } + /** * {@inheritDoc} */ diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMSnapshotProvider.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMSnapshotProvider.java index ffcba4e311c4..ac960d5c1c7e 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMSnapshotProvider.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/ha/SCMSnapshotProvider.java @@ -49,7 +49,6 @@ public class SCMSnapshotProvider { private final File scmSnapshotDir; - private final ConfigurationSource conf; private Map peerNodesMap; diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/DatanodeAdminMonitor.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/DatanodeAdminMonitor.java index 48bebd6f53aa..c7c5c313326c 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/DatanodeAdminMonitor.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/DatanodeAdminMonitor.java @@ -31,9 +31,13 @@ public interface DatanodeAdminMonitor extends Runnable { void startMonitoring(DatanodeDetails dn); + void stopMonitoring(DatanodeDetails dn); + Set getTrackedNodes(); + void setMetrics(NodeDecommissionMetrics metrics); + Map> getContainersPendingReplication(DatanodeDetails dn) throws NodeNotFoundException; } diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeDecommissionMetrics.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeDecommissionMetrics.java index 2410b198e1b6..dbe4b37d26b2 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeDecommissionMetrics.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeDecommissionMetrics.java @@ -97,7 +97,6 @@ public static final class ContainerStateInWorkflow { private static final MetricsInfo HOST_START_TIME = Interns.info("StartTimeDN", "Time at which decommissioning was started"); - public ContainerStateInWorkflow(String host, long sufficiently, long under, diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeManager.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeManager.java index c2eba8df0a23..cd64d669aab0 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeManager.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeManager.java @@ -75,7 +75,6 @@ public interface NodeManager extends StorageContainerNodeProtocol, EventHandler, NodeManagerMXBean, Closeable { - /** * Register API without a layout version info object passed in. Useful for * tests. diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeStateManager.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeStateManager.java index 793c6dc4c350..eab514f4c8a3 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeStateManager.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeStateManager.java @@ -95,7 +95,6 @@ private enum NodeLifeCycleEvent { private static final Logger LOG = LoggerFactory .getLogger(NodeStateManager.class); - /** * StateMachine for node lifecycle. */ diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeStatus.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeStatus.java index a67525accea1..6e6667e650cf 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeStatus.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/NodeStatus.java @@ -46,6 +46,7 @@ public final class NodeStatus { /** For the {@link NodeStatus} objects with {@link #opStateExpiryEpochSeconds} == 0. */ private static final Map> CONSTANTS; + static { final Map> map = new EnumMap<>(NodeOperationalState.class); for (NodeOperationalState op : NodeOperationalState.values()) { @@ -73,6 +74,7 @@ public static NodeStatus valueOf(NodeOperationalState op, NodeState health, long private static final Set MAINTENANCE_STATES = Collections.unmodifiableSet( EnumSet.of(ENTERING_MAINTENANCE, IN_MAINTENANCE)); + /** * @return the set consists of {@link NodeOperationalState#ENTERING_MAINTENANCE} * and {@link NodeOperationalState#IN_MAINTENANCE}. @@ -83,6 +85,7 @@ public static Set maintenanceStates() { private static final Set DECOMMISSION_STATES = Collections.unmodifiableSet( EnumSet.of(DECOMMISSIONING, DECOMMISSIONED)); + /** * @return the set consists of {@link NodeOperationalState#DECOMMISSIONING} * and {@link NodeOperationalState#DECOMMISSIONED}. @@ -93,6 +96,7 @@ public static Set decommissionStates() { private static final Set OUT_OF_SERVICE_STATES = Collections.unmodifiableSet( EnumSet.of(DECOMMISSIONING, DECOMMISSIONED, ENTERING_MAINTENANCE, IN_MAINTENANCE)); + /** * @return the set consists of {@link NodeOperationalState#DECOMMISSIONING}, * {@link NodeOperationalState#DECOMMISSIONED}, @@ -104,24 +108,28 @@ public static Set outOfServiceStates() { } private static final NodeStatus IN_SERVICE_HEALTHY = valueOf(IN_SERVICE, HEALTHY); + /** @return the status of {@link NodeOperationalState#IN_SERVICE} and {@link NodeState#HEALTHY}. */ public static NodeStatus inServiceHealthy() { return IN_SERVICE_HEALTHY; } private static final NodeStatus IN_SERVICE_HEALTHY_READONLY = valueOf(IN_SERVICE, HEALTHY_READONLY); + /** @return the status of {@link NodeOperationalState#IN_SERVICE} and {@link NodeState#HEALTHY_READONLY}. */ public static NodeStatus inServiceHealthyReadOnly() { return IN_SERVICE_HEALTHY_READONLY; } private static final NodeStatus IN_SERVICE_STALE = NodeStatus.valueOf(IN_SERVICE, STALE); + /** @return the status of {@link NodeOperationalState#IN_SERVICE} and {@link NodeState#STALE}. */ public static NodeStatus inServiceStale() { return IN_SERVICE_STALE; } private static final NodeStatus IN_SERVICE_DEAD = NodeStatus.valueOf(IN_SERVICE, DEAD); + /** @return the status of {@link NodeOperationalState#IN_SERVICE} and {@link NodeState#DEAD}. */ public static NodeStatus inServiceDead() { return IN_SERVICE_DEAD; diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/SCMNodeManager.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/SCMNodeManager.java index 68c1e0bbb376..87c5e5d50c47 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/SCMNodeManager.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/SCMNodeManager.java @@ -148,6 +148,7 @@ public class SCMNodeManager implements NodeManager { private static final String TOTALCAPACITY = "CAPACITY"; private static final String DNUUID = "UUID"; private static final String VERSION = "VERSION"; + /** * Constructs SCM machine Manager. */ @@ -1438,6 +1439,7 @@ private enum UsageStatics { public String getLabel() { return label; } + UsageStatics(String label) { this.label = label; } @@ -1450,9 +1452,11 @@ private enum StateStatistics { ENTERING_MAINTENANCE("EnteringMaintenance"), VOLUME_FAILURES("VolumeFailures"); private String label; + public String getLabel() { return label; } + StateStatistics(String label) { this.label = label; } @@ -1463,10 +1467,13 @@ private enum SpaceStatistics { SCM_USED("Scmused"), NON_SCM_USED("NonScmused"), REMAINING("Remaining"); + private String label; + public String getLabel() { return label; } + SpaceStatistics(String label) { this.label = label; } diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/SCMNodeStorageStatMap.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/SCMNodeStorageStatMap.java index a0bae80157e3..4a60767623a3 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/SCMNodeStorageStatMap.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/SCMNodeStorageStatMap.java @@ -51,6 +51,7 @@ public class SCMNodeStorageStatMap implements SCMNodeStorageStatMXBean { private final double criticalUtilizationThreshold; private final Map> scmNodeStorageReportMap; + /** * constructs the scmNodeStorageReportMap object. */ @@ -98,8 +99,6 @@ public List getDatanodeList( .collect(Collectors.toList()); } - - /** * Insert a new datanode into Node2Container Map. * @@ -312,7 +311,6 @@ public Set getStorageVolumes(UUID datanodeID) { return scmNodeStorageReportMap.get(datanodeID); } - /** * Truncate to 4 digits since uncontrolled precision is some times * counter intuitive to what users expect. @@ -332,6 +330,7 @@ public double getScmUsedratio(long scmUsed, long capacity) { truncateDecimals(scmUsed / (double) capacity); return scmUsedRatio; } + /** * Results possible from processing a Node report by * Node2ContainerMapper. diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/states/NodeNotFoundException.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/states/NodeNotFoundException.java index b23d98e35d8c..ee515f24182f 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/states/NodeNotFoundException.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/node/states/NodeNotFoundException.java @@ -25,7 +25,6 @@ */ public class NodeNotFoundException extends NodeException { - /** * Constructs an {@code NodeNotFoundException} with {@code null} * as its error detail message. diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/BackgroundPipelineCreator.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/BackgroundPipelineCreator.java index 744b5dc1ebb5..29cc68885970 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/BackgroundPipelineCreator.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/BackgroundPipelineCreator.java @@ -89,7 +89,6 @@ public class BackgroundPipelineCreator implements SCMService { private final long intervalInMillis; private final Clock clock; - BackgroundPipelineCreator(PipelineManager pipelineManager, ConfigurationSource conf, SCMContext scmContext, Clock clock) { this.pipelineManager = pipelineManager; diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/ECPipelineProvider.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/ECPipelineProvider.java index b28a435de058..b1698628514a 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/ECPipelineProvider.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/ECPipelineProvider.java @@ -106,7 +106,6 @@ protected Pipeline create(ECReplicationConfig replicationConfig, return dead != 0 ? dead : left.getOperationalState().compareTo(right.getOperationalState()); }; - @Override public Pipeline createForRead( ECReplicationConfig replicationConfig, diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/PipelineManager.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/PipelineManager.java index 6b3ab97505af..29566e78f68d 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/PipelineManager.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/PipelineManager.java @@ -49,7 +49,6 @@ Pipeline buildECPipeline(ReplicationConfig replicationConfig, void addEcPipeline(Pipeline pipeline) throws IOException; - Pipeline createPipeline( ReplicationConfig replicationConfig, List nodes diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/PipelineReportHandler.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/PipelineReportHandler.java index b7c295ef81aa..dc3b01682331 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/PipelineReportHandler.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/PipelineReportHandler.java @@ -141,7 +141,6 @@ protected void processPipelineReport(PipelineReport report, } } - protected void setReportedDatanode(Pipeline pipeline, DatanodeDetails dn) throws IOException { pipeline.reportDatanode(dn); diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/PipelineStateManagerImpl.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/PipelineStateManagerImpl.java index 89a7b25c724b..28d213567c75 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/PipelineStateManagerImpl.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/PipelineStateManagerImpl.java @@ -148,7 +148,6 @@ public List getPipelines() { } } - @Override public List getPipelines( ReplicationConfig replicationConfig) { diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/WritableRatisContainerProvider.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/WritableRatisContainerProvider.java index 2c15d1013e3d..a61b32892352 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/WritableRatisContainerProvider.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/WritableRatisContainerProvider.java @@ -53,7 +53,6 @@ public WritableRatisContainerProvider( this.pipelineChoosePolicy = pipelineChoosePolicy; } - @Override public ContainerInfo getContainer(final long size, ReplicationConfig repConfig, String owner, ExcludeList excludeList) diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/choose/algorithms/CapacityPipelineChoosePolicy.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/choose/algorithms/CapacityPipelineChoosePolicy.java index 89def07d6499..7497eb58f949 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/choose/algorithms/CapacityPipelineChoosePolicy.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/choose/algorithms/CapacityPipelineChoosePolicy.java @@ -108,6 +108,7 @@ static class CapacityPipelineComparator implements Comparator { CapacityPipelineComparator(CapacityPipelineChoosePolicy policy) { this.policy = policy; } + @Override public int compare(Pipeline p1, Pipeline p2) { if (p1.getId().equals(p2.getId())) { diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/leader/choose/algorithms/LeaderChoosePolicyFactory.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/leader/choose/algorithms/LeaderChoosePolicyFactory.java index 187a41638bfc..7d091a912fe4 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/leader/choose/algorithms/LeaderChoosePolicyFactory.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/pipeline/leader/choose/algorithms/LeaderChoosePolicyFactory.java @@ -41,7 +41,6 @@ public final class LeaderChoosePolicyFactory { private LeaderChoosePolicyFactory() { } - public static LeaderChoosePolicy getPolicy( ConfigurationSource conf, final NodeManager nodeManager, final PipelineStateManager pipelineStateManager) throws SCMException { diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/DataNodeSafeModeRule.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/DataNodeSafeModeRule.java index 081a9d02308d..cb487ce8a3f4 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/DataNodeSafeModeRule.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/DataNodeSafeModeRule.java @@ -95,7 +95,6 @@ public String getStatusText() { this.registeredDns, this.requiredDns); } - @Override public void refresh(boolean forceRefresh) { // Do nothing. diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/HealthyPipelineSafeModeRule.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/HealthyPipelineSafeModeRule.java index 21d09fffc14b..abc7555b98e3 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/HealthyPipelineSafeModeRule.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/HealthyPipelineSafeModeRule.java @@ -146,7 +146,6 @@ protected synchronized void process(Pipeline pipeline) { } } - @Override public synchronized void refresh(boolean forceRefresh) { if (forceRefresh) { @@ -184,7 +183,6 @@ private synchronized void initializeRule(boolean refresh) { healthyPipelineThresholdCount); } - @Override protected synchronized void cleanup() { processedPipelineIDs.clear(); diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/OneReplicaPipelineSafeModeRule.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/OneReplicaPipelineSafeModeRule.java index 0b7adc7c4ec3..fb95215bff21 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/OneReplicaPipelineSafeModeRule.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/OneReplicaPipelineSafeModeRule.java @@ -57,7 +57,6 @@ public class OneReplicaPipelineSafeModeRule extends private PipelineManager pipelineManager; private final double pipelinePercent; - public OneReplicaPipelineSafeModeRule(EventQueue eventQueue, PipelineManager pipelineManager, SCMSafeModeManager safeModeManager, ConfigurationSource configuration) { super(safeModeManager, NAME, eventQueue); diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/RatisContainerSafeModeRule.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/RatisContainerSafeModeRule.java index 4015fd81b4df..5f995baed1d8 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/RatisContainerSafeModeRule.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/RatisContainerSafeModeRule.java @@ -41,7 +41,6 @@ import org.slf4j.Logger; import org.slf4j.LoggerFactory; - /** * Class defining Safe mode exit criteria for Ratis Containers. */ diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/SCMSafeModeManager.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/SCMSafeModeManager.java index 9dd79ca81595..248103a48ead 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/SCMSafeModeManager.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/SCMSafeModeManager.java @@ -173,7 +173,6 @@ public void emitSafeModeStatus() { } } - public synchronized void validateSafeModeExitRules(String ruleName, EventPublisher eventQueue) { @@ -277,6 +276,7 @@ public boolean getInSafeMode() { } return inSafeMode.get(); } + /** * Get the safe mode status of all rules. * @@ -356,7 +356,6 @@ public DataNodeSafeModeRule getDataNodeSafeModeRule() { return (DataNodeSafeModeRule) exitRules.get(DN_EXIT_RULE); } - /** * Class used during SafeMode status event. */ diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/SafeModeMetrics.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/SafeModeMetrics.java index bcdfa81425f9..d103e1f1fca0 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/SafeModeMetrics.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/SafeModeMetrics.java @@ -31,7 +31,6 @@ public class SafeModeMetrics { private static final String SOURCE_NAME = SafeModeMetrics.class.getSimpleName(); - // These all values will be set to some values when safemode is enabled. private @Metric MutableGaugeLong numContainerWithOneReplicaReportedThreshold; @@ -118,7 +117,6 @@ MutableCounterLong getCurrentContainersWithOneReplicaReportedCount() { return currentContainersWithOneReplicaReportedCount; } - public void unRegister() { MetricsSystem ms = DefaultMetricsSystem.instance(); ms.unregisterSource(SOURCE_NAME); diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/SafeModeRuleFactory.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/SafeModeRuleFactory.java index 96c22d06e7db..8d866d748adf 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/SafeModeRuleFactory.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/safemode/SafeModeRuleFactory.java @@ -31,7 +31,6 @@ */ public final class SafeModeRuleFactory { - private final ConfigurationSource config; private final SCMContext scmContext; private final EventQueue eventQueue; diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/security/RootCARotationManager.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/security/RootCARotationManager.java index 95fc30319771..e20cf22cb18e 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/security/RootCARotationManager.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/security/RootCARotationManager.java @@ -485,6 +485,7 @@ private void cleanupAndStop(String reason) { isProcessing.set(false); processStartTime.set(null); } + /** * Calculate time before root certificate will enter 2 * expiry grace period. * @return Duration, time before certificate enters the 2 * grace diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/security/SecretKeyManagerService.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/security/SecretKeyManagerService.java index 9b540d433f32..77c44ccb3ed0 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/security/SecretKeyManagerService.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/security/SecretKeyManagerService.java @@ -52,7 +52,6 @@ public class SecretKeyManagerService implements SCMService, Runnable { private final SecretKeyManager secretKeyManager; private final SecretKeyConfig secretKeyConfig; - /** * SCMService related variables. */ diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMCertStore.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMCertStore.java index ad5ae52cb14c..da5597bf41e6 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMCertStore.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMCertStore.java @@ -213,7 +213,6 @@ public static class Builder { private SCMMetadataStore metadataStore; private SCMRatisServer scmRatisServer; - public Builder setMetadaStore(SCMMetadataStore scmMetadataStore) { this.metadataStore = scmMetadataStore; return this; diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMClientProtocolServer.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMClientProtocolServer.java index 29ff82576eb4..7bc437627aa5 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMClientProtocolServer.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMClientProtocolServer.java @@ -1543,7 +1543,6 @@ public boolean getSafeModeStatus() { return scm.getScmContext().isInSafeMode(); } - /** * Query the System for Nodes. * diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeHeartbeatDispatcher.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeHeartbeatDispatcher.java index 56590d7122e4..dc79a21f8f99 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeHeartbeatDispatcher.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeHeartbeatDispatcher.java @@ -65,7 +65,6 @@ public final class SCMDatanodeHeartbeatDispatcher { private final NodeManager nodeManager; private final EventPublisher eventPublisher; - public SCMDatanodeHeartbeatDispatcher(NodeManager nodeManager, EventPublisher eventPublisher) { Preconditions.checkNotNull(nodeManager); @@ -74,7 +73,6 @@ public SCMDatanodeHeartbeatDispatcher(NodeManager nodeManager, this.eventPublisher = eventPublisher; } - /** * Dispatches heartbeat to registered event handlers. * @@ -242,6 +240,7 @@ public static class CommandQueueReportFromDatanode extends ReportFromDatanode { private final Map commandsToBeSent; + public CommandQueueReportFromDatanode(DatanodeDetails datanodeDetails, CommandQueueReportProto report, Map commandsToBeSent) { @@ -271,7 +270,9 @@ public LayoutReportFromDatanode(DatanodeDetails datanodeDetails, */ public interface ContainerReport { DatanodeDetails getDatanodeDetails(); + ContainerReportType getType(); + void mergeReport(ContainerReport val); } diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeProtocolServer.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeProtocolServer.java index 50b6d25f079b..df615eec3bc1 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeProtocolServer.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMDatanodeProtocolServer.java @@ -437,7 +437,6 @@ public static SCMCommandProto getCommandResponse(SCMCommand cmd, } } - public void join() throws InterruptedException { LOG.trace("Join RPC server for DataNodes"); datanodeRpcServer.join(); diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMStarterInterface.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMStarterInterface.java index a5912c8ce615..33cab486a2bc 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMStarterInterface.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/SCMStarterInterface.java @@ -28,9 +28,12 @@ public interface SCMStarterInterface { void start(OzoneConfiguration conf) throws Exception; + boolean init(OzoneConfiguration conf, String clusterId) throws IOException; + boolean bootStrap(OzoneConfiguration conf) throws IOException, AuthenticationException; + String generateClusterId(); } diff --git a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/StorageContainerManager.java b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/StorageContainerManager.java index 1e1eb4cbe2eb..04bf80366ed2 100644 --- a/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/StorageContainerManager.java +++ b/hadoop-hdds/server-scm/src/main/java/org/apache/hadoop/hdds/scm/server/StorageContainerManager.java @@ -1378,12 +1378,14 @@ public static void initMetrics() { public static SCMMetrics getMetrics() { return metrics == null ? SCMMetrics.create() : metrics; } + /** * Initialize SCMPerformance metrics. */ public static void initPerfMetrics() { perfMetrics = SCMPerformanceMetrics.create(); } + /** * Return SCMPerformance metrics instance. */ @@ -2013,6 +2015,7 @@ public StatefulServiceStateManager getStatefulServiceStateManager() { public String getNamespace() { return scmHANodeDetails.getLocalNodeDetails().getServiceId(); } + /** * Get the safe mode status of all rules. * diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/HddsTestUtils.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/HddsTestUtils.java index 4029b49172d8..ba6597500efe 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/HddsTestUtils.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/HddsTestUtils.java @@ -258,18 +258,19 @@ public static StorageReportProto createStorageReport(DatanodeID nodeId, String p return createStorageReport(nodeId, path, capacity, used, remaining, type, false); } - /** - * Creates storage report with the given information. - * - * @param nodeId datanode id - * @param path storage dir - * @param capacity storage size - * @param used space used - * @param remaining space remaining - * @param type type of storage - * - * @return StorageReportProto - */ + + /** + * Creates storage report with the given information. + * + * @param nodeId datanode id + * @param path storage dir + * @param capacity storage size + * @param used space used + * @param remaining space remaining + * @param type type of storage + * + * @return StorageReportProto + */ public static StorageReportProto createStorageReport(DatanodeID nodeId, String path, long capacity, long used, long remaining, StorageTypeProto type, boolean failed) { @@ -357,7 +358,6 @@ public static ContainerReportsProto getRandomContainerReports( return getContainerReports(containerInfos); } - public static PipelineReportsProto getRandomPipelineReports() { return PipelineReportsProto.newBuilder().build(); } @@ -632,7 +632,6 @@ private static ContainerInfo.Builder getDefaultContainerInfoBuilder( .setOwner("TEST"); } - public static ContainerInfo getContainer( final HddsProtos.LifeCycleState state) { return getDefaultContainerInfoBuilder(state) @@ -756,8 +755,6 @@ public static Set getReplicasWithReplicaIndex( usedBytes, keyCount, sequenceId, Arrays.asList(datanodeDetails))); } - - public static Pipeline getRandomPipeline() { List nodes = new ArrayList<>(); nodes.add(MockDatanodeDetails.randomDatanodeDetails()); diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/TestSCMCommonPlacementPolicy.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/TestSCMCommonPlacementPolicy.java index d607d83e748d..d1a2f6e65763 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/TestSCMCommonPlacementPolicy.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/TestSCMCommonPlacementPolicy.java @@ -525,7 +525,6 @@ private static class DummyPlacementPolicy extends SCMCommonPlacementPolicy { private List racks; private int rackCnt; - /** * Creates Dummy Placement Policy with dn index to rack Mapping * in round robin fashion (rack Index = dn Index % total number of racks). @@ -560,8 +559,6 @@ private static class DummyPlacementPolicy extends SCMCommonPlacementPolicy { entry -> racks.get(entry.getValue()))); } - - @Override public DatanodeDetails chooseNode(List healthyNodes) { return healthyNodes.get(0); diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestBlockManager.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestBlockManager.java index 05eec8981bad..eb2fb3437bdc 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestBlockManager.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestBlockManager.java @@ -324,7 +324,6 @@ void testBlockDistribution() throws Exception { }); } - @Test void testBlockDistributionWithMultipleDisks() throws Exception { int threadCount = numContainerPerOwnerInPipeline * @@ -454,7 +453,6 @@ public void testAllocateOversizedBlock() { t.getMessage()); } - @Test public void testAllocateBlockFailureInSafeMode() { scm.getScmContext().updateSafeModeStatus( diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestDeletedBlockLog.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestDeletedBlockLog.java index f37fa1ac10ff..4725b30ab8f8 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestDeletedBlockLog.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestDeletedBlockLog.java @@ -434,7 +434,6 @@ public void testResetCount() throws Exception { assertEquals(30 * THREE, blocks.size()); } - @Test public void testSCMDelIteratorProgress() throws Exception { int maxRetry = conf.getInt(OZONE_SCM_BLOCK_DELETION_MAX_RETRY, 20); @@ -571,7 +570,6 @@ private void assertNoDuplicateTransactions( } } - private void assertContainsAllTransactions( DatanodeDeletedBlockTransactions transactions1, DatanodeDeletedBlockTransactions transactions2) { diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerReportHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerReportHandler.java index a3b8f2751c0f..629ef75916f1 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerReportHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerReportHandler.java @@ -303,7 +303,6 @@ public void testUnderReplicatedContainer() datanodeOne, datanodeTwo, datanodeThree) .forEach(containerStateManager::updateContainerReplica); - // SCM expects both containerOne and containerTwo to be in all the three // datanodes datanodeOne, datanodeTwo and datanodeThree @@ -360,7 +359,6 @@ public void testOverReplicatedContainer() throws NodeNotFoundException, datanodeOne, datanodeTwo, datanodeThree) .forEach(containerStateManager::updateContainerReplica); - // SCM expects both containerOne and containerTwo to be in all the three // datanodes datanodeOne, datanodeTwo and datanodeThree @@ -379,7 +377,6 @@ public void testOverReplicatedContainer() throws NodeNotFoundException, containerOne.containerID()).size()); } - @Test public void testClosingToClosed() throws NodeNotFoundException, IOException, TimeoutException { @@ -1231,7 +1228,6 @@ protected static ContainerReportsProto getContainerReportsProto( 2000000000L, 100000000L, bcsId, replicaIndex); } - protected static ContainerReportsProto getContainerReportsProto( final ContainerID containerId, final ContainerReplicaProto.State state, final String originNodeId, final long usedBytes, final long keyCount, diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/TestIncrementalContainerReportHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/TestIncrementalContainerReportHandler.java index 502c668cb159..976c4d697834 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/TestIncrementalContainerReportHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/TestIncrementalContainerReportHandler.java @@ -192,7 +192,6 @@ public void tearDown() throws Exception { } } - @Test public void testClosingToClosed() throws IOException, TimeoutException { final IncrementalContainerReportHandler reportHandler = diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/balancer/TestContainerBalancerDatanodeNodeLimit.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/balancer/TestContainerBalancerDatanodeNodeLimit.java index d5f30935ecfa..872a7469d6a9 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/balancer/TestContainerBalancerDatanodeNodeLimit.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/balancer/TestContainerBalancerDatanodeNodeLimit.java @@ -427,7 +427,6 @@ public void targetDatanodeShouldBeInServiceHealthy(@Nonnull MockedSCM mockedSCM) } } - @ParameterizedTest(name = "MockedSCM #{index}: {0}") @MethodSource("createMockedSCMs") public void selectedContainerShouldNotAlreadyHaveBeenSelected(@Nonnull MockedSCM mockedSCM) diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/TestSCMContainerPlacementRackAware.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/TestSCMContainerPlacementRackAware.java index 9497abfdee36..c1b5b78a804e 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/TestSCMContainerPlacementRackAware.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/TestSCMContainerPlacementRackAware.java @@ -86,7 +86,6 @@ public class TestSCMContainerPlacementRackAware { private SCMContainerPlacementMetrics metrics; private static final int NODE_PER_RACK = 5; - private static IntStream numDatanodes() { return IntStream.rangeClosed(3, 15); } @@ -814,6 +813,7 @@ public void chooseNodeWithOnlyExcludeAndNoUsedNodes(int datanodeCount) datanodeDetails.get(0).getUuid() && excludedNodes.get(1).getUuid() != datanodeDetails.get(1).getUuid()); } + @ParameterizedTest @ValueSource(ints = {11, 12, 13, 14, 15}) public void testNoFallbackWithUsedNodes(int datanodeCount) { diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/TestSCMContainerPlacementRackScatter.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/TestSCMContainerPlacementRackScatter.java index dbb9be27a085..ce66bf9f68c7 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/TestSCMContainerPlacementRackScatter.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/TestSCMContainerPlacementRackScatter.java @@ -893,7 +893,6 @@ public void testAllNodesOnRackExcludedReducesRackCount2() assertEquals(1, chosenNodes.size()); } - private int getRackSize(List... datanodeDetails) { Set racks = new HashSet<>(); for (List list : datanodeDetails) { diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationTestUtil.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationTestUtil.java index d8949df81871..999aadaa6e90 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationTestUtil.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/ReplicationTestUtil.java @@ -138,6 +138,7 @@ public static Set createReplicasWithSameOrigin( } return replicas; } + public static ContainerReplica createEmptyContainerReplica(ContainerID containerID, int replicaIndex, HddsProtos.NodeOperationalState opState, ContainerReplicaProto.State replicaState) { diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestContainerReplicaPendingOps.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestContainerReplicaPendingOps.java index 0f546a4e3ec4..d645beb62406 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestContainerReplicaPendingOps.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestContainerReplicaPendingOps.java @@ -447,6 +447,4 @@ public void subscribersShouldNotBeNotifiedWhenReplacingAnOpWithDuplicate() { // subscriber verifyNoMoreInteractions(subscriber1); } - - } diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestECMisReplicationHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestECMisReplicationHandler.java index e00847b66021..363097ce4001 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestECMisReplicationHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestECMisReplicationHandler.java @@ -61,7 +61,6 @@ public class TestECMisReplicationHandler extends TestMisReplicationHandler { private static final int DATA = 3; private static final int PARITY = 2; - @BeforeEach void setup(@TempDir File testDir) throws NodeNotFoundException, CommandTargetOverloadedException, NotLeaderException { diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestMisReplicationHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestMisReplicationHandler.java index 43824f8886ca..59ffab79c3df 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestMisReplicationHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestMisReplicationHandler.java @@ -132,6 +132,7 @@ static PlacementPolicy mockPlacementPolicy() { protected abstract MisReplicationHandler getMisreplicationHandler( PlacementPolicy placementPolicy, OzoneConfiguration configuration, ReplicationManager rm); + protected void testMisReplication(Set availableReplicas, List pendingOp, int maintenanceCnt, int misreplicationCount, diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestQuasiClosedStuckOverReplicationHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestQuasiClosedStuckOverReplicationHandler.java index 6d6a342f5f75..5b4b64537942 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestQuasiClosedStuckOverReplicationHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestQuasiClosedStuckOverReplicationHandler.java @@ -160,7 +160,6 @@ public void testOverloadedExceptionContinuesAndThrows() throws NotLeaderExceptio assertEquals(1, commandsSent.size()); } - private ContainerHealthResult.OverReplicatedHealthResult getOverReplicatedHealthResult() { ContainerHealthResult.OverReplicatedHealthResult healthResult = mock(ContainerHealthResult.OverReplicatedHealthResult.class); diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestQuasiClosedStuckUnderReplicationHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestQuasiClosedStuckUnderReplicationHandler.java index a43a04fa26d7..0e32e8220c4b 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestQuasiClosedStuckUnderReplicationHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestQuasiClosedStuckUnderReplicationHandler.java @@ -70,7 +70,6 @@ public class TestQuasiClosedStuckUnderReplicationHandler { private Set>> commandsSent; private QuasiClosedStuckUnderReplicationHandler handler; - @BeforeEach void setup(@TempDir File testDir) throws NodeNotFoundException, CommandTargetOverloadedException, NotLeaderException { diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestRatisMisReplicationHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestRatisMisReplicationHandler.java index 4e5d64f8d20d..88889da0b269 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestRatisMisReplicationHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestRatisMisReplicationHandler.java @@ -173,7 +173,6 @@ public void testMisReplicationWithPendingOps() pendingOp, 0, 1, 0); } - @Test public void testAllSourcesOverloaded() throws IOException { ReplicationManager replicationManager = getReplicationManager(); diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestRatisOverReplicationHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestRatisOverReplicationHandler.java index ebaad611d056..6c4fcd248eef 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestRatisOverReplicationHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/TestRatisOverReplicationHandler.java @@ -414,6 +414,7 @@ public void testPerfectlyReplicatedContainer() throws IOException { testProcessing(replicas, pendingOps, getOverReplicatedHealthResult(), 0); } + @Test public void testOverReplicationOfQuasiClosedReplicaWithWrongSequenceID() throws IOException { diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestDeletingContainerHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestDeletingContainerHandler.java index 10aadc687ca8..9a3a12c73dee 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestDeletingContainerHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestDeletingContainerHandler.java @@ -63,7 +63,6 @@ public class TestDeletingContainerHandler { private ECReplicationConfig ecReplicationConfig; private RatisReplicationConfig ratisReplicationConfig; - @BeforeEach public void setup() throws IOException { @@ -134,7 +133,6 @@ public void testCleanupIfNoReplicaExist() { cleanupIfNoReplicaExist(ecReplicationConfig, 1); } - private void cleanupIfNoReplicaExist( ReplicationConfig replicationConfig, int times) { clearInvocations(replicationManager); diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestOpenContainerHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestOpenContainerHandler.java index 14c9c4b5161f..5ebe2e8dc73c 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestOpenContainerHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestOpenContainerHandler.java @@ -152,6 +152,7 @@ public void testOpenContainerWithoutPipelineIsClosed() { .sendCloseContainerEvent(containerInfo.containerID()); assertEquals(1, request.getReport().getStat(HealthState.OPEN_WITHOUT_PIPELINE)); } + @Test public void testClosedRatisContainerReturnsFalse() { ContainerInfo containerInfo = ReplicationTestUtil.createContainerInfo( diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestQuasiClosedStuckReplicationCheck.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestQuasiClosedStuckReplicationCheck.java index 6a6dd84243d4..7fc7058577f3 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestQuasiClosedStuckReplicationCheck.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestQuasiClosedStuckReplicationCheck.java @@ -45,7 +45,6 @@ import org.junit.jupiter.api.BeforeEach; import org.junit.jupiter.api.Test; - /** * Tests for the QuasiClosedStuckReplicationCheck class. */ diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestRatisUnhealthyReplicationCheckHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestRatisUnhealthyReplicationCheckHandler.java index cb89cbb862bc..1499b43f69ce 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestRatisUnhealthyReplicationCheckHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/replication/health/TestRatisUnhealthyReplicationCheckHandler.java @@ -390,6 +390,4 @@ public void testUnderReplicatedWithQuasiClosedReplicasHavingLessSequenceId() { assertEquals(1, report.getStat(ReplicationManagerReport.HealthState.UNHEALTHY)); } - - } diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMConfiguration.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMConfiguration.java index 601cd8502c17..15715d72b618 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMConfiguration.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMConfiguration.java @@ -211,7 +211,6 @@ public void testSCMConfig() throws Exception { RaftServerConfigKeys.Log.Appender.WAIT_TIME_MIN_KEY); } - @Test public void testSamePortConfig() throws Exception { String scmServiceId = "scmserviceId"; diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestNodeDecommissionManager.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestNodeDecommissionManager.java index 63ac6b05f8f0..87616aff9fa0 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestNodeDecommissionManager.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestNodeDecommissionManager.java @@ -101,6 +101,7 @@ private ContainerInfo createMockContainer(ReplicationConfig rep, String owner) { id++; return builder.build(); } + private ContainerInfo getMockContainer(ReplicationConfig rep, ContainerID conId) { ContainerInfo.Builder builder = new ContainerInfo.Builder() .setReplicationConfig(rep) diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestNodeDecommissionMetrics.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestNodeDecommissionMetrics.java index e60de693fca6..3003f7452b75 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestNodeDecommissionMetrics.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/node/TestNodeDecommissionMetrics.java @@ -55,7 +55,6 @@ public class TestNodeDecommissionMetrics { private ReplicationManager repManager; private EventQueue eventQueue; - @BeforeEach public void setup() { conf = new OzoneConfiguration(); diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestECPipelineProvider.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestECPipelineProvider.java index 6aecd2b5fdd3..ae7eb60c4754 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestECPipelineProvider.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestECPipelineProvider.java @@ -70,6 +70,7 @@ public class TestECPipelineProvider { mock(PipelineStateManager.class); private PlacementPolicy placementPolicy = mock(PlacementPolicy.class); private long containerSizeBytes; + @BeforeEach public void setup() throws IOException, NodeNotFoundException { conf = new OzoneConfiguration(); @@ -95,7 +96,6 @@ public void setup() throws IOException, NodeNotFoundException { .thenReturn(NodeStatus.inServiceHealthy()); } - @Test public void testSimplePipelineCanBeCreatedWithIndexes() throws IOException { ECReplicationConfig ecConf = new ECReplicationConfig(3, 2); diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelineActionHandler.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelineActionHandler.java index da837db16502..3f231caec248 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelineActionHandler.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelineActionHandler.java @@ -42,7 +42,6 @@ */ public class TestPipelineActionHandler { - @Test public void testPipelineActionHandlerForValidPipeline() throws IOException { diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelineStateMap.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelineStateMap.java index a4e96a54f0fd..446d6b426085 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelineStateMap.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelineStateMap.java @@ -93,6 +93,4 @@ public void testCountPipelines() throws IOException { assertEquals(1, map.getPipelineCount(new ECReplicationConfig(3, 2), Pipeline.PipelineState.CLOSED)); } - - } diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/choose/algorithms/TestPipelineChoosePolicyFactory.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/choose/algorithms/TestPipelineChoosePolicyFactory.java index 1883d4cae683..bede4413d3ee 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/choose/algorithms/TestPipelineChoosePolicyFactory.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/pipeline/choose/algorithms/TestPipelineChoosePolicyFactory.java @@ -77,7 +77,6 @@ public void testNonDefaultPolicyEC() throws IOException { assertSame(DummyGoodImpl.class, policy.getClass()); } - /** * A dummy pipeline choose policy implementation for test with an invalid * constructor. @@ -118,7 +117,6 @@ public int choosePipelineIndex(List pipelineList, } } - @Test public void testConstructorNotFound() throws SCMException { // set a policy class which does't have the right constructor implemented diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestHealthyPipelineSafeModeRule.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestHealthyPipelineSafeModeRule.java index aa71edcc68f9..8ffdba407feb 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestHealthyPipelineSafeModeRule.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestHealthyPipelineSafeModeRule.java @@ -200,7 +200,6 @@ public void testHealthyPipelineSafeModeRuleWithPipelines() throws Exception { } } - @Test public void testHealthyPipelineSafeModeRuleWithMixedPipelines() throws Exception { diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestOneReplicaPipelineSafeModeRule.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestOneReplicaPipelineSafeModeRule.java index c89c8762b03d..8d5e005ae02e 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestOneReplicaPipelineSafeModeRule.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestOneReplicaPipelineSafeModeRule.java @@ -153,7 +153,6 @@ public void testOneReplicaPipelineRule() throws Exception { GenericTestUtils.waitFor(() -> rule.validate(), 1000, 5000); } - @Test public void testOneReplicaPipelineRuleMixedPipelines() throws Exception { diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestSCMSafeModeManager.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestSCMSafeModeManager.java index 7a43792eb5db..f43a7047c0ec 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestSCMSafeModeManager.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestSCMSafeModeManager.java @@ -205,7 +205,6 @@ public void testSafeModeExitRule() throws Exception { 100, 1000 * 5); } - private OzoneConfiguration createConf(double healthyPercent, double oneReplicaPercent) { OzoneConfiguration conf = new OzoneConfiguration(config); @@ -432,7 +431,6 @@ private void firePipelineEvent(PipelineManager pipelineManager, } } - @Test public void testDisableSafeMode() { OzoneConfiguration conf = new OzoneConfiguration(config); diff --git a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/server/TestSCMDatanodeHeartbeatDispatcher.java b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/server/TestSCMDatanodeHeartbeatDispatcher.java index 5ac3236a39ee..adf1a98c4be2 100644 --- a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/server/TestSCMDatanodeHeartbeatDispatcher.java +++ b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/server/TestSCMDatanodeHeartbeatDispatcher.java @@ -51,7 +51,6 @@ */ public class TestSCMDatanodeHeartbeatDispatcher { - @Test public void testNodeReportDispatcher() throws IOException { diff --git a/hadoop-hdds/test-utils/src/test/java/org/apache/ozone/test/GenericTestUtils.java b/hadoop-hdds/test-utils/src/test/java/org/apache/ozone/test/GenericTestUtils.java index 78d00712a544..71d116b55c17 100644 --- a/hadoop-hdds/test-utils/src/test/java/org/apache/ozone/test/GenericTestUtils.java +++ b/hadoop-hdds/test-utils/src/test/java/org/apache/ozone/test/GenericTestUtils.java @@ -245,6 +245,7 @@ public void clearOutput() { writer().getBuffer().setLength(0); } } + @Deprecated public static Logger toLog4j(org.slf4j.Logger logger) { return LogManager.getLogger(logger.getName()); diff --git a/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/ScmOption.java b/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/ScmOption.java index 57f871369006..640433c99b3f 100644 --- a/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/ScmOption.java +++ b/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/ScmOption.java @@ -41,7 +41,6 @@ public class ScmOption extends AbstractMixin { description = "The destination scm (host:port)") private String scm; - @CommandLine.Option(names = {"--service-id", "-id"}, description = "ServiceId of SCM HA Cluster") private String scmServiceId; diff --git a/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/cert/ListSubcommand.java b/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/cert/ListSubcommand.java index 8b0d4c625764..15c7b50660fd 100644 --- a/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/cert/ListSubcommand.java +++ b/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/cert/ListSubcommand.java @@ -131,7 +131,7 @@ private static class Certificate { private String validFrom; private String expiry; private Map subjectDN = new LinkedHashMap<>(); - private Map issuerDN = new LinkedHashMap<>();; + private Map issuerDN = new LinkedHashMap<>(); Certificate(X509Certificate cert) { serialNumber = cert.getSerialNumber(); diff --git a/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/container/ListSubcommand.java b/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/container/ListSubcommand.java index 57053f76e08f..ba82c8c14842 100644 --- a/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/container/ListSubcommand.java +++ b/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/container/ListSubcommand.java @@ -182,7 +182,6 @@ private void listAllContainers(ScmClient scmClient, SequenceWriter writer, } while (fetchedCount > 0); } - private static class NonClosingOutputStream extends OutputStream { private final OutputStream delegate; diff --git a/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/ListInfoSubcommand.java b/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/ListInfoSubcommand.java index e0c6b158ac5d..929873c65803 100644 --- a/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/ListInfoSubcommand.java +++ b/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/ListInfoSubcommand.java @@ -78,7 +78,6 @@ public class ListInfoSubcommand extends ScmSubcommand { private List pipelines; - @Override public void execute(ScmClient scmClient) throws IOException { pipelines = scmClient.listPipelines(); diff --git a/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/MaintenanceSubCommand.java b/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/MaintenanceSubCommand.java index baa60aebe0d7..60adb11c35b8 100644 --- a/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/MaintenanceSubCommand.java +++ b/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/MaintenanceSubCommand.java @@ -51,7 +51,6 @@ public class MaintenanceSubCommand extends ScmSubcommand { description = "Forcefully try to decommission the datanode(s)") private boolean force; - @Override public void execute(ScmClient scmClient) throws IOException { List hosts = hostNameParams.getHostNames(); diff --git a/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/UsageInfoSubcommand.java b/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/UsageInfoSubcommand.java index b196a6008d4a..514be5a62065 100644 --- a/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/UsageInfoSubcommand.java +++ b/hadoop-hdds/tools/src/main/java/org/apache/hadoop/hdds/scm/cli/datanode/UsageInfoSubcommand.java @@ -51,6 +51,7 @@ public class UsageInfoSubcommand extends ScmSubcommand { private static final NumberFormat PERCENT_FORMAT = NumberFormat.getPercentInstance(); + static { PERCENT_FORMAT.setMinimumFractionDigits(2); PERCENT_FORMAT.setMaximumFractionDigits(2); @@ -90,7 +91,6 @@ private static class ExclusiveArguments { description = "Format output as JSON") private boolean json; - @Override public void execute(ScmClient scmClient) throws IOException { List infoList; @@ -241,9 +241,11 @@ public long getOzoneUsed() { public long getRemaining() { return remaining; } + public long getCommitted() { return committed; } + public long getFreeSpaceToSpare() { return freeSpaceToSpare; } diff --git a/hadoop-hdds/tools/src/test/java/org/apache/hadoop/hdds/scm/cli/container/TestInfoSubCommand.java b/hadoop-hdds/tools/src/test/java/org/apache/hadoop/hdds/scm/cli/container/TestInfoSubCommand.java index e284acf9edd3..072babf1f01b 100644 --- a/hadoop-hdds/tools/src/test/java/org/apache/hadoop/hdds/scm/cli/container/TestInfoSubCommand.java +++ b/hadoop-hdds/tools/src/test/java/org/apache/hadoop/hdds/scm/cli/container/TestInfoSubCommand.java @@ -165,7 +165,6 @@ public void testContainersCanBeReadFromStdinJson() validateJsonMultiOutput(); } - @Test public void testMultipleContainersCanBePassedJson() throws Exception { when(scmClient.getContainerReplicas(anyLong())).thenReturn(getReplicas(true)); @@ -307,7 +306,6 @@ private void testJsonOutput() throws IOException { assertTrue(matcher.matches()); } - private List getReplicas(boolean includeIndex) { List replicas = new ArrayList<>(); int index = 1; diff --git a/hadoop-hdds/tools/src/test/java/org/apache/hadoop/hdds/scm/cli/datanode/TestDecommissionStatusSubCommand.java b/hadoop-hdds/tools/src/test/java/org/apache/hadoop/hdds/scm/cli/datanode/TestDecommissionStatusSubCommand.java index 66ced0132a4c..f7660ab10011 100644 --- a/hadoop-hdds/tools/src/test/java/org/apache/hadoop/hdds/scm/cli/datanode/TestDecommissionStatusSubCommand.java +++ b/hadoop-hdds/tools/src/test/java/org/apache/hadoop/hdds/scm/cli/datanode/TestDecommissionStatusSubCommand.java @@ -236,7 +236,6 @@ public void testIpOptionDecommissionStatusFail() throws IOException { assertFalse(m.find()); } - private List getNodeDetails(int n) { List nodesList = new ArrayList<>(); diff --git a/hadoop-ozone/cli-shell/src/main/java/org/apache/hadoop/ozone/shell/snapshot/SnapshotDiffHandler.java b/hadoop-ozone/cli-shell/src/main/java/org/apache/hadoop/ozone/shell/snapshot/SnapshotDiffHandler.java index 76a5764cf567..6b864a58c4ef 100644 --- a/hadoop-ozone/cli-shell/src/main/java/org/apache/hadoop/ozone/shell/snapshot/SnapshotDiffHandler.java +++ b/hadoop-ozone/cli-shell/src/main/java/org/apache/hadoop/ozone/shell/snapshot/SnapshotDiffHandler.java @@ -73,7 +73,6 @@ public class SnapshotDiffHandler extends Handler { hidden = true) private boolean forceFullDiff; - @CommandLine.Option(names = {"-c", "--cancel"}, description = "Request to cancel a running SnapshotDiff job. " + "If the job is not IN_PROGRESS, the request will fail.", @@ -162,7 +161,6 @@ private ObjectNode getJsonObject(SnapshotDiffReportOzone diffReportOzone) { return diffReportNode; } - private ObjectNode getJsonObject(SnapshotDiffReport.DiffReportEntry diffReportEntry) { ObjectNode diffReportNode = JsonUtils.createObjectNode(null); diffReportNode.put("diffType", diffReportEntry.getType().getLabel()); diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneBucket.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneBucket.java index 057ae61c10aa..7f956b77df52 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneBucket.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneBucket.java @@ -1459,7 +1459,6 @@ private static OzoneKey toOzoneKey(OzoneFileStatusLight status) { Collections.emptyMap()); } - /** * An Iterator to iterate over {@link OzoneKey} list. * diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneClient.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneClient.java index 5fbb34ee1a55..3b9d91bf503c 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneClient.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneClient.java @@ -96,6 +96,7 @@ protected OzoneClient(ObjectStore objectStore, // For the unit test this.conf = new OzoneConfiguration(); } + /** * Returns the object store associated with the Ozone Cluster. * @return ObjectStore diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneKeyLocation.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneKeyLocation.java index 7d5f25e7ef4d..19cdca040e6d 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneKeyLocation.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneKeyLocation.java @@ -42,6 +42,7 @@ public class OzoneKeyLocation { * KeyOffset of this key. */ private final long keyOffset; + /** * Constructs OzoneKeyLocation. */ diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneVolume.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneVolume.java index 89d9be2cb896..d742ebe8d189 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneVolume.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/OzoneVolume.java @@ -177,6 +177,7 @@ public long getQuotaInBytes() { public long getQuotaInNamespace() { return quotaInNamespace; } + /** * Returns creation time of the volume. * diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/VolumeArgs.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/VolumeArgs.java index 2ec251c264f9..cc07e60b64e8 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/VolumeArgs.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/VolumeArgs.java @@ -152,6 +152,7 @@ public VolumeArgs.Builder addMetadata(String key, String value) { metadata.put(key, value); return this; } + public VolumeArgs.Builder addAcl(OzoneAcl acl) throws IOException { if (acls == null) { diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/checksum/BaseFileChecksumHelper.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/checksum/BaseFileChecksumHelper.java index 5da795b6d9cf..be63cb11e553 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/checksum/BaseFileChecksumHelper.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/checksum/BaseFileChecksumHelper.java @@ -195,7 +195,7 @@ protected String populateBlockChecksumBuf(ByteBuffer blockChecksumByteBuffer) th "Unknown combine mode: " + getCombineMode()); } return blockChecksumForDebug; - }; + } /** * Compute block checksums block by block and append the raw bytes of the diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/checksum/ECBlockChecksumComputer.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/checksum/ECBlockChecksumComputer.java index 66edadc3247d..0453b2734d1e 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/checksum/ECBlockChecksumComputer.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/checksum/ECBlockChecksumComputer.java @@ -45,7 +45,6 @@ public class ECBlockChecksumComputer extends AbstractBlockChecksumComputer { private final OmKeyInfo keyInfo; private final long blockLength; - public ECBlockChecksumComputer( List chunkInfoList, OmKeyInfo keyInfo, long blockLength) { this.chunkInfoList = chunkInfoList; diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockDataStreamOutputEntry.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockDataStreamOutputEntry.java index 74e936467723..53c072159070 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockDataStreamOutputEntry.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockDataStreamOutputEntry.java @@ -234,7 +234,6 @@ public Builder setPipeline(Pipeline ppln) { return this; } - public Builder setLength(long len) { this.length = len; return this; diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockDataStreamOutputEntryPool.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockDataStreamOutputEntryPool.java index a5adb0dc774e..77d1fc6fce33 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockDataStreamOutputEntryPool.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockDataStreamOutputEntryPool.java @@ -220,6 +220,7 @@ long getKeyLength() { return streamEntries.stream().mapToLong( BlockDataStreamOutputEntry::getCurrentPosition).sum(); } + /** * Contact OM to get a new block. Set the new block with the index (e.g. * first block has index = 0, second has index = 1 etc.) @@ -237,7 +238,6 @@ private void allocateNewBlock() throws IOException { addKeyLocationInfo(subKeyInfo); } - void commitKey(long offset) throws IOException { if (keyArgs != null) { // in test, this could be null diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockOutputStreamEntry.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockOutputStreamEntry.java index 2d9ed0c7637f..5fc0e92fb72b 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockOutputStreamEntry.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockOutputStreamEntry.java @@ -82,7 +82,6 @@ public class BlockOutputStreamEntry extends OutputStream { */ private AtomicInteger inflightCalls = new AtomicInteger(); - BlockOutputStreamEntry(Builder b) { this.config = b.config; this.outputStream = null; diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockOutputStreamEntryPool.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockOutputStreamEntryPool.java index d5a015238a2a..3ebf5133139a 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockOutputStreamEntryPool.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/BlockOutputStreamEntryPool.java @@ -285,6 +285,7 @@ synchronized long getKeyLength() { return streamEntries.stream() .mapToLong(BlockOutputStreamEntry::getCurrentPosition).sum(); } + /** * Contact OM to get a new block. Set the new block with the index (e.g. * first block has index = 0, second has index = 1 etc.) diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/ECKeyOutputStream.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/ECKeyOutputStream.java index 588bfc2a7f30..d712c2f4c0f7 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/ECKeyOutputStream.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/ECKeyOutputStream.java @@ -643,6 +643,7 @@ private static class EOFDummyStripe extends ECChunkBuffers { private static class CheckpointDummyStripe extends ECChunkBuffers { private final long version; + CheckpointDummyStripe(long version) { super(); this.version = version; diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/KeyDataStreamOutput.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/KeyDataStreamOutput.java index 4a524d80c34f..7e4b59ec7db1 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/KeyDataStreamOutput.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/KeyDataStreamOutput.java @@ -520,7 +520,6 @@ public Builder enableUnsafeByteBufferConversion(boolean enabled) { return this; } - public Builder setReplicationConfig(ReplicationConfig replConfig) { this.replicationConfig = replConfig; return this; diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/OzoneInputStream.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/OzoneInputStream.java index 5f175896b204..2366893ef562 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/OzoneInputStream.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/io/OzoneInputStream.java @@ -36,6 +36,7 @@ public class OzoneInputStream extends InputStream implements CanUnbuffer, public OzoneInputStream() { inputStream = null; } + /** * Constructs OzoneInputStream with KeyInputStream. * diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/protocol/ClientProtocol.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/protocol/ClientProtocol.java index a1155fd8d393..edbe74205bdb 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/protocol/ClientProtocol.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/protocol/ClientProtocol.java @@ -1112,7 +1112,6 @@ OzoneKey headObject(String volumeName, String bucketName, */ void setThreadLocalS3Auth(S3Auth s3Auth); - void setIsS3Request(boolean isS3Request); /** diff --git a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/rpc/RpcClient.java b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/rpc/RpcClient.java index e32a40d81e0a..f1e26ed87e58 100644 --- a/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/rpc/RpcClient.java +++ b/hadoop-ozone/client/src/main/java/org/apache/hadoop/ozone/client/rpc/RpcClient.java @@ -1704,7 +1704,6 @@ public void renameKeys(String volumeName, String bucketName, ozoneManagerClient.renameKeys(omRenameKeys); } - @Override public List listKeys(String volumeName, String bucketName, String keyPrefix, String prevKey, @@ -2456,6 +2455,7 @@ private OzoneInputStream createInputStream( new MultipartInputStream(keyInfo.getKeyName(), cryptoInputStreams)); } } + private OzoneDataStreamOutput createDataStreamOutput(OpenKeySession openKey) throws IOException { final ReplicationConfig replicationConfig diff --git a/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/MockOmTransport.java b/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/MockOmTransport.java index 47a7256d7ea6..146b79656bd3 100644 --- a/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/MockOmTransport.java +++ b/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/MockOmTransport.java @@ -208,7 +208,6 @@ private String toOperationString(CommitKeyRequest commitKeyRequest) { return "commit"; } - private CommitKeyResponse commitKey(CommitKeyRequest commitKeyRequest) { final KeyArgs keyArgs = commitKeyRequest.getKeyArgs(); final KeyInfo openKey = diff --git a/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/TestOzoneECClient.java b/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/TestOzoneECClient.java index 4425e8821521..f89b2c1d563b 100644 --- a/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/TestOzoneECClient.java +++ b/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/TestOzoneECClient.java @@ -274,6 +274,7 @@ private void testMultipleChunksInSingleWriteOp(int numChunks) throws IOException { testMultipleChunksInSingleWriteOp(0, numChunks, numChunks); } + private void testMultipleChunksInSingleWriteOp(int offset, int bufferChunks, int numChunks) throws IOException { @@ -1117,6 +1118,7 @@ private OzoneBucket writeIntoECKey(byte[] data, String key, DefaultReplicationConfig defaultReplicationConfig) throws IOException { return writeIntoECKey(0, data.length, data, key, defaultReplicationConfig); } + private OzoneBucket writeIntoECKey(int offset, int length, byte[] data, String key, DefaultReplicationConfig defaultReplicationConfig) throws IOException { diff --git a/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/rpc/TestRpcClient.java b/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/rpc/TestRpcClient.java index 272e4946e94f..4e4efef51e1f 100644 --- a/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/rpc/TestRpcClient.java +++ b/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/rpc/TestRpcClient.java @@ -29,6 +29,7 @@ import org.junit.jupiter.api.Test; import org.junit.jupiter.params.ParameterizedTest; import org.junit.jupiter.params.provider.EnumSource; + /** * Run RPC Client tests. */ diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OmUtils.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OmUtils.java index ea03c677a067..789d26c8880b 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OmUtils.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OmUtils.java @@ -834,7 +834,6 @@ public static String normalizePathUptoBucket(String path) { return normalizedPath.toString(); } - /** * For a given service ID, return list of configured OM hosts. * @param conf configuration diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OzoneAcl.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OzoneAcl.java index 338e4972739d..57ba7b3df691 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OzoneAcl.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/OzoneAcl.java @@ -97,7 +97,6 @@ private OzoneAcl(ACLIdentityType type, String name, AclScope scope, int acls) { BitSet.valueOf(getAclByteString().asReadOnlyByteBuffer()), getType().toString(), getAclScope())); } - private static int toInt(int aclTypeOrdinal) { return 1 << aclTypeOrdinal; } diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/conf/OMClientConfig.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/conf/OMClientConfig.java index 73ba52cd6be2..2f333d1f7606 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/conf/OMClientConfig.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/conf/OMClientConfig.java @@ -57,7 +57,6 @@ public class OMClientConfig { description = "Total number of threads in pool for the Trash Emptier") private int trashEmptierPoolSize = 5; - public int getTrashEmptierPoolSize() { return trashEmptierPoolSize; } diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/IOmMetadataReader.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/IOmMetadataReader.java index c9f6ec1606d0..df10fad74e6d 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/IOmMetadataReader.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/IOmMetadataReader.java @@ -99,6 +99,7 @@ default List listStatus(OmKeyArgs args, boolean recursive, * invalid arguments */ OzoneFileStatus getFileStatus(OmKeyArgs keyArgs) throws IOException; + /** * OzoneFS api to lookup for a file. * diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/OMConfigKeys.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/OMConfigKeys.java index b45fc26462f0..200c63028b75 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/OMConfigKeys.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/OMConfigKeys.java @@ -30,6 +30,7 @@ public final class OMConfigKeys { public static final String OZONE_OM_SNAPSHOT_LOAD_NATIVE_LIB = "ozone.om.snapshot.load.native.lib"; public static final boolean OZONE_OM_SNAPSHOT_LOAD_NATIVE_LIB_DEFAULT = true; + /** * Never constructed. */ diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/OmUpgradeConfig.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/OmUpgradeConfig.java index 526ac5fb6444..50ae907f2a73 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/OmUpgradeConfig.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/OmUpgradeConfig.java @@ -75,6 +75,4 @@ public int getDefaultLayoutVersionOnInit() { public void setDefaultLayoutVersionOnInit(int defaultLayoutVersionOnInit) { this.defaultLayoutVersionOnInit = defaultLayoutVersionOnInit; } - - } diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/exceptions/OMException.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/exceptions/OMException.java index cc540c15e3b8..660b21c45bcf 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/exceptions/OMException.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/exceptions/OMException.java @@ -98,6 +98,7 @@ public OMException.ResultCodes getResult() { public String toString() { return result + " " + super.toString(); } + /** * Error codes to make it easy to decode these exceptions. */ diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/ha/OMFailoverProxyProviderBase.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/ha/OMFailoverProxyProviderBase.java index d9e739587956..356d3fb1eff5 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/ha/OMFailoverProxyProviderBase.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/ha/OMFailoverProxyProviderBase.java @@ -148,7 +148,6 @@ protected T createOMProxy(InetSocketAddress omAddress) throws IOException { ).getProxy(); } - protected synchronized boolean shouldFailover(Exception ex) { Throwable unwrappedException = HddsUtils.getUnwrappedException(ex); if (unwrappedException instanceof AccessControlException || @@ -391,12 +390,10 @@ public synchronized long getWaitTime() { return waitBetweenRetries; } - public List getOMProxies() { return new ArrayList(omProxies.values()); } - public Map> getOMProxyMap() { return omProxies; } diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmBucketArgs.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmBucketArgs.java index b9a4ff39fbc8..f0c0c2deea61 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmBucketArgs.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmBucketArgs.java @@ -140,7 +140,6 @@ public boolean hasQuotaInNamespace() { return quotaInNamespaceSet; } - /** * Returns Bucket default replication config. */ @@ -223,6 +222,7 @@ public static class Builder extends WithMetadata.Builder { private BucketEncryptionKeyInfo bekInfo; private DefaultReplicationConfig defaultReplicationConfig; private String ownerName; + /** * Constructs a builder. */ diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmBucketInfo.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmBucketInfo.java index 42aef9de6781..c64f727300cb 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmBucketInfo.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmBucketInfo.java @@ -215,7 +215,6 @@ public long getModificationTime() { return modificationTime; } - /** * Returns bucket encryption key info. * @return bucket encryption key info @@ -250,7 +249,6 @@ public String getSourceBucket() { return sourceBucket; } - public long getUsedBytes() { return usedBytes; } @@ -588,7 +586,6 @@ public BucketInfo getProtobuf() { return bib.build(); } - /** * Parses BucketInfo protobuf and creates OmBucketInfo. * @param bucketInfo diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmDeleteKeys.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmDeleteKeys.java index 3cef7b1a830a..47372c5835dd 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmDeleteKeys.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmDeleteKeys.java @@ -29,7 +29,6 @@ public class OmDeleteKeys { private List keyNames; - public OmDeleteKeys(String volume, String bucket, List keyNames) { this.volume = volume; this.bucket = bucket; diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmKeyInfo.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmKeyInfo.java index bbb91808a530..f6301797a6fa 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmKeyInfo.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmKeyInfo.java @@ -805,7 +805,6 @@ public String getObjectInfo() { '}'; } - public boolean isKeyInfoSame(OmKeyInfo omKeyInfo, boolean checkPath, boolean checkKeyLocationVersions, boolean checkModificationTime, diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmMultipartAbortInfo.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmMultipartAbortInfo.java index f28a6d7c3a72..c398aec1c9b8 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmMultipartAbortInfo.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmMultipartAbortInfo.java @@ -88,7 +88,6 @@ public OmMultipartAbortInfo build() { } } - @Override public boolean equals(Object other) { if (this == other) { diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmRangerSyncArgs.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmRangerSyncArgs.java index 765de85b3573..95125182c9ff 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmRangerSyncArgs.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OmRangerSyncArgs.java @@ -46,6 +46,7 @@ public static OmRangerSyncArgs.Builder newBuilder() { */ public static class Builder { private long newServiceVersion; + /** * Constructs a builder. */ diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFSUtils.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFSUtils.java index 3a2adfa4f880..e4116876a3c1 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFSUtils.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/OzoneFSUtils.java @@ -201,7 +201,6 @@ public static boolean isAncestorPath(String parentKey, String childKey) { return childParent == parentParent; } - /** * Verifies whether the childKey is an immediate path under the given * parentKey. diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/RepeatedOmKeyInfo.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/RepeatedOmKeyInfo.java index 27911b71ceca..5461786838e5 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/RepeatedOmKeyInfo.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/RepeatedOmKeyInfo.java @@ -93,7 +93,6 @@ public ImmutablePair getTotalSize() { return new ImmutablePair(unreplicatedSize, replicatedSize); } - // HDDS-7041. Return a new ArrayList to avoid ConcurrentModifyException public List cloneOmKeyInfoList() { return new ArrayList<>(omKeyInfoList); diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/ServiceInfo.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/ServiceInfo.java index c8b723f802fd..728f8bda06f3 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/ServiceInfo.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/ServiceInfo.java @@ -35,7 +35,6 @@ */ public final class ServiceInfo { - /** * Type of node/service. */ diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/SnapshotInfo.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/SnapshotInfo.java index 39ae76254b1e..6dcb64026a97 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/SnapshotInfo.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/helpers/SnapshotInfo.java @@ -552,6 +552,7 @@ public static String getCheckpointDirName(UUID snapshotId) { "SnapshotId is needed to create checkpoint directory"); return SEPARATOR + snapshotId; } + /** * Get the name of the checkpoint directory, (non-static). */ diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/IOzoneManagerLock.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/IOzoneManagerLock.java index 6926b7d9bf23..ebe330c67ab1 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/IOzoneManagerLock.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/IOzoneManagerLock.java @@ -30,7 +30,6 @@ OMLockDetails acquireReadLock(OzoneManagerLock.Resource resource, OMLockDetails acquireReadLocks(OzoneManagerLock.Resource resource, Collection resources); - OMLockDetails acquireWriteLock(OzoneManagerLock.Resource resource, String... resources); diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/OMLockDetails.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/OMLockDetails.java index fdb2373b9d38..fcf25ff38370 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/OMLockDetails.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/OMLockDetails.java @@ -40,7 +40,6 @@ public OMLockDetails(boolean lockAcquired) { this.lockAcquired = lockAcquired; } - enum LockOpType { WAIT, READ, diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/OzoneManagerLock.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/OzoneManagerLock.java index 9fd567344cd0..2f6d220e7afb 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/OzoneManagerLock.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/OzoneManagerLock.java @@ -728,7 +728,6 @@ short getMask() { } } - /** * Update the processing details. * diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/OzoneManagerLockUtil.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/OzoneManagerLockUtil.java index cdce302c30ba..e0c89ef227d1 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/OzoneManagerLockUtil.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/lock/OzoneManagerLockUtil.java @@ -29,7 +29,6 @@ */ final class OzoneManagerLockUtil { - private OzoneManagerLockUtil() { } diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/protocol/OzoneManagerProtocol.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/protocol/OzoneManagerProtocol.java index 840f3fa00e26..d164587ffa55 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/protocol/OzoneManagerProtocol.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/protocol/OzoneManagerProtocol.java @@ -102,7 +102,6 @@ default void createVolume(OmVolumeArgs args) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * Changes the owner of a volume. * @param volume - Name of the volume. @@ -116,7 +115,6 @@ default boolean setOwner(String volume, String owner) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * Changes the Quota on a volume. * @param volume - Name of the volume. @@ -130,7 +128,6 @@ default void setQuota(String volume, long quotaInNamespace, long quotaInBytes) "this to be implemented, as write requests use a new approach."); } - /** * Checks if the specified user can access this volume. * @param volume - volume @@ -144,7 +141,6 @@ default boolean checkVolumeAccess(String volume, OzoneAclInfo userAcl) throw new UnsupportedOperationException("This operation is not supported."); } - /** * Gets the volume information. * @param volume - Volume name. @@ -163,7 +159,6 @@ default void deleteVolume(String volume) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * Lists volumes accessible by a specific user. * @param userName - user name @@ -197,7 +192,6 @@ default void createBucket(OmBucketInfo bucketInfo) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * Gets the bucket information. * @param volumeName - Volume name. @@ -230,7 +224,6 @@ default boolean setBucketOwner(OmBucketArgs args) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * Open the given key and return an open key session. * @@ -338,7 +331,6 @@ default void renameKey(OmKeyArgs args, String toKeyName) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * Rename existing keys within a bucket. * @param omRenameKeys Includes volume, bucket, and fromKey toKey name map @@ -350,7 +342,6 @@ default void renameKeys(OmRenameKeys omRenameKeys) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * Deletes an existing key. * @@ -362,7 +353,6 @@ default void deleteKey(OmKeyArgs args) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * Deletes existing key/keys. This interface supports delete * multiple keys and a single key. Used by deleting files @@ -391,7 +381,6 @@ default Map deleteKeys(OmDeleteKeys deleteKeys, boolean quiet "this to be implemented, as write requests use a new approach."); } - /** * Deletes an existing empty bucket from volume. * @param volume - Name of the volume. @@ -403,7 +392,6 @@ default void deleteBucket(String volume, String bucket) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * Returns a list of buckets represented by {@link OmBucketInfo} * in the given volume. Argument volumeName is required, others @@ -550,8 +538,6 @@ default OmMultipartInfo initiateMultipartUpload(OmKeyArgs keyArgs) "this to be implemented, as write requests use a new approach."); } - - /** * Commit Multipart upload part file. * @param omKeyArgs @@ -565,7 +551,6 @@ default OmMultipartCommitUploadPartInfo commitMultipartUploadPart( "this to be implemented, as write requests use a new approach."); } - /** * Complete Multipart upload Request. * @param omKeyArgs @@ -580,7 +565,6 @@ default OmMultipartUploadCompleteInfo completeMultipartUpload( "this to be implemented, as write requests use a new approach."); } - /** * Abort multipart upload. * @param omKeyArgs @@ -591,7 +575,6 @@ default void abortMultipartUpload(OmKeyArgs omKeyArgs) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * Returns list of parts of a multipart upload key. * @param volumeName @@ -638,7 +621,7 @@ default S3SecretValue getS3Secret(String kerberosID, boolean createIfNotExist) throws IOException { throw new UnsupportedOperationException("OzoneManager does not require " + "this to be implemented, as write requests use a new approach"); - }; + } /** * Set secret key for accessId. @@ -927,7 +910,6 @@ default void createDirectory(OmKeyArgs args) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * OzoneFS api to creates an output stream for a file. * @@ -948,7 +930,6 @@ default OpenKeySession createFile(OmKeyArgs keyArgs, boolean overWrite, "this to be implemented, as write requests use a new approach."); } - /** * OzoneFS api to lookup for a file. * @@ -1028,7 +1009,6 @@ default boolean addAcl(OzoneObj obj, OzoneAcl acl) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * Remove acl for Ozone object. Return true if acl is removed successfully * else false. @@ -1042,7 +1022,6 @@ default boolean removeAcl(OzoneObj obj, OzoneAcl acl) throws IOException { "this to be implemented, as write requests use a new approach."); } - /** * Acls to be set for given Ozone object. This operations reset ACL for * given object to list of ACLs provided in argument. @@ -1142,6 +1121,7 @@ void setTimes(OmKeyArgs keyArgs, long mtime, long atime) throws IOException; UUID refetchSecretKey() throws IOException; + /** * Enter, leave, or get safe mode. * diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/protocolPB/OzoneManagerClientProtocol.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/protocolPB/OzoneManagerClientProtocol.java index 2ca9d60379f4..de38f275c6e3 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/protocolPB/OzoneManagerClientProtocol.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/protocolPB/OzoneManagerClientProtocol.java @@ -32,7 +32,9 @@ public interface OzoneManagerClientProtocol extends OzoneManagerProtocol { * a thread local variable. */ void setThreadLocalS3Auth(S3Auth s3Auth); + S3Auth getThreadLocalS3Auth(); + void clearThreadLocalS3Auth(); ThreadLocal getS3CredentialsProvider(); diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/protocolPB/OzoneManagerProtocolClientSideTranslatorPB.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/protocolPB/OzoneManagerProtocolClientSideTranslatorPB.java index 1c5b11fcb100..0e23fc1937b7 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/protocolPB/OzoneManagerProtocolClientSideTranslatorPB.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/om/protocolPB/OzoneManagerProtocolClientSideTranslatorPB.java @@ -1935,7 +1935,6 @@ public void transferLeadership(String newLeaderId) handleError(submitRequest(omRequest)); } - @Override public boolean triggerRangerBGSync(boolean noWait) throws IOException { RangerBGSyncRequest req = RangerBGSyncRequest.newBuilder() @@ -2101,6 +2100,7 @@ public void setThreadLocalS3Auth( S3Auth s3Auth) { this.threadLocalS3Auth.set(s3Auth); } + @Override public void clearThreadLocalS3Auth() { this.threadLocalS3Auth.remove(); @@ -2349,7 +2349,6 @@ public OpenKeySession createFile(OmKeyArgs args, OmKeyInfo.getFromProtobuf(resp.getKeyInfo()), resp.getOpenVersion()); } - @Nonnull private OMResponse handleSubmitRequestAndSCMSafeModeRetry(OMRequest omRequest) throws IOException { int retryCount = BLOCK_ALLOCATION_RETRY_COUNT; diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/protocolPB/OMPBHelper.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/protocolPB/OMPBHelper.java index 35501072e9fe..c613fb8c603e 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/protocolPB/OMPBHelper.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/protocolPB/OMPBHelper.java @@ -114,7 +114,6 @@ public static BucketEncryptionKeyInfo convert( beInfo.getKeyName()); } - public static BucketEncryptionInfoProto convert( BucketEncryptionKeyInfo beInfo) { if (beInfo == null || beInfo.getKeyName() == null) { @@ -368,7 +367,6 @@ public static CryptoProtocolVersion convert( } } - public static OMRequest processForDebug(OMRequest msg) { return msg; } diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/acl/IAccessAuthorizer.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/acl/IAccessAuthorizer.java index 92023f0b1a58..d4a1c3d87ff7 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/acl/IAccessAuthorizer.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/acl/IAccessAuthorizer.java @@ -70,6 +70,7 @@ enum ACLType { + " > 16, check the commit of this change and update the code."); } } + private static ACLType[] vals = ACLType.values(); public static int getNoOfAcls() { @@ -177,6 +178,7 @@ enum ACLIdentityType { public String toString() { return value; } + /** * String value for this Enum. */ diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/acl/OzoneObjInfo.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/acl/OzoneObjInfo.java index a9e768dc5959..36b450bec7fa 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/acl/OzoneObjInfo.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/acl/OzoneObjInfo.java @@ -232,7 +232,6 @@ public Builder setOzonePrefixPath(OzonePrefixPath ozonePrefixPathViewer) { return this; } - public OzoneObjInfo build() { return new OzoneObjInfo(resType, storeType, volumeName, bucketName, name, ozonePrefixPath); diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/acl/RequestContext.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/acl/RequestContext.java index fadca7abfd73..08724eae5ff2 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/acl/RequestContext.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/security/acl/RequestContext.java @@ -140,7 +140,6 @@ public static Builder newBuilder() { return new Builder(); } - public static RequestContext.Builder getBuilder( UserGroupInformation ugi, InetAddress remoteAddress, String hostName, ACLType aclType, String ownerName) { diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/snapshot/SnapshotDiffReportOzone.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/snapshot/SnapshotDiffReportOzone.java index b87ef9223c8a..735169175798 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/snapshot/SnapshotDiffReportOzone.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/snapshot/SnapshotDiffReportOzone.java @@ -65,13 +65,11 @@ public static Codec getDiffReportEntryCodec() { */ private final String bucketName; - /** * subsequent token for the diff report. */ private final String token; - public SnapshotDiffReportOzone(final String snapshotRoot, final String volumeName, final String bucketName, @@ -190,7 +188,6 @@ public static DiffReportEntry fromProtobufDiffReportEntry( return builder.build(); } - public static DiffReportEntry getDiffReportEntry(final DiffType type, final String sourcePath) { return getDiffReportEntry(type, sourcePath, null); diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/snapshot/SnapshotDiffResponse.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/snapshot/SnapshotDiffResponse.java index ece5e72a94ff..47b586b8dd05 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/snapshot/SnapshotDiffResponse.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/snapshot/SnapshotDiffResponse.java @@ -45,6 +45,7 @@ public static JobStatus fromProtobuf(JobStatusProto jobStatusProto) { return JobStatus.valueOf(jobStatusProto.name()); } } + /** * Snapshot diff job sub-status enum. */ diff --git a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/PayloadUtils.java b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/PayloadUtils.java index 10bd0e3b8032..27a08eb1ad64 100644 --- a/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/PayloadUtils.java +++ b/hadoop-ozone/common/src/main/java/org/apache/hadoop/ozone/util/PayloadUtils.java @@ -22,6 +22,7 @@ import org.apache.ratis.thirdparty.com.google.protobuf.ByteString; import org.apache.ratis.thirdparty.com.google.protobuf.UnsafeByteOperations; import org.apache.ratis.util.Preconditions; + /** * Utility class for payload operations. */ diff --git a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/helpers/TestOmKeyInfo.java b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/helpers/TestOmKeyInfo.java index 00ecbed52361..5a23df4d9688 100644 --- a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/helpers/TestOmKeyInfo.java +++ b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/helpers/TestOmKeyInfo.java @@ -205,7 +205,6 @@ private void createdAndTest(boolean isMPU) { assertNotEquals(key, cloneKey); } - private OmKeyLocationInfoGroup createOmKeyLocationInfoGroup(boolean isMPU) { List omKeyLocationInfos = new ArrayList<>(); omKeyLocationInfos.add(getOmKeyLocationInfo(new BlockID( diff --git a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/lock/TestOzoneManagerLock.java b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/lock/TestOzoneManagerLock.java index ca986a84639d..49377c8cf282 100644 --- a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/lock/TestOzoneManagerLock.java +++ b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/lock/TestOzoneManagerLock.java @@ -183,7 +183,6 @@ void releaseLockWithOutAcquiringLock() { () -> lock.releaseWriteLock(Resource.USER_LOCK, "user3")); } - private String[] generateResourceName(Resource resource) { if (resource == Resource.BUCKET_LOCK) { return new String[]{UUID.randomUUID().toString(), @@ -318,7 +317,6 @@ void testMultiLocksResourceParallel() throws Exception { } - @Test void testMultiLockResourceParallel() throws Exception { OzoneManagerLock lock = new OzoneManagerLock(new OzoneConfiguration()); @@ -398,7 +396,6 @@ void testLockConcurrentStats(Resource resource) throws InterruptedException { testSyntheticReadWriteLockConcurrentStats(resource, resourceName, 10, 3); } - private void testReadLockConcurrentStats(Resource resource, String[] resourceName, int threadCount) diff --git a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/protocolPB/TestS3GrpcOmTransport.java b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/protocolPB/TestS3GrpcOmTransport.java index d89fdc065854..69d69f7fcb52 100644 --- a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/protocolPB/TestS3GrpcOmTransport.java +++ b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/om/protocolPB/TestS3GrpcOmTransport.java @@ -75,7 +75,6 @@ public class TestS3GrpcOmTransport { private String serverName; - private ServiceException createNotLeaderException() { RaftPeerId raftPeerId = RaftPeerId.getRaftPeerId("testNodeId"); diff --git a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/security/TestOzoneDelegationTokenSelector.java b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/security/TestOzoneDelegationTokenSelector.java index 69569e2bd495..ffc857b0d80f 100644 --- a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/security/TestOzoneDelegationTokenSelector.java +++ b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/security/TestOzoneDelegationTokenSelector.java @@ -33,7 +33,6 @@ */ public class TestOzoneDelegationTokenSelector { - @Test public void testTokenSelector() { @@ -77,10 +76,7 @@ public void testTokenSelector() { } - private Text getService() { return new Text("om1:9862,om2:9862,om3:9862"); } - - } diff --git a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/security/acl/TestOzoneObjInfo.java b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/security/acl/TestOzoneObjInfo.java index 85f142460f2b..9c76f02f0e59 100644 --- a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/security/acl/TestOzoneObjInfo.java +++ b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/security/acl/TestOzoneObjInfo.java @@ -38,7 +38,6 @@ public class TestOzoneObjInfo { private String key = "key1"; private static final OzoneObj.StoreType STORE = OzoneObj.StoreType.OZONE; - @Test public void testGetVolumeName() { diff --git a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/util/TestRadixTree.java b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/util/TestRadixTree.java index 7aadda6de1eb..9a3071d754b9 100644 --- a/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/util/TestRadixTree.java +++ b/hadoop-ozone/common/src/test/java/org/apache/hadoop/ozone/util/TestRadixTree.java @@ -122,6 +122,4 @@ public void testRemovePrefixPath() { ROOT.removePrefixPath("/d/a"); assertEquals("/a/b/c/d", ROOT.getLongestPrefix("a/b/c/d")); } - - } diff --git a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/failure/FailureManager.java b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/failure/FailureManager.java index 2a27708ef9da..59fae25bbad7 100644 --- a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/failure/FailureManager.java +++ b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/failure/FailureManager.java @@ -43,6 +43,7 @@ public class FailureManager { private final List failures; private ScheduledFuture scheduledFuture; private final ScheduledExecutorService executorService; + public FailureManager(MiniOzoneChaosCluster cluster, Configuration conf, Set> clazzes) { diff --git a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadBucket.java b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadBucket.java index 4f0c1b921197..62a990a00861 100644 --- a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadBucket.java +++ b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadBucket.java @@ -150,7 +150,9 @@ public void execute() throws Exception { } abstract void doFsOp(Path p) throws IOException; + abstract void doBucketOp(String key) throws IOException; + abstract void doPostOp() throws IOException; @Override diff --git a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadExecutors.java b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadExecutors.java index c7c8ccf9ec9c..5e240186b7e5 100644 --- a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadExecutors.java +++ b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadExecutors.java @@ -69,7 +69,6 @@ private void load(long runTimeMillis) { } } - public void startLoad(long time) throws Exception { LOG.info("Starting {} threads for {} generators", numThreads, generators.size()); diff --git a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/ReadOnlyLoadGenerator.java b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/ReadOnlyLoadGenerator.java index 9d19f9f6c767..2e928cfb07ae 100644 --- a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/ReadOnlyLoadGenerator.java +++ b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/ReadOnlyLoadGenerator.java @@ -41,7 +41,6 @@ public void generateLoad() throws Exception { replBucket.readKey(buffer, keyName); } - @Override public void initialize() throws Exception { for (int index = 0; index < NUM_KEYS; index++) { diff --git a/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/FSOperations.java b/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/FSOperations.java index 8d8f62ba0996..55d90ef99011 100644 --- a/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/FSOperations.java +++ b/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/FSOperations.java @@ -76,6 +76,7 @@ public final class FSOperations { private FSOperations() { // not called } + /** * Set the buffer size. The size is set during the initialization of * HttpFSServerWebApp. @@ -1344,6 +1345,7 @@ public Void execute(FileSystem fs) throws IOException { public static class FSTrashRoot implements FileSystemAccess.FileSystemExecutor { private Path path; + public FSTrashRoot(String path) { this.path = new Path(path); } diff --git a/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/HttpFSParametersProvider.java b/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/HttpFSParametersProvider.java index 5ec328575954..8a27b7199476 100644 --- a/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/HttpFSParametersProvider.java +++ b/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/HttpFSParametersProvider.java @@ -137,6 +137,7 @@ public static class AccessTimeParam extends LongParam { * Parameter name. */ public static final String NAME = HttpFSConstants.ACCESS_TIME_PARAM; + /** * Constructor. */ @@ -192,6 +193,7 @@ public static class NoRedirectParam extends BooleanParam { * Parameter name. */ public static final String NAME = "noredirect"; + /** * Constructor. */ @@ -211,6 +213,7 @@ public static class OperationParam * Parameter name. */ public static final String NAME = HttpFSConstants.OP_PARAM; + /** * Constructor. */ @@ -406,7 +409,6 @@ public static class PermissionParam extends ShortParam { */ public static final String NAME = HttpFSConstants.PERMISSION_PARAM; - /** * Constructor. */ @@ -428,7 +430,6 @@ public static class UnmaskedPermissionParam extends ShortParam { public static final String NAME = HttpFSConstants.UNMASKED_PERMISSION_PARAM; - /** * Constructor. */ diff --git a/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/HttpFSServer.java b/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/HttpFSServer.java index f71b3d2b53da..4623dee4832a 100644 --- a/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/HttpFSServer.java +++ b/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/HttpFSServer.java @@ -101,6 +101,7 @@ public class HttpFSServer { enum AccessMode { READWRITE, WRITEONLY, READONLY; } + private static final Logger AUDIT_LOG = LoggerFactory.getLogger("httpfsaudit"); private static final Logger LOG = LoggerFactory.getLogger(HttpFSServer.class); diff --git a/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/HttpFSServerWebApp.java b/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/HttpFSServerWebApp.java index b5a1736b0edc..cb946f57676c 100644 --- a/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/HttpFSServerWebApp.java +++ b/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/fs/http/server/HttpFSServerWebApp.java @@ -115,6 +115,7 @@ private static void setMetrics(Configuration config) { FSOperations.setBufferSize(config); DefaultMetricsSystem.initialize("HttpFSServer"); } + /** * Returns HttpFSServer server singleton, configuration and services are * accessible through it. diff --git a/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/lib/service/instrumentation/InstrumentationService.java b/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/lib/service/instrumentation/InstrumentationService.java index 69652ef8f9ed..28b87518f348 100644 --- a/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/lib/service/instrumentation/InstrumentationService.java +++ b/hadoop-ozone/httpfsgateway/src/main/java/org/apache/ozone/lib/service/instrumentation/InstrumentationService.java @@ -419,6 +419,4 @@ public void run() { public Map> getSnapshot() { return all; } - - } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractContractCreateTest.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractContractCreateTest.java index 521895766a69..87fcc721b86b 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractContractCreateTest.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractContractCreateTest.java @@ -360,7 +360,6 @@ public void testCreateUnderFileSubdir() throws Throwable { child); } - @Test public void testMkdirUnderFile() throws Throwable { describe("Verify that it is forbidden to create file/dir"); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractContractGetFileStatusTest.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractContractGetFileStatusTest.java index 15c9599073c1..3c62094393f6 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractContractGetFileStatusTest.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractContractGetFileStatusTest.java @@ -552,7 +552,6 @@ private int verifyFileStats(RemoteIterator results) return count; } - @Test public void testListStatusFiltering() throws Throwable { describe("Call listStatus() against paths and directories with filtering"); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractFSContractTestBase.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractFSContractTestBase.java index 6211f546f447..0f5d4d4b49d0 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractFSContractTestBase.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractFSContractTestBase.java @@ -346,7 +346,6 @@ protected void assertIsDirectory(Path path) throws IOException { ContractTestUtils.assertIsDirectory(fileSystem, path); } - /** * Assert that a file exists and whose {@link FileStatus} entry * declares that this is a file and not a symlink or directory. diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/ContractTestUtils.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/ContractTestUtils.java index a4d7c02b4a47..1d3fcf4ff501 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/ContractTestUtils.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/ContractTestUtils.java @@ -1174,7 +1174,6 @@ public static void returnBuffersToPoolPostRead(List fileRanges, } } - /** * Assert that the data read matches the dataset at the given offset. * This helps verify that the seek process is moving the read pointer @@ -1476,7 +1475,6 @@ public static boolean collectionsEquivalentNoDuplicates(Collection left, !containsDuplicates(left) && !containsDuplicates(right); } - /** * Predicate to test for a collection of paths containing duplicate entries. * @param paths collection of paths @@ -1657,7 +1655,6 @@ public static void assertCapabilities( } } - /** * Custom assert to verify capabilities supported by * an object through {@link StreamCapabilities}. @@ -1761,7 +1758,6 @@ public static long readStream(InputStream in) { } } - /** * Results of recursive directory creation/scan operations. */ @@ -1772,7 +1768,6 @@ public static final class TreeScanResults { private final List directories = new ArrayList<>(); private final List other = new ArrayList<>(); - public TreeScanResults() { } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractOzoneFileSystemTest.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractOzoneFileSystemTest.java index b9464402d62a..f3218e4c163e 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractOzoneFileSystemTest.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractOzoneFileSystemTest.java @@ -2020,7 +2020,6 @@ private void verifyOwnerGroup(FileStatus fileStatus) { assertEquals(owner, fileStatus.getGroup()); } - @Test public void testDirectory() throws IOException { assumeFalse(FILE_SYSTEM_OPTIMIZED.equals(getBucketLayout())); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTest.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTest.java index 167066f30080..d86e2f3b1121 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTest.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTest.java @@ -1981,7 +1981,6 @@ private void checkInvalidPath(Path path) { assertThat(exception.getMessage()).contains("Invalid path Name"); } - @Test void testRenameFile() throws Exception { final String dir = "/dir" + RandomUtils.secure().randomInt(0, 1000); @@ -2212,7 +2211,6 @@ void testUnbuffer() throws IOException { } - @Test void testCreateAndCheckECFileDiskUsage() throws Exception { String key = "eckeytest"; @@ -2235,7 +2233,6 @@ void testCreateAndCheckECFileDiskUsage() throws Exception { ofs.delete(filePath, true); } - @Test void testCreateAndCheckRatisFileDiskUsage() throws Exception { String key = "ratiskeytest"; @@ -2372,7 +2369,6 @@ void testFileSystemDeclaresCapability() throws Throwable { assertHasPathCapabilities(fs, getBucketPath(), FS_CHECKSUMS); } - @Test void testSnapshotDiff() throws Exception { OzoneBucket bucket1 = diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTestWithFSO.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTestWithFSO.java index 787d3fde3b51..55c4ccc47555 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTestWithFSO.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTestWithFSO.java @@ -110,6 +110,7 @@ void testRenameDir() throws Exception { // cleanup getFs().delete(dest, true); } + /** * Cannot rename a directory to its own subdirectory. */ diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestHSync.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestHSync.java index f25e47cdfed5..4493c90696d2 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestHSync.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestHSync.java @@ -456,7 +456,6 @@ public void testO3fsHSync(boolean incrementalChunkList) throws Exception { } } - @ParameterizedTest @ValueSource(booleans = {false, true}) public void testOfsHSync(boolean incrementalChunkList) throws Exception { @@ -821,7 +820,6 @@ static void runTestHSync(FileSystem fs, Path file, }, 500, 3000); } - public static Stream concurrentExceptionHandling() { return Stream.of( Arguments.of(4, 1), @@ -1571,9 +1569,11 @@ private Map getAllDeletedKeys(Table { Collection peers = nodeManager.getPeerList(dn); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelineClose.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelineClose.java index fda3e47c64af..292c20958e81 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelineClose.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipelineClose.java @@ -83,6 +83,7 @@ public class TestPipelineClose { private PipelineManager pipelineManager; private long pipelineDestroyTimeoutInMillis; + /** * Create a MiniDFSCluster for testing. * diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestSCMSafeModeWithPipelineRules.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestSCMSafeModeWithPipelineRules.java index cc7419648a42..f6e33b352626 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestSCMSafeModeWithPipelineRules.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/safemode/TestSCMSafeModeWithPipelineRules.java @@ -79,7 +79,6 @@ public void setup(int numDatanodes) throws Exception { pipelineManager = scm.getPipelineManager(); } - @Test void testScmSafeMode() throws Exception { int datanodeCount = 6; @@ -169,7 +168,6 @@ public void tearDown() { } } - private void waitForRatis3NodePipelines(int numPipelines) throws TimeoutException, InterruptedException { GenericTestUtils.waitFor(() -> pipelineManager diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/upgrade/TestHDDSUpgrade.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/upgrade/TestHDDSUpgrade.java index 0704932744a6..547ea0dfe863 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/upgrade/TestHDDSUpgrade.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/upgrade/TestHDDSUpgrade.java @@ -210,7 +210,6 @@ private void loadSCMState() { scmVersionManager = scm.getLayoutVersionManager(); } - /* * helper function to create a Key. */ diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestSecureOzoneCluster.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestSecureOzoneCluster.java index 7cdf6eb2db47..c33f1712e822 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestSecureOzoneCluster.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestSecureOzoneCluster.java @@ -997,6 +997,7 @@ void testCertificateRotation() throws Exception { 1000, certificateLifetime * 1000); } } + /** * Test unexpected SCMGetCertResponseProto returned from SCM. */ diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java index de72e2ddfb47..0b26710b7e83 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java @@ -313,6 +313,7 @@ public static OzoneClient getClient() { public static MiniOzoneCluster getCluster() { return cluster; } + /** * Test OM Proxy Provider. */ @@ -674,7 +675,6 @@ public void testCreateBucketWithMetadata() } - @Test public void testCreateBucket() throws IOException { @@ -2078,7 +2078,6 @@ public void testPutKeyRatisThreeNodes() throws IOException { } } - @Test @Flaky("HDDS-10886") public void testPutKeyRatisThreeNodesParallel() throws IOException, @@ -2137,7 +2136,6 @@ public void testPutKeyRatisThreeNodesParallel() throws IOException, .isLessThanOrEqualTo(0); } - @Test public void testReadKeyWithVerifyChecksumFlagEnable() throws Exception { String volumeName = UUID.randomUUID().toString(); @@ -2152,7 +2150,6 @@ public void testReadKeyWithVerifyChecksumFlagEnable() throws Exception { } - @Test public void testReadKeyWithVerifyChecksumFlagDisable() throws Exception { String volumeName = UUID.randomUUID().toString(); @@ -2193,7 +2190,6 @@ private void createAndCorruptKey(String volumeName, String bucketName, corruptData(cluster, container, key); } - private void readCorruptedKey(String volumeName, String bucketName, String keyName, boolean verifyChecksum) { try { @@ -3008,7 +3004,6 @@ void testInitiateMultipartUpload(ReplicationConfig replicationConfig) assertNotNull(multipartInfo.getUploadID()); } - @Test public void testInitiateMultipartUploadWithDefaultReplication() throws IOException { @@ -3310,6 +3305,7 @@ public void testClientLeakDetector() throws Exception { GenericTestUtils.waitFor(() -> ozoneClientFactoryLogCapturer.getOutput() .contains("is not closed properly"), 100, 2000); } + @Test public void testMultipartUploadOwner() throws Exception { // Save the old user, and switch to the old user after test @@ -3424,6 +3420,7 @@ public void testMultipartUploadWithPartsLessThanMinSize() throws Exception { () -> completeMultipartUpload(bucket, keyName, uploadID, eTagsMaps)); } + @Test public void testMultipartUploadWithPartsMisMatchWithListSizeDifferent() throws Exception { @@ -3685,7 +3682,6 @@ void testCommitPartAfterCompleteUpload() throws Exception { assertEquals(NO_SUCH_MULTIPART_UPLOAD_ERROR, ex.getResult()); } - @Test public void testAbortUploadSuccessWithOutAnyParts() throws Exception { String volumeName = UUID.randomUUID().toString(); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBlockDataStreamOutput.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBlockDataStreamOutput.java index 2f5aa1ff9a15..8a01464b5295 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBlockDataStreamOutput.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBlockDataStreamOutput.java @@ -84,7 +84,7 @@ public class TestBlockDataStreamOutput { private static final int BLOCK_SIZE = 2 * MAX_FLUSH_SIZE; private static final String VOLUME_NAME = "testblockoutputstream"; private static final String BUCKET_NAME = VOLUME_NAME; - private static String keyString = UUID.randomUUID().toString();; + private static String keyString = UUID.randomUUID().toString(); private static final DatanodeVersion DN_OLD_VERSION = DatanodeVersion.SEPARATE_RATIS_PORTS_AVAILABLE; static MiniOzoneCluster createCluster() throws IOException, diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestContainerStateMachineFailures.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestContainerStateMachineFailures.java index 94114825fa61..63eaf0d9680d 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestContainerStateMachineFailures.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestContainerStateMachineFailures.java @@ -261,7 +261,6 @@ public void testContainerStateMachineCloseOnMissingPipeline() key.close(); } - @Test @Flaky("HDDS-12215") public void testContainerStateMachineRestartWithDNChangePipeline() diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestFailureHandlingByClient.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestFailureHandlingByClient.java index 30446af41b0e..9fcb126be475 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestFailureHandlingByClient.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestFailureHandlingByClient.java @@ -359,7 +359,6 @@ public void testWriteSmallFile() throws Exception { validateData(keyName, data.getBytes(UTF_8)); } - @Test public void testContainerExclusionWithClosedContainerException() throws Exception { @@ -479,7 +478,6 @@ public void testDatanodeExclusionWithMajorityCommit(RaftProtos.ReplicationLevel validateData(keyName, data.concat(data).concat(data).getBytes(UTF_8)); } - @Test public void testPipelineExclusionWithPipelineFailure() throws Exception { startCluster(); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneAtRestEncryption.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneAtRestEncryption.java index eda5d100c223..07c3ad5b3070 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneAtRestEncryption.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneAtRestEncryption.java @@ -242,7 +242,6 @@ private KMSClientProvider getKMSClientProvider() { return lbkmscp.getProviders()[0]; } - @ParameterizedTest @EnumSource void testPutKeyWithEncryption(BucketLayout bucketLayout) throws Exception { diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/read/TestChunkInputStream.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/read/TestChunkInputStream.java index 972eae13d7cb..730b76ecf0a3 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/read/TestChunkInputStream.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/read/TestChunkInputStream.java @@ -57,7 +57,6 @@ void testAll(ContainerLayoutVersion layout) throws Exception { } } - /** * Test to verify that data read from chunks is stored in a list of buffers * with max capacity equal to the bytes per checksum. diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReplication.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReplication.java index c976c59a8a61..125166daa378 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReplication.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReplication.java @@ -237,7 +237,6 @@ private static OmKeyLocationInfo lookupKeyFirstLocation(MiniOzoneCluster cluster return locations.getLocationList().get(0); } - public void assertState(MiniOzoneCluster cluster, Map expectedReplicaMap) throws IOException { OmKeyLocationInfo keyLocation = lookupKeyFirstLocation(cluster); @@ -269,7 +268,6 @@ private void mockContainerProtocolCalls(final MockedStatic keyList = new ArrayList<>(); - @BeforeAll public static void init() throws Exception { conf = new OzoneConfiguration(); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestKeyManagerImpl.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestKeyManagerImpl.java index 6b95b46f0a8b..f1568407a74d 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestKeyManagerImpl.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestKeyManagerImpl.java @@ -264,6 +264,7 @@ private static void mockContainerClient() { HddsWhiteboxTestUtils.setInternalState(om, "scmClient", scmClient); } + private static void mockBlockClient() { ScmClient scmClient = new ScmClient(mockScmBlockLocationProtocol, null, conf); @@ -272,6 +273,7 @@ private static void mockBlockClient() { HddsWhiteboxTestUtils.setInternalState(om, "scmClient", scmClient); } + private static void createBucket(String volumeName, String bucketName, boolean isVersionEnabled) throws IOException { @@ -1474,7 +1476,6 @@ public void testRefreshPipeline() throws Exception { } - @Test public void testRefreshPipelineException() throws Exception { @@ -1626,6 +1627,7 @@ private Pipeline getRandomPipeline() { .setNodes(new ArrayList<>()) .build(); } + /** * Creates a depth two directory. * diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestListKeysWithFSO.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestListKeysWithFSO.java index 269a9c4e77f2..fd76512181b9 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestListKeysWithFSO.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestListKeysWithFSO.java @@ -576,7 +576,6 @@ private static void buildNameSpaceTree2(OzoneBucket ozoneBucket) createAndAssertKeys(ozoneBucket, keys); } - private static List getExpectedKeyList(String keyPrefix, String startKey, OzoneBucket legacyBucket, boolean shallow) throws Exception { diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMRatisSnapshots.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMRatisSnapshots.java index 1b876dbc2ca8..c1b2d002e11c 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMRatisSnapshots.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMRatisSnapshots.java @@ -515,9 +515,11 @@ public void testInstallIncrementalSnapshot(@TempDir Path tempDir) static class IncrementData { private List keys; private SnapshotInfo snapshotInfo; + public List getKeys() { return keys; } + public SnapshotInfo getSnapshotInfo() { return snapshotInfo; } @@ -1135,6 +1137,7 @@ private static class SnapshotMaxSizeInjector extends FaultInjector { private final File snapshotDir; private final List> sstSetList; private final Path tempDir; + SnapshotMaxSizeInjector(OzoneManager om, File snapshotDir, List> sstSetList, Path tempDir) { this.om = om; diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMUpgradeFinalization.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMUpgradeFinalization.java index 4787f2064e5b..4019010b026d 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMUpgradeFinalization.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOMUpgradeFinalization.java @@ -60,10 +60,12 @@ class TestOMUpgradeFinalization { public void setup() throws Exception { AuditLogTestUtils.truncateAuditLogFile(); } + @AfterAll public static void shutdown() { AuditLogTestUtils.deleteAuditLogFile(); } + @Test void testOMUpgradeFinalizationWithOneOMDown() throws Exception { OzoneConfiguration conf = new OzoneConfiguration(); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOmContainerLocationCache.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOmContainerLocationCache.java index c5930fb8bd8e..0ea807d74442 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOmContainerLocationCache.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOmContainerLocationCache.java @@ -242,7 +242,6 @@ private static ArgumentMatcher matchEmptyPipeline() { && argument.getNodes().isEmpty(); } - private static ArgumentMatcher matchPipeline(DatanodeDetails dn) { return argument -> argument != null && !argument.getNodes().isEmpty() && argument.getNodes().get(0).getUuid().equals(dn.getUuid()); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOmMetrics.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOmMetrics.java index 7dec6a385268..cb594e963e3c 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOmMetrics.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOmMetrics.java @@ -910,12 +910,14 @@ private OmVolumeArgs createVolumeArgs() { .setAdminName("dummyAdmin") .build(); } + private OmBucketArgs getBucketArgs(OmBucketInfo info) { return new OmBucketArgs.Builder() .setVolumeName(info.getVolumeName()) .setBucketName(info.getBucketName()) .build(); } + private OmBucketInfo createBucketInfo(boolean isEcBucket) throws IOException { OmVolumeArgs volumeArgs = createVolumeArgs(); writeClient.createVolume(volumeArgs); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHA.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHA.java index 6089f0c65cac..4f291b7de715 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHA.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHA.java @@ -89,7 +89,6 @@ public abstract class TestOzoneManagerHA { private static final Duration RETRY_CACHE_DURATION = Duration.ofSeconds(30); private static OzoneClient client; - public MiniOzoneHAClusterImpl getCluster() { return cluster; } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHAWithAllRunning.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHAWithAllRunning.java index f0c121138708..ce4c778ddb64 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHAWithAllRunning.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHAWithAllRunning.java @@ -244,7 +244,6 @@ public void testAllVolumeOperations() throws Exception { () -> objectStore.deleteVolume(volumeName)); } - @Test public void testAllBucketOperations() throws Exception { String volumeName = "volume" + RandomStringUtils.secure().nextNumeric(5); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHAWithStoppedNodes.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHAWithStoppedNodes.java index 065835532aec..89ca385bf8b2 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHAWithStoppedNodes.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHAWithStoppedNodes.java @@ -96,6 +96,7 @@ public class TestOzoneManagerHAWithStoppedNodes extends TestOzoneManagerHA { private static final org.slf4j.Logger LOG = LoggerFactory.getLogger( TestOzoneManagerHAWithStoppedNodes.class); + /** * After restarting OMs we need to wait * for a leader to be elected and ready. @@ -466,6 +467,7 @@ public void testKeyDeletion() throws Exception { } }); } + /** * 1. Stop one of the OM * 2. make a call to OM, this will make failover attempts to find new node. diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerRestart.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerRestart.java index 6634c96fcd64..8b709d5e4173 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerRestart.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerRestart.java @@ -118,7 +118,6 @@ public void testRestartOMWithVolumeOperation() throws Exception { } - @Test public void testRestartOMWithBucketOperation() throws Exception { String volumeName = "volume" + RandomStringUtils.secure().nextNumeric(5); @@ -149,7 +148,6 @@ public void testRestartOMWithBucketOperation() throws Exception { assertEquals(bucketName, ozoneBucket.getName()); } - @Test public void testRestartOMWithKeyOperation() throws Exception { String volumeName = "volume" + RandomStringUtils.secure().nextNumeric(5); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOmSnapshot.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOmSnapshot.java index ce144d4c703a..4df8c0eddb37 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOmSnapshot.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOmSnapshot.java @@ -1558,6 +1558,7 @@ private SnapshotDiffReportOzone getSnapDiffReport(String volume, return response.getSnapshotDiffReport(); } + @Test public void testSnapDiffNoSnapshot() throws Exception { String volume = "vol-" + counter.incrementAndGet(); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOzoneManagerHASnapshot.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOzoneManagerHASnapshot.java index ac2330172eb3..3d4bc628e2e1 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOzoneManagerHASnapshot.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOzoneManagerHASnapshot.java @@ -266,7 +266,6 @@ public void testSnapshotChainManagerRestore() throws Exception { .isSnapshotChainCorrupted()); } - private void createFileKey(OzoneBucket bucket, String keyName) throws IOException { byte[] value = RandomStringUtils.secure().nextAscii(10240).getBytes(UTF_8); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOzoneManagerSnapshotAcl.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOzoneManagerSnapshotAcl.java index fd195949672f..662c328a84d3 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOzoneManagerSnapshotAcl.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOzoneManagerSnapshotAcl.java @@ -438,7 +438,6 @@ public void testLookupKeyWithNotAllowedUserForPrefixAcl(BucketLayout bucketLayou assertDoesNotThrow(() -> ozoneManager.lookupKey(keyArgs)); } - private void setup(BucketLayout bucketLayout) throws IOException { UserGroupInformation.setLoginUser(UGI1); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestSnapshotDeletingServiceIntegrationTest.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestSnapshotDeletingServiceIntegrationTest.java index 87dd31c2178b..140f4ae427f2 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestSnapshotDeletingServiceIntegrationTest.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestSnapshotDeletingServiceIntegrationTest.java @@ -773,7 +773,6 @@ private synchronized void createSnapshotDataForBucket(OzoneBucket bucket) throws bucket.getName())); } - /* Flow ---- @@ -844,7 +843,6 @@ private synchronized void createSnapshotFSODataForBucket(OzoneBucket bucket) thr assertTableRowCount(snapshotInfoTable, countMap.get(snapshotInfoTable.getName()) + 2); } - private void verifySnapshotChain(SnapshotInfo deletedSnapshot, String nextSnapshot) throws Exception { diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/recon/TestReconInsightsForDeletedDirectories.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/recon/TestReconInsightsForDeletedDirectories.java index 069fc8576724..14b11c74fd2e 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/recon/TestReconInsightsForDeletedDirectories.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/recon/TestReconInsightsForDeletedDirectories.java @@ -486,7 +486,6 @@ private void syncDataFromOM() throws IOException { impl.syncDataFromOM(); } - private static BucketLayout getFSOBucketLayout() { return BucketLayout.FILE_SYSTEM_OPTIMIZED; } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/shell/TestOzoneShellHA.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/shell/TestOzoneShellHA.java index 5ab9061f6f04..b93111b8173f 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/shell/TestOzoneShellHA.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/shell/TestOzoneShellHA.java @@ -1263,7 +1263,6 @@ public void testDeleteTrashNoSkipTrash() throws Exception { } - @Test @SuppressWarnings("methodlength") public void testShQuota() throws Exception { @@ -1849,7 +1848,6 @@ public void testCreateBucketWithECReplicationConfigWithoutReplicationParam() { } } - @Test public void testKeyDeleteOrSkipTrashWhenTrashEnableFSO() throws IOException { diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/tools/contract/AbstractContractDistCpTest.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/tools/contract/AbstractContractDistCpTest.java index cdeea68e640c..c7e9fa3bbca5 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/tools/contract/AbstractContractDistCpTest.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/tools/contract/AbstractContractDistCpTest.java @@ -399,7 +399,6 @@ private Path modifySourceDirectories() throws IOException { return inputFileNew1; } - @Test public void testTrackDeepDirectoryStructureToRemote() throws Exception { describe("copy a deep directory structure from local to remote"); @@ -828,6 +827,7 @@ private Job runDistCpDirectWrite(final Path srcDir, final Path destDir) Collections.singletonList(srcDir), destDir) .withDirectWrite(true))); } + /** * Run distcp srcDir destDir. * @param srcDir local source directory diff --git a/hadoop-ozone/interface-storage/src/main/java/org/apache/hadoop/ozone/om/OMMetadataManager.java b/hadoop-ozone/interface-storage/src/main/java/org/apache/hadoop/ozone/om/OMMetadataManager.java index 02ef9bf12a3a..5a8942fe6b24 100644 --- a/hadoop-ozone/interface-storage/src/main/java/org/apache/hadoop/ozone/om/OMMetadataManager.java +++ b/hadoop-ozone/interface-storage/src/main/java/org/apache/hadoop/ozone/om/OMMetadataManager.java @@ -165,7 +165,6 @@ String getOzoneKeyFSO(String volumeName, */ String getOzoneDirKey(String volume, String bucket, String key); - /** * Returns the DB key name of a open key in OM metadata store. Should be * #open# prefix followed by actual key name. @@ -467,6 +466,7 @@ String getMultipartKeyFSO(String volume, String bucket, String key, String Table getSnapshotRenamedTable(); Table getCompactionLogTable(); + /** * Gets the OM Meta table. * @return meta table reference. diff --git a/hadoop-ozone/interface-storage/src/test/java/org/apache/hadoop/ozone/om/helpers/TestOmPrefixInfo.java b/hadoop-ozone/interface-storage/src/test/java/org/apache/hadoop/ozone/om/helpers/TestOmPrefixInfo.java index 215488694ebe..4b8679408649 100644 --- a/hadoop-ozone/interface-storage/src/test/java/org/apache/hadoop/ozone/om/helpers/TestOmPrefixInfo.java +++ b/hadoop-ozone/interface-storage/src/test/java/org/apache/hadoop/ozone/om/helpers/TestOmPrefixInfo.java @@ -81,7 +81,6 @@ private OmPrefixInfo getOmPrefixInfoForTest(String path, .build(); } - @Test public void testCopyObject() { String testPath = "/my/custom/path"; diff --git a/hadoop-ozone/mini-cluster/src/main/java/org/apache/hadoop/ozone/MiniOzoneCluster.java b/hadoop-ozone/mini-cluster/src/main/java/org/apache/hadoop/ozone/MiniOzoneCluster.java index 6c1a70923917..3b376ceefa28 100644 --- a/hadoop-ozone/mini-cluster/src/main/java/org/apache/hadoop/ozone/MiniOzoneCluster.java +++ b/hadoop-ozone/mini-cluster/src/main/java/org/apache/hadoop/ozone/MiniOzoneCluster.java @@ -193,6 +193,7 @@ void restartHddsDatanode(int i, boolean waitForDatanode) */ void restartHddsDatanode(DatanodeDetails dn, boolean waitForDatanode) throws InterruptedException, TimeoutException, IOException; + /** * Shutdown a particular HddsDatanode. * @@ -408,6 +409,7 @@ interface DatanodeFactory extends CheckedFunction getExpiredMultipartUploads( */ KeyDeletingService getDeletingService(); - OmMultipartUploadList listMultipartUploads(String volumeName, String bucketName, String prefix, String keyMarker, String uploadIdMarker, int maxUploads, boolean withPagination) throws OMException; diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java index 26d9b9b21aee..ec448dabf239 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/KeyManagerImpl.java @@ -467,6 +467,7 @@ public EncryptedKeyVersion run() throws IOException { Preconditions.checkNotNull(edek); return edek; } + @Override public OmKeyInfo lookupKey(OmKeyArgs args, ResolvedBucket bucket, String clientAddress) throws IOException { @@ -606,6 +607,7 @@ private void addBlockToken4Read(OmKeyInfo value) throws IOException { } } } + /** * Refresh pipeline info in OM by asking SCM. * @param keyList a list of OmKeyInfo @@ -1541,6 +1543,7 @@ private OmKeyInfo createDirectoryKey(OmKeyInfo keyInfo, String keyName) .setOwnerName(keyInfo.getOwnerName()) .build(); } + /** * OzoneFS api to lookup for a file. * @@ -2057,6 +2060,7 @@ private static List toNodeUuid(Collection nodes) { } return nodeSet; } + private void slimLocationVersion(OmKeyInfo... keyInfos) { if (keyInfos != null) { for (OmKeyInfo keyInfo : keyInfos) { diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMDBCheckpointServlet.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMDBCheckpointServlet.java index 727eb5e710cd..6117a7e373bf 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMDBCheckpointServlet.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMDBCheckpointServlet.java @@ -252,11 +252,11 @@ public DBCheckpoint getCheckpoint(Path tmpdir, boolean flush) throws IOException return dbCheckpoint; } - // Convenience class for keeping track of the tmp dirs. static class DirectoryData { private final File originalDir; private final File tmpDir; + DirectoryData(Path tmpdir, String dirStr) throws IOException { originalDir = new File(dirStr); tmpDir = new File(tmpdir.toString(), getOriginalDir().getName()); diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMMetrics.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMMetrics.java index 89c306fe25a5..c294325ef2b9 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMMetrics.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMMetrics.java @@ -359,7 +359,6 @@ public long getNumKeys() { return numKeys.value(); } - public void incNumVolumeCreates() { numVolumeOps.incr(); numVolumeCreates.incr(); @@ -555,6 +554,7 @@ public void decNumSnapshotDeleted() { public int getNumSnapshotCacheSize() { return numSnapshotCacheSize.value(); } + public void incNumSnapshotCacheSize() { numSnapshotCacheSize.incr(); } @@ -584,6 +584,7 @@ public void incNumListMultipartUploads() { public void incNumAbortMultipartUploadFails() { numAbortMultipartUploadFails.incr(); } + public void incNumListMultipartUploadParts() { numKeyOps.incr(); numListMultipartUploadParts.incr(); diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMMultiTenantManager.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMMultiTenantManager.java index f0a5d68a1fc9..b2dbba1d323e 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMMultiTenantManager.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMMultiTenantManager.java @@ -136,6 +136,7 @@ boolean isTenantAdmin(UserGroupInformation callerUgi, String tenantId, */ TenantUserList listUsersInTenant(String tenantID, String prefix) throws IOException; + /** * Given an access ID return its corresponding tenant. * @param accessID diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMPerformanceMetrics.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMPerformanceMetrics.java index 78d10acd6d75..19a77fea7a44 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMPerformanceMetrics.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMPerformanceMetrics.java @@ -61,7 +61,6 @@ public static void unregister() { @Metric(about = "resolveBucketLink latency nanoseconds") private MutableRate lookupResolveBucketLatencyNs; - @Metric(about = "Overall getKeyInfo in nanoseconds") private MutableRate getKeyInfoLatencyNs; @@ -169,7 +168,6 @@ MutableRate getLookupRefreshLocationLatencyNs() { return lookupRefreshLocationLatencyNs; } - MutableRate getLookupGenerateBlockTokenLatencyNs() { return lookupGenerateBlockTokenLatencyNs; } diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMStarterInterface.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMStarterInterface.java index 1800440c2218..94be4139bc2a 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMStarterInterface.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OMStarterInterface.java @@ -28,10 +28,13 @@ public interface OMStarterInterface { void start(OzoneConfiguration conf) throws IOException, AuthenticationException; + boolean init(OzoneConfiguration conf) throws IOException, AuthenticationException; + void bootstrap(OzoneConfiguration conf, boolean force) throws IOException, AuthenticationException; + void startAndCancelPrepare(OzoneConfiguration conf) throws IOException, AuthenticationException; } diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmMetadataManagerImpl.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmMetadataManagerImpl.java index d5fe33958aab..01d187fffd69 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmMetadataManagerImpl.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmMetadataManagerImpl.java @@ -398,7 +398,6 @@ private OmMetadataManagerImpl(OzoneConfiguration conf, File dir, String name) perfMetrics = null; } - // metadata constructor for snapshots OmMetadataManagerImpl(OzoneConfiguration conf, String snapshotDirName, boolean isSnapshotInCache, int maxOpenFiles) throws IOException { @@ -2148,6 +2147,7 @@ public String getRenameKey(String volumeName, String bucketName, renameKey.append(OM_KEY_PREFIX).append(objectID); return renameKey.toString(); } + @Override public String getMultipartKey(long volumeId, long bucketId, long parentID, String fileName, diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmSnapshot.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmSnapshot.java index bd9a3bc764e9..426aa3000445 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmSnapshot.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmSnapshot.java @@ -205,7 +205,6 @@ private OzoneObj normalizeOzoneObj(OzoneObj o) { } - // Remove snapshot indicator from keyname private String normalizeKeyName(String keyname) { if (keyname == null) { diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmSnapshotManager.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmSnapshotManager.java index 8b27bac88364..6629e911af98 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmSnapshotManager.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmSnapshotManager.java @@ -695,7 +695,6 @@ public static boolean areSnapshotChangesFlushedToDB(OMMetadataManager metadataMa return true; } - /** * Returns OmSnapshot object and skips active check. * This should only be used for API calls initiated by background service e.g. purgeKeys, purgeSnapshot, diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmSnapshotMetrics.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmSnapshotMetrics.java index de56e39e8263..3fe8c154cd3d 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmSnapshotMetrics.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OmSnapshotMetrics.java @@ -162,7 +162,6 @@ public void incNumGetAcl() { private @Metric MutableCounterLong numFSOps; - private @Metric MutableCounterLong numGetObjectTagging; private @Metric diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneConfigUtil.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneConfigUtil.java index e5433e32892f..3e2905629661 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneConfigUtil.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneConfigUtil.java @@ -30,6 +30,7 @@ public final class OzoneConfigUtil { static final Logger LOG = LoggerFactory.getLogger(OzoneConfigUtil.class); + private OzoneConfigUtil() { } diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneManager.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneManager.java index b244153c0623..0624f0ee0e92 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneManager.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneManager.java @@ -365,7 +365,6 @@ public final class OzoneManager extends ServiceRuntimeInfoImpl "persists, the service may be down. Possible cause: only one OM is up, or " + "other OMs are unable to respond to Ratis leader vote messages"; - // This is set for read requests when OMRequest has S3Authentication set, // and it is reset when read request is processed. private static final ThreadLocal S3_AUTH = @@ -1759,6 +1758,7 @@ public OMMetrics getMetrics() { public OMPerformanceMetrics getPerfMetrics() { return perfMetrics; } + public DeletingServiceMetrics getDeletionMetrics() { return omDeletionMetrics; } @@ -2720,8 +2720,6 @@ public boolean checkAcls(ResourceType resType, StoreType storeType, return omMetadataReader.checkAcls(obj, context, throwIfPermissionDenied); } - - /** * Return true if Ozone acl's are enabled, else false. * @@ -4325,6 +4323,7 @@ public String getComponent() { public long getMaxUserVolumeCount() { return config.getMaxUserVolumeCount(); } + /** * Return true, if the current OM node is leader and in ready state to * process the requests. @@ -4881,7 +4880,6 @@ public Map getObjectTagging(final OmKeyArgs args) } } - /** * Write down Layout version of a finalized feature to DB on finalization. * @param lvm OMLayoutVersionManager diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneManagerUtils.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneManagerUtils.java index e2fb89e16199..10331c99dfad 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneManagerUtils.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneManagerUtils.java @@ -86,8 +86,6 @@ private static void reportNotFound(OMMetadataManager metaMgr, OMException.ResultCodes.BUCKET_NOT_FOUND); } - - /** * Get bucket layout for the given volume and bucket name. * diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneTrash.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneTrash.java index 3136beb709ce..b0e95652b670 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneTrash.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/OzoneTrash.java @@ -29,11 +29,13 @@ public class OzoneTrash extends Trash { private TrashPolicy trashPolicy; + public OzoneTrash(FileSystem fs, Configuration conf, OzoneManager om) throws IOException { super(fs, conf); this.trashPolicy = new TrashPolicyOzone(fs, conf, om); } + @Override public Runnable getEmptier() throws IOException { return this.trashPolicy.getEmptier(); diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/PrefixManagerImpl.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/PrefixManagerImpl.java index 7ae293971da8..50c1574ecb20 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/PrefixManagerImpl.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/PrefixManagerImpl.java @@ -90,7 +90,6 @@ private void loadPrefixTree() { } } - @Override public OMMetadataManager getMetadataManager() { return metadataManager; diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ResolvedBucket.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ResolvedBucket.java index a34758b32e8b..4e976e1a2764 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ResolvedBucket.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ResolvedBucket.java @@ -139,7 +139,6 @@ public boolean isDangling() { return realVolume == null || realBucket == null; } - public Map audit() { return audit(new LinkedHashMap<>()); } diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/SstFilteringService.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/SstFilteringService.java index a4bc9e4f0d8d..99d749a9fb56 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/SstFilteringService.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/SstFilteringService.java @@ -120,14 +120,12 @@ public void resume() { running.set(true); } - private class SstFilteringTask implements BackgroundTask { private boolean isSnapshotDeleted(SnapshotInfo snapshotInfo) { return snapshotInfo == null || snapshotInfo.getSnapshotStatus() == SnapshotInfo.SnapshotStatus.SNAPSHOT_DELETED; } - /** * Marks the snapshot as SSTFiltered by creating a file in snapshot directory. * @param snapshotInfo snapshotInfo diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/TrashOzoneFileSystem.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/TrashOzoneFileSystem.java index 723b977111f2..f794fae7a77f 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/TrashOzoneFileSystem.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/TrashOzoneFileSystem.java @@ -204,7 +204,6 @@ public FileStatus[] listStatus(Path path) throws IOException { return fileStatuses.toArray(new FileStatus[0]); } - /** * converts OzoneFileStatus object to FileStatus. */ @@ -244,7 +243,6 @@ public boolean mkdirs(Path path, "fs.mkdirs() not implemented in TrashOzoneFileSystem"); } - @Override public FileStatus getFileStatus(Path path) throws IOException { ozoneManager.getMetrics().incNumGetFileStatus(); @@ -539,7 +537,6 @@ private class DeleteIterator extends OzoneListingIterator { private final boolean recursive; private List keysList; - DeleteIterator(Path f, boolean recursive) throws IOException { super(f); @@ -617,7 +614,4 @@ OzoneManagerProtocolProtos.UserInfo getUserInfo() throws IOException { return userInfo.build(); } - - - } diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/TrashPolicyOzone.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/TrashPolicyOzone.java index 34813bb9a1df..47547af8cc13 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/TrashPolicyOzone.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/TrashPolicyOzone.java @@ -103,7 +103,6 @@ protected class Emptier implements Runnable { // same as checkpoint interval private long emptierInterval; - private ThreadPoolExecutor executor; Emptier(OzoneConfiguration conf, long emptierInterval, @@ -215,6 +214,7 @@ private Runnable getEmptierTask(Path trashRootPath, TrashPolicyOzone trash, private long ceiling(long time, long interval) { return floor(time, interval) + interval; } + private long floor(long time, long interval) { return (time / interval) * interval; } diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ha/OMHANodeDetails.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ha/OMHANodeDetails.java index 3686006f1478..af93c7a9ec9b 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ha/OMHANodeDetails.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ha/OMHANodeDetails.java @@ -84,7 +84,6 @@ public Map getPeerNodesMap() { return peerNodesMap; } - /** * Inspects and loads OM node configurations. * diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/helpers/OmFSOFile.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/helpers/OmFSOFile.java index 534f75080f0b..8cc88a23bb1c 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/helpers/OmFSOFile.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/helpers/OmFSOFile.java @@ -36,7 +36,6 @@ public final class OmFSOFile { private long bucketId; private long parentID; - @SuppressWarnings("checkstyle:parameternumber") private OmFSOFile(String volumeName, String bucketName, String keyName, OMMetadataManager omMetadataManager, String fileName, diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ratis_snapshot/OmRatisSnapshotProvider.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ratis_snapshot/OmRatisSnapshotProvider.java index 35ae4a60fec0..a343232c39e3 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ratis_snapshot/OmRatisSnapshotProvider.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/ratis_snapshot/OmRatisSnapshotProvider.java @@ -91,7 +91,6 @@ public OmRatisSnapshotProvider(File snapshotDir, this.connectionFactory = connectionFactory; } - public OmRatisSnapshotProvider(MutableConfigurationSource conf, File omRatisSnapshotDir, Map peerNodeDetails) { super(omRatisSnapshotDir, OM_DB_NAME); diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/OMClientRequest.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/OMClientRequest.java index 9753fa3e0a0e..592e375ae229 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/OMClientRequest.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/OMClientRequest.java @@ -98,6 +98,7 @@ public OMClientRequest(OMRequest omRequest) { this.omRequest = omRequest; this.omLockDetails.clear(); } + /** * Perform pre-execute steps on a OMRequest. * @@ -511,7 +512,6 @@ public Map buildVolumeAuditMap(String volume) { return auditMap; } - public static String validateAndNormalizeKey(boolean enableFileSystemPaths, String keyName) throws OMException { if (enableFileSystemPaths) { diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/bucket/OMBucketCreateRequest.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/bucket/OMBucketCreateRequest.java index 53d0323399a2..e1f15a8246e5 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/bucket/OMBucketCreateRequest.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/bucket/OMBucketCreateRequest.java @@ -83,6 +83,7 @@ public class OMBucketCreateRequest extends OMClientRequest { public OMBucketCreateRequest(OMRequest omRequest) { super(omRequest); } + @Override public OMRequest preExecute(OzoneManager ozoneManager) throws IOException { @@ -311,7 +312,6 @@ private boolean isECBucket(BucketInfo bucketInfo) { .getDefaultReplicationConfig().hasEcReplicationConfig(); } - /** * Add default acls for bucket. These acls are inherited from volume * default acl list. diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/bucket/acl/OMBucketAclRequest.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/bucket/acl/OMBucketAclRequest.java index 15724b605c6a..609f5390c0cb 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/bucket/acl/OMBucketAclRequest.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/bucket/acl/OMBucketAclRequest.java @@ -170,7 +170,6 @@ public OMClientResponse validateAndUpdateCache(OzoneManager ozoneManager, Execut */ abstract String getPath(); - /** * Get the Bucket object Info from the request. * @return OzoneObjInfo diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/key/OMKeyRenameRequest.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/key/OMKeyRenameRequest.java index d59c5f37b8ad..645bfc78546d 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/key/OMKeyRenameRequest.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/key/OMKeyRenameRequest.java @@ -112,7 +112,6 @@ protected KeyArgs resolveBucketAndCheckAcls(KeyArgs keyArgs, return resolvedArgs; } - @Override @SuppressWarnings("methodlength") public OMClientResponse validateAndUpdateCache(OzoneManager ozoneManager, ExecutionContext context) { diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/s3/multipart/S3MultipartUploadCompleteRequestWithFSO.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/s3/multipart/S3MultipartUploadCompleteRequestWithFSO.java index 738fb860fbbe..ff967e731172 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/s3/multipart/S3MultipartUploadCompleteRequestWithFSO.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/s3/multipart/S3MultipartUploadCompleteRequestWithFSO.java @@ -99,7 +99,6 @@ protected void addMultiPartToCache( } - @Override protected OmKeyInfo getOmKeyInfoFromKeyTable(String dbOzoneFileKey, String keyName, OMMetadataManager omMetadataManager) throws IOException { diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/snapshot/OMSnapshotRenameRequest.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/snapshot/OMSnapshotRenameRequest.java index 5a7bc8dcb9d0..060e1d6ec828 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/snapshot/OMSnapshotRenameRequest.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/snapshot/OMSnapshotRenameRequest.java @@ -108,7 +108,6 @@ public OMRequest preExecute(OzoneManager ozoneManager) throws IOException { return omRequestBuilder.build(); } - @Override public OMClientResponse validateAndUpdateCache(OzoneManager ozoneManager, ExecutionContext context) { boolean acquiredBucketLock = false; diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/util/OMMultipartUploadUtils.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/util/OMMultipartUploadUtils.java index 9bee8b144036..75ebcd64ddbd 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/util/OMMultipartUploadUtils.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/util/OMMultipartUploadUtils.java @@ -32,7 +32,6 @@ */ public final class OMMultipartUploadUtils { - private OMMultipartUploadUtils() { } @@ -95,7 +94,6 @@ public static String getMultipartOpenKey(String volumeName, } } - /** * Check whether key's isMultipartKey flag is set. * @param openKeyInfo open key diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/validation/VersionExtractor.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/validation/VersionExtractor.java index cd898744ae7e..b8a9f8f6ce51 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/validation/VersionExtractor.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/validation/VersionExtractor.java @@ -60,5 +60,6 @@ public Class getVersionClass() { }; public abstract Versioned extractVersion(OMRequest req, ValidationContext ctx); + public abstract Class getVersionClass(); } diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/volume/OMVolumeRequest.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/volume/OMVolumeRequest.java index 74ccae327b13..116758c71ed4 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/volume/OMVolumeRequest.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/request/volume/OMVolumeRequest.java @@ -74,7 +74,6 @@ protected PersistedUserVolumeInfo delVolumeFromOwnerList( return newVolList; } - /** * Add volume to user volume list. This method should be called after * acquiring user lock. diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/response/OMClientResponse.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/response/OMClientResponse.java index a7aeff8d1b32..39c09f6f4b71 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/response/OMClientResponse.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/response/OMClientResponse.java @@ -88,7 +88,6 @@ public CompletableFuture getFlushFuture() { return flushFuture; } - public OMLockDetails getOmLockDetails() { return omLockDetails; } diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/response/key/AbstractOMKeyDeleteResponse.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/response/key/AbstractOMKeyDeleteResponse.java index 86f4965d9ba5..f3896b284b32 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/response/key/AbstractOMKeyDeleteResponse.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/response/key/AbstractOMKeyDeleteResponse.java @@ -130,7 +130,6 @@ protected void addDeletionToBatch( } } - @Override public abstract void addToDBBatch(OMMetadataManager omMetadataManager, BatchOperation batchOperation) throws IOException; diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/response/upgrade/OMFinalizeUpgradeResponse.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/response/upgrade/OMFinalizeUpgradeResponse.java index 2347f56a4aa0..6b62dab349b2 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/response/upgrade/OMFinalizeUpgradeResponse.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/response/upgrade/OMFinalizeUpgradeResponse.java @@ -37,6 +37,7 @@ public class OMFinalizeUpgradeResponse extends OMClientResponse { private static final Logger LOG = LoggerFactory.getLogger(OMFinalizeUpgradeResponse.class); private int layoutVersionToWrite = -1; + public OMFinalizeUpgradeResponse( OzoneManagerProtocolProtos.OMResponse omResponse, int layoutVersionToWrite) { diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/service/QuotaRepairTask.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/service/QuotaRepairTask.java index b35a95a2e899..99cebe9a6351 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/service/QuotaRepairTask.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/service/QuotaRepairTask.java @@ -75,6 +75,7 @@ public class QuotaRepairTask { private static final AtomicLong RUN_CNT = new AtomicLong(0); private final OzoneManager om; private ExecutorService executor; + public QuotaRepairTask(OzoneManager ozoneManager) { this.om = ozoneManager; } @@ -96,6 +97,7 @@ public CompletableFuture repair(List buckets) throws IOExceptio public static String getStatus() { return REPAIR_STATUS.toString(); } + private boolean repairTask(List buckets) { LOG.info("Starting quota repair task {}", REPAIR_STATUS); OMMetadataManager activeMetaManager = null; diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/snapshot/RocksDbPersistentSet.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/snapshot/RocksDbPersistentSet.java index 34fed72d214a..53d15b9f88d1 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/snapshot/RocksDbPersistentSet.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/snapshot/RocksDbPersistentSet.java @@ -45,7 +45,6 @@ public RocksDbPersistentSet(ManagedRocksDB db, this.entryType = entryType; } - @Override public void add(E entry) { try { diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/snapshot/SnapshotDiffManager.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/snapshot/SnapshotDiffManager.java index 285874a86788..fcb29baefd9e 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/snapshot/SnapshotDiffManager.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/snapshot/SnapshotDiffManager.java @@ -1485,6 +1485,7 @@ private boolean isObjectModified(String fromObjectName, String toObjectName, toObject.getClass().getName()); } } + private boolean areAclsSame(OmDirectoryInfo fromObject, OmDirectoryInfo toObject) { return fromObject.getAcls().equals(toObject.getAcls()); diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/upgrade/QuotaRepairUpgradeAction.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/upgrade/QuotaRepairUpgradeAction.java index 5a5aa0875022..8637713cb4bd 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/upgrade/QuotaRepairUpgradeAction.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/om/upgrade/QuotaRepairUpgradeAction.java @@ -34,6 +34,7 @@ @UpgradeActionOm(type = ON_FINALIZE, feature = QUOTA) public class QuotaRepairUpgradeAction implements OmUpgradeAction { private static final Logger LOG = LoggerFactory.getLogger(QuotaRepairUpgradeAction.class); + @Override public void execute(OzoneManager arg) throws Exception { boolean enabled = arg.getConfiguration().getBoolean( diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/protocolPB/OzoneManagerProtocolServerSideTranslatorPB.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/protocolPB/OzoneManagerProtocolServerSideTranslatorPB.java index 1cf6c1d5690c..251e81e83ed3 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/protocolPB/OzoneManagerProtocolServerSideTranslatorPB.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/protocolPB/OzoneManagerProtocolServerSideTranslatorPB.java @@ -70,7 +70,6 @@ public class OzoneManagerProtocolServerSideTranslatorPB implements OzoneManagerP private OMRequest lastRequestToSubmit; - /** * Constructs an instance of the server handler. * diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/protocolPB/OzoneManagerRequestHandler.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/protocolPB/OzoneManagerRequestHandler.java index 2810f737388f..b631552cc122 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/protocolPB/OzoneManagerRequestHandler.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/protocolPB/OzoneManagerRequestHandler.java @@ -179,7 +179,6 @@ public class OzoneManagerRequestHandler implements RequestHandler { private final OzoneManager impl; private FaultInjector injector; - public OzoneManagerRequestHandler(OzoneManager om) { this.impl = om; } @@ -488,6 +487,7 @@ protected Status exceptionToResponseStatus(IOException ex) { return Status.INTERNAL_ERROR; } } + /** * Validates that the incoming OM request has required parameters. * TODO: Add more validation checks before writing the request to Ratis log. @@ -1114,7 +1114,6 @@ public static OMResponse disallowGetFileStatusWithECReplicationConfig( return resp; } - @RequestFeatureValidator( conditions = ValidationCondition.OLDER_CLIENT_REQUESTS, processingPhase = RequestProcessingPhase.POST_PROCESS, @@ -1573,6 +1572,7 @@ private OzoneManagerProtocolProtos.GetQuotaRepairStatusResponse getQuotaRepairSt .setStatus(impl.getQuotaRepairStatus()) .build(); } + private OzoneManagerProtocolProtos.StartQuotaRepairResponse startQuotaRepair( OzoneManagerProtocolProtos.StartQuotaRepairRequest req) throws IOException { impl.startQuotaRepair(req.getBucketsList()); diff --git a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/security/OzoneSecretStore.java b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/security/OzoneSecretStore.java index 456510fb5fd7..fad7721e01cd 100644 --- a/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/security/OzoneSecretStore.java +++ b/hadoop-ozone/ozone-manager/src/main/java/org/apache/hadoop/ozone/security/OzoneSecretStore.java @@ -36,6 +36,7 @@ public class OzoneSecretStore implements Closeable { private static final Logger LOG = LoggerFactory .getLogger(OzoneSecretStore.class); private OMMetadataManager omMetadataManager; + @Override public void close() throws IOException { if (omMetadataManager != null) { @@ -47,12 +48,12 @@ public void close() throws IOException { } } - /** * Support class to maintain state of OzoneSecretStore. */ public static class OzoneManagerSecretState { private Map tokenState = new HashMap<>(); + public Map getTokenState() { return tokenState; } diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/OmTestManagers.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/OmTestManagers.java index c2cf77fb8033..154f608d3248 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/OmTestManagers.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/OmTestManagers.java @@ -57,27 +57,35 @@ public final class OmTestManagers { public OzoneManager getOzoneManager() { return om; } + public OzoneManagerProtocol getWriteClient() { return writeClient; } + public BucketManager getBucketManager() { return bucketManager; } + public VolumeManager getVolumeManager() { return volumeManager; } + public PrefixManager getPrefixManager() { return prefixManager; } + public OMMetadataManager getMetadataManager() { return metadataManager; } + public KeyManager getKeyManager() { return keyManager; } + public ScmBlockLocationProtocol getScmBlockClient() { return scmBlockClient; } + public OzoneClient getRpcClient() { return rpcClient; } diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestBucketManagerImpl.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestBucketManagerImpl.java index 3dfe9de0203b..0ec33ca14117 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestBucketManagerImpl.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestBucketManagerImpl.java @@ -143,7 +143,6 @@ void testCreateEncryptedBucket() throws Exception { bucketInfo.getEncryptionKeyInfo().getKeyName()); } - @Test public void testCreateBucket() throws Exception { String volume = volumeName(); diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestOmMetadataManager.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestOmMetadataManager.java index 6bb8b2b8ce23..0f7401fb8728 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestOmMetadataManager.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestOmMetadataManager.java @@ -97,7 +97,6 @@ public class TestOmMetadataManager { @TempDir private File folder; - @BeforeEach public void setup() throws Exception { ozoneConfiguration = new OzoneConfiguration(); @@ -1030,7 +1029,6 @@ public void testListSnapshot() throws Exception { assertNull(lastSnapshot); } - @ParameterizedTest @MethodSource("listSnapshotWithInvalidPathCases") public void testListSnapshotWithInvalidPath(String volume, diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestOmSnapshotManager.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestOmSnapshotManager.java index d372382510a9..7cb1e2e38ab8 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestOmSnapshotManager.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestOmSnapshotManager.java @@ -330,7 +330,6 @@ public void testHardLinkCreation() throws IOException { getINode(f1FileLink.toPath()), "link matches original file"); } - @Test public void testGetSnapshotInfo() throws IOException { SnapshotInfo s1 = createSnapshotInfo("vol", "buck"); @@ -353,6 +352,7 @@ public void testGetSnapshotInfo() throws IOException { () -> om.getOmSnapshotManager().getSnapshot(s2.getSnapshotId())); assertEquals(OMException.ResultCodes.FILE_NOT_FOUND, ome.getResult()); } + /* * Test that exclude list is generated correctly. */ diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHttpServer.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHttpServer.java index 9cc925cbb920..a9c1cf2e2055 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHttpServer.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestOzoneManagerHttpServer.java @@ -51,6 +51,7 @@ public class TestOzoneManagerHttpServer { private static OzoneConfiguration conf; private static URLConnectionFactory connectionFactory; private static File ozoneMetadataDirectory; + public static Collection policy() { Object[][] params = new Object[][] { {HttpConfig.Policy.HTTP_ONLY}, diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestScmClient.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestScmClient.java index fdbad6119d7b..750fb9bf6014 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestScmClient.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestScmClient.java @@ -97,6 +97,7 @@ true, newHashSet(3L, 4L)), true, newHashSet(1L, 3L, 4L)) ); } + @ParameterizedTest @MethodSource("getContainerLocationsTestCases") public void testGetContainerLocations(String testCaseName, diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/lock/TestOzoneLockProvider.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/lock/TestOzoneLockProvider.java index 577a09fbe519..a1b46df3d7d8 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/lock/TestOzoneLockProvider.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/lock/TestOzoneLockProvider.java @@ -49,6 +49,7 @@ public static Collection data() { new Object[]{false, true}, new Object[]{false, false}); } + private boolean keyPathLockEnabled; private boolean enableFileSystemPaths; diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ratis/TestOzoneManagerRatisServer.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ratis/TestOzoneManagerRatisServer.java index a86162d3ee3b..f832c61d9e8c 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ratis/TestOzoneManagerRatisServer.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ratis/TestOzoneManagerRatisServer.java @@ -236,6 +236,4 @@ public void verifyRaftGroupIdGenerationWithCustomOmServiceId(@TempDir Path metaD assertEquals(raftGroupId.toByteString().size(), 16); newOmRatisServer.stop(); } - - } diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/OMRequestTestUtils.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/OMRequestTestUtils.java index 60990b34f352..7b22c26c2bb4 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/OMRequestTestUtils.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/OMRequestTestUtils.java @@ -821,7 +821,6 @@ public static OMRequest createSetVolumePropertyRequest(String volumeName, .setSetVolumePropertyRequest(setVolumePropertyRequest).build(); } - /** * Create OMRequest for set volume property request with quota set. * @param volumeName diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/TestNormalizePaths.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/TestNormalizePaths.java index 0889a532727f..59aa6f22737d 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/TestNormalizePaths.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/TestNormalizePaths.java @@ -83,8 +83,6 @@ private void checkInvalidPath(String keyName) { assertThat(ex.getMessage()).contains("Invalid KeyPath"); } - - @Test public void testNormalizePathsDisable() throws OMException { diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMAllocateBlockRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMAllocateBlockRequest.java index e125b6122a00..1318e5b0645f 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMAllocateBlockRequest.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMAllocateBlockRequest.java @@ -225,7 +225,6 @@ protected OMRequest doPreExecute(OMRequest originalOMRequest) return modifiedOmRequest; } - protected OMRequest createAllocateBlockRequest() { KeyArgs keyArgs = KeyArgs.newBuilder() diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMDirectoriesPurgeRequestAndResponse.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMDirectoriesPurgeRequestAndResponse.java index e1ae9ef3ccd9..928c375c706a 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMDirectoriesPurgeRequestAndResponse.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMDirectoriesPurgeRequestAndResponse.java @@ -143,6 +143,7 @@ private OMRequest createPurgeKeysRequest(String fromSnapshot, String purgeDelete .build(); return omRequest; } + private OzoneManagerProtocolProtos.PurgePathRequest wrapPurgeRequest( final long volumeId, final long bucketId, final String purgeDeletedDir, final List purgeDeletedFiles, final List markDirsAsDeleted) { diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCommitRequestWithFSO.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCommitRequestWithFSO.java index 6ba5a778d9c7..07ddf966a83b 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCommitRequestWithFSO.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCommitRequestWithFSO.java @@ -44,6 +44,7 @@ public class TestOMKeyCommitRequestWithFSO extends TestOMKeyCommitRequest { private long getVolumeID() throws IOException { return omMetadataManager.getVolumeId(volumeName); } + private long getBucketID() throws java.io.IOException { String bucketKey = omMetadataManager.getBucketKey(volumeName, bucketName); OmBucketInfo omBucketInfo = diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCreateRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCreateRequest.java index 049d39d205a9..2a7632705178 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCreateRequest.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyCreateRequest.java @@ -346,8 +346,6 @@ public void testValidateAndUpdateCacheWithNoSuchMultipartUploadError( assertNull(omKeyInfo); } - - @ParameterizedTest @MethodSource("data") public void testValidateAndUpdateCacheWithVolumeNotFound( @@ -388,7 +386,6 @@ public void testValidateAndUpdateCacheWithVolumeNotFound( } - @ParameterizedTest @MethodSource("data") public void testValidateAndUpdateCacheWithBucketNotFound( @@ -431,7 +428,6 @@ public void testValidateAndUpdateCacheWithBucketNotFound( } - @ParameterizedTest @MethodSource("data") public void testValidateAndUpdateCacheWithInvalidPath( @@ -567,7 +563,6 @@ public void testCreationWithoutMetadataFollowedByOverwriteWithMetadata( verifyMetadataInResponse(overwriteResponse, overwriteMetadata); } - private void verifyMetadataInResponse(OMClientResponse response, Map expectedMetadata) { // Extract metadata from the response @@ -1050,7 +1045,6 @@ protected void addToKeyTable(String keyName) throws Exception { keyName.substring(1), 0L, RatisReplicationConfig.getInstance(THREE), omMetadataManager); } - private void checkNotAValidPath(String keyName) throws IOException { OMRequest omRequest = createKeyRequest(false, 0, keyName); OMKeyCreateRequest omKeyCreateRequest = getOMKeyCreateRequest(omRequest); diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyRenameRequestWithFSO.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyRenameRequestWithFSO.java index 107fba591547..20d142fdadb8 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyRenameRequestWithFSO.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyRenameRequestWithFSO.java @@ -49,6 +49,7 @@ public class TestOMKeyRenameRequestWithFSO extends TestOMKeyRenameRequest { private OmKeyInfo fromKeyParentInfo; private OmKeyInfo toKeyParentInfo; + @Override @BeforeEach public void createParentKey() throws Exception { diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyRequest.java index 334659105c09..dae04b2e0ed7 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyRequest.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeyRequest.java @@ -264,7 +264,6 @@ protected OzoneConfiguration getOzoneConfiguration() { return new OzoneConfiguration(); } - /** * Verify path in open key table. Also, it returns OMKeyInfo for the given * key path. diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeysDeleteRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeysDeleteRequest.java index 874e6f0349bc..4f5989e386d0 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeysDeleteRequest.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/key/TestOMKeysDeleteRequest.java @@ -43,7 +43,6 @@ */ public class TestOMKeysDeleteRequest extends TestOMKeyRequest { - private List deleteKeyList; private OMRequest omRequest; diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3ExpiredMultipartUploadsAbortRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3ExpiredMultipartUploadsAbortRequest.java index c6a10d3a3b07..fb23385b7c8c 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3ExpiredMultipartUploadsAbortRequest.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3ExpiredMultipartUploadsAbortRequest.java @@ -336,7 +336,6 @@ private void abortExpiredMPUsFromCache(String volumeName, String bucketName, .collect(Collectors.toList())); } - /** * Runs the validate and update cache step of * {@link S3ExpiredMultipartUploadsAbortRequest} to mark the MPUs @@ -508,7 +507,6 @@ private List createMPUsWithFSO(String volume, String bucket, return mpuKeys; } - /** * Make MPUs with same key name and randomized upload ID for LEGACY/OBS * bucket. diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3InitiateMultipartUploadRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3InitiateMultipartUploadRequest.java index 6b92cbc16c26..742d17a87b06 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3InitiateMultipartUploadRequest.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3InitiateMultipartUploadRequest.java @@ -50,7 +50,6 @@ public void testPreExecute() throws Exception { UUID.randomUUID().toString(), UUID.randomUUID().toString()); } - @Test public void testValidateAndUpdateCache() throws Exception { String volumeName = UUID.randomUUID().toString(); @@ -114,7 +113,6 @@ public void testValidateAndUpdateCache() throws Exception { } - @Test public void testValidateAndUpdateCacheWithBucketNotFound() throws Exception { String volumeName = UUID.randomUUID().toString(); diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3MultipartRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3MultipartRequest.java index 60af5dd55471..a28836782882 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3MultipartRequest.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3MultipartRequest.java @@ -112,7 +112,6 @@ public void setup() throws Exception { when(ozoneManager.getConfiguration()).thenReturn(ozoneConfiguration); } - @AfterEach public void stop() { omMetrics.unRegister(); @@ -283,7 +282,6 @@ protected OMRequest doPreExecuteCompleteMPU( } - /** * Perform preExecute of Initiate Multipart upload request for given * volume, bucket and key name. diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3MultipartUploadAbortRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3MultipartUploadAbortRequest.java index c98e51c36812..99e91eda3b1e 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3MultipartUploadAbortRequest.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3MultipartUploadAbortRequest.java @@ -33,7 +33,6 @@ */ public class TestS3MultipartUploadAbortRequest extends TestS3MultipartRequest { - @Test public void testPreExecute() throws IOException { String volumeName = UUID.randomUUID().toString(); @@ -121,7 +120,6 @@ public void testValidateAndUpdateCacheMultipartNotFound() throws Exception { } - @Test public void testValidateAndUpdateCacheVolumeNotFound() throws Exception { String volumeName = UUID.randomUUID().toString(); diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3MultipartUploadCommitPartRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3MultipartUploadCommitPartRequest.java index 01b984be0dd0..80283e3bc4ce 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3MultipartUploadCommitPartRequest.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/multipart/TestS3MultipartUploadCommitPartRequest.java @@ -189,7 +189,6 @@ public void testValidateAndUpdateCacheKeyNotFound() throws Exception { } - @Test public void testValidateAndUpdateCacheBucketFound() throws Exception { String volumeName = UUID.randomUUID().toString(); diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/security/TestS3GetSecretRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/security/TestS3GetSecretRequest.java index cdeabde7fe31..b41bc687807c 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/security/TestS3GetSecretRequest.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/security/TestS3GetSecretRequest.java @@ -226,7 +226,6 @@ private OMRequest s3GetSecretRequest(String userPrincipalStr) { ).build(); } - @Test public void testS3CacheRecordsTransactionIndex() throws IOException { // Create secrets for "alice" and "carol". @@ -545,7 +544,6 @@ public void testGetSecretWithTenant() throws IOException { assertNull(s3GetSecretResponse.getS3SecretValue()); } - private S3Secret processSuccessSecretRequest( String userPrincipalId, int txLogIndex, diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/tagging/TestS3PutObjectTaggingRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/tagging/TestS3PutObjectTaggingRequest.java index 3898f4ab0599..e2b71715e1ef 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/tagging/TestS3PutObjectTaggingRequest.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/tagging/TestS3PutObjectTaggingRequest.java @@ -175,7 +175,6 @@ public void testValidateAndUpdateCacheEmptyTagSet() throws Exception { assertEquals(tags.size(), updatedKeyInfo.getTags().size()); } - protected OMRequest doPreExecute(String volumeName, String bucketName, String keyName, diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/tagging/TestS3PutObjectTaggingRequestWithFSO.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/tagging/TestS3PutObjectTaggingRequestWithFSO.java index 2b0075fcf910..86ec90537f92 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/tagging/TestS3PutObjectTaggingRequestWithFSO.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/s3/tagging/TestS3PutObjectTaggingRequestWithFSO.java @@ -90,6 +90,4 @@ public BucketLayout getBucketLayout() { protected S3PutObjectTaggingRequest getPutObjectTaggingRequest(OMRequest originalRequest) { return new S3PutObjectTaggingRequestWithFSO(originalRequest, getBucketLayout()); } - - } diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/volume/TestOMVolumeSetOwnerRequest.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/volume/TestOMVolumeSetOwnerRequest.java index d81266141e5d..99106e43c115 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/volume/TestOMVolumeSetOwnerRequest.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/request/volume/TestOMVolumeSetOwnerRequest.java @@ -55,7 +55,6 @@ public void testPreExecute() throws Exception { assertNotEquals(modifiedRequest, originalRequest); } - @Test public void testValidateAndUpdateCacheSuccess() throws Exception { String volumeName = UUID.randomUUID().toString(); @@ -122,7 +121,6 @@ public void testValidateAndUpdateCacheSuccess() throws Exception { } - @Test public void testValidateAndUpdateCacheWithVolumeNotFound() throws Exception { @@ -173,7 +171,6 @@ public void testInvalidRequest() throws Exception { omResponse.getStatus()); } - @Test public void testOwnSameVolumeTwice() throws Exception { String volumeName = UUID.randomUUID().toString(); diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/TestCleanupTableInfo.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/TestCleanupTableInfo.java index 7def54ae5e8b..0fc3320dcf86 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/TestCleanupTableInfo.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/TestCleanupTableInfo.java @@ -154,7 +154,6 @@ public void checkAnnotationAndTableName() { reset(om); } - private Set> responseClasses() { Reflections reflections = new Reflections(OM_RESPONSE_PACKAGE); return reflections.getSubTypesOf(OMClientResponse.class); diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/bucket/TestOMBucketCreateResponse.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/bucket/TestOMBucketCreateResponse.java index bb285bca4b0b..7ae5f878a268 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/bucket/TestOMBucketCreateResponse.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/bucket/TestOMBucketCreateResponse.java @@ -95,6 +95,4 @@ public void testAddToDBBatch() throws Exception { bucketName), keyValue.getKey()); assertEquals(omBucketInfo, keyValue.getValue()); } - - } diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/bucket/TestOMBucketDeleteResponse.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/bucket/TestOMBucketDeleteResponse.java index e3fdcfaab88f..3699b91cd275 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/bucket/TestOMBucketDeleteResponse.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/bucket/TestOMBucketDeleteResponse.java @@ -48,7 +48,6 @@ public class TestOMBucketDeleteResponse { private OMMetadataManager omMetadataManager; private BatchOperation batchOperation; - @BeforeEach public void setup() throws Exception { OzoneConfiguration ozoneConfiguration = new OzoneConfiguration(); diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/bucket/TestOMBucketSetPropertyResponse.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/bucket/TestOMBucketSetPropertyResponse.java index 5ae50499f794..562549357646 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/bucket/TestOMBucketSetPropertyResponse.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/bucket/TestOMBucketSetPropertyResponse.java @@ -64,7 +64,6 @@ public void tearDown() { } } - @Test public void testAddToDBBatch() throws Exception { String volumeName = UUID.randomUUID().toString(); @@ -95,6 +94,4 @@ public void testAddToDBBatch() throws Exception { bucketName), keyValue.getKey()); assertEquals(omBucketInfo, keyValue.getValue()); } - - } diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/file/TestOMDirectoryCreateResponseWithFSO.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/file/TestOMDirectoryCreateResponseWithFSO.java index 2358112a6153..37939a9358dc 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/file/TestOMDirectoryCreateResponseWithFSO.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/file/TestOMDirectoryCreateResponseWithFSO.java @@ -131,6 +131,7 @@ private void addVolumeToDB(String volumeName) throws IOException { new CacheKey<>(omMetadataManager.getVolumeKey(volumeName)), CacheValue.get(1, volumeArgs)); } + private void addBucketToDB(String volumeName, String bucketName) throws IOException { final OmBucketInfo omBucketInfo = OmBucketInfo.newBuilder() diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeyDeleteResponse.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeyDeleteResponse.java index f332918be9df..4cbfed0789ae 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeyDeleteResponse.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeyDeleteResponse.java @@ -128,7 +128,6 @@ public void testAddToDBBatchWithNonEmptyBlocks() throws Exception { assertThat(rangeKVs.size()).isGreaterThan(0); } - @Test public void testAddToDBBatchWithErrorResponse() throws Exception { OmKeyInfo omKeyInfo = getOmKeyInfo(); diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeyRenameResponse.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeyRenameResponse.java index 1e8771c10a3d..aa9e69466959 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeyRenameResponse.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeyRenameResponse.java @@ -40,6 +40,7 @@ public class TestOMKeyRenameResponse extends TestOMKeyResponse { protected OmKeyInfo fromKeyParent; protected OmKeyInfo toKeyParent; protected OmBucketInfo bucketInfo; + @Test public void testAddToDBBatch() throws Exception { OMResponse omResponse = diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeyRenameResponseWithFSO.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeyRenameResponseWithFSO.java index 9cfa2f60c7fa..c054eb10f54c 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeyRenameResponseWithFSO.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeyRenameResponseWithFSO.java @@ -76,6 +76,7 @@ protected OMKeyRenameResponse getOMKeyRenameResponse(OMResponse response, getDBKeyName(toKeyInfo), fromKeyParent, toKeyParent, toKeyInfo, bucketInfo, false, getBucketLayout()); } + protected void createParent() { long bucketId = random.nextLong(); String fromKeyParentName = UUID.randomUUID().toString(); @@ -96,6 +97,7 @@ protected void createParent() { String bucketName = UUID.randomUUID().toString(); bucketInfo = TestOMResponseUtils.createBucket(volumeName, bucketName); } + @Override public BucketLayout getBucketLayout() { return BucketLayout.FILE_SYSTEM_OPTIMIZED; diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeysDeleteResponse.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeysDeleteResponse.java index 3057114e4c28..229f4cb459bf 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeysDeleteResponse.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/key/TestOMKeysDeleteResponse.java @@ -42,7 +42,6 @@ */ public class TestOMKeysDeleteResponse extends TestOMKeyResponse { - private List omKeyInfoList = new ArrayList<>(); private List ozoneKeys = new ArrayList<>(); diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3ExpiredMultipartUploadsAbortResponse.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3ExpiredMultipartUploadsAbortResponse.java index 3d2c35fbfeca..e8de58041b90 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3ExpiredMultipartUploadsAbortResponse.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3ExpiredMultipartUploadsAbortResponse.java @@ -218,7 +218,6 @@ public void testAddToDBBatchWithErrorResponse( } - /** * Constructs an {@link S3ExpiredMultipartUploadsAbortResponse} to abort * MPus in (@code mpusToAbort}, with the completion status set to diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3MultipartResponse.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3MultipartResponse.java index fa8e79edadbe..63372776aa9d 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3MultipartResponse.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3MultipartResponse.java @@ -80,8 +80,6 @@ public void tearDown() { } } - - public S3InitiateMultipartUploadResponse createS3InitiateMPUResponse( String volumeName, String bucketName, String keyName, String multipartUploadID) throws IOException { diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3MultipartUploadAbortResponse.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3MultipartUploadAbortResponse.java index 68aa749cbdd6..7b9de57e099c 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3MultipartUploadAbortResponse.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3MultipartUploadAbortResponse.java @@ -37,7 +37,6 @@ public class TestS3MultipartUploadAbortResponse extends TestS3MultipartResponse { - @Test public void testAddDBToBatch() throws Exception { diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3MultipartUploadCompleteResponseWithFSO.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3MultipartUploadCompleteResponseWithFSO.java index cd693fbe4eda..ca83f2abb361 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3MultipartUploadCompleteResponseWithFSO.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/s3/multipart/TestS3MultipartUploadCompleteResponseWithFSO.java @@ -136,8 +136,6 @@ public void testAddDBToBatch() throws Exception { omMetadataManager.getDeletedTable())); } - - @Test // similar to testAddDBToBatch(), but omBucketInfo is null public void testAddDBToBatchWithNullBucketInfo() throws Exception { diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/volume/TestOMVolumeResponse.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/volume/TestOMVolumeResponse.java index 220492141da1..e8d1707bbe9d 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/volume/TestOMVolumeResponse.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/response/volume/TestOMVolumeResponse.java @@ -56,6 +56,7 @@ public void tearDown() { protected OMMetadataManager getOmMetadataManager() { return omMetadataManager; } + protected BatchOperation getBatchOperation() { return batchOperation; } diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestOpenKeyCleanupService.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestOpenKeyCleanupService.java index 97fa7db3c02f..5d5677699c88 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestOpenKeyCleanupService.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestOpenKeyCleanupService.java @@ -457,7 +457,6 @@ private List getKeyInfo(BucketLayout bucketLayout, boolean openKey) { return omKeyInfo; } - void waitForOpenKeyCleanup(boolean hsync, BucketLayout layout) throws Exception { GenericTestUtils.waitFor(() -> 0 == getExpiredOpenKeys(hsync, layout), diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestSnapshotDeletingService.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestSnapshotDeletingService.java index 052292c5e623..df7dfb7ce48b 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestSnapshotDeletingService.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestSnapshotDeletingService.java @@ -54,11 +54,10 @@ public class TestSnapshotDeletingService { private SnapshotChainManager chainManager; @Mock private OmMetadataManagerImpl omMetadataManager; - private final OzoneConfiguration conf = new OzoneConfiguration();; + private final OzoneConfiguration conf = new OzoneConfiguration(); private final long sdsRunInterval = Duration.ofMillis(1000).toMillis(); private final long sdsServiceTimeout = Duration.ofSeconds(10).toMillis(); - private static Stream testCasesForIgnoreSnapshotGc() throws IOException { SnapshotInfo flushedSnapshot = SnapshotInfo.newBuilder().setSstFiltered(true) .setLastTransactionInfo(TransactionInfo.valueOf(1, 1).toByteString()) diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestSnapshotDiffCleanupService.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestSnapshotDiffCleanupService.java index 577aa1a1a295..98e7ac62cf95 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestSnapshotDiffCleanupService.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestSnapshotDiffCleanupService.java @@ -364,7 +364,6 @@ private void assertJobInPurgedTable(String jobKey, assertEquals(expectedEntriesCount, actualEntriesCount); } - private void assertNumberOfEntriesInTable(ColumnFamilyHandle table, long expectedCount) { int count = 0; diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOmSnapshotUtils.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOmSnapshotUtils.java index 85223a36fb04..9b402b8dca67 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOmSnapshotUtils.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/snapshot/TestOmSnapshotUtils.java @@ -30,6 +30,7 @@ import java.util.stream.Collectors; import org.junit.jupiter.api.Test; import org.junit.jupiter.api.io.TempDir; + /** * Class to test snapshot utilities. */ diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/upgrade/TestOMUpgradeFinalizer.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/upgrade/TestOMUpgradeFinalizer.java index ac23e47e365e..687bd9ca4a8e 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/upgrade/TestOMUpgradeFinalizer.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/upgrade/TestOMUpgradeFinalizer.java @@ -215,8 +215,6 @@ public void testFinalizationWithFailingUpgradeAction() throws Exception { assertFalse(status.msgs().isEmpty()); } - - private VerificationMode once() { return times(1); } diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/protocolPB/TestOzoneManagerRequestHandler.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/protocolPB/TestOzoneManagerRequestHandler.java index 5b289c1b8213..91b99cfe295e 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/protocolPB/TestOzoneManagerRequestHandler.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/protocolPB/TestOzoneManagerRequestHandler.java @@ -43,7 +43,6 @@ */ public class TestOzoneManagerRequestHandler { - private OzoneManagerRequestHandler getRequestHandler(int limitListKeySize) { OmConfig config = OzoneConfiguration.newInstanceOf(OmConfig.class); config.setMaxListSize(limitListKeySize); diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneTokenIdentifier.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneTokenIdentifier.java index b5b81ca5d304..ab02e7520670 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneTokenIdentifier.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneTokenIdentifier.java @@ -194,7 +194,6 @@ public void testSymmetricTokenPerf() { testSymmetricTokenPerfHelper(hmacSHA256, 1024); } - public void testSymmetricTokenPerfHelper(String hmacAlgorithm, int keyLen) { final int testTokenCount = 1000; List tokenIds = new ArrayList<>(); @@ -252,7 +251,6 @@ public void testReadWriteInProtobuf(@TempDir Path baseDir) throws IOException { } } - public OzoneTokenIdentifier getIdentifierInst() { OzoneTokenIdentifier id = new OzoneTokenIdentifier(); id.setOwner(new Text("User1")); diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/acl/TestOzoneNativeAuthorizer.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/acl/TestOzoneNativeAuthorizer.java index 48b978064eaa..fbee5cdad08c 100644 --- a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/acl/TestOzoneNativeAuthorizer.java +++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/acl/TestOzoneNativeAuthorizer.java @@ -310,7 +310,6 @@ public void testCheckAccessForPrefix( resetAclsAndValidateAccess(prefixObj, ANONYMOUS, writeClient); } - private void setVolumeAcl(List ozoneAcls) throws IOException { OzoneNativeAclTestUtil.setVolumeAcl(metadataManager, vol, ozoneAcls); } diff --git a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicOzoneClientAdapterImpl.java b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicOzoneClientAdapterImpl.java index 3ac6d0db745e..1abddfc8afc6 100644 --- a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicOzoneClientAdapterImpl.java +++ b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicOzoneClientAdapterImpl.java @@ -427,7 +427,6 @@ public FileStatusAdapter getFileStatus(String key, URI uri, } } - @Override public Iterator listKeys(String pathKey) throws IOException { incrementCounter(Statistic.OBJECTS_LIST, 1); diff --git a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicOzoneFileSystem.java b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicOzoneFileSystem.java index 51cdf6cd2963..8e1588889010 100644 --- a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicOzoneFileSystem.java +++ b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicOzoneFileSystem.java @@ -993,6 +993,7 @@ public void setTimes(Path f, long mtime, long atime) throws IOException { String key = pathToKey(qualifiedPath); adapter.setTimes(key, mtime, atime); } + /** * A private class implementation for iterating list of file status. * diff --git a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicRootedOzoneFileSystem.java b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicRootedOzoneFileSystem.java index c135e78a2325..2e05b8e401d1 100644 --- a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicRootedOzoneFileSystem.java +++ b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/BasicRootedOzoneFileSystem.java @@ -136,7 +136,6 @@ public class BasicRootedOzoneFileSystem extends FileSystem { private static final int PATH_DEPTH_TO_BUCKET = 2; private OzoneConfiguration ozoneConfiguration; - @Override public void initialize(URI name, Configuration conf) throws IOException { super.initialize(name, conf); @@ -611,6 +610,7 @@ private class DeleteIteratorWithFSO extends OzoneListingIterator { private final BasicRootedOzoneClientAdapterImpl adapterImpl; private boolean recursive; private Path f; + DeleteIteratorWithFSO(Path f, boolean recursive) throws IOException { super(f, true); @@ -666,7 +666,6 @@ OzoneListingIterator getDeleteIterator() } } - /** * Deletes the children of the input dir path by iterating though the * DeleteIterator. @@ -933,7 +932,6 @@ private FileStatus[] convertFileStatusArr( return fileStatuses; } - private List listStatusAdapter(Path f, boolean lite) throws IOException { incrementCounter(Statistic.INVOCATION_LIST_STATUS, 1); statistics.incrementReadOps(1); @@ -1420,7 +1418,6 @@ private abstract class OzoneListingIterator { this(path, false); } - /** * The output of processKey determines if further iteration through the * keys should be done or not. diff --git a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/CapableOzoneFSDataStreamOutput.java b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/CapableOzoneFSDataStreamOutput.java index 12f2f5633b6b..8ea9509e429c 100644 --- a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/CapableOzoneFSDataStreamOutput.java +++ b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/CapableOzoneFSDataStreamOutput.java @@ -35,6 +35,7 @@ public class CapableOzoneFSDataStreamOutput extends OzoneFSDataStreamOutput implements StreamCapabilities { private final boolean isHsyncEnabled; + public CapableOzoneFSDataStreamOutput(OzoneFSDataStreamOutput outputStream, boolean enabled) { super(outputStream.getByteBufferStreamOutput()); diff --git a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/CapableOzoneFSOutputStream.java b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/CapableOzoneFSOutputStream.java index feefb584e5c2..23ce9d9bb735 100644 --- a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/CapableOzoneFSOutputStream.java +++ b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/CapableOzoneFSOutputStream.java @@ -38,6 +38,7 @@ public class CapableOzoneFSOutputStream extends OzoneFSOutputStream implements StreamCapabilities { private final boolean isHsyncEnabled; + public CapableOzoneFSOutputStream(OzoneFSOutputStream outputStream, boolean enabled) { super(outputStream.getWrappedOutputStream()); diff --git a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/FileStatusAdapter.java b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/FileStatusAdapter.java index db40448941e0..cf858bb335bb 100644 --- a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/FileStatusAdapter.java +++ b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/FileStatusAdapter.java @@ -77,7 +77,6 @@ public FileStatusAdapter(long length, long diskConsumed, Path path, this.isErasureCoded = isErasureCoded; } - public Path getPath() { return path; } diff --git a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/OzoneClientUtils.java b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/OzoneClientUtils.java index 2239874c048a..a885116902b4 100644 --- a/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/OzoneClientUtils.java +++ b/hadoop-ozone/ozonefs-common/src/main/java/org/apache/hadoop/fs/ozone/OzoneClientUtils.java @@ -56,9 +56,11 @@ public final class OzoneClientUtils { static final Logger LOG = LoggerFactory.getLogger(OzoneClientUtils.class); + private OzoneClientUtils() { // Not used. } + public static BucketLayout resolveLinkBucketLayout(OzoneBucket bucket, ObjectStore objectStore, Set recoverLeaseTraced(f)); } + private boolean recoverLeaseTraced(final Path f) throws IOException { GlobalTracer.get().activeSpan().setTag("path", f.toString()); statistics.incrementWriteOps(1); diff --git a/hadoop-ozone/recon-codegen/src/main/java/org/apache/ozone/recon/codegen/JooqCodeGenerator.java b/hadoop-ozone/recon-codegen/src/main/java/org/apache/ozone/recon/codegen/JooqCodeGenerator.java index 72559a27ae28..123dd97db1de 100644 --- a/hadoop-ozone/recon-codegen/src/main/java/org/apache/ozone/recon/codegen/JooqCodeGenerator.java +++ b/hadoop-ozone/recon-codegen/src/main/java/org/apache/ozone/recon/codegen/JooqCodeGenerator.java @@ -112,6 +112,7 @@ private void generateSourceCode(String outputDir) throws Exception { */ static class LocalDataSourceProvider implements Provider { private static EmbeddedDataSource dataSource; + static { try { SqlDbUtils.createNewDerbyDatabase(JDBC_URL, RECON_SCHEMA_NAME); diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconConstants.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconConstants.java index 3b19e42d3935..a7a2a1daf80d 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconConstants.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconConstants.java @@ -73,7 +73,6 @@ private ReconConstants() { public static final String TOTAL_USED_BYTES = "TOTAL_USED_BYTES"; public static final String STAGING = ".staging_"; - // 1125899906842624L = 1PB public static final long MAX_FILE_SIZE_UPPER_BOUND = 1125899906842624L; // 1024 = 1KB @@ -83,7 +82,6 @@ private ReconConstants() { (double) MAX_FILE_SIZE_UPPER_BOUND / MIN_FILE_SIZE_UPPER_BOUND) / Math.log(2)) + 1; - // 1125899906842624L = 1PB public static final long MAX_CONTAINER_SIZE_UPPER_BOUND = 1125899906842624L; // 536870912L = 512MB diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconUtils.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconUtils.java index fa8b908b2330..86035c45df10 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconUtils.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/ReconUtils.java @@ -174,7 +174,6 @@ public void untarCheckpointFile(File tarFile, Path destPath) Archiver.extract(tarFile, destPath); } - /** * Constructs the full path of a key from its OmKeyInfo using a bottom-up approach, starting from the leaf node. * @@ -226,7 +225,6 @@ public static String constructFullPath(String keyName, long initialParentId, Str return fullPath.toString(); } - /** * Constructs the prefix path to a key from its key name and parent ID using a bottom-up approach, starting from the * leaf node. @@ -571,7 +569,6 @@ public static long getContainerSizeUpperBound(long containerSize) { return (long) Math.pow(2, (29 + binIndex)); } - public static int getFileSizeBinIndex(long fileSize) { Preconditions.checkArgument(fileSize >= 0, "fileSize = %s < 0", fileSize); diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/ClusterStateEndpoint.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/ClusterStateEndpoint.java index bb60b271fa61..3fdaca164192 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/ClusterStateEndpoint.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/ClusterStateEndpoint.java @@ -72,8 +72,6 @@ public class ClusterStateEndpoint { private OzoneConfiguration ozoneConfiguration; private final ContainerHealthSchemaManager containerHealthSchemaManager; - - @Inject ClusterStateEndpoint(OzoneStorageContainerManager reconSCM, GlobalStatsDao globalStatsDao, diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/ContainerEndpoint.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/ContainerEndpoint.java index 17382f0634b8..8ccfdbf82795 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/ContainerEndpoint.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/ContainerEndpoint.java @@ -139,7 +139,6 @@ public static DataFilter fromValue(String value) { } } - @Inject public ContainerEndpoint(OzoneStorageContainerManager reconSCM, ContainerHealthSchemaManager containerHealthSchemaManager, diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/NSSummaryEndpoint.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/NSSummaryEndpoint.java index 0da84ab066b2..3a994df8b6c4 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/NSSummaryEndpoint.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/NSSummaryEndpoint.java @@ -51,6 +51,7 @@ public class NSSummaryEndpoint { private final ReconOMMetadataManager omMetadataManager; private final OzoneStorageContainerManager reconSCM; + @Inject public NSSummaryEndpoint(ReconNamespaceSummaryManager namespaceSummaryManager, ReconOMMetadataManager omMetadataManager, diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/OMDBInsightEndpoint.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/OMDBInsightEndpoint.java index c07d841417a5..f3e0fd0a8b74 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/OMDBInsightEndpoint.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/OMDBInsightEndpoint.java @@ -105,7 +105,6 @@ public class OMDBInsightEndpoint { private ReconNamespaceSummaryManagerImpl reconNamespaceSummaryManager; private final OzoneStorageContainerManager reconSCM; - @Inject public OMDBInsightEndpoint(OzoneStorageContainerManager reconSCM, ReconOMMetadataManager omMetadataManager, @@ -559,7 +558,6 @@ private void createKeysSummaryForDeletedKey(Map keysSummary) { keysSummary.put("totalUnreplicatedDataSize", unreplicatedSizeDeleted); } - private void getPendingForDeletionDirInfo( int limit, String prevKey, KeyInsightInfoResponse pendingForDeletionKeyInfo) { @@ -1113,7 +1111,6 @@ public void searchKeysInFSO(ParamInfo paramInfo, List retrieveKeysFromTable(fileTable, paramInfo, results); } - /** * Converts a startPrefix path into an objectId path for FSO buckets, using IDs. *

@@ -1329,7 +1326,6 @@ private String createPath(OmKeyInfo omKeyInfo) { omKeyInfo.getBucketName() + OM_KEY_PREFIX + omKeyInfo.getKeyName(); } - @VisibleForTesting public GlobalStatsDao getDao() { return this.globalStatsDao; diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/UtilizationEndpoint.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/UtilizationEndpoint.java index 175a610e0554..d01c91a0db29 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/UtilizationEndpoint.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/UtilizationEndpoint.java @@ -58,6 +58,7 @@ public class UtilizationEndpoint { private ContainerCountBySizeDao containerCountBySizeDao; private static final Logger LOG = LoggerFactory .getLogger(UtilizationEndpoint.class); + @Inject public UtilizationEndpoint(FileCountBySizeDao fileCountBySizeDao, ContainerCountBySizeDao containerCountBySizeDao, diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/DatanodeMetadata.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/DatanodeMetadata.java index a3ea202b9af0..78133f5a8895 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/DatanodeMetadata.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/DatanodeMetadata.java @@ -279,6 +279,7 @@ public Builder setNetworkLocation(String networkLocation) { this.networkLocation = networkLocation; return this; } + /** * Constructs DatanodeMetadata. * diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/DatanodesResponse.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/DatanodesResponse.java index ce479b9bf34b..ac0c0585e46c 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/DatanodesResponse.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/DatanodesResponse.java @@ -47,7 +47,6 @@ public class DatanodesResponse { @JsonInclude(JsonInclude.Include.NON_EMPTY) private Map failedNodeErrorResponseMap; - public DatanodesResponse() { this(0, new ArrayList<>()); } diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/KeysResponse.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/KeysResponse.java index 14b7bf91422a..45ff76909a35 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/KeysResponse.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/KeysResponse.java @@ -52,6 +52,7 @@ public long getTotalCount() { public Collection getKeys() { return keys; } + public String getLastKey() { return lastKey; } diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/ListKeysResponse.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/ListKeysResponse.java index c1e16b58df7b..ed875f61d4cb 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/ListKeysResponse.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/ListKeysResponse.java @@ -51,7 +51,6 @@ public class ListKeysResponse { @JsonProperty("keys") private List keys; - public ListKeysResponse() { this.status = ResponseStatus.OK; this.path = ""; diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/NamespaceSummaryResponse.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/NamespaceSummaryResponse.java index 51b702dbcab4..4731da64504c 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/NamespaceSummaryResponse.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/NamespaceSummaryResponse.java @@ -111,7 +111,6 @@ public static final class Builder { private ObjectDBInfo objectDBInfo; private ResponseStatus status; - public Builder() { // Default values this.path = StringUtils.EMPTY; diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/ObjectDBInfo.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/ObjectDBInfo.java index 7067073b3971..8530d6c4cea6 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/ObjectDBInfo.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/ObjectDBInfo.java @@ -69,7 +69,6 @@ public ObjectDBInfo(OmPrefixInfo omPrefixInfo) { this.setMetadata(omPrefixInfo.getMetadata()); } - public Map getMetadata() { return metadata; } diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/PipelineMetadata.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/PipelineMetadata.java index 4cd7a6b38957..ec2a514e6bb2 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/PipelineMetadata.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/PipelineMetadata.java @@ -106,7 +106,6 @@ public int getContainers() { return containers; } - /** * Returns new builder class that builds a PipelineMetadata. * diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/PipelinesResponse.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/PipelinesResponse.java index 914053e16e92..30c173706e23 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/PipelinesResponse.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/api/types/PipelinesResponse.java @@ -37,7 +37,6 @@ public class PipelinesResponse { @JsonProperty("pipelines") private Collection pipelines; - public PipelinesResponse() { this(0, new ArrayList<>()); } diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/heatmap/HeatMapUtil.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/heatmap/HeatMapUtil.java index 30776ab51027..c58b775b59c1 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/heatmap/HeatMapUtil.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/heatmap/HeatMapUtil.java @@ -115,7 +115,6 @@ private void updateBucketSize(EntityReadAccessHeatMapResponse bucket, bucket.setSize(bucket.getSize() + keySize); } - private void addPrefixPathInfoToBucket( EntityReadAccessHeatMapResponse rootEntity, String[] split, EntityReadAccessHeatMapResponse bucket, diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/heatmap/IHeatMapProvider.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/heatmap/IHeatMapProvider.java index 995115b0833d..fe55f8431b9e 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/heatmap/IHeatMapProvider.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/heatmap/IHeatMapProvider.java @@ -72,5 +72,5 @@ default InetSocketAddress getSolrAddress() { default HealthCheckResponse doHeatMapHealthCheck() { return new HealthCheckResponse.Builder("Healthy", Response.Status.OK.getStatusCode()).build(); - }; + } } diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/scm/ReconStorageConfig.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/scm/ReconStorageConfig.java index d9c40079e099..d988d8b73cdf 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/scm/ReconStorageConfig.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/scm/ReconStorageConfig.java @@ -90,6 +90,4 @@ public String getReconCertSerialId() { public void unsetReconCertSerialId() { getStorageInfo().unsetProperty(RECON_CERT_SERIAL_ID); } - - } diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/spi/ReconContainerMetadataManager.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/spi/ReconContainerMetadataManager.java index 4b0b7df42f1c..61631fa73338 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/spi/ReconContainerMetadataManager.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/spi/ReconContainerMetadataManager.java @@ -187,7 +187,6 @@ Map getKeyPrefixesForContainer( Map getContainers(int limit, long prevContainer) throws IOException; - SeekableIterator getContainersIterator() throws IOException; /** diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/ContainerSizeCountTask.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/ContainerSizeCountTask.java index da8002e88c04..2478ddbbe956 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/ContainerSizeCountTask.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/ContainerSizeCountTask.java @@ -200,7 +200,6 @@ public void processContainers(List containers) { } } - /** * Populate DB with the counts of container sizes calculated * using the dao. @@ -354,7 +353,6 @@ private static ContainerSizeCountKey getContainerSizeCountKey( ReconUtils.getContainerSizeUpperBound(containerSize)); } - /** * The ContainerSizeCountKey class is a simple key class that has a single * field, containerSizeUpperBound, which is a Long representing the upper diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTask.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTask.java index aa0193dc1fb5..9b28f669195e 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTask.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTask.java @@ -43,7 +43,6 @@ import org.slf4j.Logger; import org.slf4j.LoggerFactory; - /** * Task to query data from OMDB and write into Recon RocksDB. * Reprocess() will take a snapshots on OMDB, and iterate the keyTable, diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTaskDbEventHandler.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTaskDbEventHandler.java index 23bee1574d98..755d966b8328 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTaskDbEventHandler.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTaskDbEventHandler.java @@ -17,7 +17,6 @@ package org.apache.hadoop.ozone.recon.tasks; - import java.io.IOException; import java.util.Map; import org.apache.hadoop.hdds.utils.db.RDBBatchOperation; diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTaskWithFSO.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTaskWithFSO.java index 9de9bfaab38c..9b865279d992 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTaskWithFSO.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTaskWithFSO.java @@ -230,6 +230,4 @@ public boolean reprocessWithFSO(OMMetadataManager omMetadataManager) { LOG.debug("Completed a reprocess run of NSSummaryTaskWithFSO"); return true; } - - } diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTaskWithOBS.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTaskWithOBS.java index 1c53e6271128..56d393ae3393 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTaskWithOBS.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/NSSummaryTaskWithOBS.java @@ -50,7 +50,6 @@ public class NSSummaryTaskWithOBS extends NSSummaryTaskDbEventHandler { private final long nsSummaryFlushToDBMaxThreshold; - public NSSummaryTaskWithOBS( ReconNamespaceSummaryManager reconNamespaceSummaryManager, ReconOMMetadataManager reconOMMetadataManager, @@ -60,7 +59,6 @@ public NSSummaryTaskWithOBS( this.nsSummaryFlushToDBMaxThreshold = nsSummaryFlushToDBMaxThreshold; } - public boolean reprocessWithOBS(OMMetadataManager omMetadataManager) { Map nsSummaryMap = new HashMap<>(); @@ -220,7 +218,6 @@ public Pair processWithOBS(OMUpdateEventBatch events, return new ImmutablePair<>(seekPos, true); } - /** * KeyTable entries don't have the parentId set. * In order to reuse the existing methods that rely on diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/OmTableHandler.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/OmTableHandler.java index f1071c4f937b..401a3c7dc099 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/OmTableHandler.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/OmTableHandler.java @@ -96,7 +96,6 @@ Triple getTableSizeAndCount( TableIterator> iterator) throws IOException; - /** * Returns the count key for the given table. * diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/OmTableInsightTask.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/OmTableInsightTask.java index 34fd1af6d695..52bd4759b868 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/OmTableInsightTask.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/tasks/OmTableInsightTask.java @@ -65,7 +65,6 @@ public class OmTableInsightTask implements ReconOmTask { private Map unReplicatedSizeMap; private Map replicatedSizeMap; - @Inject public OmTableInsightTask(GlobalStatsDao globalStatsDao, Configuration sqlConfiguration, @@ -234,7 +233,6 @@ private void handlePutEvent(OMDBUpdateEvent event, } } - private void handleDeleteEvent(OMDBUpdateEvent event, String tableName) { OmTableHandler tableHandler = tableHandlers.get(tableName); @@ -249,7 +247,6 @@ private void handleDeleteEvent(OMDBUpdateEvent event, } } - private void handleUpdateEvent(OMDBUpdateEvent event, String tableName) { diff --git a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/upgrade/ReconLayoutVersionManager.java b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/upgrade/ReconLayoutVersionManager.java index 6889b335fc89..a18a383acc2e 100644 --- a/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/upgrade/ReconLayoutVersionManager.java +++ b/hadoop-ozone/recon/src/main/java/org/apache/hadoop/ozone/recon/upgrade/ReconLayoutVersionManager.java @@ -116,7 +116,6 @@ public void finalizeLayoutFeatures(ReconStorageContainerManagerFacade scmFacade) } } - /** * Returns a list of ReconLayoutFeature objects that are registered for finalization. */ diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/OMMetadataManagerTestUtils.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/OMMetadataManagerTestUtils.java index 94927148a6f5..d61240155c03 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/OMMetadataManagerTestUtils.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/OMMetadataManagerTestUtils.java @@ -67,6 +67,7 @@ public final class OMMetadataManagerTestUtils { public static final String TEST_USER = "TestUser"; private static OzoneConfiguration configuration; + private OMMetadataManagerTestUtils() { } @@ -382,7 +383,6 @@ public static void writeOpenKeyToOm(OMMetadataManager omMetadataManager, .put(openKey, omKeyInfo); } - /** * Writes deleted key information to the Ozone Manager metadata table. * @param omMetadataManager the Ozone Manager metadata manager diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/ReconTestInjector.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/ReconTestInjector.java index 6c1542bbc00f..68c79b751d8c 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/ReconTestInjector.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/ReconTestInjector.java @@ -223,7 +223,6 @@ public OzoneConfiguration getTestOzoneConfiguration( return configuration; } - /** * Builder for Recon Test Injector. */ diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/TestReconUtils.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/TestReconUtils.java index 5cfd5f26fd47..6f613894dd82 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/TestReconUtils.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/TestReconUtils.java @@ -211,7 +211,6 @@ private static ContainerInfo.Builder getDefaultContainerInfoBuilder( .setOwner("TEST"); } - public static ContainerInfo getContainer( final HddsProtos.LifeCycleState state) { return getDefaultContainerInfoBuilder(state) diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestClusterStateEndpoint.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestClusterStateEndpoint.java index 4d22a3cb614a..a27d33940ee4 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestClusterStateEndpoint.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestClusterStateEndpoint.java @@ -81,7 +81,6 @@ public TestClusterStateEndpoint() { super(); } - @BeforeEach public void setUp() throws Exception { reconOMMetadataManager = getTestReconOmMetadataManager( @@ -121,7 +120,6 @@ public void setUp() throws Exception { reconPipelineManager.addPipeline(pipeline); } - @Test public void testGetContainerCounts() throws Exception { putContainerInfos(NUM_OPEN_CONTAINERS, @@ -162,7 +160,6 @@ public void testScmAndOmServiceId() { assertEquals("omServiceId", clusterStateResponse.getOmServiceId()); } - ContainerInfo newContainerInfo(long containerId, HddsProtos.LifeCycleState state) { return new ContainerInfo.Builder() diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestContainerEndpoint.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestContainerEndpoint.java index 4db2a276c8a1..c99081593625 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestContainerEndpoint.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestContainerEndpoint.java @@ -964,7 +964,6 @@ public void testUnhealthyContainersFilteredResponse() assertEquals(0, records.size()); } - @Test public void testUnhealthyContainersInvalidState() { WebApplicationException e = assertThrows(WebApplicationException.class, @@ -1079,7 +1078,6 @@ private void createNegativeSizeUnhealthyRecords(int negativeSize) { } } - private void createUnhealthyRecords(int missing, int overRep, int underRep, int misRep) { int cid = 0; @@ -1328,7 +1326,6 @@ public void testGetContainerInsightsNonSCMContainers() assertEquals("OM", containerDiscrepancyInfo.getExistsAt()); } - @Test public void testGetContainerInsightsNonSCMContainersWithPrevKey() throws IOException, TimeoutException { @@ -1479,7 +1476,6 @@ public void testGetContainerInsightsNonOMContainersWithPrevKey() assertEquals("SCM", containerDiscrepancyInfo2.getExistsAt()); } - @Test public void testContainerMissingFilter() throws IOException, TimeoutException { @@ -1531,7 +1527,6 @@ public void testContainerMissingFilter() assertThat(missingContainerIdsSCM).contains(2L); } - @Test public void testGetOmContainersDeletedInSCM() throws Exception { Map omContainers = @@ -1618,7 +1613,6 @@ public void testGetOmContainersDeletedInSCMPagination() throws Exception { assertEquals(1, containerDiscrepancyInfoList.size()); } - @Test public void testGetOmContainersDeletedInSCMLimitParam() throws Exception { Map omContainers = @@ -1817,6 +1811,4 @@ public void testDuplicateFSOKeysForContainerEndpoint() throws IOException { } } } - - } diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestDeletedKeysSearchEndpoint.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestDeletedKeysSearchEndpoint.java index be2b0a799547..5191808df7d9 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestDeletedKeysSearchEndpoint.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestDeletedKeysSearchEndpoint.java @@ -108,7 +108,6 @@ public void setUp() throws Exception { populateOMDB(); } - private static OMMetadataManager initializeNewOmMetadataManager(File omDbDir) throws IOException { OzoneConfiguration omConfiguration = new OzoneConfiguration(); omConfiguration.set(OZONE_OM_DB_DIRS, omDbDir.getAbsolutePath()); @@ -442,7 +441,6 @@ public void testPrevKeyProvidedStartPrefixProvided() throws IOException { assertEquals("fileb3", result.getRepeatedOmKeyInfoList().get(0).getOmKeyInfoList().get(0).getKeyName()); } - /** * Populates the OMDB with a set of deleted keys for testing purposes. * This diagram is for reference: diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryDiskUsageOrdering.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryDiskUsageOrdering.java index 3fffbd62c8d6..f5842131644d 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryDiskUsageOrdering.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryDiskUsageOrdering.java @@ -73,6 +73,7 @@ public class TestNSSummaryDiskUsageOrdering { private static final String ROOT_PATH = "/"; private static final String TEST_USER = "TestUser"; private OMMetadataManager omMetadataManager; + @BeforeEach public void setUp() throws Exception { ozoneConfiguration = new OzoneConfiguration(); diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryEndpointWithFSO.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryEndpointWithFSO.java index 30d786918b4b..084dfd11744d 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryEndpointWithFSO.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryEndpointWithFSO.java @@ -678,7 +678,6 @@ public void testQuotaUsage() throws Exception { invalidResObj.getResponseCode()); } - @Test public void testFileSizeDist() throws Exception { checkFileSizeDist(ROOT_PATH, 2, 3, 4, 1); @@ -848,7 +847,6 @@ public void testLoggingWhenParentIdIsNegative() throws IOException { "for path construction.", loggedMessage); } - /** * Write directories and keys info into OM DB. * @throws Exception diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryEndpointWithLegacy.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryEndpointWithLegacy.java index 3a37c45b5f3a..abb2869bc4ec 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryEndpointWithLegacy.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryEndpointWithLegacy.java @@ -670,7 +670,6 @@ public void testQuotaUsage() throws Exception { invalidResObj.getResponseCode()); } - @Test public void testFileSizeDist() throws Exception { checkFileSizeDist(ROOT_PATH, 2, 3, 4, 1); diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryEndpointWithOBSAndLegacy.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryEndpointWithOBSAndLegacy.java index a4fa0e220d29..00196d5f5e39 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryEndpointWithOBSAndLegacy.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestNSSummaryEndpointWithOBSAndLegacy.java @@ -273,7 +273,6 @@ public class TestNSSummaryEndpointWithOBSAndLegacy { FILE9_SIZE_WITH_REPLICA + FILE10_SIZE_WITH_REPLICA; - private static final long MULTI_BLOCK_TOTAL_SIZE_WITH_REPLICA_UNDER_KEY = FILE4_SIZE_WITH_REPLICA; @@ -339,7 +338,6 @@ public class TestNSSummaryEndpointWithOBSAndLegacy { private static final long BUCKET_FOUR_DATA_SIZE = FILE_ELEVEN_SIZE; - @BeforeEach public void setUp() throws Exception { conf = new OzoneConfiguration(); @@ -846,7 +844,6 @@ public void testQuotaUsage() throws Exception { invalidResObj.getResponseCode()); } - @Test public void testFileSizeDist() throws Exception { checkFileSizeDist(ROOT_PATH, 2, 3, 3, 1); @@ -1205,7 +1202,6 @@ private OmKeyLocationInfoGroup getLocationInfoGroup1() { return new OmKeyLocationInfoGroup(0L, locationInfoList); } - private OmKeyLocationInfoGroup getLocationInfoGroup2() { List locationInfoList = new ArrayList<>(); BlockID block4 = new BlockID(CONTAINER_FOUR_ID, 0L); diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestOmDBInsightEndPoint.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestOmDBInsightEndPoint.java index eb45c6abe847..0848b40e50a6 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestOmDBInsightEndPoint.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestOmDBInsightEndPoint.java @@ -1389,7 +1389,6 @@ public void testGetDeletedKeysWithBothPrevKeyAndStartPrefixProvided() keyInsightInfoResp.getRepeatedOmKeyInfoList().get(3).getOmKeyInfoList().get(0).getKeyName()); } - private OmKeyInfo getOmKeyInfo(String volumeName, String bucketName, String keyName, boolean isFile) { return new OmKeyInfo.Builder() diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestOpenKeysSearchEndpoint.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestOpenKeysSearchEndpoint.java index 0897f548630a..024cee560ba2 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestOpenKeysSearchEndpoint.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/api/TestOpenKeysSearchEndpoint.java @@ -345,7 +345,6 @@ public void testKeyLevelSearchUnderDirectory() throws IOException { "Expected a message indicating no keys were found"); } - @Test public void testSearchUnderNestedDirectory() throws IOException { Response response = omdbInsightEndpoint diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/fsck/TestContainerHealthTask.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/fsck/TestContainerHealthTask.java index bdd5ca35b296..558b3cbebbf6 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/fsck/TestContainerHealthTask.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/fsck/TestContainerHealthTask.java @@ -748,7 +748,6 @@ public Set replicasToCopyToFixMisreplication( return Collections.emptySet(); } - @Override public Set replicasToRemoveToFixOverreplication( Set replicas, int expectedCountPerUniqueReplica) { diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/fsck/TestContainerHealthTaskRecordGenerator.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/fsck/TestContainerHealthTaskRecordGenerator.java index db95d551062c..128893367d89 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/fsck/TestContainerHealthTaskRecordGenerator.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/fsck/TestContainerHealthTaskRecordGenerator.java @@ -178,7 +178,6 @@ public void testNegativeSizeRecordNotInsertedButLogged() { UnHealthyContainerStates.NEGATIVE_SIZE).getOrDefault(CONTAINER_COUNT, 0L)); } - @Test public void testUnderReplicatedRecordRetainedAndUpdated() { // under replicated container diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/persistence/TestReconWithDifferentSqlDBs.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/persistence/TestReconWithDifferentSqlDBs.java index 55223c692b1e..6c8963a882fc 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/persistence/TestReconWithDifferentSqlDBs.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/persistence/TestReconWithDifferentSqlDBs.java @@ -43,6 +43,7 @@ public class TestReconWithDifferentSqlDBs { @TempDir private static Path temporaryFolder; + public static Stream parametersSource() throws IOException { return Stream.of( new AbstractReconSqlDBTest.DerbyDataSourceConfigurationProvider( diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/persistence/TestSchemaVersionTableDefinition.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/persistence/TestSchemaVersionTableDefinition.java index 2731bef7335e..0804089afc8f 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/persistence/TestSchemaVersionTableDefinition.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/persistence/TestSchemaVersionTableDefinition.java @@ -79,7 +79,6 @@ public void testSchemaVersionTableCreation() throws Exception { assertEquals(expectedPairs, actualPairs, "Column definitions do not match expected values."); } - @Test public void testSchemaVersionCRUDOperations() throws SQLException { Connection connection = getConnection(); diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/scm/TestReconPipelineManager.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/scm/TestReconPipelineManager.java index 6f64cbf31b85..39b2c9bdcf34 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/scm/TestReconPipelineManager.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/scm/TestReconPipelineManager.java @@ -230,7 +230,6 @@ public void testDuplicatePipelineHandling() throws IOException { }, "Exception was thrown when adding a duplicate pipeline."); } - @Test public void testStubbedReconPipelineFactory() throws IOException { diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/spi/impl/TestOzoneManagerServiceProviderImpl.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/spi/impl/TestOzoneManagerServiceProviderImpl.java index 085f9c12cea1..75e52a105424 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/spi/impl/TestOzoneManagerServiceProviderImpl.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/spi/impl/TestOzoneManagerServiceProviderImpl.java @@ -330,7 +330,6 @@ public void testGetOzoneManagerDBSnapshot(@TempDir File dirReconMetadata) } } - static RocksDatabase getRocksDatabase(OMMetadataManager om) { return ((RDBStore)om.getStore()).getDb(); } diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestContainerSizeCountTask.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestContainerSizeCountTask.java index 6e444717b290..c224a8c0aa62 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestContainerSizeCountTask.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestContainerSizeCountTask.java @@ -85,7 +85,6 @@ public void setUp() { dslContext.truncate(CONTAINER_COUNT_BY_SIZE); } - @Test public void testProcess() { // mock a container with invalid used bytes diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestFileSizeCountTask.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestFileSizeCountTask.java index 15f0f0f093bb..1c1343f5051e 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestFileSizeCountTask.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestFileSizeCountTask.java @@ -496,7 +496,6 @@ public void testProcessAtScale() { .getCount().longValue()); } - @Test public void testTruncateTableExceptionPropagation() { // Mock DSLContext and FileCountBySizeDao diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestNSSummaryTaskWithLegacyOBSLayout.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestNSSummaryTaskWithLegacyOBSLayout.java index 48054d1eed39..f8ecd911c5d8 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestNSSummaryTaskWithLegacyOBSLayout.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestNSSummaryTaskWithLegacyOBSLayout.java @@ -95,7 +95,6 @@ public class TestNSSummaryTaskWithLegacyOBSLayout { private static final long KEY_SIX_OBJECT_ID = 10L; private static final long KEY_SEVEN_OBJECT_ID = 11L; - private static final long KEY_ONE_SIZE = 500L; private static final long KEY_TWO_OLD_SIZE = 1025L; private static final long KEY_TWO_UPDATE_SIZE = 1023L; @@ -356,7 +355,6 @@ public void testProcessForSize() throws IOException { nsSummaryForBucket2.getSizeOfFiles()); } - @Test public void testProcessFileBucketSize() { int[] fileDistBucket1 = nsSummaryForBucket1.getFileSizeBucket(); diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestNSSummaryTaskWithOBS.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestNSSummaryTaskWithOBS.java index 0b41162f837f..b275e1b66ee5 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestNSSummaryTaskWithOBS.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestNSSummaryTaskWithOBS.java @@ -93,7 +93,6 @@ public class TestNSSummaryTaskWithOBS { private static final long KEY_SIX_OBJECT_ID = 10L; private static final long KEY_SEVEN_OBJECT_ID = 11L; - private static final long KEY_ONE_SIZE = 500L; private static final long KEY_TWO_OLD_SIZE = 1025L; private static final long KEY_TWO_UPDATE_SIZE = 1023L; @@ -350,7 +349,6 @@ public void testProcessForSize() throws IOException { nsSummaryForBucket2.getSizeOfFiles()); } - @Test public void testProcessFileBucketSize() { int[] fileDistBucket1 = nsSummaryForBucket1.getFileSizeBucket(); diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestOmTableInsightTask.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestOmTableInsightTask.java index e889195a9583..307643cf80d0 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestOmTableInsightTask.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestOmTableInsightTask.java @@ -102,7 +102,6 @@ public class TestOmTableInsightTask extends AbstractReconSqlDBTest { private static final String DIR_TWO = "dir2"; private static final String DIR_THREE = "dir3"; - private static final long VOL_OBJECT_ID = 0L; private static final long BUCKET_ONE_OBJECT_ID = 1L; private static final long BUCKET_TWO_OBJECT_ID = 2L; diff --git a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestReconTaskControllerImpl.java b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestReconTaskControllerImpl.java index c7cb961ee513..1cc0ac3bdc34 100644 --- a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestReconTaskControllerImpl.java +++ b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/tasks/TestReconTaskControllerImpl.java @@ -200,7 +200,6 @@ public void testBadBehavedTaskIsIgnored() throws Exception { assertEquals(Long.valueOf(0L), dbRecord.getLastUpdatedSeqNumber()); } - @Test public void testReInitializeTasks() throws Exception { diff --git a/hadoop-ozone/s3-secret-store/src/main/java/org/apache/hadoop/ozone/s3/remote/vault/VaultS3SecretStoreBuilder.java b/hadoop-ozone/s3-secret-store/src/main/java/org/apache/hadoop/ozone/s3/remote/vault/VaultS3SecretStoreBuilder.java index 0e7f56016bab..c53e2db2313b 100644 --- a/hadoop-ozone/s3-secret-store/src/main/java/org/apache/hadoop/ozone/s3/remote/vault/VaultS3SecretStoreBuilder.java +++ b/hadoop-ozone/s3-secret-store/src/main/java/org/apache/hadoop/ozone/s3/remote/vault/VaultS3SecretStoreBuilder.java @@ -99,7 +99,6 @@ public VaultS3SecretStoreBuilder setSecretPath(String secretPath) { return this; } - public VaultS3SecretStoreBuilder setNameSpace(String nameSpace) { this.nameSpace = nameSpace; return this; diff --git a/hadoop-ozone/s3-secret-store/src/main/java/org/apache/hadoop/ozone/s3/remote/vault/auth/AppRoleAuth.java b/hadoop-ozone/s3-secret-store/src/main/java/org/apache/hadoop/ozone/s3/remote/vault/auth/AppRoleAuth.java index 8facb5334ff1..35e15f3b7da9 100644 --- a/hadoop-ozone/s3-secret-store/src/main/java/org/apache/hadoop/ozone/s3/remote/vault/auth/AppRoleAuth.java +++ b/hadoop-ozone/s3-secret-store/src/main/java/org/apache/hadoop/ozone/s3/remote/vault/auth/AppRoleAuth.java @@ -45,7 +45,6 @@ public AppRoleAuth(@Nullable String path, String roleId, String secretId) { this.secretId = secretId; } - @Override public Vault auth(VaultConfig config) throws VaultException { Vault vault = new Vault(config); diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/Gateway.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/Gateway.java index 0f89284bf5da..9126d8ed4a2b 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/Gateway.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/Gateway.java @@ -70,8 +70,6 @@ public static void main(String[] args) throws Exception { new Gateway().run(args); } - - @Override public Void call() throws Exception { OzoneConfiguration ozoneConfiguration = getOzoneConf(); diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/OzoneClientCache.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/OzoneClientCache.java index 5c836587244f..ba7038d025f4 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/OzoneClientCache.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/OzoneClientCache.java @@ -160,7 +160,6 @@ private void setCertificate(String omServiceID, } } - @PreDestroy public void destroy() throws IOException { OzoneClientCache.closeClient(); diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/TracingFilter.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/TracingFilter.java index 93f505bce2da..dd9d7cdf170f 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/TracingFilter.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/TracingFilter.java @@ -47,7 +47,6 @@ public class TracingFilter implements ContainerRequestFilter, @Context private ResourceInfo resourceInfo; - @Override public void filter(ContainerRequestContext requestContext) { finishAndCloseActiveSpan(); diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/VirtualHostStyleFilter.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/VirtualHostStyleFilter.java index 93893153c0a8..662fa67cd1ef 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/VirtualHostStyleFilter.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/VirtualHostStyleFilter.java @@ -122,7 +122,6 @@ public void setConfiguration(OzoneConfiguration config) { this.conf = config; } - /** * This method finds the longest match with the domain name. * @param host @@ -151,6 +150,4 @@ private String checkHostWithoutPort(String host) { return host; } } - - } diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/CopyObjectResponse.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/CopyObjectResponse.java index b5aa8bea6d37..c6a1c5e15cf8 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/CopyObjectResponse.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/CopyObjectResponse.java @@ -42,7 +42,6 @@ public class CopyObjectResponse { @XmlElement(name = OzoneConsts.ETAG) private String eTag; - public Instant getLastModified() { return lastModified; } @@ -58,6 +57,4 @@ public String getETag() { public void setETag(String tag) { this.eTag = tag; } - - } diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/EndpointBase.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/EndpointBase.java index df6fbfa0d804..69d2ed1a79c2 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/EndpointBase.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/EndpointBase.java @@ -485,7 +485,6 @@ public AuditMessage buildAuditMessageForFailure(AuditAction op, return builder.build(); } - @VisibleForTesting public void setClient(OzoneClient ozoneClient) { this.client = ozoneClient; diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/ListPartsResponse.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/ListPartsResponse.java index 7396dbdaaf48..e04c9944a407 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/ListPartsResponse.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/ListPartsResponse.java @@ -156,7 +156,6 @@ public static class Part { @XmlElement(name = OzoneConsts.ETAG) private String eTag; - @XmlElement(name = "Size") private long size; diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/ObjectEndpoint.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/ObjectEndpoint.java index 870f8ce03bc1..21718295b9c1 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/ObjectEndpoint.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/ObjectEndpoint.java @@ -1490,7 +1490,6 @@ private Response deleteObjectTagging(OzoneVolume volume, String bucketName, Stri return Response.noContent().build(); } - @VisibleForTesting public void setOzoneConfiguration(OzoneConfiguration config) { this.ozoneConfiguration = config; diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/S3Acl.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/S3Acl.java index 6666dc8975be..de92c315a82a 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/S3Acl.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/endpoint/S3Acl.java @@ -66,6 +66,7 @@ enum ACLType { public String getValue() { return value; } + /** * String value for this Enum. */ @@ -78,7 +79,6 @@ public String getValue() { value = val; } - public static ACLType getType(String typeStr) { for (ACLType type: ACLType.values()) { if (type.getValue().equals(typeStr)) { diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/exception/BadRequestExceptionMapper.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/exception/BadRequestExceptionMapper.java index 5d3a4acd9123..2218544a37c6 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/exception/BadRequestExceptionMapper.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/exception/BadRequestExceptionMapper.java @@ -33,6 +33,7 @@ public class BadRequestExceptionMapper implements private static final Logger LOG = LoggerFactory.getLogger(BadRequestExceptionMapper.class); + @Override public Response toResponse(BadRequestException exception) { if (LOG.isDebugEnabled()) { diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/exception/OS3Exception.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/exception/OS3Exception.java index b36be3688576..00b36427d434 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/exception/OS3Exception.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/exception/OS3Exception.java @@ -46,6 +46,7 @@ public class OS3Exception extends Exception { mapper.registerModule(new JaxbAnnotationModule()); mapper.enable(SerializationFeature.INDENT_OUTPUT); } + @XmlElement(name = "Code") private String code; diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/metrics/S3GatewayMetrics.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/metrics/S3GatewayMetrics.java index e9cf741ab059..54a7263d644f 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/metrics/S3GatewayMetrics.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/metrics/S3GatewayMetrics.java @@ -65,7 +65,6 @@ public final class S3GatewayMetrics implements Closeable, MetricsSource { private @Metric MutableCounterLong listMultipartUploadsFailure; private @Metric MutableCounterLong listKeyCount; - // RootEndpoint private @Metric MutableCounterLong listS3BucketsSuccess; private @Metric MutableCounterLong listS3BucketsFailure; diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/signature/SignatureInfo.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/signature/SignatureInfo.java index 09dfd8e0192c..730481bce214 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/signature/SignatureInfo.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/signature/SignatureInfo.java @@ -53,7 +53,6 @@ public class SignatureInfo { private String unfilteredURI = null; - private String stringToSign = null; public SignatureInfo() { } diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/util/RangeHeader.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/util/RangeHeader.java index 22e42e7051a5..58812c020b51 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/util/RangeHeader.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/util/RangeHeader.java @@ -28,7 +28,6 @@ public class RangeHeader { private boolean readFull; private boolean inValidRange; - /** * Construct RangeHeader object. * @param startOffset diff --git a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/util/S3Consts.java b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/util/S3Consts.java index f137fa7f5d7f..3204abeebab3 100644 --- a/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/util/S3Consts.java +++ b/hadoop-ozone/s3gateway/src/main/java/org/apache/hadoop/ozone/s3/util/S3Consts.java @@ -50,7 +50,6 @@ private S3Consts() { public static final String STREAMING_AWS4_ECDSA_P256_SHA256_PAYLOAD_TRAILER = "STREAMING-AWS4-ECDSA-P256-SHA256-PAYLOAD-TRAILER"; - // Constants related to Range Header public static final String COPY_SOURCE_IF_PREFIX = "x-amz-copy-source-if-"; public static final String COPY_SOURCE_IF_MODIFIED_SINCE = @@ -64,7 +63,6 @@ private S3Consts() { public static final String ACCEPT_RANGE_HEADER = "Accept-Ranges"; public static final String CONTENT_RANGE_HEADER = "Content-Range"; - public static final Pattern RANGE_HEADER_MATCH_PATTERN = Pattern.compile("bytes=(?[0-9]*)-(?[0-9]*)"); @@ -77,7 +75,6 @@ private S3Consts() { public static final String CUSTOM_METADATA_HEADER_PREFIX = "x-amz-meta-"; public static final String CUSTOM_METADATA_COPY_DIRECTIVE_HEADER = "x-amz-metadata-directive"; - public static final String DECODED_CONTENT_LENGTH_HEADER = "x-amz-decoded-content-length"; @@ -95,7 +92,6 @@ private S3Consts() { public static final Pattern TAG_REGEX_PATTERN = Pattern.compile("^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-]*)$"); public static final String MP_PARTS_COUNT = "x-amz-mp-parts-count"; - /** * Copy directive for metadata and tags. */ diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestAbortMultipartUpload.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestAbortMultipartUpload.java index 289e5ff0e1ab..9c46a718508f 100644 --- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestAbortMultipartUpload.java +++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestAbortMultipartUpload.java @@ -37,7 +37,6 @@ */ public class TestAbortMultipartUpload { - @Test public void testAbortMultipartUpload() throws Exception { diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestBucketDelete.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestBucketDelete.java index b06e714d03d8..c32b0e0ffd5c 100644 --- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestBucketDelete.java +++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestBucketDelete.java @@ -79,7 +79,6 @@ public void testDeleteWithNoSuchBucket() throws Exception { fail("testDeleteWithNoSuchBucket failed"); } - @Test public void testDeleteWithBucketNotEmpty() throws Exception { try { diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestBucketList.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestBucketList.java index b3589afbfd46..ad8b47754559 100644 --- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestBucketList.java +++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestBucketList.java @@ -116,7 +116,6 @@ public void listSubDir() throws OS3Exception, IOException { } - @Test public void listObjectOwner() throws OS3Exception, IOException { @@ -395,7 +394,6 @@ public void listWithContinuationTokenFail() throws IOException { assertEquals("Invalid Argument", e.getErrorMessage()); } - @Test public void testStartAfter() throws IOException, OS3Exception { BucketEndpoint getBucket = new BucketEndpoint(); diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestCompleteMultipartUploadRequestUnmarshaller.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestCompleteMultipartUploadRequestUnmarshaller.java index f88f4f2ed470..d3615bb5c360 100644 --- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestCompleteMultipartUploadRequestUnmarshaller.java +++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestCompleteMultipartUploadRequestUnmarshaller.java @@ -37,6 +37,7 @@ public class TestCompleteMultipartUploadRequestUnmarshaller { private static String part1 = UUID.randomUUID().toString(); private static String part2 = UUID.randomUUID().toString(); + @Test public void fromStreamWithNamespace() throws IOException { //GIVEN diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestListParts.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestListParts.java index 3093571042b1..0800b81bb7e9 100644 --- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestListParts.java +++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestListParts.java @@ -42,7 +42,6 @@ */ public class TestListParts { - private ObjectEndpoint rest; private String uploadID; @@ -132,6 +131,4 @@ public void testListPartsWithUnknownUploadID() throws Exception { ex.getErrorMessage()); } } - - } diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadComplete.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadComplete.java index 6050053b412c..46a141df74b5 100644 --- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadComplete.java +++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadComplete.java @@ -196,7 +196,6 @@ public void testMultipartWithCustomMetadata() throws Exception { assertEquals("custom-value2", headResponse.getHeaderString(CUSTOM_METADATA_HEADER_PREFIX + "custom-key2")); } - @Test public void testMultipartInvalidPartOrderError() throws Exception { diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadWithCopy.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadWithCopy.java index cc78b39336e5..f1321820f15a 100644 --- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadWithCopy.java +++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestMultipartUploadWithCopy.java @@ -78,6 +78,7 @@ public class TestMultipartUploadWithCopy { private static final String UNPARSABLE_TIME_STR = "Unparsable time string"; private static final String ERROR_CODE = S3ErrorTable.PRECOND_FAILED.getCode(); + @BeforeAll public static void setUp() throws Exception { CLIENT.getObjectStore().createS3Bucket(OzoneConsts.S3_BUCKET); @@ -290,6 +291,7 @@ public String toString() { + " ErrorCode:" + this.errorCode; } } + @Test public void testMultipartTSHeaders() throws Exception { for (CopyIfTimestampTestCase t : CopyIfTimestampTestCase.values()) { diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestPartUpload.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestPartUpload.java index 41b4044f1ecc..7a3faf6bd53a 100644 --- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestPartUpload.java +++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestPartUpload.java @@ -78,7 +78,6 @@ public void setUp() throws Exception { .build(); } - @Test public void testPartUpload() throws Exception { @@ -132,7 +131,6 @@ public void testPartUploadWithOverride() throws Exception { } - @Test public void testPartUploadWithIncorrectUploadID() throws Exception { OS3Exception ex = assertThrows(OS3Exception.class, () -> { diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestPermissionCheck.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestPermissionCheck.java index 443686551d76..0e4319d6956b 100644 --- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestPermissionCheck.java +++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/endpoint/TestPermissionCheck.java @@ -141,6 +141,7 @@ public void testDeleteBucket() throws IOException { bucketEndpoint.delete("bucketName")); assertEquals(HTTP_FORBIDDEN, e.getHttpCode()); } + @Test public void testListMultiUpload() throws IOException { when(objectStore.getS3Bucket(anyString())).thenReturn(bucket); diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/metrics/TestS3GatewayMetrics.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/metrics/TestS3GatewayMetrics.java index 5abc3ea77c60..53a8736bddec 100644 --- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/metrics/TestS3GatewayMetrics.java +++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/metrics/TestS3GatewayMetrics.java @@ -76,7 +76,6 @@ public class TestS3GatewayMetrics { private S3GatewayMetrics metrics; private ContainerRequestContext context; - @BeforeEach public void setup() throws Exception { clientStub = new OzoneClientStub(); @@ -327,7 +326,6 @@ public void testCreateKeyFailure() throws Exception { assertEquals(1L, curMetric - oriMetric); } - @Test public void testDeleteKeySuccess() throws Exception { long oriMetric = metrics.getDeleteKeySuccess(); @@ -424,7 +422,6 @@ public void testAbortMultiPartUploadFailure() throws Exception { assertEquals(1L, curMetric - oriMetric); } - @Test public void testCompleteMultiPartUploadSuccess() throws Exception { diff --git a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/signature/TestStringToSignProducer.java b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/signature/TestStringToSignProducer.java index d8044f36c00d..cbce030ef69f 100644 --- a/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/signature/TestStringToSignProducer.java +++ b/hadoop-ozone/s3gateway/src/test/java/org/apache/hadoop/ozone/s3/signature/TestStringToSignProducer.java @@ -185,6 +185,7 @@ private static Stream testValidateRequestHeadersInput() { arguments(headersMap6, S3_AUTHINFO_CREATION_ERROR.getCode()) ); } + @ParameterizedTest @MethodSource("testValidateRequestHeadersInput") public void testValidateRequestHeaders( diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/containerlog/parser/ContainerDatanodeDatabase.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/containerlog/parser/ContainerDatanodeDatabase.java index de3ab8439979..19f7402518ad 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/containerlog/parser/ContainerDatanodeDatabase.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/containerlog/parser/ContainerDatanodeDatabase.java @@ -33,12 +33,10 @@ import org.slf4j.LoggerFactory; import org.sqlite.SQLiteConfig; - /** * Handles creation and interaction with the database. * Provides methods for table creation, log data insertion, and index setup. */ - public class ContainerDatanodeDatabase { private static Map queries; diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/RocksDBUtils.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/RocksDBUtils.java index a1eb6a7cd4b9..5d79c3a67a26 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/RocksDBUtils.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/RocksDBUtils.java @@ -72,6 +72,4 @@ public static T getValue(ManagedRocksDB db, StringCodec.get().toPersistedFormat(key)); return bytes != null ? codec.fromPersistedFormat(bytes) : null; } - - } diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/audit/parser/common/DatabaseHelper.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/audit/parser/common/DatabaseHelper.java index ef9bb652eebb..8b68f396f69e 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/audit/parser/common/DatabaseHelper.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/audit/parser/common/DatabaseHelper.java @@ -48,9 +48,11 @@ public final class DatabaseHelper { private DatabaseHelper() { //Never constructed } + static { loadProperties(); } + private static final Logger LOG = LoggerFactory.getLogger(DatabaseHelper.class); private static Map properties; diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/datanode/container/ExportSubcommand.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/datanode/container/ExportSubcommand.java index 91ef00d217f0..d06633f14408 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/datanode/container/ExportSubcommand.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/datanode/container/ExportSubcommand.java @@ -61,7 +61,6 @@ public class ExportSubcommand implements Callable { description = "Count of containers to export") private long containerCount = 1; - @Override public Void call() throws Exception { parent.loadContainersFromVolumes(); diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/ldb/DBScanner.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/ldb/DBScanner.java index 9ca84de17621..3f95a92db0d7 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/ldb/DBScanner.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/ldb/DBScanner.java @@ -667,7 +667,6 @@ public static ObjectWriter getWriter() { } } - private class Task implements Callable { private final DBColumnFamilyDefinition dbColumnFamilyDefinition; diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/replicas/Checksums.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/replicas/Checksums.java index 38f6f0103017..f9d983ccef6d 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/replicas/Checksums.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/replicas/Checksums.java @@ -58,7 +58,6 @@ public class Checksums implements ReplicaVerifier { private static final String JSON_PROPERTY_REPLICA_UUID = "uuid"; private static final String JSON_PROPERTY_REPLICA_EXCEPTION = "exception"; - private String outputDir; private OzoneClient client; diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/replicas/ReplicasVerify.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/replicas/ReplicasVerify.java index 752b5f99e9b7..2aa72fa56700 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/replicas/ReplicasVerify.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/replicas/ReplicasVerify.java @@ -65,6 +65,7 @@ static class Verification { private boolean doExecuteChecksums; } + private List replicaVerifiers; @Override diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/replicas/chunk/ContainerChunkInfo.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/replicas/chunk/ContainerChunkInfo.java index a5988cac73e9..bf1e22e64206 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/replicas/chunk/ContainerChunkInfo.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/debug/replicas/chunk/ContainerChunkInfo.java @@ -84,7 +84,6 @@ public ChunkType getChunkType() { return chunkType; } - @Override public String toString() { return "Container{" diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeChunkGenerator.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeChunkGenerator.java index 199af1ea1b92..ca9c17519135 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeChunkGenerator.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeChunkGenerator.java @@ -100,11 +100,9 @@ public class DatanodeChunkGenerator extends BaseFreonGenerator implements private ByteString dataToWrite; private ChecksumData checksumProtobuf; - @Override public Void call() throws Exception { - OzoneConfiguration ozoneConf = createOzoneConfiguration(); if (OzoneSecurityUtil.isSecurityEnabled(ozoneConf)) { throw new IllegalArgumentException( @@ -181,7 +179,6 @@ private boolean arePipelinesOrDatanodesProvided() { return !(pipelineIds.equals("") && datanodes.equals("")); } - private void runTest() throws IOException { diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeChunkValidator.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeChunkValidator.java index dd6d05bf10cc..b17a6b54fbfd 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeChunkValidator.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeChunkValidator.java @@ -76,7 +76,6 @@ public class DatanodeChunkValidator extends BaseFreonGenerator private Checksum checksum; private ContainerProtos.ChecksumData checksumProtobuf; - @Override public Void call() throws Exception { @@ -132,7 +131,6 @@ private void readReference() throws IOException { checksumReference = computeChecksum(response); } - private void validateChunk(long stepNo) throws Exception { ContainerCommandRequestProto request = createReadChunkRequest(stepNo); @@ -201,6 +199,4 @@ private ChecksumData computeChecksum(ContainerCommandResponseProto response) readChunk.getDataBuffers().getBuffersList()); } } - - } diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeSimulationState.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeSimulationState.java index 1e3bae684d19..b9809677855f 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeSimulationState.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeSimulationState.java @@ -70,7 +70,6 @@ class DatanodeSimulationState { private Map containers = new HashMap<>(); - // indicate if this node is in read-only mode, no pipeline should be created. private volatile boolean readOnly = false; diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/FreonHttpServer.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/FreonHttpServer.java index a67aba98ef4c..ba49d9c82159 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/FreonHttpServer.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/FreonHttpServer.java @@ -30,7 +30,6 @@ public FreonHttpServer(MutableConfigurationSource conf) throws IOException { super(conf, "freon"); } - @Override protected String getHttpAddressKey() { return OzoneConfigKeys.OZONE_FREON_HTTP_ADDRESS_KEY; } diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/HsyncGenerator.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/HsyncGenerator.java index 62f949150907..55bd72617a23 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/HsyncGenerator.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/HsyncGenerator.java @@ -91,7 +91,6 @@ public class HsyncGenerator extends BaseFreonGenerator implements Callable public HsyncGenerator() { } - @VisibleForTesting HsyncGenerator(OzoneConfiguration ozoneConfiguration) { this.configuration = ozoneConfiguration; diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/KeyGeneratorUtil.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/KeyGeneratorUtil.java index 831a3a85ef37..581dab076103 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/KeyGeneratorUtil.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/KeyGeneratorUtil.java @@ -31,6 +31,7 @@ public class KeyGeneratorUtil { public String generatePureIndexKeyName(int number) { return String.valueOf(number); } + public Function pureIndexKeyNameFunc() { return number -> String.valueOf(number); } diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/OzoneClientKeyReadWriteListOps.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/OzoneClientKeyReadWriteListOps.java index 2051aeec25f0..a820bbe9c7a4 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/OzoneClientKeyReadWriteListOps.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/OzoneClientKeyReadWriteListOps.java @@ -143,6 +143,7 @@ public class OzoneClientKeyReadWriteListOps extends BaseFreonGenerator LoggerFactory.getLogger(OzoneClientKeyReadWriteListOps.class); private static final AtomicLong NEXT_NUMBER = new AtomicLong(); + /** * Task type of read task, or write task. */ @@ -151,8 +152,8 @@ public enum TaskType { WRITE_TASK, LIST_TASK } - private KeyGeneratorUtil kg; + private KeyGeneratorUtil kg; @Override public Void call() throws Exception { diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/RandomKeyGenerator.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/RandomKeyGenerator.java index e1df0460e067..db335ac23c1a 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/RandomKeyGenerator.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/RandomKeyGenerator.java @@ -612,6 +612,7 @@ int getNumberOfVolumesCleaned() { int getNumberOfBucketsCleaned() { return numberOfBucketsCleaned.get(); } + /** * Returns true if random validation of write is enabled. * diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/RangeKeysGenerator.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/RangeKeysGenerator.java index 44156dc091ff..a2f91b70467e 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/RangeKeysGenerator.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/RangeKeysGenerator.java @@ -71,7 +71,6 @@ public class RangeKeysGenerator extends BaseFreonGenerator defaultValue = "0") private int startIndex; - @CommandLine.Option(names = {"-k", "--key-encode"}, description = "The algorithm to generate key names. " + "Options are pureIndex, md5", @@ -101,7 +100,6 @@ public class RangeKeysGenerator extends BaseFreonGenerator private ContentGenerator contentGenerator; private Timer timer; - @Override public Void call() throws Exception { init(); @@ -150,7 +148,6 @@ public void generateRangeKeys(long count) throws Exception { }); } - public void loopRunner(Function keyNameGeneratorfunc, OzoneClient client, int start, int end) throws Exception { diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/StreamingGenerator.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/StreamingGenerator.java index 4e6e5f584628..52456bf6c2d7 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/StreamingGenerator.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/StreamingGenerator.java @@ -76,7 +76,6 @@ public class StreamingGenerator extends BaseFreonGenerator private Timer timer; - @Override public Void call() throws Exception { init(); diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/containergenerator/GeneratorScm.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/containergenerator/GeneratorScm.java index 548777744c98..0e9aa8330986 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/containergenerator/GeneratorScm.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/containergenerator/GeneratorScm.java @@ -70,7 +70,6 @@ public Void call() throws Exception { return null; } - private void writeScmData(long index) throws Exception { timer.time((Callable) () -> { long containerId = getContainerIdOffset() + index; diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/repair/scm/cert/RecoverSCMCertificate.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/repair/scm/cert/RecoverSCMCertificate.java index 987c930e32be..f6a9759a3e96 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/repair/scm/cert/RecoverSCMCertificate.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/repair/scm/cert/RecoverSCMCertificate.java @@ -148,7 +148,6 @@ private static X509Certificate getRootCertificate( return cert.get(); } - private static X509Certificate getSubCertificate( Map allCerts, String hostName) throws Exception { Optional cert = allCerts.values().stream().filter( diff --git a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/utils/FormattingCLIUtils.java b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/utils/FormattingCLIUtils.java index bcb55e7c6581..36e729f2742a 100644 --- a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/utils/FormattingCLIUtils.java +++ b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/utils/FormattingCLIUtils.java @@ -214,6 +214,7 @@ public String render() { private static class TableRow { private TableRowType tableRowType; private List data; + TableRow(TableRowType tableRowType, List data) { this.tableRowType = tableRowType; this.data = data; diff --git a/hadoop-ozone/tools/src/test/java/org/apache/hadoop/ozone/debug/audit/parser/TestAuditParser.java b/hadoop-ozone/tools/src/test/java/org/apache/hadoop/ozone/debug/audit/parser/TestAuditParser.java index 9e2c74fa6e91..9e8167c66b56 100644 --- a/hadoop-ozone/tools/src/test/java/org/apache/hadoop/ozone/debug/audit/parser/TestAuditParser.java +++ b/hadoop-ozone/tools/src/test/java/org/apache/hadoop/ozone/debug/audit/parser/TestAuditParser.java @@ -63,6 +63,7 @@ public class TestAuditParser { .getClassLoader().getResource("testaudit.log").getPath(); private static final String LOGS1 = TestAuditParser.class .getClassLoader().getResource("testloadaudit.log").getPath(); + /** * Creates output directory which will be used by the test-cases. * If a test-case needs a separate directory, it has to create a random diff --git a/hadoop-ozone/tools/src/test/java/org/apache/hadoop/ozone/freon/TestContentGenerator.java b/hadoop-ozone/tools/src/test/java/org/apache/hadoop/ozone/freon/TestContentGenerator.java index 0bbbfe2f2b44..d2b89b9d56f4 100644 --- a/hadoop-ozone/tools/src/test/java/org/apache/hadoop/ozone/freon/TestContentGenerator.java +++ b/hadoop-ozone/tools/src/test/java/org/apache/hadoop/ozone/freon/TestContentGenerator.java @@ -93,6 +93,7 @@ class SyncableByteArrayOutputStream extends ByteArrayOutputStream @Override public void hflush() throws IOException { } + @Override public void hsync() throws IOException { } @@ -102,6 +103,7 @@ public boolean hasCapability(String capability) { return true; } } + ContentGenerator generator = new ContentGenerator(20, 8, 3, ContentGenerator.SyncOptions.HSYNC); diff --git a/hadoop-ozone/tools/src/test/java/org/apache/hadoop/ozone/genconf/TestGenerateOzoneRequiredConfigurations.java b/hadoop-ozone/tools/src/test/java/org/apache/hadoop/ozone/genconf/TestGenerateOzoneRequiredConfigurations.java index b21b0caaf339..16477d6acd46 100644 --- a/hadoop-ozone/tools/src/test/java/org/apache/hadoop/ozone/genconf/TestGenerateOzoneRequiredConfigurations.java +++ b/hadoop-ozone/tools/src/test/java/org/apache/hadoop/ozone/genconf/TestGenerateOzoneRequiredConfigurations.java @@ -57,6 +57,7 @@ public class TestGenerateOzoneRequiredConfigurations { private static final PrintStream OLD_OUT = System.out; private static final PrintStream OLD_ERR = System.err; private static final String DEFAULT_ENCODING = UTF_8.name(); + /** * Creates output directory which will be used by the test-cases. * If a test-case needs a separate directory, it has to create a random