diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFSInputStream.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFSInputStream.java index 23d71047ef88..2e00e8c2c622 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFSInputStream.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFSInputStream.java @@ -52,9 +52,10 @@ import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Test; import org.junit.jupiter.api.Timeout; -import org.junit.jupiter.api.Assertions; import static org.apache.hadoop.hdds.StringUtils.string2Bytes; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertArrayEquals; /** * Test OzoneFSInputStream by reading through multiple interfaces. @@ -147,11 +148,11 @@ public void testO3FSSingleByteRead() throws IOException { break; } value[i] = (byte) val; - Assertions.assertEquals(value[i], data[i], "value mismatch at:" + i); + assertEquals(value[i], data[i], "value mismatch at:" + i); i++; } - Assertions.assertEquals(i, data.length); - Assertions.assertArrayEquals(value, data); + assertEquals(i, data.length); + assertArrayEquals(value, data); } } @@ -169,8 +170,8 @@ public void testO3FSMultiByteRead() throws IOException { System.arraycopy(tmp, 0, value, i * tmp.length, tmp.length); i++; } - Assertions.assertEquals((long) i * tmp.length, data.length); - Assertions.assertArrayEquals(value, data); + assertEquals((long) i * tmp.length, data.length); + assertArrayEquals(value, data); } } @@ -181,12 +182,12 @@ public void testO3FSByteBufferRead() throws IOException { ByteBuffer buffer = ByteBuffer.allocate(1024 * 1024); int byteRead = inputStream.read(buffer); - Assertions.assertEquals(byteRead, 1024 * 1024); + assertEquals(byteRead, 1024 * 1024); byte[] value = new byte[1024 * 1024]; System.arraycopy(data, 0, value, 0, value.length); - Assertions.assertArrayEquals(value, buffer.array()); + assertArrayEquals(value, buffer.array()); } } @@ -208,7 +209,7 @@ public void testSequenceFileReaderSync() throws IOException { in.sync(0); blockStart = in.getPosition(); // The behavior should be consistent with HDFS - Assertions.assertEquals(srcfile.length(), blockStart); + assertEquals(srcfile.length(), blockStart); in.close(); } @@ -230,7 +231,7 @@ public void testSequenceFileReaderSyncEC() throws IOException { in.sync(0); blockStart = in.getPosition(); // The behavior should be consistent with HDFS - Assertions.assertEquals(srcfile.length(), blockStart); + assertEquals(srcfile.length(), blockStart); in.close(); } } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFileSystemPrefixParser.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFileSystemPrefixParser.java index 972ad7dd2e78..37116f33e272 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFileSystemPrefixParser.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFileSystemPrefixParser.java @@ -35,11 +35,12 @@ import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Test; import org.junit.jupiter.api.Timeout; -import org.junit.jupiter.api.Assertions; import java.io.IOException; import java.net.URI; +import static org.junit.jupiter.api.Assertions.assertEquals; + /** * Test Ozone Prefix Parser. */ @@ -115,18 +116,12 @@ public void testPrefixParsePath() throws Exception { private void assertPrefixStats(PrefixParser parser, int volumeCount, int bucketCount, int intermediateDirCount, int nonExistentDirCount, int fileCount, int dirCount) { - Assertions.assertEquals(volumeCount, - parser.getParserStats(PrefixParser.Types.VOLUME)); - Assertions.assertEquals(bucketCount, - parser.getParserStats(PrefixParser.Types.BUCKET)); - Assertions.assertEquals(intermediateDirCount, - parser.getParserStats(PrefixParser.Types.INTERMEDIATE_DIRECTORY)); - Assertions.assertEquals(nonExistentDirCount, - parser.getParserStats(PrefixParser.Types.NON_EXISTENT_DIRECTORY)); - Assertions.assertEquals(fileCount, - parser.getParserStats(PrefixParser.Types.FILE)); - Assertions.assertEquals(dirCount, - parser.getParserStats(PrefixParser.Types.DIRECTORY)); + assertEquals(volumeCount, parser.getParserStats(PrefixParser.Types.VOLUME)); + assertEquals(bucketCount, parser.getParserStats(PrefixParser.Types.BUCKET)); + assertEquals(intermediateDirCount, parser.getParserStats(PrefixParser.Types.INTERMEDIATE_DIRECTORY)); + assertEquals(nonExistentDirCount, parser.getParserStats(PrefixParser.Types.NON_EXISTENT_DIRECTORY)); + assertEquals(fileCount, parser.getParserStats(PrefixParser.Types.FILE)); + assertEquals(dirCount, parser.getParserStats(PrefixParser.Types.DIRECTORY)); } private void testPrefixParseWithInvalidPaths() throws Exception { diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFileSystemWithStreaming.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFileSystemWithStreaming.java index 5d068546828f..9d5009eeb30c 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFileSystemWithStreaming.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFileSystemWithStreaming.java @@ -37,7 +37,6 @@ import org.apache.hadoop.ozone.client.io.SelectorOutputStream; import org.apache.hadoop.ozone.om.helpers.BucketLayout; import org.junit.jupiter.api.AfterAll; -import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Test; import org.junit.jupiter.api.Timeout; @@ -55,6 +54,9 @@ import static org.apache.hadoop.ozone.om.OMConfigKeys.OZONE_OM_ADDRESS_KEY; import static org.apache.hadoop.ozone.om.OMConfigKeys.OZONE_OM_RATIS_ENABLE_KEY; import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertNotNull; +import static org.junit.jupiter.api.Assertions.assertNull; +import static org.junit.jupiter.api.Assertions.assertInstanceOf; /** * Ozone file system tests with Streaming. @@ -152,7 +154,7 @@ static void createFile(FileSystem fs, Path path, boolean overwrite, final OutputStream wrapped = out.getWrappedStream(); LOG.info("wrapped: {}", wrapped.getClass()); - Assertions.assertEquals(SelectorOutputStream.class, wrapped.getClass()); + assertEquals(SelectorOutputStream.class, wrapped.getClass()); final SelectorOutputStream selector = (SelectorOutputStream) wrapped; final boolean belowThreshold = data.length <= AUTO_THRESHOLD; LOG.info("data.length={}, threshold={}, belowThreshold? {}", @@ -161,13 +163,12 @@ static void createFile(FileSystem fs, Path path, boolean overwrite, out.close(); final OutputStream underlying = selector.getUnderlying(); - Assertions.assertNotNull(underlying); + assertNotNull(underlying); LOG.info("underlying after close: {}", underlying.getClass()); if (belowThreshold) { - Assertions.assertTrue(underlying instanceof OzoneFSOutputStream); + assertInstanceOf(OzoneFSOutputStream.class, underlying); } else { - Assertions.assertEquals(OzoneFSDataStreamOutput.class, - underlying.getClass()); + assertEquals(OzoneFSDataStreamOutput.class, underlying.getClass()); } } @@ -177,10 +178,10 @@ static void assertUnderlying(SelectorOutputStream selector, LOG.info("underlying before close: {}", underlying != null ? underlying.getClass() : null); if (belowThreshold) { - Assertions.assertNull(underlying); + assertNull(underlying); } else { - Assertions.assertNotNull(underlying); - Assertions.assertEquals(OzoneFSDataStreamOutput.class, + assertNotNull(underlying); + assertEquals(OzoneFSDataStreamOutput.class, underlying.getClass()); } } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFsSnapshot.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFsSnapshot.java index d74c77d3435c..6d537d03ca13 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFsSnapshot.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFsSnapshot.java @@ -27,8 +27,6 @@ import java.util.UUID; import java.util.stream.Stream; -import com.google.common.base.Strings; - import org.apache.commons.io.FileUtils; import org.apache.commons.lang3.RandomStringUtils; import org.apache.hadoop.hdds.conf.OzoneConfiguration; @@ -39,7 +37,6 @@ import org.apache.hadoop.util.ToolRunner; import org.apache.ozone.test.GenericTestUtils; import org.junit.jupiter.api.AfterAll; -import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Test; import org.junit.jupiter.api.Timeout; @@ -56,6 +53,8 @@ import static org.apache.hadoop.ozone.OzoneConsts.OM_SNAPSHOT_INDICATOR; import static org.apache.hadoop.ozone.om.OmSnapshotManager.getSnapshotPath; import static org.assertj.core.api.Assertions.assertThat; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertNotNull; /** * Test client-side CRUD snapshot operations with Ozone Manager. @@ -120,13 +119,13 @@ private static void createVolBuckKey() // Create volume and bucket int res = ToolRunner.run(shell, new String[]{"-mkdir", "-p", BUCKET_PATH}); - Assertions.assertEquals(0, res); + assertEquals(0, res); // Create key res = ToolRunner.run(shell, new String[]{"-touch", KEY_PATH}); - Assertions.assertEquals(0, res); + assertEquals(0, res); // List the bucket to make sure that bucket exists. res = ToolRunner.run(shell, new String[]{"-ls", BUCKET_PATH}); - Assertions.assertEquals(0, res); + assertEquals(0, res); } @@ -137,12 +136,12 @@ void testCreateSnapshotDuplicateName() throws Exception { int res = ToolRunner.run(shell, new String[]{"-createSnapshot", BUCKET_PATH, snapshotName}); // Asserts that create request succeeded - Assertions.assertEquals(0, res); + assertEquals(0, res); res = ToolRunner.run(shell, new String[]{"-createSnapshot", BUCKET_PATH, snapshotName}); // Asserts that create request fails since snapshot name provided twice - Assertions.assertEquals(1, res); + assertEquals(1, res); } @Test @@ -162,19 +161,19 @@ void testCreateSnapshotWithSubDirInput() throws Exception { int res = ToolRunner.run(shell, new String[] { "-mkdir", "-p", dirPath}); - Assertions.assertEquals(0, res); + assertEquals(0, res); try (GenericTestUtils.SystemOutCapturer capture = new GenericTestUtils.SystemOutCapturer()) { res = ToolRunner.run(shell, new String[] { "-createSnapshot", dirPath, snapshotName}); // Asserts that create request succeeded - Assertions.assertEquals(0, res); + assertEquals(0, res); String expectedSnapshotPath = Paths.get( BUCKET_PATH, OM_SNAPSHOT_INDICATOR, snapshotName).toString(); String out = capture.getOutput().trim(); - Assertions.assertTrue(out.endsWith(expectedSnapshotPath)); + assertThat(out).endsWith(expectedSnapshotPath); } } @@ -192,7 +191,7 @@ void testCreateSnapshotSuccess(String snapshotName) int res = ToolRunner.run(shell, new String[]{"-createSnapshot", BUCKET_PATH, snapshotName}); // Asserts that create request succeeded - Assertions.assertEquals(0, res); + assertEquals(0, res); SnapshotInfo snapshotInfo = ozoneManager .getMetadataManager() @@ -202,7 +201,7 @@ void testCreateSnapshotSuccess(String snapshotName) // Assert that snapshot exists in RocksDB. // We can't use list or valid if snapshot directory exists because DB // transaction might not be flushed by the time. - Assertions.assertNotNull(snapshotInfo); + assertNotNull(snapshotInfo); } private static Stream createSnapshotFailureScenarios() { @@ -252,8 +251,7 @@ void testCreateSnapshotFailure(String description, String errorMessage = execShellCommandAndGetOutput(expectedResponse, new String[]{"-createSnapshot", paramBucketPath, snapshotName}); - Assertions.assertTrue(errorMessage - .contains(expectedMessage)); + assertThat(errorMessage).contains(expectedMessage); } /** @@ -291,7 +289,7 @@ void testFsLsSnapshot(@TempDir Path tempDir) throws Exception { int res = ToolRunner.run(shell, new String[]{"-deleteSnapshot", BUCKET_PATH, snapshotName1}); // Asserts that delete request succeeded - Assertions.assertEquals(0, res); + assertEquals(0, res); // Wait for the snapshot to be marked deleted. GenericTestUtils.waitFor(() -> { @@ -339,24 +337,22 @@ void testDeleteBucketWithSnapshot() throws Exception { String deleteKeyOut = execShellCommandAndGetOutput(0, new String[]{"-rm", "-r", "-skipTrash", KEY_PATH}); - Assertions.assertTrue(deleteKeyOut - .contains("Deleted " + BUCKET_PATH)); + assertThat(deleteKeyOut).contains("Deleted " + BUCKET_PATH); // Delete bucket should fail due to existing snapshot String deleteBucketOut = execShellCommandAndGetOutput(1, new String[]{"-rm", "-r", "-skipTrash", BUCKET_PATH}); - Assertions.assertTrue(deleteBucketOut - .contains(BUCKET + " can't be deleted when it has snapshots")); + assertThat(deleteBucketOut).contains(BUCKET + " can't be deleted when it has snapshots"); // Key shouldn't exist under bucket String listKeyOut = execShellCommandAndGetOutput(0, new String[]{"-ls", BUCKET_PATH}); - Assertions.assertTrue(Strings.isNullOrEmpty(listKeyOut)); + assertThat(listKeyOut).isNullOrEmpty(); // Key should still exist under snapshot String listSnapKeyOut = execShellCommandAndGetOutput(0, new String[]{"-ls", snapshotPath}); - Assertions.assertTrue(listSnapKeyOut.contains(snapshotKeyPath)); + assertThat(listSnapKeyOut).contains(snapshotKeyPath); } @Test @@ -366,7 +362,7 @@ void testSnapshotDeleteSuccess() throws Exception { int res = ToolRunner.run(shell, new String[]{"-deleteSnapshot", BUCKET_PATH, snapshotName}); // Asserts that delete request succeeded - Assertions.assertEquals(0, res); + assertEquals(0, res); // Wait for the snapshot to be marked deleted. GenericTestUtils.waitFor(() -> { @@ -417,8 +413,7 @@ void testSnapshotDeleteFailure(String description, String errorMessage = execShellCommandAndGetOutput(expectedResponse, new String[]{"-deleteSnapshot", paramBucketPath, snapshotName}); - Assertions.assertTrue(errorMessage - .contains(expectedMessage), errorMessage); + assertThat(errorMessage).contains(expectedMessage); } /** @@ -438,7 +433,7 @@ private String execShellCommandAndGetOutput( // Execute command int res = ToolRunner.run(shell, args); - Assertions.assertEquals(response, res); + assertEquals(response, res); // Store command output to a string, // if command should succeed then @@ -467,7 +462,7 @@ private String createSnapshot() throws Exception { int res = ToolRunner.run(shell, new String[]{"-createSnapshot", BUCKET_PATH, snapshotName}); // Asserts that create request succeeded - Assertions.assertEquals(0, res); + assertEquals(0, res); OzoneConfiguration conf = ozoneManager.getConfiguration(); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/upgrade/TestScmHAFinalization.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/upgrade/TestScmHAFinalization.java index f9f6871f546b..87ccb466ee5f 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/upgrade/TestScmHAFinalization.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/upgrade/TestScmHAFinalization.java @@ -37,7 +37,6 @@ import org.apache.ozone.test.GenericTestUtils; import org.apache.ozone.test.tag.Flaky; import org.junit.jupiter.api.AfterEach; -import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.Test; import org.junit.jupiter.params.ParameterizedTest; import org.junit.jupiter.params.provider.Arguments; @@ -56,6 +55,11 @@ import java.util.stream.Stream; import static org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos.ContainerDataProto.State.CLOSED; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertFalse; +import static org.junit.jupiter.api.Assertions.assertNotEquals; +import static org.junit.jupiter.api.Assertions.assertTrue; +import static org.junit.jupiter.api.Assertions.fail; /** * Tests upgrade finalization failure scenarios and corner cases specific to SCM @@ -167,7 +171,7 @@ public void testFinalizationWithLeaderChange( // Make sure the original SCM leader is not the leader anymore. StorageContainerManager newLeaderScm = cluster.getActiveSCM(); - Assertions.assertNotEquals(newLeaderScm.getSCMNodeId(), + assertNotEquals(newLeaderScm.getSCMNodeId(), oldLeaderScm.getSCMNodeId()); // Resume finalization from the new leader. @@ -288,7 +292,7 @@ public void testSnapshotFinalization() throws Exception { inactiveScm, 0, NUM_DATANODES); // Use log to verify a snapshot was installed. - Assertions.assertTrue(logCapture.getOutput().contains("New SCM snapshot " + + assertTrue(logCapture.getOutput().contains("New SCM snapshot " + "received with metadata layout version")); } @@ -319,35 +323,31 @@ private void checkMidFinalizationConditions( for (StorageContainerManager scm: scms) { switch (haltingPoint) { case BEFORE_PRE_FINALIZE_UPGRADE: - Assertions.assertFalse( - scm.getPipelineManager().isPipelineCreationFrozen()); - Assertions.assertEquals( + assertFalse(scm.getPipelineManager().isPipelineCreationFrozen()); + assertEquals( scm.getScmContext().getFinalizationCheckpoint(), FinalizationCheckpoint.FINALIZATION_REQUIRED); break; case AFTER_PRE_FINALIZE_UPGRADE: - Assertions.assertTrue( - scm.getPipelineManager().isPipelineCreationFrozen()); - Assertions.assertEquals( + assertTrue(scm.getPipelineManager().isPipelineCreationFrozen()); + assertEquals( scm.getScmContext().getFinalizationCheckpoint(), FinalizationCheckpoint.FINALIZATION_STARTED); break; case AFTER_COMPLETE_FINALIZATION: - Assertions.assertFalse( - scm.getPipelineManager().isPipelineCreationFrozen()); - Assertions.assertEquals( + assertFalse(scm.getPipelineManager().isPipelineCreationFrozen()); + assertEquals( scm.getScmContext().getFinalizationCheckpoint(), FinalizationCheckpoint.MLV_EQUALS_SLV); break; case AFTER_POST_FINALIZE_UPGRADE: - Assertions.assertFalse( - scm.getPipelineManager().isPipelineCreationFrozen()); - Assertions.assertEquals( + assertFalse(scm.getPipelineManager().isPipelineCreationFrozen()); + assertEquals( scm.getScmContext().getFinalizationCheckpoint(), FinalizationCheckpoint.FINALIZATION_COMPLETE); break; default: - Assertions.fail("Unknown halting point in test: " + haltingPoint); + fail("Unknown halting point in test: " + haltingPoint); } } } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/OzoneTestUtils.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/OzoneTestUtils.java index d89e6a6c3600..adf4d5ab9551 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/OzoneTestUtils.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/OzoneTestUtils.java @@ -35,8 +35,10 @@ import org.apache.ozone.test.GenericTestUtils; import org.apache.ozone.test.LambdaTestUtils.VoidCallable; import org.apache.ratis.util.function.CheckedConsumer; -import org.junit.jupiter.api.Assertions; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertFalse; +import static org.junit.jupiter.api.Assertions.assertThrows; /** * Helper class for Tests. @@ -92,7 +94,7 @@ public static void closeContainers( .updateContainerState(ContainerID.valueOf(blockID.getContainerID()), HddsProtos.LifeCycleEvent.CLOSE); } - Assertions.assertFalse(scm.getContainerManager() + assertFalse(scm.getContainerManager() .getContainer(ContainerID.valueOf(blockID.getContainerID())) .isOpen()); }, omKeyLocationInfoGroups); @@ -140,14 +142,10 @@ public static void performOperationOnKeyContainers( public static void expectOmException( OMException.ResultCodes code, - VoidCallable eval) - throws Exception { - try { - eval.call(); - Assertions.fail("OMException is expected"); - } catch (OMException ex) { - Assertions.assertEquals(code, ex.getResult()); - } + VoidCallable eval) { + + OMException ex = assertThrows(OMException.class, () -> eval.call(), "OMException is expected"); + assertEquals(code, ex.getResult()); } /** diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokensCLI.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokensCLI.java index 9a98a0a1897e..53356087e696 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokensCLI.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokensCLI.java @@ -71,11 +71,13 @@ import static org.apache.hadoop.ozone.om.OMConfigKeys.OZONE_OM_KERBEROS_PRINCIPAL_KEY; import static org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod.KERBEROS; import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertNotEquals; + import org.junit.jupiter.api.AfterAll; import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Test; import org.junit.jupiter.api.Timeout; -import org.junit.jupiter.api.Assertions; + /** * Integration test class to verify block token CLI commands functionality in a * secure cluster. @@ -264,7 +266,7 @@ public void testRotateKeySCMAdminCommandUtil(String[] args) { // rotating. String currentKey = getScmSecretKeyManager().getCurrentSecretKey().toString(); - Assertions.assertEquals(initialKey, currentKey); + assertEquals(initialKey, currentKey); // Rotate the secret key. ozoneAdmin.execute(args); @@ -280,9 +282,9 @@ public void testRotateKeySCMAdminCommandUtil(String[] args) { // Otherwise, both keys should be the same. if (isForceFlagPresent(args) || shouldRotate(getScmSecretKeyManager().getCurrentSecretKey())) { - Assertions.assertNotEquals(initialKey, newKey); + assertNotEquals(initialKey, newKey); } else { - Assertions.assertEquals(initialKey, newKey); + assertEquals(initialKey, newKey); } } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestContainerOperations.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestContainerOperations.java index 56e744886b82..d50c75b0da4f 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestContainerOperations.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestContainerOperations.java @@ -40,9 +40,11 @@ import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Test; import org.junit.jupiter.api.Timeout; -import org.junit.jupiter.api.Assertions; + import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertFalse; import static org.junit.jupiter.api.Assertions.assertTrue; +import static org.junit.jupiter.api.Assertions.fail; /** * This class tests container operations (TODO currently only supports create) @@ -93,13 +95,13 @@ public void testCreate() throws Exception { public void testGetPipeline() throws Exception { try { storageClient.getPipeline(PipelineID.randomId().getProtobuf()); - Assertions.fail("Get Pipeline should fail"); + fail("Get Pipeline should fail"); } catch (Exception e) { assertTrue( SCMHAUtils.unwrapException(e) instanceof PipelineNotFoundException); } - Assertions.assertFalse(storageClient.listPipelines().isEmpty()); + assertFalse(storageClient.listPipelines().isEmpty()); } @Test diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestContainerStateMachineIdempotency.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestContainerStateMachineIdempotency.java index 6e76a86dd947..37d0703e9fd8 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestContainerStateMachineIdempotency.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestContainerStateMachineIdempotency.java @@ -41,11 +41,11 @@ import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Test; import org.junit.jupiter.api.Timeout; -import org.junit.jupiter.api.Assertions; import java.io.IOException; import java.util.concurrent.TimeUnit; import static java.nio.charset.StandardCharsets.UTF_8; +import static org.junit.jupiter.api.Assertions.fail; /** * Tests the idempotent operations in ContainerStateMachine. @@ -115,7 +115,7 @@ public void testContainerStateMachineIdempotency() throws Exception { ContainerProtocolCalls.closeContainer(client, containerID, null); ContainerProtocolCalls.closeContainer(client, containerID, null); } catch (IOException ioe) { - Assertions.fail("Container operation failed" + ioe); + fail("Container operation failed" + ioe); } xceiverClientManager.releaseClient(client, false); } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMiniOzoneOMHACluster.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMiniOzoneOMHACluster.java index a3e0be5a85d3..76a304086b71 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMiniOzoneOMHACluster.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMiniOzoneOMHACluster.java @@ -29,12 +29,13 @@ import static org.apache.hadoop.ozone.OzoneConfigKeys.OZONE_ACL_ENABLED; import static org.apache.hadoop.ozone.OzoneConfigKeys.OZONE_ADMINISTRATORS_WILDCARD; +import static org.junit.jupiter.api.Assertions.assertNotNull; +import static org.junit.jupiter.api.Assertions.assertTrue; import org.junit.jupiter.api.AfterEach; import org.junit.jupiter.api.BeforeEach; import org.junit.jupiter.api.Test; import org.junit.jupiter.api.Timeout; -import org.junit.jupiter.api.Assertions; /** * This class tests MiniOzoneHAClusterImpl. @@ -52,7 +53,7 @@ public class TestMiniOzoneOMHACluster { /** * Create a MiniOzoneHAClusterImpl for testing. * - * @throws IOException + * @throws Exception */ @BeforeEach public void init() throws Exception { @@ -91,9 +92,8 @@ public void testGetOMLeader() throws InterruptedException, TimeoutException { ozoneManager.set(om); return om != null; }, 100, 120000); - Assertions.assertNotNull(ozoneManager, "Timed out waiting OM leader election to finish: " + assertNotNull(ozoneManager, "Timed out waiting OM leader election to finish: " + "no leader or more than one leader."); - Assertions.assertTrue(ozoneManager.get().isLeaderReady(), - "Should have gotten the leader!"); + assertTrue(ozoneManager.get().isLeaderReady(), "Should have gotten the leader!"); } } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMultipartObjectGet.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMultipartObjectGet.java index cec90067dacb..0dae8a8b0dc6 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMultipartObjectGet.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMultipartObjectGet.java @@ -31,7 +31,6 @@ import org.junit.jupiter.api.AfterAll; import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Test; -import org.mockito.Mockito; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -53,6 +52,7 @@ import static org.apache.hadoop.ozone.s3.util.S3Consts.STORAGE_CLASS_HEADER; import static org.junit.jupiter.api.Assertions.assertEquals; import static org.junit.jupiter.api.Assertions.assertNotNull; +import static org.mockito.Mockito.mock; import static org.mockito.Mockito.when; /** @@ -89,13 +89,13 @@ public static void init() throws Exception { client = cluster.newClient(); client.getObjectStore().createS3Bucket(BUCKET); - headers = Mockito.mock(HttpHeaders.class); + headers = mock(HttpHeaders.class); when(headers.getHeaderString(STORAGE_CLASS_HEADER)).thenReturn( "STANDARD"); - context = Mockito.mock(ContainerRequestContext.class); - Mockito.when(context.getUriInfo()).thenReturn(Mockito.mock(UriInfo.class)); - Mockito.when(context.getUriInfo().getQueryParameters()) + context = mock(ContainerRequestContext.class); + when(context.getUriInfo()).thenReturn(mock(UriInfo.class)); + when(context.getUriInfo().getQueryParameters()) .thenReturn(new MultivaluedHashMap<>()); REST.setHeaders(headers); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/debug/TestLDBCli.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/debug/TestLDBCli.java index 0d65d81c5ed6..a5264fd0b4e0 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/debug/TestLDBCli.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/debug/TestLDBCli.java @@ -38,7 +38,6 @@ import org.apache.hadoop.ozone.om.request.OMRequestTestUtils; import org.jetbrains.annotations.NotNull; import org.junit.jupiter.api.AfterEach; -import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.BeforeEach; import org.junit.jupiter.api.Named; import org.junit.jupiter.api.Test; @@ -62,6 +61,8 @@ import java.util.stream.Stream; import static java.nio.charset.StandardCharsets.UTF_8; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertTrue; /** * This class tests `ozone debug ldb` CLI that reads from a RocksDB directory. @@ -218,7 +219,7 @@ void testLDBScan( int exitCode = cmd.execute(completeScanArgs.toArray(new String[0])); // Check exit code. Print stderr if not expected int expectedExitCode = expectedExitCodeStderrPair.getLeft(); - Assertions.assertEquals(expectedExitCode, exitCode, stderr.toString()); + assertEquals(expectedExitCode, exitCode, stderr.toString()); // Construct expected result map given test param input Map> expectedMap; @@ -235,7 +236,7 @@ void testLDBScan( // Check stderr final String stderrShouldContain = expectedExitCodeStderrPair.getRight(); - Assertions.assertTrue(stderr.toString().contains(stderrShouldContain)); + assertTrue(stderr.toString().contains(stderrShouldContain)); } @Test @@ -251,13 +252,13 @@ void testScanOfPipelinesWhenNoData() throws IOException { int exitCode = cmd.execute(completeScanArgs.toArray(new String[0])); // Check exit code. Print stderr if not expected - Assertions.assertEquals(0, exitCode, stderr.toString()); + assertEquals(0, exitCode, stderr.toString()); // Check stdout - Assertions.assertEquals("{ }\n", stdout.toString()); + assertEquals("{ }\n", stdout.toString()); // Check stderr - Assertions.assertEquals("", stderr.toString()); + assertEquals("", stderr.toString()); } /** @@ -271,7 +272,7 @@ private void assertContents(Map expected, String actualStr) Map> actualMap = MAPPER.readValue( actualStr, new TypeReference>>() { }); - Assertions.assertEquals(expected, actualMap); + assertEquals(expected, actualMap); } /** diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestDataValidate.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestDataValidate.java index 17d82fcc8e64..28cc863c26d5 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestDataValidate.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestDataValidate.java @@ -23,12 +23,15 @@ import org.apache.hadoop.hdds.protocol.proto.HddsProtos; import org.apache.hadoop.hdds.ratis.conf.RatisClientConfig; import org.apache.hadoop.ozone.MiniOzoneCluster; -import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.Test; import picocli.CommandLine; import java.time.Duration; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertNotEquals; +import static org.junit.jupiter.api.Assertions.assertTrue; + /** * Tests Freon, with MiniOzoneCluster and validate data. */ @@ -76,10 +79,10 @@ public void ratisTestLargeKey() { "--validate-writes" ); - Assertions.assertEquals(1, randomKeyGenerator.getNumberOfVolumesCreated()); - Assertions.assertEquals(1, randomKeyGenerator.getNumberOfBucketsCreated()); - Assertions.assertEquals(1, randomKeyGenerator.getNumberOfKeysAdded()); - Assertions.assertEquals(0, randomKeyGenerator.getUnsuccessfulValidationCount()); + assertEquals(1, randomKeyGenerator.getNumberOfVolumesCreated()); + assertEquals(1, randomKeyGenerator.getNumberOfBucketsCreated()); + assertEquals(1, randomKeyGenerator.getNumberOfKeysAdded()); + assertEquals(0, randomKeyGenerator.getUnsuccessfulValidationCount()); } @Test @@ -95,14 +98,12 @@ public void validateWriteTest() { "--validate-writes" ); - Assertions.assertEquals(2, randomKeyGenerator.getNumberOfVolumesCreated()); - Assertions.assertEquals(10, randomKeyGenerator.getNumberOfBucketsCreated()); - Assertions.assertEquals(100, randomKeyGenerator.getNumberOfKeysAdded()); - Assertions.assertTrue(randomKeyGenerator.getValidateWrites()); - Assertions.assertNotEquals(0, randomKeyGenerator.getTotalKeysValidated()); - Assertions.assertNotEquals(0, randomKeyGenerator - .getSuccessfulValidationCount()); - Assertions.assertEquals(0, randomKeyGenerator - .getUnsuccessfulValidationCount()); + assertEquals(2, randomKeyGenerator.getNumberOfVolumesCreated()); + assertEquals(10, randomKeyGenerator.getNumberOfBucketsCreated()); + assertEquals(100, randomKeyGenerator.getNumberOfKeysAdded()); + assertTrue(randomKeyGenerator.getValidateWrites()); + assertNotEquals(0, randomKeyGenerator.getTotalKeysValidated()); + assertNotEquals(0, randomKeyGenerator.getSuccessfulValidationCount()); + assertEquals(0, randomKeyGenerator.getUnsuccessfulValidationCount()); } } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestFreonWithDatanodeFastRestart.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestFreonWithDatanodeFastRestart.java index 12b00a0c8cc0..d530a6312c82 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestFreonWithDatanodeFastRestart.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestFreonWithDatanodeFastRestart.java @@ -28,14 +28,15 @@ import org.apache.ratis.statemachine.impl.SingleFileSnapshotInfo; import org.junit.jupiter.api.AfterAll; -import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Test; import org.junit.jupiter.api.Timeout; import picocli.CommandLine; - import java.util.concurrent.TimeUnit; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertTrue; + /** * Tests Freon with Datanode restarts without waiting for pipeline to close. */ @@ -86,14 +87,13 @@ public void testRestart() throws Exception { String expectedSnapFile = storage.getSnapshotFile(termIndexBeforeRestart.getTerm(), termIndexBeforeRestart.getIndex()).getAbsolutePath(); - Assertions.assertEquals(expectedSnapFile, - snapshotInfo.getFile().getPath().toString()); - Assertions.assertEquals(termInSnapshot, termIndexBeforeRestart); + assertEquals(expectedSnapFile, snapshotInfo.getFile().getPath().toString()); + assertEquals(termInSnapshot, termIndexBeforeRestart); // After restart the term index might have progressed to apply pending // transactions. TermIndex termIndexAfterRestart = sm.getLastAppliedTermIndex(); - Assertions.assertTrue(termIndexAfterRestart.getIndex() >= + assertTrue(termIndexAfterRestart.getIndex() >= termIndexBeforeRestart.getIndex()); // TODO: fix me // Give some time for the datanode to register again with SCM. @@ -119,10 +119,10 @@ private void startFreon() { "--validate-writes" ); - Assertions.assertEquals(1, randomKeyGenerator.getNumberOfVolumesCreated()); - Assertions.assertEquals(1, randomKeyGenerator.getNumberOfBucketsCreated()); - Assertions.assertEquals(1, randomKeyGenerator.getNumberOfKeysAdded()); - Assertions.assertEquals(0, randomKeyGenerator.getUnsuccessfulValidationCount()); + assertEquals(1, randomKeyGenerator.getNumberOfVolumesCreated()); + assertEquals(1, randomKeyGenerator.getNumberOfBucketsCreated()); + assertEquals(1, randomKeyGenerator.getNumberOfKeysAdded()); + assertEquals(0, randomKeyGenerator.getUnsuccessfulValidationCount()); } private StateMachine getStateMachine() throws Exception { diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestFreonWithPipelineDestroy.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestFreonWithPipelineDestroy.java index c0b055a8b205..d78beff7e78b 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestFreonWithPipelineDestroy.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestFreonWithPipelineDestroy.java @@ -31,14 +31,14 @@ import org.apache.hadoop.ozone.MiniOzoneCluster; import org.apache.hadoop.ozone.container.common.transport.server.XceiverServerSpi; import org.junit.jupiter.api.AfterAll; -import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Test; import picocli.CommandLine; - import java.time.Duration; import java.util.concurrent.TimeUnit; +import static org.junit.jupiter.api.Assertions.assertEquals; + /** * Tests Freon with Pipeline destroy. */ @@ -114,11 +114,10 @@ private void startFreon() { "--validate-writes" ); - Assertions.assertEquals(1, randomKeyGenerator.getNumberOfVolumesCreated()); - Assertions.assertEquals(1, randomKeyGenerator.getNumberOfBucketsCreated()); - Assertions.assertEquals(1, randomKeyGenerator.getNumberOfKeysAdded()); - Assertions.assertEquals(0, - randomKeyGenerator.getUnsuccessfulValidationCount()); + assertEquals(1, randomKeyGenerator.getNumberOfVolumesCreated()); + assertEquals(1, randomKeyGenerator.getNumberOfBucketsCreated()); + assertEquals(1, randomKeyGenerator.getNumberOfKeysAdded()); + assertEquals(0, randomKeyGenerator.getUnsuccessfulValidationCount()); } private void destroyPipeline() throws Exception { diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestHadoopDirTreeGenerator.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestHadoopDirTreeGenerator.java index afda0368f02e..651e5350cff8 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestHadoopDirTreeGenerator.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestHadoopDirTreeGenerator.java @@ -34,7 +34,6 @@ import org.apache.ozone.test.GenericTestUtils; import org.apache.ratis.server.RaftServer; import org.apache.ratis.server.raftlog.RaftLog; -import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.BeforeEach; import org.junit.jupiter.api.Test; import org.slf4j.Logger; @@ -47,6 +46,9 @@ import java.net.URI; import java.util.ArrayList; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertFalse; + /** * Test for HadoopDirTreeGenerator. */ @@ -157,7 +159,7 @@ private void verifyDirTree(String volumeName, String bucketName, int depth, int actualDepth = traverseToLeaf(fileSystem, fileStatus.getPath(), 1, depth, span, fileCount, StorageSize.parse(perFileSize, StorageUnit.BYTES)); - Assertions.assertEquals(depth, actualDepth, "Mismatch depth in a path"); + assertEquals(depth, actualDepth, "Mismatch depth in a path"); } } @@ -179,16 +181,16 @@ private int traverseToLeaf(FileSystem fs, Path dirPath, int depth, return traverseToLeaf(fs, fileStatus.getPath(), depth, expectedDepth, expectedSpanCnt, expectedFileCnt, perFileSize); } else { - Assertions.assertEquals(perFileSize.toBytes(), fileStatus.getLen(), "Mismatches file len"); + assertEquals(perFileSize.toBytes(), fileStatus.getLen(), "Mismatches file len"); String fName = fileStatus.getPath().getName(); - Assertions.assertFalse(files.contains(fName), actualNumFiles + "actualNumFiles:" + fName + + assertFalse(files.contains(fName), actualNumFiles + "actualNumFiles:" + fName + ", fName:" + expectedFileCnt + ", expectedFileCnt:" + depth + ", depth:"); files.add(fName); actualNumFiles++; } } - Assertions.assertEquals(expectedFileCnt, actualNumFiles, "Mismatches files count in a directory"); + assertEquals(expectedFileCnt, actualNumFiles, "Mismatches files count in a directory"); return depth; } @@ -200,7 +202,7 @@ private int verifyActualSpan(int expectedSpanCnt, ++actualSpan; } } - Assertions.assertEquals(expectedSpanCnt, actualSpan, "Mismatches subdirs count in a directory"); + assertEquals(expectedSpanCnt, actualSpan, "Mismatches subdirs count in a directory"); return actualSpan; } } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestHadoopNestedDirGenerator.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestHadoopNestedDirGenerator.java index 7c75858269c0..4411c0d2ea56 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestHadoopNestedDirGenerator.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestHadoopNestedDirGenerator.java @@ -32,7 +32,6 @@ import org.apache.ratis.server.RaftServer; import org.apache.ratis.server.raftlog.RaftLog; import java.util.LinkedList; -import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.BeforeEach; import org.junit.jupiter.api.Test; import org.slf4j.Logger; @@ -44,6 +43,7 @@ import java.net.URI; import static org.apache.ozone.test.GenericTestUtils.getTempPath; +import static org.junit.jupiter.api.Assertions.assertEquals; /** * Test for HadoopNestedDirGenerator. @@ -142,7 +142,7 @@ private void verifyDirTree(String volumeName, String bucketName, // verify the num of peer directories and span directories p = depthBFS(fileSystem, fileStatuses, span, actualDepth); int actualSpan = spanCheck(fileSystem, span, p); - Assertions.assertEquals(span, actualSpan, "Mismatch span in a path"); + assertEquals(span, actualSpan, "Mismatch span in a path"); } } @@ -182,7 +182,7 @@ private Path depthBFS(FileSystem fs, FileStatus[] fileStatuses, p = f.getPath().getParent(); } } - Assertions.assertEquals(depth, actualDepth, "Mismatch depth in a path"); + assertEquals(depth, actualDepth, "Mismatch depth in a path"); return p; } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestOMSnapshotDAG.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestOMSnapshotDAG.java index 88fb0107969c..bca21aebd1ac 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestOMSnapshotDAG.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestOMSnapshotDAG.java @@ -46,7 +46,6 @@ import org.apache.ratis.server.RaftServer; import org.apache.ratis.server.raftlog.RaftLog; import org.junit.jupiter.api.AfterAll; -import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Test; import org.slf4j.Logger; @@ -70,6 +69,10 @@ import static org.apache.hadoop.ozone.OzoneConsts.OM_KEY_PREFIX; import static org.apache.hadoop.ozone.OzoneConsts.OM_SNAPSHOT_DIFF_DIR; import static org.apache.hadoop.ozone.om.snapshot.SnapshotUtils.getColumnFamilyToKeyPrefixMap; +import static org.assertj.core.api.Assertions.assertThat; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertNotNull; + /** * Tests Freon, with MiniOzoneCluster. @@ -178,9 +181,8 @@ public void testDAGReconstruction() "--validate-writes" ); - Assertions.assertEquals(500L, randomKeyGenerator.getNumberOfKeysAdded()); - Assertions.assertEquals(500L, - randomKeyGenerator.getSuccessfulValidationCount()); + assertEquals(500L, randomKeyGenerator.getNumberOfKeysAdded()); + assertEquals(500L, randomKeyGenerator.getSuccessfulValidationCount()); List volList = cluster.getOzoneManager() .listAllVolumes("", "", 2); @@ -263,7 +265,7 @@ public void testDAGReconstruction() // Same snapshot. Result should be empty list List sstDiffList22 = differ.getSSTDiffList(snap2, snap2); - Assertions.assertTrue(sstDiffList22.isEmpty()); + assertThat(sstDiffList22).isEmpty(); snapDB1.close(); snapDB2.close(); snapDB3.close(); @@ -292,13 +294,13 @@ public void testDAGReconstruction() ((RDBStore)((OmSnapshot)snapDB3.get()) .getMetadataManager().getStore()).getDb().getManagedRocksDb()); List sstDiffList21Run2 = differ.getSSTDiffList(snap2, snap1); - Assertions.assertEquals(sstDiffList21, sstDiffList21Run2); + assertEquals(sstDiffList21, sstDiffList21Run2); List sstDiffList32Run2 = differ.getSSTDiffList(snap3, snap2); - Assertions.assertEquals(sstDiffList32, sstDiffList32Run2); + assertEquals(sstDiffList32, sstDiffList32Run2); List sstDiffList31Run2 = differ.getSSTDiffList(snap3, snap1); - Assertions.assertEquals(sstDiffList31, sstDiffList31Run2); + assertEquals(sstDiffList31, sstDiffList31Run2); snapDB1.close(); snapDB2.close(); snapDB3.close(); @@ -324,9 +326,8 @@ public void testSkipTrackingWithZeroSnapshot() { "--validate-writes" ); - Assertions.assertEquals(1000L, randomKeyGenerator.getNumberOfKeysAdded()); - Assertions.assertEquals(1000L, - randomKeyGenerator.getSuccessfulValidationCount()); + assertEquals(1000L, randomKeyGenerator.getNumberOfKeysAdded()); + assertEquals(1000L, randomKeyGenerator.getSuccessfulValidationCount()); String omMetadataDir = cluster.getOzoneManager().getConfiguration().get(OZONE_METADATA_DIRS); @@ -338,7 +339,7 @@ public void testSkipTrackingWithZeroSnapshot() { if (fileList != null) { for (File file : fileList) { if (file != null && file.isFile() && file.getName().endsWith(".log")) { - Assertions.assertEquals(0L, file.length()); + assertEquals(0L, file.length()); } } } @@ -346,8 +347,8 @@ public void testSkipTrackingWithZeroSnapshot() { Path sstBackupPath = Paths.get(omMetadataDir, OM_SNAPSHOT_DIFF_DIR, DB_COMPACTION_SST_BACKUP_DIR); fileList = sstBackupPath.toFile().listFiles(); - Assertions.assertNotNull(fileList); - Assertions.assertEquals(0L, fileList.length); + assertNotNull(fileList); + assertEquals(0L, fileList.length); } } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestOmBucketReadWriteKeyOps.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestOmBucketReadWriteKeyOps.java index ba141f7c5ebe..ad10c4d47651 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestOmBucketReadWriteKeyOps.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/freon/TestOmBucketReadWriteKeyOps.java @@ -33,7 +33,6 @@ import org.apache.ozone.test.GenericTestUtils; import org.apache.ratis.server.RaftServer; import org.apache.ratis.server.raftlog.RaftLog; -import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.BeforeEach; import org.junit.jupiter.api.Test; import org.slf4j.Logger; @@ -45,6 +44,9 @@ import java.io.IOException; import java.util.Iterator; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertTrue; + /** * Test for OmBucketReadWriteKeyOps. */ @@ -193,7 +195,7 @@ private void verifyKeyCreation(int expectedCount, OzoneBucket bucket, ozoneKeyIterator.next(); ++actual; } - Assertions.assertEquals(expectedCount, actual, "Mismatch Count!"); + assertEquals(expectedCount, actual, "Mismatch Count!"); } private void verifyOMLockMetrics(OMLockMetrics omLockMetrics) { @@ -204,7 +206,7 @@ private void verifyOMLockMetrics(OMLockMetrics omLockMetrics) { omLockMetrics.getLongestReadLockWaitingTimeMs()); int readWaitingSamples = Integer.parseInt(readLockWaitingTimeMsStat.split(" ")[2]); - Assertions.assertTrue(readWaitingSamples > 0, "Read Lock Waiting Samples should be positive"); + assertTrue(readWaitingSamples > 0, "Read Lock Waiting Samples should be positive"); String readLockHeldTimeMsStat = omLockMetrics.getReadLockHeldTimeMsStat(); LOG.info("Read Lock Held Time Stat: " + readLockHeldTimeMsStat); @@ -212,7 +214,7 @@ private void verifyOMLockMetrics(OMLockMetrics omLockMetrics) { omLockMetrics.getLongestReadLockHeldTimeMs()); int readHeldSamples = Integer.parseInt(readLockHeldTimeMsStat.split(" ")[2]); - Assertions.assertTrue(readHeldSamples > 0, "Read Lock Held Samples should be positive"); + assertTrue(readHeldSamples > 0, "Read Lock Held Samples should be positive"); String writeLockWaitingTimeMsStat = omLockMetrics.getWriteLockWaitingTimeMsStat(); @@ -221,7 +223,7 @@ private void verifyOMLockMetrics(OMLockMetrics omLockMetrics) { omLockMetrics.getLongestWriteLockWaitingTimeMs()); int writeWaitingSamples = Integer.parseInt(writeLockWaitingTimeMsStat.split(" ")[2]); - Assertions.assertTrue(writeWaitingSamples > 0, "Write Lock Waiting Samples should be positive"); + assertTrue(writeWaitingSamples > 0, "Write Lock Waiting Samples should be positive"); String writeLockHeldTimeMsStat = omLockMetrics.getWriteLockHeldTimeMsStat(); LOG.info("Write Lock Held Time Stat: " + writeLockHeldTimeMsStat); @@ -229,7 +231,7 @@ private void verifyOMLockMetrics(OMLockMetrics omLockMetrics) { omLockMetrics.getLongestWriteLockHeldTimeMs()); int writeHeldSamples = Integer.parseInt(writeLockHeldTimeMsStat.split(" ")[2]); - Assertions.assertTrue(writeHeldSamples > 0, "Write Lock Held Samples should be positive"); + assertTrue(writeHeldSamples > 0, "Write Lock Held Samples should be positive"); } private static class ParameterBuilder { diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/parser/TestOzoneHARatisLogParser.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/parser/TestOzoneHARatisLogParser.java index 9902187be665..e5349b09d00f 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/parser/TestOzoneHARatisLogParser.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/parser/TestOzoneHARatisLogParser.java @@ -34,7 +34,6 @@ import org.apache.ozone.test.GenericTestUtils; import org.apache.ozone.test.tag.Flaky; import org.junit.jupiter.api.AfterEach; -import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.BeforeEach; import org.junit.jupiter.api.Test; import org.junit.jupiter.api.Timeout; @@ -46,6 +45,9 @@ import static java.nio.charset.StandardCharsets.UTF_8; import static org.apache.hadoop.hdds.HddsConfigKeys.OZONE_METADATA_DIRS; +import static org.assertj.core.api.Assertions.assertThat; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertNotNull; /** * Test Ozone OM and SCM HA Ratis log parser. @@ -113,20 +115,20 @@ void testRatisLogParsing() throws Exception { File omMetaDir = new File(ozoneConfiguration.get(OZONE_METADATA_DIRS), "ratis"); - Assertions.assertTrue(omMetaDir.isDirectory()); + assertThat(omMetaDir).isDirectory(); String[] ratisDirs = omMetaDir.list(); - Assertions.assertNotNull(ratisDirs); - Assertions.assertEquals(1, ratisDirs.length); + assertNotNull(ratisDirs); + assertEquals(1, ratisDirs.length); File groupDir = new File(omMetaDir, ratisDirs[0]); - Assertions.assertNotNull(groupDir); - Assertions.assertTrue(groupDir.isDirectory()); + assertNotNull(groupDir); + assertThat(groupDir).isDirectory(); File currentDir = new File(groupDir, "current"); File logFile = new File(currentDir, "log_inprogress_0"); GenericTestUtils.waitFor(logFile::exists, 100, 15000); - Assertions.assertTrue(logFile.isFile()); + assertThat(logFile).isFile(); OMRatisLogParser omRatisLogParser = new OMRatisLogParser(); omRatisLogParser.setSegmentFile(logFile); @@ -135,27 +137,26 @@ void testRatisLogParsing() throws Exception { // Not checking total entry count, because of not sure of exact count of // metadata entry changes. - Assertions.assertTrue(out.toString(UTF_8.name()) - .contains("Num Total Entries:")); + assertThat(out.toString(UTF_8.name())).contains("Num Total Entries:"); out.reset(); // Now check for SCM. File scmMetadataDir = new File(SCMHAUtils.getRatisStorageDir(leaderSCMConfig)); - Assertions.assertTrue(scmMetadataDir.isDirectory()); + assertThat(scmMetadataDir).isDirectory(); ratisDirs = scmMetadataDir.list(); - Assertions.assertNotNull(ratisDirs); - Assertions.assertEquals(1, ratisDirs.length); + assertNotNull(ratisDirs); + assertEquals(1, ratisDirs.length); groupDir = new File(scmMetadataDir, ratisDirs[0]); - Assertions.assertNotNull(groupDir); - Assertions.assertTrue(groupDir.isDirectory()); + assertNotNull(groupDir); + assertThat(groupDir).isDirectory(); currentDir = new File(groupDir, "current"); logFile = new File(currentDir, "log_inprogress_1"); GenericTestUtils.waitFor(logFile::exists, 100, 15000); - Assertions.assertTrue(logFile.isFile()); + assertThat(logFile).isFile(); SCMRatisLogParser scmRatisLogParser = new SCMRatisLogParser(); scmRatisLogParser.setSegmentFile(logFile); @@ -163,7 +164,6 @@ void testRatisLogParsing() throws Exception { // Not checking total entry count, because of not sure of exact count of // metadata entry changes. - Assertions.assertTrue(out.toString(UTF_8.name()) - .contains("Num Total Entries:")); + assertThat(out.toString(UTF_8.name())).contains("Num Total Entries:"); } } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/recon/TestReconInsightsForDeletedDirectories.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/recon/TestReconInsightsForDeletedDirectories.java index 74751dde6def..4c059be1b542 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/recon/TestReconInsightsForDeletedDirectories.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/recon/TestReconInsightsForDeletedDirectories.java @@ -47,7 +47,6 @@ import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.AfterAll; import org.junit.jupiter.api.AfterEach; -import org.junit.jupiter.api.Assertions; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -62,6 +61,9 @@ import static org.apache.hadoop.ozone.om.OMConfigKeys.OZONE_DIR_DELETING_SERVICE_INTERVAL; import static org.apache.hadoop.ozone.om.OMConfigKeys.OZONE_PATH_DELETING_LIMIT_PER_TASK; import static org.apache.hadoop.ozone.om.OMConfigKeys.OZONE_OM_RATIS_ENABLE_KEY; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertFalse; +import static org.junit.jupiter.api.Assertions.fail; import static org.mockito.Mockito.mock; /** @@ -132,7 +134,7 @@ public void cleanup() { fs.delete(fileStatus.getPath(), true); } } catch (IOException ex) { - Assertions.fail("Failed to cleanup files."); + fail("Failed to cleanup files."); } } @@ -201,7 +203,7 @@ public void testGetDeletedDirectoryInfo() } if (directoryObjectId == null) { - Assertions.fail("directoryObjectId is null. Test case cannot proceed."); + fail("directoryObjectId is null. Test case cannot proceed."); } else { // Retrieve Namespace Summary for dir1 from Recon. ReconNamespaceSummaryManagerImpl namespaceSummaryManager = @@ -210,8 +212,8 @@ public void testGetDeletedDirectoryInfo() NSSummary summary = namespaceSummaryManager.getNSSummary(directoryObjectId); // Assert that the directory dir1 has 10 sub-files and size of 1000 bytes. - Assertions.assertEquals(10, summary.getNumOfFiles()); - Assertions.assertEquals(10, summary.getSizeOfFiles()); + assertEquals(10, summary.getNumOfFiles()); + assertEquals(10, summary.getSizeOfFiles()); } // Delete the entire directory dir1. @@ -238,7 +240,7 @@ public void testGetDeletedDirectoryInfo() KeyInsightInfoResponse entity = (KeyInsightInfoResponse) deletedDirInfo.getEntity(); // Assert the size of deleted directory is 10. - Assertions.assertEquals(10, entity.getUnreplicatedDataSize()); + assertEquals(10, entity.getUnreplicatedDataSize()); // Cleanup the tables. cleanupTables(); @@ -327,7 +329,7 @@ public void testGetDeletedDirectoryInfoForNestedDirectories() KeyInsightInfoResponse entity = (KeyInsightInfoResponse) deletedDirInfo.getEntity(); // Assert the size of deleted directory is 3. - Assertions.assertEquals(3, entity.getUnreplicatedDataSize()); + assertEquals(3, entity.getUnreplicatedDataSize()); // Cleanup the tables. cleanupTables(); @@ -368,7 +370,7 @@ public void testGetDeletedDirectoryInfoWithMultipleSubdirectories() fs.delete(rootDir, true); // Verify that the directory is deleted - Assertions.assertFalse(fs.exists(rootDir), "Directory was not deleted"); + assertFalse(fs.exists(rootDir), "Directory was not deleted"); // Sync data from Ozone Manager to Recon. syncDataFromOM(); @@ -389,7 +391,7 @@ public void testGetDeletedDirectoryInfoWithMultipleSubdirectories() KeyInsightInfoResponse entity = (KeyInsightInfoResponse) deletedDirInfo.getEntity(); // Assert the size of deleted directory is 100. - Assertions.assertEquals(100, entity.getUnreplicatedDataSize()); + assertEquals(100, entity.getUnreplicatedDataSize()); // Cleanup the tables. cleanupTables(); @@ -471,7 +473,7 @@ private boolean assertTableRowCount(int expectedCount, LOG.info("{} actual row count={}, expectedCount={}", table.getName(), count, expectedCount); } catch (IOException ex) { - Assertions.fail("Test failed with: " + ex); + fail("Test failed with: " + ex); } return count == expectedCount; }