diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTest.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTest.java index d27622592244..f8b9c26d9117 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTest.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTest.java @@ -62,7 +62,6 @@ import java.util.Collection; import java.util.Collections; import java.util.EnumSet; -import java.util.HashMap; import java.util.Iterator; import java.util.LinkedList; import java.util.List; @@ -113,7 +112,6 @@ import org.apache.hadoop.ozone.client.OzoneKeyDetails; import org.apache.hadoop.ozone.client.OzoneVolume; import org.apache.hadoop.ozone.client.VolumeArgs; -import org.apache.hadoop.ozone.client.io.OzoneOutputStream; import org.apache.hadoop.ozone.client.protocol.ClientProtocol; import org.apache.hadoop.ozone.om.OMConfigKeys; import org.apache.hadoop.ozone.om.OMMetrics; @@ -342,12 +340,9 @@ void testListStatusWithIntermediateDirWithECEnabled() String key = "object-dir/object-name1"; // write some test data into bucket - try (OzoneOutputStream outputStream = objectStore.getVolume(volumeName). - getBucket(bucketName).createKey(key, 1, - new ECReplicationConfig("RS-3-2-1024k"), - new HashMap<>())) { - outputStream.write(RandomUtils.nextBytes(1)); - } + TestDataUtil.createKey(objectStore.getVolume(volumeName).getBucket(bucketName), + key, new ECReplicationConfig("RS-3-2-1024k"), + RandomUtils.secure().randomBytes(1)); List dirs = Arrays.asList(volumeName, bucketName, "object-dir", "object-name1"); @@ -1495,13 +1490,10 @@ void testSymlinkList() throws Exception { // add key in source bucket final String key = "object-dir/object-name1"; - try (OzoneOutputStream outputStream = objectStore.getVolume(srcVolume) - .getBucket(srcBucket) - .createKey(key, 1)) { - outputStream.write(RandomUtils.nextBytes(1)); - } - assertEquals(objectStore.getVolume(srcVolume) - .getBucket(srcBucket).getKey(key).getName(), key); + TestDataUtil.createKey(objectStore.getVolume(srcVolume).getBucket(srcBucket), + key, RandomUtils.secure().randomBytes(1)); + assertEquals(key, objectStore.getVolume(srcVolume) + .getBucket(srcBucket).getKey(key).getName()); // test ls -R /destVol/destBucket, srcBucket with key (non-empty) try (GenericTestUtils.SystemOutCapturer capture = @@ -1546,13 +1538,10 @@ void testSymlinkPosixDelete() throws Exception { // add key to srcBucket final String key = "object-dir/object-name1"; - try (OzoneOutputStream outputStream = objectStore.getVolume(srcVolume) - .getBucket(srcBucket) - .createKey(key, 1)) { - outputStream.write(RandomUtils.nextBytes(1)); - } - assertEquals(objectStore.getVolume(srcVolume) - .getBucket(srcBucket).getKey(key).getName(), key); + TestDataUtil.createKey(objectStore.getVolume(srcVolume).getBucket(srcBucket), + key, RandomUtils.secure().randomBytes(1)); + assertEquals(key, objectStore.getVolume(srcVolume). + getBucket(srcBucket).getKey(key).getName()); // test symlink -rm destVol/destBucket -> srcVol/srcBucket // should delete only link, srcBucket and key unaltered @@ -2231,12 +2220,9 @@ void testCreateAndCheckECFileDiskUsage() throws Exception { Path bucketPathTest = new Path(volPathTest, bucketName); // write some test data into bucket - try (OzoneOutputStream outputStream = objectStore.getVolume(volumeName). - getBucket(bucketName).createKey(key, 1, - new ECReplicationConfig("RS-3-2-1024k"), - new HashMap<>())) { - outputStream.write(RandomUtils.nextBytes(1)); - } + TestDataUtil.createKey(objectStore.getVolume(volumeName). + getBucket(bucketName), key, new ECReplicationConfig("RS-3-2-1024k"), + RandomUtils.secure().randomBytes(1)); // make sure the disk usage matches the expected value Path filePath = new Path(bucketPathTest, key); ContentSummary contentSummary = ofs.getContentSummary(filePath); @@ -2258,13 +2244,10 @@ void testCreateAndCheckRatisFileDiskUsage() throws Exception { Path filePathTest = new Path(bucketPathTest, key); // write some test data into bucket - try (OzoneOutputStream outputStream = objectStore.getVolume(volumeName). - getBucket(bucketName).createKey(key, 1, - RatisReplicationConfig.getInstance( - HddsProtos.ReplicationFactor.THREE), - new HashMap<>())) { - outputStream.write(RandomUtils.nextBytes(1)); - } + TestDataUtil.createKey(objectStore. + getVolume(volumeName).getBucket(bucketName), key, + RatisReplicationConfig.getInstance(HddsProtos.ReplicationFactor.THREE), + RandomUtils.secure().randomBytes(1)); // make sure the disk usage matches the expected value ContentSummary contentSummary = ofs.getContentSummary(filePathTest); long length = contentSummary.getLength(); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestStorageContainerManagerHAWithAllRunning.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestStorageContainerManagerHAWithAllRunning.java index 1447e377b464..a882de3ff697 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestStorageContainerManagerHAWithAllRunning.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestStorageContainerManagerHAWithAllRunning.java @@ -25,9 +25,7 @@ import static org.junit.jupiter.api.Assertions.assertNotNull; import static org.junit.jupiter.api.Assertions.assertTrue; -import java.io.OutputStream; import java.time.Instant; -import java.util.HashMap; import java.util.HashSet; import java.util.List; import java.util.Set; @@ -38,6 +36,7 @@ import org.apache.hadoop.hdds.scm.ha.SCMHAMetrics; import org.apache.hadoop.hdds.scm.ha.SCMRatisServerImpl; import org.apache.hadoop.hdds.scm.server.StorageContainerManager; +import org.apache.hadoop.ozone.TestDataUtil; import org.apache.hadoop.ozone.client.ObjectStore; import org.apache.hadoop.ozone.client.OzoneBucket; import org.apache.hadoop.ozone.client.OzoneClient; @@ -129,9 +128,7 @@ private void doPutKey() throws Exception { byte[] bytes = value.getBytes(UTF_8); RatisReplicationConfig replication = RatisReplicationConfig.getInstance(HddsProtos.ReplicationFactor.ONE); - try (OutputStream out = bucket.createKey(keyName, bytes.length, replication, new HashMap<>())) { - out.write(bytes); - } + TestDataUtil.createKey(bucket, keyName, replication, bytes); OzoneKey key = bucket.getKey(keyName); assertEquals(keyName, key.getName()); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBCSID.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBCSID.java index bccbee395418..e5d772b20d0e 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBCSID.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBCSID.java @@ -27,7 +27,6 @@ import static org.junit.jupiter.api.Assertions.assertEquals; import java.io.IOException; -import java.util.HashMap; import java.util.List; import java.util.concurrent.TimeUnit; import org.apache.hadoop.hdds.client.RatisReplicationConfig; @@ -38,10 +37,10 @@ import org.apache.hadoop.hdds.protocol.proto.HddsProtos; import org.apache.hadoop.hdds.utils.IOUtils; import org.apache.hadoop.ozone.MiniOzoneCluster; +import org.apache.hadoop.ozone.TestDataUtil; import org.apache.hadoop.ozone.client.ObjectStore; import org.apache.hadoop.ozone.client.OzoneClient; import org.apache.hadoop.ozone.client.OzoneClientFactory; -import org.apache.hadoop.ozone.client.io.OzoneOutputStream; import org.apache.hadoop.ozone.om.helpers.OmKeyArgs; import org.apache.hadoop.ozone.om.helpers.OmKeyInfo; import org.apache.hadoop.ozone.om.helpers.OmKeyLocationInfo; @@ -104,13 +103,9 @@ public static void shutdown() { @Test public void testBCSID() throws Exception { - OzoneOutputStream key = - objectStore.getVolume(volumeName).getBucket(bucketName) - .createKey("ratis", 1024, - ReplicationConfig.fromTypeAndFactor(ReplicationType.RATIS, - ReplicationFactor.ONE), new HashMap<>()); - key.write("ratis".getBytes(UTF_8)); - key.close(); + TestDataUtil.createKey(objectStore.getVolume(volumeName).getBucket(bucketName), + "ratis", ReplicationConfig.fromTypeAndFactor(ReplicationType.RATIS, + ReplicationFactor.ONE), "ratis".getBytes(UTF_8)); // get the name of a valid container. OmKeyArgs keyArgs = new OmKeyArgs.Builder().setVolumeName(volumeName). diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestHybridPipelineOnDatanode.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestHybridPipelineOnDatanode.java index 2d7928b476e6..08e8ddba47c8 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestHybridPipelineOnDatanode.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestHybridPipelineOnDatanode.java @@ -26,7 +26,6 @@ import static org.junit.jupiter.api.Assertions.assertSame; import java.io.IOException; -import java.util.HashMap; import java.util.List; import java.util.UUID; import org.apache.hadoop.hdds.client.ReplicationConfig; @@ -39,6 +38,7 @@ import org.apache.hadoop.hdds.scm.pipeline.PipelineID; import org.apache.hadoop.hdds.utils.IOUtils; import org.apache.hadoop.ozone.MiniOzoneCluster; +import org.apache.hadoop.ozone.TestDataUtil; import org.apache.hadoop.ozone.client.ObjectStore; import org.apache.hadoop.ozone.client.OzoneBucket; import org.apache.hadoop.ozone.client.OzoneClient; @@ -47,7 +47,6 @@ import org.apache.hadoop.ozone.client.OzoneKeyDetails; import org.apache.hadoop.ozone.client.OzoneVolume; import org.apache.hadoop.ozone.client.io.OzoneInputStream; -import org.apache.hadoop.ozone.client.io.OzoneOutputStream; import org.junit.jupiter.api.AfterAll; import org.junit.jupiter.api.BeforeAll; import org.junit.jupiter.api.Test; @@ -111,22 +110,16 @@ public void testHybridPipelineOnDatanode() throws IOException { String keyName1 = UUID.randomUUID().toString(); // Write data into a key - OzoneOutputStream out = bucket - .createKey(keyName1, data.length, - ReplicationConfig.fromTypeAndFactor(ReplicationType.RATIS, - ReplicationFactor.ONE), new HashMap<>()); - out.write(value.getBytes(UTF_8)); - out.close(); + TestDataUtil.createKey(bucket, keyName1, + ReplicationConfig.fromTypeAndFactor(ReplicationType.RATIS, + ReplicationFactor.ONE), value.getBytes(UTF_8)); String keyName2 = UUID.randomUUID().toString(); // Write data into a key - out = bucket - .createKey(keyName2, data.length, - ReplicationConfig.fromTypeAndFactor(ReplicationType.RATIS, - ReplicationFactor.THREE), new HashMap<>()); - out.write(value.getBytes(UTF_8)); - out.close(); + TestDataUtil.createKey(bucket, keyName2, + ReplicationConfig.fromTypeAndFactor(ReplicationType.RATIS, + ReplicationFactor.THREE), value.getBytes(UTF_8)); // We need to find the location of the chunk file corresponding to the // data we just wrote. diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneAtRestEncryption.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneAtRestEncryption.java index 35d4dae2d2fa..eda5d100c223 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneAtRestEncryption.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneAtRestEncryption.java @@ -86,6 +86,7 @@ import org.apache.hadoop.ozone.ClientConfigForTesting; import org.apache.hadoop.ozone.MiniOzoneCluster; import org.apache.hadoop.ozone.OzoneConsts; +import org.apache.hadoop.ozone.TestDataUtil; import org.apache.hadoop.ozone.client.BucketArgs; import org.apache.hadoop.ozone.client.ObjectStore; import org.apache.hadoop.ozone.client.OzoneBucket; @@ -298,22 +299,19 @@ static void createAndVerifyKeyData(OzoneBucket bucket) throws Exception { Instant testStartTime = getTestStartTime(); String keyName = UUID.randomUUID().toString(); String value = "sample value"; - try (OzoneOutputStream out = bucket.createKey(keyName, - value.getBytes(StandardCharsets.UTF_8).length, + + TestDataUtil.createKey(bucket, keyName, ReplicationConfig.fromTypeAndFactor(RATIS, ONE), - new HashMap<>())) { - out.write(value.getBytes(StandardCharsets.UTF_8)); - } + value.getBytes(StandardCharsets.UTF_8)); + verifyKeyData(bucket, keyName, value, testStartTime); OzoneKeyDetails key1 = bucket.getKey(keyName); // Overwrite the key - try (OzoneOutputStream out = bucket.createKey(keyName, - value.getBytes(StandardCharsets.UTF_8).length, + TestDataUtil.createKey(bucket, keyName, ReplicationConfig.fromTypeAndFactor(RATIS, ONE), - new HashMap<>())) { - out.write(value.getBytes(StandardCharsets.UTF_8)); - } + value.getBytes(StandardCharsets.UTF_8)); + OzoneKeyDetails key2 = bucket.getKey(keyName); assertNotEquals(key1.getFileEncryptionInfo().toString(), key2.getFileEncryptionInfo().toString()); } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneRpcClientWithKeyLatestVersion.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneRpcClientWithKeyLatestVersion.java index 50ea1c0eee4a..fa6bf2274c0a 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneRpcClientWithKeyLatestVersion.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneRpcClientWithKeyLatestVersion.java @@ -25,8 +25,6 @@ import java.io.IOException; import java.io.InputStream; -import java.io.OutputStream; -import java.util.HashMap; import java.util.List; import java.util.UUID; import org.apache.commons.io.IOUtils; @@ -34,6 +32,7 @@ import org.apache.hadoop.hdds.client.RatisReplicationConfig; import org.apache.hadoop.hdds.client.ReplicationConfig; import org.apache.hadoop.hdds.conf.OzoneConfiguration; +import org.apache.hadoop.ozone.TestDataUtil; import org.apache.hadoop.ozone.client.BucketArgs; import org.apache.hadoop.ozone.client.ObjectStore; import org.apache.hadoop.ozone.client.OzoneBucket; @@ -105,10 +104,7 @@ private void createAndOverwriteKey(OzoneBucket bucket, String key, private static void writeKey(OzoneBucket bucket, String key, byte[] content, ReplicationConfig replication) throws IOException { - try (OutputStream out = bucket.createKey(key, content.length, replication, - new HashMap<>())) { - out.write(content); - } + TestDataUtil.createKey(bucket, key, replication, content); } public static void assertKeyContent(OzoneBucket bucket, String key, diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestReadRetries.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestReadRetries.java index 98d4d2bba615..06cd29529eac 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestReadRetries.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestReadRetries.java @@ -25,8 +25,6 @@ import static org.junit.jupiter.api.Assertions.assertTrue; import java.io.IOException; -import java.io.OutputStream; -import java.util.HashMap; import java.util.List; import java.util.UUID; import org.apache.commons.lang3.RandomUtils; @@ -39,6 +37,7 @@ import org.apache.hadoop.hdds.scm.pipeline.Pipeline; import org.apache.hadoop.hdds.scm.server.StorageContainerManager; import org.apache.hadoop.ozone.MiniOzoneCluster; +import org.apache.hadoop.ozone.TestDataUtil; import org.apache.hadoop.ozone.client.ObjectStore; import org.apache.hadoop.ozone.client.OzoneBucket; import org.apache.hadoop.ozone.client.OzoneClient; @@ -81,10 +80,8 @@ void testPutKeyAndGetKeyThreeNodes() throws Exception { String keyName = "a/b/c/" + UUID.randomUUID(); byte[] content = RandomUtils.nextBytes(128); - try (OutputStream out = bucket.createKey(keyName, content.length, - RatisReplicationConfig.getInstance(THREE), new HashMap<>())) { - out.write(content); - } + TestDataUtil.createKey(bucket, keyName, + RatisReplicationConfig.getInstance(THREE), content); // First, confirm the key info from the client matches the info in OM. OmKeyArgs keyArgs = new OmKeyArgs.Builder() diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestSecureOzoneRpcClient.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestSecureOzoneRpcClient.java index 0ce1457bdf67..6645687e8d1c 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestSecureOzoneRpcClient.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestSecureOzoneRpcClient.java @@ -59,6 +59,7 @@ import org.apache.hadoop.ozone.ClientVersion; import org.apache.hadoop.ozone.MiniOzoneCluster; import org.apache.hadoop.ozone.OzoneConfigKeys; +import org.apache.hadoop.ozone.TestDataUtil; import org.apache.hadoop.ozone.client.BucketArgs; import org.apache.hadoop.ozone.client.OzoneBucket; import org.apache.hadoop.ozone.client.OzoneKey; @@ -159,10 +160,7 @@ private void testPutKeySuccessWithBlockTokenWithBucketLayout( String keyName = UUID.randomUUID().toString(); long committedBytes = ozoneManager.getMetrics().getDataCommittedBytes(); - try (OzoneOutputStream out = bucket.createKey(keyName, - value.getBytes(UTF_8).length, replication, new HashMap<>())) { - out.write(value.getBytes(UTF_8)); - } + TestDataUtil.createKey(bucket, keyName, replication, value.getBytes(UTF_8)); assertEquals(committedBytes + value.getBytes(UTF_8).length, ozoneManager.getMetrics().getDataCommittedBytes()); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReplication.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReplication.java index 9f3f67324b67..635993cc8732 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReplication.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReplication.java @@ -19,7 +19,6 @@ import static java.nio.charset.StandardCharsets.UTF_8; import static java.util.Arrays.asList; -import static java.util.Collections.emptyMap; import static org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos.ContainerType.KeyValueContainer; import static org.apache.hadoop.hdds.protocol.proto.HddsProtos.ReplicationFactor.THREE; import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_CONTAINER_PLACEMENT_EC_IMPL_KEY; @@ -40,11 +39,9 @@ import com.google.common.base.Functions; import com.google.common.collect.ImmutableMap; import java.io.IOException; -import java.io.OutputStream; import java.time.Duration; import java.util.ArrayList; import java.util.Arrays; -import java.util.HashMap; import java.util.LinkedList; import java.util.List; import java.util.Map; @@ -69,6 +66,7 @@ import org.apache.hadoop.hdds.scm.storage.ContainerProtocolCalls; import org.apache.hadoop.ozone.HddsDatanodeService; import org.apache.hadoop.ozone.MiniOzoneCluster; +import org.apache.hadoop.ozone.TestDataUtil; import org.apache.hadoop.ozone.client.ObjectStore; import org.apache.hadoop.ozone.client.OzoneBucket; import org.apache.hadoop.ozone.client.OzoneClient; @@ -193,10 +191,9 @@ private void createTestData(OzoneClient client) throws IOException { OzoneBucket bucket = volume.getBucket(BUCKET); - try (OutputStream out = bucket.createKey(KEY, 0, - RatisReplicationConfig.getInstance(THREE), emptyMap())) { - out.write("Hello".getBytes(UTF_8)); - } + TestDataUtil.createKey(bucket, KEY, + RatisReplicationConfig.getInstance(THREE), + "Hello".getBytes(UTF_8)); } private byte[] createTestData(OzoneClient client, int size) throws IOException { @@ -205,13 +202,12 @@ private byte[] createTestData(OzoneClient client, int size) throws IOException { OzoneVolume volume = objectStore.getVolume(VOLUME); volume.createBucket(BUCKET); OzoneBucket bucket = volume.getBucket(BUCKET); - try (OutputStream out = bucket.createKey(KEY, 0, new ECReplicationConfig("RS-3-2-1k"), - new HashMap<>())) { - byte[] b = new byte[size]; - b = RandomUtils.secure().randomBytes(b.length); - out.write(b); - return b; - } + + byte[] b = new byte[size]; + b = RandomUtils.secure().randomBytes(b.length); + TestDataUtil.createKey(bucket, KEY, + new ECReplicationConfig("RS-3-2-1k"), b); + return b; } private static List lookupKey(MiniOzoneCluster cluster) diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReportHandling.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReportHandling.java index 1aed5b76d3d4..5327fc78bf03 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReportHandling.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReportHandling.java @@ -18,7 +18,6 @@ package org.apache.hadoop.ozone.container; import static java.nio.charset.StandardCharsets.UTF_8; -import static java.util.Collections.emptyMap; import static org.apache.hadoop.hdds.protocol.proto.HddsProtos.ReplicationFactor.THREE; import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_DEADNODE_INTERVAL; import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_STALENODE_INTERVAL; @@ -30,7 +29,6 @@ import static org.junit.jupiter.api.Assertions.assertTrue; import java.io.IOException; -import java.io.OutputStream; import java.nio.file.Path; import java.nio.file.Paths; import java.util.List; @@ -43,6 +41,7 @@ import org.apache.hadoop.hdds.scm.container.ContainerManager; import org.apache.hadoop.ozone.HddsDatanodeService; import org.apache.hadoop.ozone.MiniOzoneCluster; +import org.apache.hadoop.ozone.TestDataUtil; import org.apache.hadoop.ozone.client.ObjectStore; import org.apache.hadoop.ozone.client.OzoneBucket; import org.apache.hadoop.ozone.client.OzoneClient; @@ -151,10 +150,8 @@ private void createTestData(OzoneClient client) throws IOException { OzoneBucket bucket = volume.getBucket(BUCKET); - try (OutputStream out = bucket.createKey(KEY, 0, - RatisReplicationConfig.getInstance(THREE), emptyMap())) { - out.write("Hello".getBytes(UTF_8)); - } + TestDataUtil.createKey(bucket, KEY, + RatisReplicationConfig.getInstance(THREE), "Hello".getBytes(UTF_8)); } } diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReportHandlingWithHA.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReportHandlingWithHA.java index 357945a3fa5f..d9337ef716b1 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReportHandlingWithHA.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/TestContainerReportHandlingWithHA.java @@ -18,7 +18,6 @@ package org.apache.hadoop.ozone.container; import static java.nio.charset.StandardCharsets.UTF_8; -import static java.util.Collections.emptyMap; import static org.apache.hadoop.hdds.protocol.proto.HddsProtos.ReplicationFactor.THREE; import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_DEADNODE_INTERVAL; import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_STALENODE_INTERVAL; @@ -30,7 +29,6 @@ import static org.junit.jupiter.api.Assertions.assertTrue; import java.io.IOException; -import java.io.OutputStream; import java.nio.file.Path; import java.nio.file.Paths; import java.util.List; @@ -46,6 +44,7 @@ import org.apache.hadoop.ozone.HddsDatanodeService; import org.apache.hadoop.ozone.MiniOzoneCluster; import org.apache.hadoop.ozone.MiniOzoneHAClusterImpl; +import org.apache.hadoop.ozone.TestDataUtil; import org.apache.hadoop.ozone.client.ObjectStore; import org.apache.hadoop.ozone.client.OzoneBucket; import org.apache.hadoop.ozone.client.OzoneClient; @@ -156,10 +155,8 @@ private void createTestData(OzoneClient client) throws IOException { OzoneBucket bucket = volume.getBucket(BUCKET); - try (OutputStream out = bucket.createKey(KEY, 0, - RatisReplicationConfig.getInstance(THREE), emptyMap())) { - out.write("Hello".getBytes(UTF_8)); - } + TestDataUtil.createKey(bucket, KEY, + RatisReplicationConfig.getInstance(THREE), "Hello".getBytes(UTF_8)); } private static void waitForContainerStateInAllSCMs(MiniOzoneHAClusterImpl cluster, ContainerID containerID, diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/dn/volume/TestDatanodeHddsVolumeFailureDetection.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/dn/volume/TestDatanodeHddsVolumeFailureDetection.java index fdea9054ce7a..7d3eaffaeced 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/dn/volume/TestDatanodeHddsVolumeFailureDetection.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/dn/volume/TestDatanodeHddsVolumeFailureDetection.java @@ -17,7 +17,6 @@ package org.apache.hadoop.ozone.dn.volume; -import static java.util.Collections.emptyMap; import static org.apache.commons.io.IOUtils.readFully; import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_DATANODE_RATIS_VOLUME_FREE_SPACE_MIN; import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_CONTAINER_SIZE; @@ -31,7 +30,6 @@ import java.io.File; import java.io.IOException; import java.io.InputStream; -import java.io.OutputStream; import java.nio.file.Path; import java.nio.file.Paths; import java.time.Duration; @@ -282,10 +280,7 @@ private static long createKey(OzoneBucket bucket, String key) byte[] bytes = RandomUtils.nextBytes(KEY_SIZE); RatisReplicationConfig replication = RatisReplicationConfig.getInstance(ReplicationFactor.ONE); - try (OutputStream out = bucket.createKey(key, bytes.length, replication, - emptyMap())) { - out.write(bytes); - } + TestDataUtil.createKey(bucket, key, replication, bytes); OzoneKeyDetails keyDetails = bucket.getKey(key); assertEquals(key, keyDetails.getName()); return keyDetails.getOzoneKeyLocations().get(0).getContainerID(); diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestBucket.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestBucket.java index 57fe728a7322..b042d0be7ec7 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestBucket.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestBucket.java @@ -18,17 +18,16 @@ package org.apache.hadoop.ozone.om; import static java.nio.charset.StandardCharsets.UTF_8; -import static java.util.Collections.emptyMap; import static org.apache.commons.lang3.RandomStringUtils.randomNumeric; import static org.apache.hadoop.hdds.protocol.proto.HddsProtos.ReplicationFactor.THREE; import static org.junit.jupiter.api.Assertions.assertArrayEquals; import java.io.IOException; -import java.io.OutputStream; import java.util.UUID; import java.util.concurrent.ThreadLocalRandom; import org.apache.hadoop.hdds.client.RatisReplicationConfig; import org.apache.hadoop.hdds.client.ReplicationConfig; +import org.apache.hadoop.ozone.TestDataUtil; import org.apache.hadoop.ozone.client.ObjectStore; import org.apache.hadoop.ozone.client.OzoneBucket; import org.apache.hadoop.ozone.client.OzoneClient; @@ -79,9 +78,7 @@ public byte[] writeKey(String key, ReplicationConfig repConfig, int len) public void writeKey(String key, ReplicationConfig repConfig, byte[] inputData) throws IOException { - try (OutputStream out = bucket.createKey(key, 0, repConfig, emptyMap())) { - out.write(inputData); - } + TestDataUtil.createKey(bucket, key, repConfig, inputData); } public byte[] writeRandomBytes(String keyName, int dataLength) diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/recon/TestReconAndAdminContainerCLI.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/recon/TestReconAndAdminContainerCLI.java index 738bdef6f22a..3f7ea803775e 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/recon/TestReconAndAdminContainerCLI.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/recon/TestReconAndAdminContainerCLI.java @@ -18,7 +18,6 @@ package org.apache.hadoop.ozone.recon; import static java.nio.charset.StandardCharsets.UTF_8; -import static java.util.Collections.emptyMap; import static java.util.concurrent.TimeUnit.SECONDS; import static org.apache.hadoop.hdds.HddsConfigKeys.HDDS_COMMAND_STATUS_REPORT_INTERVAL; import static org.apache.hadoop.hdds.HddsConfigKeys.HDDS_CONTAINER_REPORT_INTERVAL; @@ -38,7 +37,6 @@ import static org.junit.jupiter.api.Assertions.assertNotNull; import java.io.IOException; -import java.io.OutputStream; import java.time.Duration; import java.util.ArrayList; import java.util.Arrays; @@ -419,10 +417,7 @@ private static OmKeyInfo createTestKey(String keyName, ReplicationConfig replicationConfig) throws IOException { byte[] textBytes = "Testing".getBytes(UTF_8); - try (OutputStream out = ozoneBucket.createKey(keyName, - textBytes.length, replicationConfig, emptyMap())) { - out.write(textBytes); - } + TestDataUtil.createKey(ozoneBucket, keyName, replicationConfig, textBytes); OmKeyArgs keyArgs = new OmKeyArgs.Builder() .setVolumeName(ozoneBucket.getVolumeName())