From 356d5ccc8ad1e37b047d69702c758010a6d7c9d1 Mon Sep 17 00:00:00 2001 From: Chia-Chuan Yu Date: Thu, 20 Feb 2025 22:29:00 +0800 Subject: [PATCH 1/2] HDDS-12387. Cleanup TestContainerOperations --- .../hdds/scm/TestContainerOperations.java | 17 +++++++---------- 1 file changed, 7 insertions(+), 10 deletions(-) diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestContainerOperations.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestContainerOperations.java index 6d462de6a065..f1784b0082b1 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestContainerOperations.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestContainerOperations.java @@ -70,7 +70,6 @@ public abstract class TestContainerOperations implements NonHATests.TestCase { private static final int CONTAINER_LIST_LIMIT = 1; private ScmClient storageClient; - private StorageContainerLocationProtocolClientSideTranslatorPB storageContainerLocationClient; private XceiverClientManager xceiverClientManager; @BeforeAll @@ -78,25 +77,23 @@ void setup() throws Exception { OzoneConfiguration clientConf = new OzoneConfiguration(cluster().getConf()); clientConf.setInt(ScmConfigKeys.OZONE_SCM_CONTAINER_LIST_MAX_COUNT, CONTAINER_LIST_LIMIT); storageClient = new ContainerOperationClient(clientConf); - storageContainerLocationClient = cluster().getStorageContainerLocationClient(); xceiverClientManager = new XceiverClientManager(cluster().getConf()); } @AfterAll void cleanup() { - IOUtils.closeQuietly(storageContainerLocationClient, xceiverClientManager); + IOUtils.closeQuietly(xceiverClientManager); } @Test void testContainerStateMachineIdempotency() throws Exception { - ContainerWithPipeline container = storageContainerLocationClient - .allocateContainer(HddsProtos.ReplicationType.RATIS, - HddsProtos.ReplicationFactor.ONE, OzoneConsts.OZONE); + ContainerWithPipeline container = storageClient.createContainer(HddsProtos + .ReplicationType.RATIS, HddsProtos.ReplicationFactor + .ONE, OzoneConsts.OZONE); long containerID = container.getContainerInfo().getContainerID(); Pipeline pipeline = container.getPipeline(); XceiverClientSpi client = xceiverClientManager.acquireClient(pipeline); - //create the container - ContainerProtocolCalls.createContainer(client, containerID, null); + // call create Container again BlockID blockID = ContainerTestHelper.getTestBlockID(containerID); byte[] data = @@ -227,7 +224,7 @@ public void testNodeOperationalStates() throws Exception { DatanodeDetails node = nm.getAllNodes().get(0); nm.setNodeOperationalState(node, DECOMMISSIONING); - // All nodes should be returned as they are all in service + // Nodes not in DECOMMISSIONING state should be returned as they are in service int nodeCount = storageClient.queryNode(IN_SERVICE, HEALTHY, HddsProtos.QueryScope.CLUSTER, "").size(); assertEquals(numOfDatanodes - 1, nodeCount); @@ -263,7 +260,7 @@ public void testNodeOperationalStates() throws Exception { nodeCount = storageClient.queryNode(s, HEALTHY, HddsProtos.QueryScope.CLUSTER, "").size(); if (s == IN_SERVICE) { - assertEquals(5, nodeCount); + assertEquals(cluster().getHddsDatanodes().size(), nodeCount); } else { assertEquals(1, nodeCount); } From 110c68d52e56f39341a101925c0db30c0f12109b Mon Sep 17 00:00:00 2001 From: Chia-Chuan Yu Date: Thu, 20 Feb 2025 22:31:05 +0800 Subject: [PATCH 2/2] Removed imports --- .../java/org/apache/hadoop/hdds/scm/TestContainerOperations.java | 1 - 1 file changed, 1 deletion(-) diff --git a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestContainerOperations.java b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestContainerOperations.java index f1784b0082b1..68235657a18a 100644 --- a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestContainerOperations.java +++ b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestContainerOperations.java @@ -46,7 +46,6 @@ import org.apache.hadoop.hdds.scm.pipeline.Pipeline; import org.apache.hadoop.hdds.scm.pipeline.PipelineID; import org.apache.hadoop.hdds.scm.pipeline.PipelineNotFoundException; -import org.apache.hadoop.hdds.scm.protocolPB.StorageContainerLocationProtocolClientSideTranslatorPB; import org.apache.hadoop.hdds.scm.server.StorageContainerManager; import org.apache.hadoop.hdds.scm.storage.ContainerProtocolCalls; import org.apache.hadoop.hdds.utils.IOUtils;