Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -327,12 +327,6 @@ public final class OzoneConfigKeys {
"ozone.key.preallocation.max.blocks";
public static final int OZONE_KEY_PREALLOCATION_BLOCKS_MAX_DEFAULT
= 64;

public static final String OZONE_BLOCK_DELETING_LIMIT_PER_CONTAINER =
"ozone.block.deleting.limit.per.task";
public static final int OZONE_BLOCK_DELETING_LIMIT_PER_CONTAINER_DEFAULT
= 1000;

public static final String HDDS_CONTAINER_RATIS_ENABLED_KEY
= ScmConfigKeys.HDDS_CONTAINER_RATIS_ENABLED_KEY;
public static final boolean HDDS_CONTAINER_RATIS_ENABLED_DEFAULT
Expand Down
13 changes: 1 addition & 12 deletions hadoop-hdds/common/src/main/resources/ozone-default.xml
Original file line number Diff line number Diff line change
Expand Up @@ -426,16 +426,6 @@
or be directly or indirectly in a group defined in this property.
</description>
</property>
<property>
<name>ozone.block.deleting.limit.per.task</name>
<value>1000</value>
<tag>OZONE, PERFORMANCE, SCM</tag>
<description>A maximum number of blocks to be deleted by block deleting
service per
time interval. This property is used to throttle the actual number of
block deletions on a data node per container.
</description>
</property>
<property>
<name>ozone.block.deleting.service.interval</name>
<value>1m</value>
Expand All @@ -453,8 +443,7 @@
<description>A timeout value of block deletion service. If this is set
greater than 0,
the service will stop waiting for the block deleting completion after this
time. If timeout happens to a large proportion of block deletion, this
needs to be increased with ozone.block.deleting.limit.per.task. This
time. This
setting supports multiple time unit suffixes as described in
dfs.heartbeat.interval. If no suffix is specified, then milliseconds is
assumed.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -95,7 +95,6 @@

import static java.nio.charset.StandardCharsets.UTF_8;
import static org.apache.commons.lang3.RandomStringUtils.randomAlphanumeric;
import static org.apache.hadoop.ozone.OzoneConfigKeys.OZONE_BLOCK_DELETING_LIMIT_PER_CONTAINER;
import static org.apache.hadoop.ozone.OzoneConfigKeys.OZONE_BLOCK_DELETING_SERVICE_INTERVAL;
import static org.apache.hadoop.ozone.OzoneConsts.SCHEMA_V1;
import static org.apache.hadoop.ozone.OzoneConsts.SCHEMA_V2;
Expand Down Expand Up @@ -769,7 +768,6 @@ public void testShutdownService(ContainerTestVersionInfo versionInfo)
setLayoutAndSchemaForTest(versionInfo);
conf.setTimeDuration(OZONE_BLOCK_DELETING_SERVICE_INTERVAL, 500,
TimeUnit.MILLISECONDS);
conf.setInt(OZONE_BLOCK_DELETING_LIMIT_PER_CONTAINER, 10);

ContainerSet containerSet = new ContainerSet(1000);
// Create 1 container with 100 blocks
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,6 @@
import org.apache.hadoop.hdds.scm.ScmConfigKeys;
import org.apache.hadoop.hdds.utils.MetadataKeyFilters.KeyPrefixFilter;
import org.apache.hadoop.hdds.utils.db.Table;
import org.apache.hadoop.ozone.OzoneConfigKeys;
import org.apache.hadoop.ozone.OzoneConsts;
import org.apache.hadoop.ozone.container.common.helpers.BlockData;
import org.apache.hadoop.ozone.container.common.helpers.ChunkInfoList;
Expand Down Expand Up @@ -554,8 +553,6 @@ public void testReadDeletedBlocks(String schemaVersion) throws Exception {

private void runBlockDeletingService(KeyValueHandler keyValueHandler)
throws Exception {
conf.setInt(OzoneConfigKeys.OZONE_BLOCK_DELETING_LIMIT_PER_CONTAINER, 2);

OzoneContainer container = makeMockOzoneContainer(keyValueHandler);

BlockDeletingServiceTestImpl service =
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,6 @@
import org.apache.hadoop.hdds.scm.container.common.helpers.StorageContainerException;
import org.apache.hadoop.hdds.utils.db.BatchOperation;
import org.apache.hadoop.hdds.utils.db.Table;
import org.apache.hadoop.ozone.OzoneConfigKeys;
import org.apache.hadoop.ozone.OzoneConsts;
import org.apache.hadoop.ozone.common.ChunkBuffer;
import org.apache.hadoop.ozone.container.ContainerTestHelper;
Expand Down Expand Up @@ -223,9 +222,6 @@ public void testReadMetadata() throws IOException {
@Test
public void testDeleteViaTransation() throws IOException, TimeoutException,
InterruptedException {
conf.setInt(OzoneConfigKeys.OZONE_BLOCK_DELETING_LIMIT_PER_CONTAINER,
BLOCKS_PER_CONTAINER);

// create a container of schema v2
KeyValueContainer container = createTestContainer();
assertEquals(container.getContainerData().getSchemaVersion(),
Expand Down
Loading