Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -27,10 +27,11 @@
import java.util.LinkedList;
import java.util.List;
import java.util.Map;
import java.util.concurrent.locks.ReentrantLock;
import java.util.concurrent.locks.Lock;
import java.util.function.Consumer;
import java.util.function.Function;

import com.google.common.util.concurrent.Striped;
import org.apache.hadoop.hdds.client.BlockID;
import org.apache.hadoop.hdds.conf.ConfigurationSource;
import org.apache.hadoop.hdds.conf.StorageUnit;
Expand All @@ -48,6 +49,7 @@
import org.apache.hadoop.hdds.scm.ByteStringConversion;
import org.apache.hadoop.hdds.scm.ScmConfigKeys;
import org.apache.hadoop.hdds.scm.container.common.helpers.StorageContainerException;
import org.apache.hadoop.hdds.utils.HddsServerUtil;
import org.apache.hadoop.ozone.OzoneConfigKeys;
import org.apache.hadoop.ozone.common.Checksum;
import org.apache.hadoop.ozone.common.ChunkBuffer;
Expand Down Expand Up @@ -77,7 +79,6 @@
import org.apache.hadoop.ozone.container.keyvalue.interfaces.BlockManager;
import org.apache.hadoop.ozone.container.keyvalue.interfaces.ChunkManager;
import org.apache.hadoop.ozone.container.upgrade.VersionedDatanodeFeatures;
import org.apache.hadoop.util.AutoCloseableLock;

import com.google.common.annotations.VisibleForTesting;
import com.google.common.base.Preconditions;
Expand Down Expand Up @@ -123,8 +124,8 @@ public class KeyValueHandler extends Handler {
private final Function<ByteBuffer, ByteString> byteBufferToByteString;
private final boolean validateChunkChecksumData;

// A lock that is held during container creation.
private final AutoCloseableLock containerCreationLock;
// A striped lock that is held during container creation.
private final Striped<Lock> containerCreationLocks;

public KeyValueHandler(ConfigurationSource config, String datanodeId,
ContainerSet contSet, VolumeSet volSet, ContainerMetrics metrics,
Expand All @@ -146,9 +147,17 @@ public KeyValueHandler(ConfigurationSource config, String datanodeId,
maxContainerSize = (long) config.getStorageSize(
ScmConfigKeys.OZONE_SCM_CONTAINER_SIZE,
ScmConfigKeys.OZONE_SCM_CONTAINER_SIZE_DEFAULT, StorageUnit.BYTES);
// this handler lock is used for synchronizing createContainer Requests,
// so using a fair lock here.
containerCreationLock = new AutoCloseableLock(new ReentrantLock(true));
// this striped handler lock is used for synchronizing createContainer
// Requests.
final int threadCountPerDisk = conf.getInt(
OzoneConfigKeys
.DFS_CONTAINER_RATIS_NUM_WRITE_CHUNK_THREADS_PER_VOLUME_KEY,
OzoneConfigKeys
.DFS_CONTAINER_RATIS_NUM_WRITE_CHUNK_THREADS_PER_VOLUME_DEFAULT);
final int numberOfDisks =
HddsServerUtil.getDatanodeStorageDirs(conf).size();
containerCreationLocks = Striped.lazyWeakLock(
threadCountPerDisk * numberOfDisks);

boolean isUnsafeByteBufferConversionEnabled =
conf.getBoolean(
Expand Down Expand Up @@ -268,7 +277,9 @@ ContainerCommandResponseProto handleCreateContainer(
newContainerData, conf);

boolean created = false;
try (AutoCloseableLock l = containerCreationLock.acquire()) {
Lock containerIdLock = containerCreationLocks.get(containerID);
containerIdLock.lock();
try {
if (containerSet.getContainer(containerID) == null) {
newContainer.create(volumeSet, volumeChoosingPolicy, clusterId);
created = containerSet.addContainer(newContainer);
Expand All @@ -280,6 +291,8 @@ ContainerCommandResponseProto handleCreateContainer(
}
} catch (StorageContainerException ex) {
return ContainerUtils.logAndReturnError(LOG, ex, request);
} finally {
containerIdLock.unlock();
}

if (created) {
Expand Down