Skip to content
Merged
2 changes: 2 additions & 0 deletions hadoop-ozone/dist/src/main/compose/common/ec-test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -30,3 +30,5 @@ docker-compose up -d --no-recreate --scale datanode=3
execute_robot_test scm -v PREFIX:${prefix} -N read-3-datanodes ec/read.robot
docker-compose up -d --no-recreate --scale datanode=5
execute_robot_test scm -v container:1 -v count:5 -N EC-recovery replication/wait.robot
docker-compose up -d --no-recreate --scale datanode=9
execute_robot_test scm -N S3-EC-Storage ec/awss3ecstorage.robot
90 changes: 90 additions & 0 deletions hadoop-ozone/dist/src/main/smoketest/ec/awss3ecstorage.robot
Original file line number Diff line number Diff line change
@@ -0,0 +1,90 @@
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

*** Settings ***
Documentation S3 gateway test with aws cli with STANDARD_IA storage class
Library OperatingSystem
Library String
Resource ../commonlib.robot
Resource ../s3/commonawslib.robot
Resource ../s3/mpu_lib.robot
Resource ../ozone-lib/shell.robot
Test Timeout 5 minutes
Suite Setup Setup EC Multipart Tests
Suite Teardown Teardown EC Multipart Tests
Test Setup Generate random prefix

*** Keywords ***
Setup EC Multipart Tests
Setup s3 tests
Create Random File KB 1023 /tmp/1mb

Teardown EC Multipart Tests
Remove Files /tmp/1mb

*** Variables ***
${ENDPOINT_URL} http://s3g:9878
${BUCKET} generated

*** Test Cases ***

Put Object with STANDARD_IA storage class
${file_checksum} = Execute md5sum /tmp/1mb | awk '{print $1}'

${result} = Execute AWSS3ApiCli put-object --bucket ${BUCKET} --key ${PREFIX}/ecKey32 --body /tmp/1mb --storage-class STANDARD_IA
${eTag} = Execute echo '${result}' | jq -r '.ETag'
Should Be Equal ${eTag} \"${file_checksum}\"
Verify Key EC Replication Config /s3v/${BUCKET}/${PREFIX}/ecKey32 RS 3 2 1048576

${result} = Execute AWSS3ApiCli put-object --bucket ${BUCKET} --key ${PREFIX}/ecKey63 --body /tmp/1mb --storage-class STANDARD_IA --metadata="storage-config=rs-6-3-1024k"
${eTag} = Execute echo '${result}' | jq -r '.ETag'
Should Be Equal ${eTag} \"${file_checksum}\"
Verify Key EC Replication Config /s3v/${BUCKET}/${PREFIX}/ecKey63 RS 6 3 1048576

Test multipart upload with STANDARD_IA storage
${uploadID} = Initiate MPU ${BUCKET} ${PREFIX}/ecmultipartKey32 0 --storage-class STANDARD_IA
${eTag1} = Upload MPU part ${BUCKET} ${PREFIX}/ecmultipartKey32 ${uploadID} 1 /tmp/1mb
${result} = Execute AWSS3APICli list-parts --bucket ${BUCKET} --key ${PREFIX}/ecmultipartKey32 --upload-id ${uploadID}
${part1} = Execute echo '${result}' | jq -r '.Parts[0].ETag'
Should Be equal ${part1} ${eTag1}
Should contain ${result} STANDARD_IA
Complete MPU ${BUCKET} ${PREFIX}/ecmultipartKey32 ${uploadID} {ETag=${eTag1},PartNumber=1}
Verify Key EC Replication Config /s3v/${BUCKET}/${PREFIX}/ecmultipartKey32 RS 3 2 1048576

${uploadID} = Initiate MPU ${BUCKET} ${PREFIX}/ecmultipartKey63 0 --storage-class STANDARD_IA --metadata="storage-config=rs-6-3-1024k"
${eTag1} = Upload MPU part ${BUCKET} ${PREFIX}/ecmultipartKey63 ${uploadID} 1 /tmp/part1
${result} = Execute AWSS3APICli list-parts --bucket ${BUCKET} --key ${PREFIX}/ecmultipartKey63 --upload-id ${uploadID}
${part1} = Execute echo '${result}' | jq -r '.Parts[0].ETag'
Should Be equal ${part1} ${eTag1}
Should contain ${result} STANDARD_IA
Complete MPU ${BUCKET} ${PREFIX}/ecmultipartKey63 ${uploadID} {ETag=${eTag1},PartNumber=1}
Verify Key EC Replication Config /s3v/${BUCKET}/${PREFIX}/ecmultipartKey63 RS 6 3 1048576

Copy Object change storage class to STANDARD_IA
${file_checksum} = Execute md5sum /tmp/1mb | awk '{print $1}'
${result} = Execute AWSS3ApiCli put-object --bucket ${BUCKET} --key ${PREFIX}/copyobject/Key1 --body /tmp/1mb
${eTag} = Execute echo '${result}' | jq -r '.ETag'
Should Be Equal ${eTag} \"${file_checksum}\"

${result} = Execute AWSS3APICli copy-object --storage-class STANDARD_IA --bucket ${BUCKET} --key ${PREFIX}/copyobject/Key1 --copy-source ${BUCKET}/${PREFIX}/copyobject/Key1
Should contain ${result} ETag
${eTag} = Execute echo '${result}' | jq -r '.CopyObjectResult.ETag'
Should Be Equal ${eTag} \"${file_checksum}\"

${result} = Execute AWSS3APICli copy-object --storage-class STANDARD_IA --metadata="storage-config=rs-6-3-1024k" --bucket ${BUCKET} --key ${PREFIX}/copyobject/Key1 --copy-source ${BUCKET}/${PREFIX}/copyobject/Key1
Should contain ${result} ETag
${eTag} = Execute echo '${result}' | jq -r '.CopyObjectResult.ETag'
Should Be Equal ${eTag} \"${file_checksum}\"
## TODO: Verify Key EC Replication Config when we support changing storage class
4 changes: 4 additions & 0 deletions hadoop-ozone/dist/src/main/smoketest/s3/awss3.robot
Original file line number Diff line number Diff line change
Expand Up @@ -46,6 +46,10 @@ File upload and directory list
Should not contain ${result} testfile
Should not contain ${result} dir1
Should contain ${result} file
# Verify S3 storage class if file is replicated or erasure coded.
${result} = Execute AWSS3CliDebug ls s3://${BUCKET}/dir1/dir2/file
Run Keyword If '${BUCKET}' == 'generated' Should contain ${result} STANDARD
Run Keyword If '${BUCKET}' == 'erasure' Should contain ${result} STANDARD_IA

File upload with special chars
Execute date > /tmp/testfile
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -53,7 +53,6 @@
import javax.ws.rs.core.MediaType;
import javax.ws.rs.core.Response;
import org.apache.commons.lang3.StringUtils;
import org.apache.hadoop.hdds.client.ReplicationType;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.hadoop.ozone.OzoneAcl;
import org.apache.hadoop.ozone.audit.S3GAction;
Expand Down Expand Up @@ -738,12 +737,8 @@ private void addKey(ListObjectResponse response, OzoneKey next) {
if (eTag != null) {
keyMetadata.setETag(ObjectEndpoint.wrapInQuotes(eTag));
}
if (next.getReplicationType().toString().equals(ReplicationType
.STAND_ALONE.toString())) {
keyMetadata.setStorageClass(S3StorageType.REDUCED_REDUNDANCY.toString());
} else {
keyMetadata.setStorageClass(S3StorageType.STANDARD.toString());
}
keyMetadata.setStorageClass(S3StorageType.fromReplicationConfig(
next.getReplicationConfig()).toString());
keyMetadata.setLastModified(next.getModificationTime());
String ownerName = next.getOwner();
String displayName = ownerName;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,7 @@
import static org.apache.hadoop.ozone.s3.exception.S3ErrorTable.newError;
import static org.apache.hadoop.ozone.s3.util.S3Consts.AWS_TAG_PREFIX;
import static org.apache.hadoop.ozone.s3.util.S3Consts.CUSTOM_METADATA_HEADER_PREFIX;
import static org.apache.hadoop.ozone.s3.util.S3Consts.STORAGE_CONFIG_HEADER;
import static org.apache.hadoop.ozone.s3.util.S3Consts.TAG_HEADER;
import static org.apache.hadoop.ozone.s3.util.S3Consts.TAG_KEY_LENGTH_LIMIT;
import static org.apache.hadoop.ozone.s3.util.S3Consts.TAG_NUM_LIMIT;
Expand Down Expand Up @@ -97,7 +98,7 @@ public abstract class EndpointBase implements Auditor {
private ContainerRequestContext context;

private Set<String> excludeMetadataFields =
new HashSet<>(Arrays.asList(OzoneConsts.GDPR_FLAG));
new HashSet<>(Arrays.asList(OzoneConsts.GDPR_FLAG, STORAGE_CONFIG_HEADER));
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

what does this mean

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We store any x-amz-meta-key=value for an object in KeyArgs.metadata (or ex., Etag). ^^This will exclude storing EC storage config in the metadata because it will be available in replication config. We only want to use this header to determine EC replication config when client uses STANDARD_IA storage class.

private static final Logger LOG =
LoggerFactory.getLogger(EndpointBase.class);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -49,12 +49,14 @@
import static org.apache.hadoop.ozone.s3.util.S3Consts.COPY_SOURCE_IF_MODIFIED_SINCE;
import static org.apache.hadoop.ozone.s3.util.S3Consts.COPY_SOURCE_IF_UNMODIFIED_SINCE;
import static org.apache.hadoop.ozone.s3.util.S3Consts.CUSTOM_METADATA_COPY_DIRECTIVE_HEADER;
import static org.apache.hadoop.ozone.s3.util.S3Consts.CUSTOM_METADATA_HEADER_PREFIX;
import static org.apache.hadoop.ozone.s3.util.S3Consts.CopyDirective;
import static org.apache.hadoop.ozone.s3.util.S3Consts.DECODED_CONTENT_LENGTH_HEADER;
import static org.apache.hadoop.ozone.s3.util.S3Consts.MP_PARTS_COUNT;
import static org.apache.hadoop.ozone.s3.util.S3Consts.RANGE_HEADER;
import static org.apache.hadoop.ozone.s3.util.S3Consts.RANGE_HEADER_SUPPORTED_UNIT;
import static org.apache.hadoop.ozone.s3.util.S3Consts.STORAGE_CLASS_HEADER;
import static org.apache.hadoop.ozone.s3.util.S3Consts.STORAGE_CONFIG_HEADER;
import static org.apache.hadoop.ozone.s3.util.S3Consts.TAG_COUNT_HEADER;
import static org.apache.hadoop.ozone.s3.util.S3Consts.TAG_DIRECTIVE_HEADER;
import static org.apache.hadoop.ozone.s3.util.S3Utils.urlDecode;
Expand Down Expand Up @@ -230,7 +232,7 @@ public Response put(
boolean auditSuccess = true;
PerformanceStringBuilder perf = new PerformanceStringBuilder();

String copyHeader = null, storageType = null;
String copyHeader = null, storageType = null, storageConfig = null;
DigestInputStream digestInputStream = null;
try {
if (aclMarker != null) {
Expand All @@ -256,12 +258,13 @@ public Response put(

copyHeader = headers.getHeaderString(COPY_SOURCE_HEADER);
storageType = headers.getHeaderString(STORAGE_CLASS_HEADER);
storageConfig = headers.getHeaderString(CUSTOM_METADATA_HEADER_PREFIX + STORAGE_CONFIG_HEADER);
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

curious: how will a client add the header to the request? is it supported by s3 cli or hadoop aws cloud connector?

i wonder where we add this information in the developer doc.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

All s3 clients should support it. They would just need to add the "x-amz-metadata-storage-class=" to their requests.
This is demonstrated in the acceptance test with AWS cli which does the above underneath.

boolean storageTypeDefault = StringUtils.isEmpty(storageType);

// Normal put object
OzoneBucket bucket = volume.getBucket(bucketName);
ReplicationConfig replicationConfig =
getReplicationConfig(bucket, storageType);
getReplicationConfig(bucket, storageType, storageConfig);

boolean enableEC = false;
if ((replicationConfig != null &&
Expand Down Expand Up @@ -817,14 +820,15 @@ public Response initializeMultipartUpload(
try {
OzoneBucket ozoneBucket = getBucket(bucket);
String storageType = headers.getHeaderString(STORAGE_CLASS_HEADER);
String storageConfig = headers.getHeaderString(CUSTOM_METADATA_HEADER_PREFIX + STORAGE_CONFIG_HEADER);

Map<String, String> customMetadata =
getCustomMetadataFromHeaders(headers.getRequestHeaders());

Map<String, String> tags = getTaggingFromHeaders(headers);

ReplicationConfig replicationConfig =
getReplicationConfig(ozoneBucket, storageType);
getReplicationConfig(ozoneBucket, storageType, storageConfig);

OmMultipartInfo multipartInfo =
ozoneBucket.initiateMultipartUpload(key, replicationConfig, customMetadata, tags);
Expand Down Expand Up @@ -857,21 +861,20 @@ public Response initializeMultipartUpload(
}

private ReplicationConfig getReplicationConfig(OzoneBucket ozoneBucket,
String storageType) throws OS3Exception {
if (StringUtils.isEmpty(storageType)) {
S3StorageType defaultStorageType = S3StorageType.getDefault(ozoneConfiguration);
storageType = (defaultStorageType != null ? defaultStorageType.toString() : null);
}
String storageType, String storageConfig) throws OS3Exception {

ReplicationConfig clientConfiguredReplicationConfig = null;
String replication = ozoneConfiguration.get(OZONE_REPLICATION);

if (replication != null) {
clientConfiguredReplicationConfig = ReplicationConfig.parse(
ReplicationType.valueOf(ozoneConfiguration
.get(OZONE_REPLICATION_TYPE, OZONE_REPLICATION_TYPE_DEFAULT)),
replication, ozoneConfiguration);
ReplicationType replicationType = ReplicationType.valueOf(
ozoneConfiguration.get(OZONE_REPLICATION_TYPE, OZONE_REPLICATION_TYPE_DEFAULT));
clientConfiguredReplicationConfig =
(replicationType == ReplicationType.EC) ?
new ECReplicationConfig(replication) : ReplicationConfig.parse(
replicationType, replication, ozoneConfiguration);
}
return S3Utils.resolveS3ClientSideReplicationConfig(storageType,
return S3Utils.resolveS3ClientSideReplicationConfig(storageType, storageConfig,
clientConfiguredReplicationConfig, ozoneBucket.getReplicationConfig());
}

Expand Down Expand Up @@ -974,9 +977,10 @@ private Response createMultipartKey(OzoneVolume volume, String bucket,

copyHeader = headers.getHeaderString(COPY_SOURCE_HEADER);
String storageType = headers.getHeaderString(STORAGE_CLASS_HEADER);
String storageConfig = headers.getHeaderString(CUSTOM_METADATA_HEADER_PREFIX + STORAGE_CONFIG_HEADER);
final OzoneBucket ozoneBucket = volume.getBucket(bucket);
ReplicationConfig replicationConfig =
getReplicationConfig(ozoneBucket, storageType);
getReplicationConfig(ozoneBucket, storageType, storageConfig);

boolean enableEC = false;
if ((replicationConfig != null &&
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -150,6 +150,12 @@ public final class S3ErrorTable {
HTTP_FORBIDDEN
);

public static final OS3Exception INVALID_STORAGE_CLASS = new OS3Exception(
"InvalidStorageClass", "The storage class that you specified is not valid. " +
"Provide a supported storage class[STANDARD|REDUCED_REDUNDANCY|STANDARD_IA] or " +
"a valid custom EC storage config for if using STANDARD_IA.",
HTTP_BAD_REQUEST);

private static Function<Exception, OS3Exception> generateInternalError =
e -> new OS3Exception("InternalError", e.getMessage(), HTTP_INTERNAL_ERROR);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -69,6 +69,7 @@ public final class S3Consts {
// Constants related to custom metadata
public static final String CUSTOM_METADATA_HEADER_PREFIX = "x-amz-meta-";
public static final String CUSTOM_METADATA_COPY_DIRECTIVE_HEADER = "x-amz-metadata-directive";
public static final String STORAGE_CONFIG_HEADER = "storage-config";

public static final String DECODED_CONTENT_LENGTH_HEADER =
"x-amz-decoded-content-length";
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -18,67 +18,34 @@
package org.apache.hadoop.ozone.s3.util;

import org.apache.hadoop.hdds.client.ECReplicationConfig;
import org.apache.hadoop.hdds.client.RatisReplicationConfig;
import org.apache.hadoop.hdds.client.ReplicationConfig;
import org.apache.hadoop.hdds.client.ReplicationFactor;
import org.apache.hadoop.hdds.client.ReplicationType;
import org.apache.hadoop.hdds.conf.ConfigurationSource;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
import org.apache.hadoop.ozone.OzoneConfigKeys;

/**
* Maps S3 storage class values to Ozone replication values.
*/

public enum S3StorageType {

REDUCED_REDUNDANCY(ReplicationType.RATIS, ReplicationFactor.ONE),
STANDARD(ReplicationType.RATIS, ReplicationFactor.THREE);
REDUCED_REDUNDANCY(RatisReplicationConfig.getInstance(HddsProtos.ReplicationFactor.ONE)),
STANDARD(
RatisReplicationConfig.getInstance(HddsProtos.ReplicationFactor.THREE)),
STANDARD_IA(new ECReplicationConfig(ECReplicationConfig.EcCodec.RS + "-3-2-1024k"));

private final ReplicationType type;
private final ReplicationFactor factor;
private final ReplicationConfig replicationConfig;

S3StorageType(
ReplicationType type,
ReplicationFactor factor) {
this.type = type;
this.factor = factor;
S3StorageType(ReplicationConfig replicationConfig) {
this.replicationConfig = replicationConfig;
}

public ReplicationFactor getFactor() {
return factor;
}

public ReplicationType getType() {
return type;
}

/**
* Get default S3StorageType for a new key to be uploaded.
* This should align to the ozone cluster configuration.
* @param config OzoneConfiguration
* @return S3StorageType which wraps ozone replication type and factor
*/
public static S3StorageType getDefault(ConfigurationSource config) {
String replicationString = config.get(OzoneConfigKeys.OZONE_REPLICATION);
ReplicationFactor configFactor;
if (replicationString == null) {
// if no config is set then let server take decision
return null;
}
try {
configFactor = ReplicationFactor.valueOf(
Integer.parseInt(replicationString));
} catch (NumberFormatException ex) {
// conservatively defaults to STANDARD on wrong config value
return STANDARD;
}
return configFactor == ReplicationFactor.ONE
? REDUCED_REDUNDANCY : STANDARD;
public ReplicationConfig getReplicationConfig() {
return replicationConfig;
}

public static S3StorageType fromReplicationConfig(ReplicationConfig config) {
if (config instanceof ECReplicationConfig) {
return S3StorageType.STANDARD;
if (config.getReplicationType() == HddsProtos.ReplicationType.EC) {
return STANDARD_IA;
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

map EC type to STANDARD_IA.

Copy link
Contributor Author

@SaketaChalamchala SaketaChalamchala May 14, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is done here. STANDARD_IA is mapped to EC/RS-3-2-1024k by default.

}
if (config.getReplicationType() == HddsProtos.ReplicationType.STAND_ALONE ||
config.getRequiredNodes() == 1) {
Expand Down
Loading