diff --git a/.all-contributorsrc b/.all-contributorsrc
index 9aad5c2d7b7..dac08ca7d53 100644
--- a/.all-contributorsrc
+++ b/.all-contributorsrc
@@ -823,6 +823,24 @@
"contributions": [
"code"
]
+ },
+ {
+ "login": "L-Applin",
+ "name": "Olivier L Applin",
+ "avatar_url": "https://avatars.githubusercontent.com/u/16511950?v=4",
+ "profile": "http://applin.ca",
+ "contributions": [
+ "code"
+ ]
+ },
+ {
+ "login": "breader124",
+ "name": "Adrian Chlebosz",
+ "avatar_url": "https://avatars.githubusercontent.com/u/36669019?v=4",
+ "profile": "https://github.com/breader124",
+ "contributions": [
+ "code"
+ ]
}
],
"contributorsPerLine": 7,
diff --git a/.changes/2.20.109.json b/.changes/2.20.109.json
new file mode 100644
index 00000000000..643bd370bd3
--- /dev/null
+++ b/.changes/2.20.109.json
@@ -0,0 +1,42 @@
+{
+ "version": "2.20.109",
+ "date": "2023-07-21",
+ "entries": [
+ {
+ "type": "bugfix",
+ "category": "S3 Transfer Manager",
+ "contributor": "",
+ "description": "Fix a bug where the SSE-C parameters are not copied to the CompleteMultipartUpload request when transforming to a multipart copy."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Elemental MediaConvert",
+ "contributor": "",
+ "description": "This release includes improvements to Preserve 444 handling, compatibility of HEVC sources without frame rates, and general improvements to MP4 outputs."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Glue",
+ "contributor": "",
+ "description": "This release adds support for AWS Glue Crawler with Apache Hudi Tables, allowing Crawlers to discover Hudi Tables in S3 and register them in Glue Data Catalog for query engines to query against."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Relational Database Service",
+ "contributor": "",
+ "description": "Adds support for the DBSystemID parameter of CreateDBInstance to RDS Custom for Oracle."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon WorkSpaces",
+ "contributor": "",
+ "description": "Fixed VolumeEncryptionKey descriptions"
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.110.json b/.changes/2.20.110.json
new file mode 100644
index 00000000000..5861938b340
--- /dev/null
+++ b/.changes/2.20.110.json
@@ -0,0 +1,54 @@
+{
+ "version": "2.20.110",
+ "date": "2023-07-24",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "AWS CloudFormation",
+ "contributor": "",
+ "description": "This release supports filtering by DRIFT_STATUS for existing API ListStackInstances and adds support for a new API ListStackInstanceResourceDrifts. Customers can now view resource drift information from their StackSet management accounts."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Cost Explorer Service",
+ "contributor": "",
+ "description": "This release introduces the new API 'GetSavingsPlanPurchaseRecommendationDetails', which retrieves the details for a Savings Plan recommendation. It also updates the existing API 'GetSavingsPlansPurchaseRecommendation' to include the recommendation detail ID."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Glue",
+ "contributor": "",
+ "description": "Added support for Data Preparation Recipe node in Glue Studio jobs"
+ },
+ {
+ "type": "feature",
+ "category": "AmazonApiGatewayV2",
+ "contributor": "",
+ "description": "Documentation updates for Amazon API Gateway."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Chime SDK Media Pipelines",
+ "contributor": "",
+ "description": "AWS Media Pipeline compositing enhancement and Media Insights Pipeline auto language identification."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Elastic Compute Cloud",
+ "contributor": "",
+ "description": "Add \"disabled\" enum value to SpotInstanceState."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon QuickSight",
+ "contributor": "",
+ "description": "This release launches new Snapshot APIs for CSV and PDF exports, adds support for info icon for filters and parameters in Exploration APIs, adds modeled exception to the DeleteAccountCustomization API, and introduces AttributeAggregationFunction's ability to add UNIQUE_VALUE aggregation in tooltips."
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.111.json b/.changes/2.20.111.json
new file mode 100644
index 00000000000..f6337e44b74
--- /dev/null
+++ b/.changes/2.20.111.json
@@ -0,0 +1,90 @@
+{
+ "version": "2.20.111",
+ "date": "2023-07-25",
+ "entries": [
+ {
+ "type": "bugfix",
+ "category": "AWS IAM Policy Builder",
+ "contributor": "",
+ "description": "Fixed bug where actions were written instead of resources."
+ },
+ {
+ "type": "feature",
+ "category": "AWSBillingConductor",
+ "contributor": "",
+ "description": "Added support for Auto-Assocate Billing Groups for CreateBillingGroup, UpdateBillingGroup, and ListBillingGroups."
+ },
+ {
+ "type": "feature",
+ "category": "AWS DataSync",
+ "contributor": "",
+ "description": "AWS DataSync now supports Microsoft Azure Blob Storage locations."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Lambda",
+ "contributor": "",
+ "description": "Add Python 3.11 (python3.11) support to AWS Lambda"
+ },
+ {
+ "type": "feature",
+ "category": "AWS SecurityHub",
+ "contributor": "",
+ "description": "Add support for CONTAINS and NOT_CONTAINS comparison operators for Automation Rules string filters and map filters"
+ },
+ {
+ "type": "feature",
+ "category": "AWS Security Token Service",
+ "contributor": "",
+ "description": "API updates for the AWS Security Token Service"
+ },
+ {
+ "type": "feature",
+ "category": "AWS Transfer Family",
+ "contributor": "",
+ "description": "This release adds support for SFTP Connectors."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Connect Customer Profiles",
+ "contributor": "",
+ "description": "Amazon Connect Customer Profiles now supports rule-based resolution to match and merge similar profiles into unified profiles, helping companies deliver faster and more personalized customer service by providing access to relevant customer information for agents and automated experiences."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Connect Wisdom Service",
+ "contributor": "",
+ "description": "This release added two new data types: AssistantIntegrationConfiguration, and SessionIntegrationConfiguration to support Wisdom integration with Amazon Connect Chat"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon DynamoDB",
+ "contributor": "",
+ "description": "Documentation updates for DynamoDB"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Elastic Compute Cloud",
+ "contributor": "",
+ "description": "This release adds an instance's peak and baseline network bandwidth as well as the memory sizes of an instance's inference accelerators to DescribeInstanceTypes."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Relational Database Service",
+ "contributor": "",
+ "description": "This release adds support for monitoring storage optimization progress on the DescribeDBInstances API."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon SageMaker Service",
+ "contributor": "",
+ "description": "Mark ContentColumn and TargetLabelColumn as required Targets in TextClassificationJobConfig in CreateAutoMLJobV2API"
+ },
+ {
+ "type": "feature",
+ "category": "EMR Serverless",
+ "contributor": "",
+ "description": "This release adds support for publishing application logs to CloudWatch."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.112.json b/.changes/2.20.112.json
new file mode 100644
index 00000000000..b324d8bea52
--- /dev/null
+++ b/.changes/2.20.112.json
@@ -0,0 +1,72 @@
+{
+ "version": "2.20.112",
+ "date": "2023-07-26",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "AWS Cloud Control API",
+ "contributor": "",
+ "description": "Updates the documentation for CreateResource."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Elemental MediaConvert",
+ "contributor": "",
+ "description": "This release includes general updates to user documentation."
+ },
+ {
+ "type": "feature",
+ "category": "AWS EntityResolution",
+ "contributor": "",
+ "description": "AWS Entity Resolution can effectively match a source record from a customer relationship management (CRM) system with a source record from a marketing system containing campaign information."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Glue",
+ "contributor": "",
+ "description": "Release Glue Studio Snowflake Connector Node for SDK/CLI"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon HealthLake",
+ "contributor": "",
+ "description": "Updating the HealthLake service documentation."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Managed Blockchain Query",
+ "contributor": "",
+ "description": "Amazon Managed Blockchain (AMB) Query provides serverless access to standardized, multi-blockchain datasets with developer-friendly APIs."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Omics",
+ "contributor": "",
+ "description": "The service is renaming as a part of AWS Health."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Polly",
+ "contributor": "",
+ "description": "Amazon Polly adds 1 new voice - Lisa (nl-BE)"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Route 53",
+ "contributor": "",
+ "description": "Update that corrects the documents for received feedback."
+ },
+ {
+ "type": "feature",
+ "category": "OpenSearch Service Serverless",
+ "contributor": "",
+ "description": "This release adds new collection type VectorSearch."
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.113.json b/.changes/2.20.113.json
new file mode 100644
index 00000000000..fc5fa7c5ce3
--- /dev/null
+++ b/.changes/2.20.113.json
@@ -0,0 +1,36 @@
+{
+ "version": "2.20.113",
+ "date": "2023-07-27",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "Amazon Elastic Block Store",
+ "contributor": "",
+ "description": "SDK and documentation updates for Amazon Elastic Block Store API"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Elastic Compute Cloud",
+ "contributor": "",
+ "description": "SDK and documentation updates for Amazon Elastic Block Store APIs"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Elastic Kubernetes Service",
+ "contributor": "",
+ "description": "Add multiple customer error code to handle customer caused failure when managing EKS node groups"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon SageMaker Service",
+ "contributor": "",
+ "description": "Expose ProfilerConfig attribute in SageMaker Search API response."
+ },
+ {
+ "type": "feature",
+ "category": "Auto Scaling",
+ "contributor": "",
+ "description": "This release updates validation for instance types used in the AllowedInstanceTypes and ExcludedInstanceTypes parameters of the InstanceRequirements property of a MixedInstancesPolicy."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.114.json b/.changes/2.20.114.json
new file mode 100644
index 00000000000..d03a88e1077
--- /dev/null
+++ b/.changes/2.20.114.json
@@ -0,0 +1,12 @@
+{
+ "version": "2.20.114",
+ "date": "2023-07-27",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "Amazon Simple Queue Service",
+ "contributor": "",
+ "description": "Documentation changes related to SQS APIs."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.115.json b/.changes/2.20.115.json
new file mode 100644
index 00000000000..5b4f41b6d87
--- /dev/null
+++ b/.changes/2.20.115.json
@@ -0,0 +1,48 @@
+{
+ "version": "2.20.115",
+ "date": "2023-07-28",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "AWS CloudFormation",
+ "contributor": "",
+ "description": "This SDK release is for the feature launch of AWS CloudFormation RetainExceptOnCreate. It adds a new parameter retainExceptOnCreate in the following APIs: CreateStack, UpdateStack, RollbackStack, ExecuteChangeSet."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon CloudFront",
+ "contributor": "",
+ "description": "Add a new JavaScript runtime version for CloudFront Functions."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon CloudWatch Application Insights",
+ "contributor": "",
+ "description": "This release enable customer to add/remove/update more than one workload for a component"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Connect Service",
+ "contributor": "",
+ "description": "This release adds support for new number types."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Pinpoint",
+ "contributor": "",
+ "description": "Added support for sending push notifications using the FCM v1 API with json credentials. Amazon Pinpoint customers can now deliver messages to Android devices using both FCM v1 API and the legacy FCM/GCM API"
+ },
+ {
+ "type": "feature",
+ "category": "Managed Streaming for Kafka",
+ "contributor": "",
+ "description": "Amazon MSK has introduced new versions of ListClusterOperations and DescribeClusterOperation APIs. These v2 APIs provide information and insights into the ongoing operations of both MSK Provisioned and MSK Serverless clusters."
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.116.json b/.changes/2.20.116.json
new file mode 100644
index 00000000000..4f5480b6a4f
--- /dev/null
+++ b/.changes/2.20.116.json
@@ -0,0 +1,78 @@
+{
+ "version": "2.20.116",
+ "date": "2023-07-31",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "AWS Amplify UI Builder",
+ "contributor": "",
+ "description": "Amplify Studio releases GraphQL support for codegen job action."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Clean Rooms Service",
+ "contributor": "",
+ "description": "This release introduces custom SQL queries - an expanded set of SQL you can run. This release adds analysis templates, a new resource for storing pre-defined custom SQL queries ahead of time. This release also adds the Custom analysis rule, which lets you approve analysis templates for querying."
+ },
+ {
+ "type": "feature",
+ "category": "AWS CodeStar connections",
+ "contributor": "",
+ "description": "New integration with the Gitlab provider type."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon EventBridge Scheduler",
+ "contributor": "",
+ "description": "This release introduces automatic deletion of schedules in EventBridge Scheduler. If configured, EventBridge Scheduler automatically deletes a schedule after the schedule has completed its last invocation."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Lookout for Equipment",
+ "contributor": "",
+ "description": "This release includes new import resource, model versioning and resource policy features."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Omics",
+ "contributor": "",
+ "description": "Add CreationType filter for ListReadSets"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Relational Database Service",
+ "contributor": "",
+ "description": "This release adds support for Aurora MySQL local write forwarding, which allows for forwarding of write operations from reader DB instances to the writer DB instance."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Route 53",
+ "contributor": "",
+ "description": "Amazon Route 53 now supports the Israel (Tel Aviv) Region (il-central-1) for latency records, geoproximity records, and private DNS for Amazon VPCs in that region."
+ },
+ {
+ "type": "feature",
+ "category": "Auto Scaling",
+ "contributor": "",
+ "description": "You can now configure an instance refresh to set its status to 'failed' when it detects that a specified CloudWatch alarm has gone into the ALARM state. You can also choose to roll back the instance refresh automatically when the alarm threshold is met."
+ },
+ {
+ "type": "feature",
+ "category": "Elastic Disaster Recovery Service",
+ "contributor": "",
+ "description": "Add support for in-aws right sizing"
+ },
+ {
+ "type": "feature",
+ "category": "Inspector2",
+ "contributor": "",
+ "description": "This release adds 1 new API: BatchGetFindingDetails to retrieve enhanced vulnerability intelligence details for findings."
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.117.json b/.changes/2.20.117.json
new file mode 100644
index 00000000000..6875a98fc8c
--- /dev/null
+++ b/.changes/2.20.117.json
@@ -0,0 +1,54 @@
+{
+ "version": "2.20.117",
+ "date": "2023-08-01",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "AWS Batch",
+ "contributor": "",
+ "description": "This release adds support for price capacity optimized allocation strategy for Spot Instances."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Database Migration Service",
+ "contributor": "",
+ "description": "Adding new API describe-engine-versions which provides information about the lifecycle of a replication instance's version."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Elemental MediaLive",
+ "contributor": "",
+ "description": "AWS Elemental Link devices now report their Availability Zone. Link devices now support the ability to change their Availability Zone."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon CloudWatch Internet Monitor",
+ "contributor": "",
+ "description": "This release adds a new feature for Amazon CloudWatch Internet Monitor that enables customers to set custom thresholds, for performance and availability drops, for impact limited to a single city-network to trigger creation of a health event."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Polly",
+ "contributor": "",
+ "description": "Amazon Polly adds new French Belgian voice - Isabelle. Isabelle is available as Neural voice only."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Relational Database Service",
+ "contributor": "",
+ "description": "Added support for deleted clusters PiTR."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon SageMaker Service",
+ "contributor": "",
+ "description": "Add Stairs TrafficPattern and FlatInvocations to RecommendationJobStoppingConditions"
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.118.json b/.changes/2.20.118.json
new file mode 100644
index 00000000000..f9c9494a13d
--- /dev/null
+++ b/.changes/2.20.118.json
@@ -0,0 +1,42 @@
+{
+ "version": "2.20.118",
+ "date": "2023-08-02",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "AWS Budgets",
+ "contributor": "",
+ "description": "As part of CAE tagging integration we need to update our budget names regex filter to prevent customers from using \"/action/\" in their budget names."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Glue",
+ "contributor": "",
+ "description": "This release includes additional Glue Streaming KAKFA SASL property types."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Resilience Hub",
+ "contributor": "",
+ "description": "Drift Detection capability added when applications policy has moved from a meet to breach state. Customers will be able to exclude operational recommendations and receive credit in their resilience score. Customers can now add ARH permissions to an existing or new role."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Cognito Identity Provider",
+ "contributor": "",
+ "description": "New feature that logs Cognito user pool error messages to CloudWatch logs."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon SageMaker Service",
+ "contributor": "",
+ "description": "SageMaker Inference Recommender introduces a new API GetScalingConfigurationRecommendation to recommend auto scaling policies based on completed Inference Recommender jobs."
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.119.json b/.changes/2.20.119.json
new file mode 100644
index 00000000000..d737645c704
--- /dev/null
+++ b/.changes/2.20.119.json
@@ -0,0 +1,42 @@
+{
+ "version": "2.20.119",
+ "date": "2023-08-03",
+ "entries": [
+ {
+ "type": "bugfix",
+ "category": "Amazon S3",
+ "contributor": "",
+ "description": "Add `Expect: 100-continue` header for `UploadPartRequest` so that a upload part request can fail faster if there is a server error."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Cloud9",
+ "contributor": "",
+ "description": "Updated the deprecation date for Amazon Linux. Doc only update."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Database Migration Service",
+ "contributor": "",
+ "description": "The release makes public API for DMS Schema Conversion feature."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Elastic Compute Cloud",
+ "contributor": "",
+ "description": "This release adds new parameter isPrimaryIPv6 to allow assigning an IPv6 address as a primary IPv6 address to a network interface which cannot be changed to give equivalent functionality available for network interfaces with primary IPv4 address."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon SageMaker Service",
+ "contributor": "",
+ "description": "Amazon SageMaker now supports running training jobs on p5.48xlarge instance types."
+ },
+ {
+ "type": "feature",
+ "category": "Auto Scaling",
+ "contributor": "",
+ "description": "Documentation changes related to Amazon EC2 Auto Scaling APIs."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.120.json b/.changes/2.20.120.json
new file mode 100644
index 00000000000..f8f88ca6099
--- /dev/null
+++ b/.changes/2.20.120.json
@@ -0,0 +1,42 @@
+{
+ "version": "2.20.120",
+ "date": "2023-08-04",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "AWS Certificate Manager Private Certificate Authority",
+ "contributor": "",
+ "description": "Documentation correction for AWS Private CA"
+ },
+ {
+ "type": "feature",
+ "category": "AWS DataSync",
+ "contributor": "",
+ "description": "Display cloud storage used capacity at a cluster level."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Connect Service",
+ "contributor": "",
+ "description": "Added a new API UpdateRoutingProfileAgentAvailabilityTimer to update agent availability timer of a routing profile."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon EC2 Container Service",
+ "contributor": "",
+ "description": "This is a documentation update to address various tickets."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon SageMaker Service",
+ "contributor": "",
+ "description": "Including DataCaptureConfig key in the Amazon Sagemaker Search's transform job object"
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.121.json b/.changes/2.20.121.json
new file mode 100644
index 00000000000..9750c52dfb2
--- /dev/null
+++ b/.changes/2.20.121.json
@@ -0,0 +1,42 @@
+{
+ "version": "2.20.121",
+ "date": "2023-08-07",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "Amazon Detective",
+ "contributor": "",
+ "description": "Updated the email validation regex to be in line with the TLD name specifications."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Interactive Video Service RealTime",
+ "contributor": "",
+ "description": "Add QUOTA_EXCEEDED and PUBLISHER_NOT_FOUND to EventErrorCode for stage health events."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Kinesis Video Streams",
+ "contributor": "",
+ "description": "This release enables minimum of Images SamplingInterval to be as low as 200 milliseconds in Kinesis Video Stream Image feature."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Kinesis Video Streams Archived Media",
+ "contributor": "",
+ "description": "This release enables minimum of Images SamplingInterval to be as low as 200 milliseconds in Kinesis Video Stream Image feature."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Rekognition",
+ "contributor": "",
+ "description": "This release adds code snippets for Amazon Rekognition Custom Labels."
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.122.json b/.changes/2.20.122.json
new file mode 100644
index 00000000000..8a3dd6ab9d1
--- /dev/null
+++ b/.changes/2.20.122.json
@@ -0,0 +1,30 @@
+{
+ "version": "2.20.122",
+ "date": "2023-08-08",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "AWS Backup",
+ "contributor": "",
+ "description": "This release introduces a new logically air-gapped vault (Preview) in AWS Backup that stores immutable backup copies, which are locked by default and isolated with encryption using AWS owned keys. Logically air-gapped vault (Preview) allows secure recovery of application data across accounts."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Service Catalog",
+ "contributor": "",
+ "description": "Introduce support for HashiCorp Terraform Cloud in Service Catalog by addying TERRAFORM_CLOUD product type in CreateProduct and CreateProvisioningArtifact API."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon ElastiCache",
+ "contributor": "",
+ "description": "Added support for cluster mode in online migration and test migration API"
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.123.json b/.changes/2.20.123.json
new file mode 100644
index 00000000000..48a9bf7369b
--- /dev/null
+++ b/.changes/2.20.123.json
@@ -0,0 +1,48 @@
+{
+ "version": "2.20.123",
+ "date": "2023-08-09",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "AWS Global Accelerator",
+ "contributor": "",
+ "description": "Documentation update for dualstack EC2 endpoint support"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Chime SDK Voice",
+ "contributor": "",
+ "description": "Updating CreatePhoneNumberOrder, UpdatePhoneNumber and BatchUpdatePhoneNumbers APIs, adding phone number name"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon FSx",
+ "contributor": "",
+ "description": "For FSx for Lustre, add new data repository task type, RELEASE_DATA_FROM_FILESYSTEM, to release files that have been archived to S3. For FSx for Windows, enable support for configuring and updating SSD IOPS, and for updating storage type. For FSx for OpenZFS, add new deployment type, MULTI_AZ_1."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon GuardDuty",
+ "contributor": "",
+ "description": "Added autoEnable ALL to UpdateOrganizationConfiguration and DescribeOrganizationConfiguration APIs."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon S3",
+ "contributor": "",
+ "description": "Allow users to configure upload threshold size for AWS CRT-based S3 client via `S3CrtAsyncClientBuilder#thresholdInBytes`."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon SageMaker Service",
+ "contributor": "",
+ "description": "This release adds support for cross account access for SageMaker Model Cards through AWS RAM."
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.124.json b/.changes/2.20.124.json
new file mode 100644
index 00000000000..bb85d621d24
--- /dev/null
+++ b/.changes/2.20.124.json
@@ -0,0 +1,60 @@
+{
+ "version": "2.20.124",
+ "date": "2023-08-10",
+ "entries": [
+ {
+ "type": "bugfix",
+ "category": "AWS DynamoDB Enhanced Client",
+ "contributor": "",
+ "description": "Fix for Issue [#4156](https://github.com/aws/aws-sdk-java-v2/issues/4156) : Single quotes in toJson conversions for EnhancedDocuments are no longer being escaped."
+ },
+ {
+ "type": "bugfix",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Fixed bug where limit was not copied over when cloning ByteBuffer using immutableCopyOf()"
+ },
+ {
+ "type": "feature",
+ "category": "AWS CloudTrail",
+ "contributor": "",
+ "description": "Documentation updates for CloudTrail."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Secrets Manager",
+ "contributor": "",
+ "description": "Add additional InvalidRequestException to list of possible exceptions for ListSecret."
+ },
+ {
+ "type": "feature",
+ "category": "AWS Transfer Family",
+ "contributor": "",
+ "description": "Documentation updates for AW Transfer Family"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Connect Service",
+ "contributor": "",
+ "description": "This release adds APIs to provision agents that are global / available in multiple AWS regions and distribute them across these regions by percentage."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Omics",
+ "contributor": "",
+ "description": "This release adds instanceType to GetRunTask & ListRunTasks responses."
+ },
+ {
+ "type": "feature",
+ "category": "Elastic Load Balancing",
+ "contributor": "",
+ "description": "This release enables configuring security groups for Network Load Balancers"
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/2.20.125.json b/.changes/2.20.125.json
new file mode 100644
index 00000000000..be4785cc08c
--- /dev/null
+++ b/.changes/2.20.125.json
@@ -0,0 +1,48 @@
+{
+ "version": "2.20.125",
+ "date": "2023-08-11",
+ "entries": [
+ {
+ "type": "feature",
+ "category": "AWS Config",
+ "contributor": "",
+ "description": "Updated ResourceType enum with new resource types onboarded by AWS Config in July 2023."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Elastic Compute Cloud",
+ "contributor": "",
+ "description": "Amazon EC2 P5 instances, powered by the latest NVIDIA H100 Tensor Core GPUs, deliver the highest performance in EC2 for deep learning (DL) and HPC applications. M7i-flex and M7i instances are next-generation general purpose instances powered by custom 4th Generation Intel Xeon Scalable processors."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon QuickSight",
+ "contributor": "",
+ "description": "New Authentication method for Account subscription - IAM Identity Center. Hierarchy layout support, default column width support and related style properties for pivot table visuals. Non-additive topic field aggregations for Topic API"
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Simple Email Service",
+ "contributor": "",
+ "description": "Doc only updates to include: 1) Clarified which part of an email address where it's okay to have Punycode when it contains non-ASCII characters for the SendRawEmail action and other actions where this is applicable. 2) Updated S3Action description with new MB max bucket size from 30 to 40."
+ },
+ {
+ "type": "feature",
+ "category": "Amazon Simple Workflow Service",
+ "contributor": "",
+ "description": "This release adds new API parameters to override workflow task list for workflow executions."
+ },
+ {
+ "type": "feature",
+ "category": "AmplifyBackend",
+ "contributor": "",
+ "description": "Adds sensitive trait to required input shapes."
+ },
+ {
+ "type": "feature",
+ "category": "AWS SDK for Java v2",
+ "contributor": "",
+ "description": "Updated endpoint and partition metadata."
+ }
+ ]
+}
\ No newline at end of file
diff --git a/.changes/next-release/bugfix-S3TransferManager-d9c09d4.json b/.changes/next-release/bugfix-S3TransferManager-d9c09d4.json
deleted file mode 100644
index 0b93e7f4040..00000000000
--- a/.changes/next-release/bugfix-S3TransferManager-d9c09d4.json
+++ /dev/null
@@ -1,6 +0,0 @@
-{
- "type": "bugfix",
- "category": "S3 Transfer Manager",
- "contributor": "",
- "description": "Fix a bug where the SSE-C parameters are not copied to the CompleteMultipartUpload request when transforming to a multipart copy."
-}
diff --git a/CHANGELOG.md b/CHANGELOG.md
index 6f7570d660d..7e409994228 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -1,3 +1,511 @@
+# __2.20.125__ __2023-08-11__
+## __AWS Config__
+ - ### Features
+ - Updated ResourceType enum with new resource types onboarded by AWS Config in July 2023.
+
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+## __Amazon Elastic Compute Cloud__
+ - ### Features
+ - Amazon EC2 P5 instances, powered by the latest NVIDIA H100 Tensor Core GPUs, deliver the highest performance in EC2 for deep learning (DL) and HPC applications. M7i-flex and M7i instances are next-generation general purpose instances powered by custom 4th Generation Intel Xeon Scalable processors.
+
+## __Amazon QuickSight__
+ - ### Features
+ - New Authentication method for Account subscription - IAM Identity Center. Hierarchy layout support, default column width support and related style properties for pivot table visuals. Non-additive topic field aggregations for Topic API
+
+## __Amazon Simple Email Service__
+ - ### Features
+ - Doc only updates to include: 1) Clarified which part of an email address where it's okay to have Punycode when it contains non-ASCII characters for the SendRawEmail action and other actions where this is applicable. 2) Updated S3Action description with new MB max bucket size from 30 to 40.
+
+## __Amazon Simple Workflow Service__
+ - ### Features
+ - This release adds new API parameters to override workflow task list for workflow executions.
+
+## __AmplifyBackend__
+ - ### Features
+ - Adds sensitive trait to required input shapes.
+
+# __2.20.124__ __2023-08-10__
+## __AWS CloudTrail__
+ - ### Features
+ - Documentation updates for CloudTrail.
+
+## __AWS DynamoDB Enhanced Client__
+ - ### Bugfixes
+ - Fix for Issue [#4156](https://github.com/aws/aws-sdk-java-v2/issues/4156) : Single quotes in toJson conversions for EnhancedDocuments are no longer being escaped.
+
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+ - ### Bugfixes
+ - Fixed bug where limit was not copied over when cloning ByteBuffer using immutableCopyOf()
+
+## __AWS Secrets Manager__
+ - ### Features
+ - Add additional InvalidRequestException to list of possible exceptions for ListSecret.
+
+## __AWS Transfer Family__
+ - ### Features
+ - Documentation updates for AW Transfer Family
+
+## __Amazon Connect Service__
+ - ### Features
+ - This release adds APIs to provision agents that are global / available in multiple AWS regions and distribute them across these regions by percentage.
+
+## __Amazon Omics__
+ - ### Features
+ - This release adds instanceType to GetRunTask & ListRunTasks responses.
+
+## __Elastic Load Balancing__
+ - ### Features
+ - This release enables configuring security groups for Network Load Balancers
+
+# __2.20.123__ __2023-08-09__
+## __AWS Global Accelerator__
+ - ### Features
+ - Documentation update for dualstack EC2 endpoint support
+
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+## __Amazon Chime SDK Voice__
+ - ### Features
+ - Updating CreatePhoneNumberOrder, UpdatePhoneNumber and BatchUpdatePhoneNumbers APIs, adding phone number name
+
+## __Amazon FSx__
+ - ### Features
+ - For FSx for Lustre, add new data repository task type, RELEASE_DATA_FROM_FILESYSTEM, to release files that have been archived to S3. For FSx for Windows, enable support for configuring and updating SSD IOPS, and for updating storage type. For FSx for OpenZFS, add new deployment type, MULTI_AZ_1.
+
+## __Amazon GuardDuty__
+ - ### Features
+ - Added autoEnable ALL to UpdateOrganizationConfiguration and DescribeOrganizationConfiguration APIs.
+
+## __Amazon S3__
+ - ### Features
+ - Allow users to configure upload threshold size for AWS CRT-based S3 client via `S3CrtAsyncClientBuilder#thresholdInBytes`.
+
+## __Amazon SageMaker Service__
+ - ### Features
+ - This release adds support for cross account access for SageMaker Model Cards through AWS RAM.
+
+# __2.20.122__ __2023-08-08__
+## __AWS Backup__
+ - ### Features
+ - This release introduces a new logically air-gapped vault (Preview) in AWS Backup that stores immutable backup copies, which are locked by default and isolated with encryption using AWS owned keys. Logically air-gapped vault (Preview) allows secure recovery of application data across accounts.
+
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+## __AWS Service Catalog__
+ - ### Features
+ - Introduce support for HashiCorp Terraform Cloud in Service Catalog by addying TERRAFORM_CLOUD product type in CreateProduct and CreateProvisioningArtifact API.
+
+## __Amazon ElastiCache__
+ - ### Features
+ - Added support for cluster mode in online migration and test migration API
+
+# __2.20.121__ __2023-08-07__
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+## __Amazon Detective__
+ - ### Features
+ - Updated the email validation regex to be in line with the TLD name specifications.
+
+## __Amazon Interactive Video Service RealTime__
+ - ### Features
+ - Add QUOTA_EXCEEDED and PUBLISHER_NOT_FOUND to EventErrorCode for stage health events.
+
+## __Amazon Kinesis Video Streams__
+ - ### Features
+ - This release enables minimum of Images SamplingInterval to be as low as 200 milliseconds in Kinesis Video Stream Image feature.
+
+## __Amazon Kinesis Video Streams Archived Media__
+ - ### Features
+ - This release enables minimum of Images SamplingInterval to be as low as 200 milliseconds in Kinesis Video Stream Image feature.
+
+## __Amazon Rekognition__
+ - ### Features
+ - This release adds code snippets for Amazon Rekognition Custom Labels.
+
+# __2.20.120__ __2023-08-04__
+## __AWS Certificate Manager Private Certificate Authority__
+ - ### Features
+ - Documentation correction for AWS Private CA
+
+## __AWS DataSync__
+ - ### Features
+ - Display cloud storage used capacity at a cluster level.
+
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+## __Amazon Connect Service__
+ - ### Features
+ - Added a new API UpdateRoutingProfileAgentAvailabilityTimer to update agent availability timer of a routing profile.
+
+## __Amazon EC2 Container Service__
+ - ### Features
+ - This is a documentation update to address various tickets.
+
+## __Amazon SageMaker Service__
+ - ### Features
+ - Including DataCaptureConfig key in the Amazon Sagemaker Search's transform job object
+
+# __2.20.119__ __2023-08-03__
+## __AWS Cloud9__
+ - ### Features
+ - Updated the deprecation date for Amazon Linux. Doc only update.
+
+## __AWS Database Migration Service__
+ - ### Features
+ - The release makes public API for DMS Schema Conversion feature.
+
+## __Amazon Elastic Compute Cloud__
+ - ### Features
+ - This release adds new parameter isPrimaryIPv6 to allow assigning an IPv6 address as a primary IPv6 address to a network interface which cannot be changed to give equivalent functionality available for network interfaces with primary IPv4 address.
+
+## __Amazon S3__
+ - ### Bugfixes
+ - Add `Expect: 100-continue` header for `UploadPartRequest` so that a upload part request can fail faster if there is a server error.
+
+## __Amazon SageMaker Service__
+ - ### Features
+ - Amazon SageMaker now supports running training jobs on p5.48xlarge instance types.
+
+## __Auto Scaling__
+ - ### Features
+ - Documentation changes related to Amazon EC2 Auto Scaling APIs.
+
+# __2.20.118__ __2023-08-02__
+## __AWS Budgets__
+ - ### Features
+ - As part of CAE tagging integration we need to update our budget names regex filter to prevent customers from using "/action/" in their budget names.
+
+## __AWS Glue__
+ - ### Features
+ - This release includes additional Glue Streaming KAKFA SASL property types.
+
+## __AWS Resilience Hub__
+ - ### Features
+ - Drift Detection capability added when applications policy has moved from a meet to breach state. Customers will be able to exclude operational recommendations and receive credit in their resilience score. Customers can now add ARH permissions to an existing or new role.
+
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+## __Amazon Cognito Identity Provider__
+ - ### Features
+ - New feature that logs Cognito user pool error messages to CloudWatch logs.
+
+## __Amazon SageMaker Service__
+ - ### Features
+ - SageMaker Inference Recommender introduces a new API GetScalingConfigurationRecommendation to recommend auto scaling policies based on completed Inference Recommender jobs.
+
+# __2.20.117__ __2023-08-01__
+## __AWS Batch__
+ - ### Features
+ - This release adds support for price capacity optimized allocation strategy for Spot Instances.
+
+## __AWS Database Migration Service__
+ - ### Features
+ - Adding new API describe-engine-versions which provides information about the lifecycle of a replication instance's version.
+
+## __AWS Elemental MediaLive__
+ - ### Features
+ - AWS Elemental Link devices now report their Availability Zone. Link devices now support the ability to change their Availability Zone.
+
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+## __Amazon CloudWatch Internet Monitor__
+ - ### Features
+ - This release adds a new feature for Amazon CloudWatch Internet Monitor that enables customers to set custom thresholds, for performance and availability drops, for impact limited to a single city-network to trigger creation of a health event.
+
+## __Amazon Polly__
+ - ### Features
+ - Amazon Polly adds new French Belgian voice - Isabelle. Isabelle is available as Neural voice only.
+
+## __Amazon Relational Database Service__
+ - ### Features
+ - Added support for deleted clusters PiTR.
+
+## __Amazon SageMaker Service__
+ - ### Features
+ - Add Stairs TrafficPattern and FlatInvocations to RecommendationJobStoppingConditions
+
+# __2.20.116__ __2023-07-31__
+## __AWS Amplify UI Builder__
+ - ### Features
+ - Amplify Studio releases GraphQL support for codegen job action.
+
+## __AWS Clean Rooms Service__
+ - ### Features
+ - This release introduces custom SQL queries - an expanded set of SQL you can run. This release adds analysis templates, a new resource for storing pre-defined custom SQL queries ahead of time. This release also adds the Custom analysis rule, which lets you approve analysis templates for querying.
+
+## __AWS CodeStar connections__
+ - ### Features
+ - New integration with the Gitlab provider type.
+
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+## __Amazon EventBridge Scheduler__
+ - ### Features
+ - This release introduces automatic deletion of schedules in EventBridge Scheduler. If configured, EventBridge Scheduler automatically deletes a schedule after the schedule has completed its last invocation.
+
+## __Amazon Lookout for Equipment__
+ - ### Features
+ - This release includes new import resource, model versioning and resource policy features.
+
+## __Amazon Omics__
+ - ### Features
+ - Add CreationType filter for ListReadSets
+
+## __Amazon Relational Database Service__
+ - ### Features
+ - This release adds support for Aurora MySQL local write forwarding, which allows for forwarding of write operations from reader DB instances to the writer DB instance.
+
+## __Amazon Route 53__
+ - ### Features
+ - Amazon Route 53 now supports the Israel (Tel Aviv) Region (il-central-1) for latency records, geoproximity records, and private DNS for Amazon VPCs in that region.
+
+## __Auto Scaling__
+ - ### Features
+ - You can now configure an instance refresh to set its status to 'failed' when it detects that a specified CloudWatch alarm has gone into the ALARM state. You can also choose to roll back the instance refresh automatically when the alarm threshold is met.
+
+## __Elastic Disaster Recovery Service__
+ - ### Features
+ - Add support for in-aws right sizing
+
+## __Inspector2__
+ - ### Features
+ - This release adds 1 new API: BatchGetFindingDetails to retrieve enhanced vulnerability intelligence details for findings.
+
+# __2.20.115__ __2023-07-28__
+## __AWS CloudFormation__
+ - ### Features
+ - This SDK release is for the feature launch of AWS CloudFormation RetainExceptOnCreate. It adds a new parameter retainExceptOnCreate in the following APIs: CreateStack, UpdateStack, RollbackStack, ExecuteChangeSet.
+
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+## __Amazon CloudFront__
+ - ### Features
+ - Add a new JavaScript runtime version for CloudFront Functions.
+
+## __Amazon CloudWatch Application Insights__
+ - ### Features
+ - This release enable customer to add/remove/update more than one workload for a component
+
+## __Amazon Connect Service__
+ - ### Features
+ - This release adds support for new number types.
+
+## __Amazon Pinpoint__
+ - ### Features
+ - Added support for sending push notifications using the FCM v1 API with json credentials. Amazon Pinpoint customers can now deliver messages to Android devices using both FCM v1 API and the legacy FCM/GCM API
+
+## __Managed Streaming for Kafka__
+ - ### Features
+ - Amazon MSK has introduced new versions of ListClusterOperations and DescribeClusterOperation APIs. These v2 APIs provide information and insights into the ongoing operations of both MSK Provisioned and MSK Serverless clusters.
+
+# __2.20.114__ __2023-07-27__
+## __Amazon Simple Queue Service__
+ - ### Features
+ - Documentation changes related to SQS APIs.
+
+# __2.20.113__ __2023-07-27__
+## __Amazon Elastic Block Store__
+ - ### Features
+ - SDK and documentation updates for Amazon Elastic Block Store API
+
+## __Amazon Elastic Compute Cloud__
+ - ### Features
+ - SDK and documentation updates for Amazon Elastic Block Store APIs
+
+## __Amazon Elastic Kubernetes Service__
+ - ### Features
+ - Add multiple customer error code to handle customer caused failure when managing EKS node groups
+
+## __Amazon SageMaker Service__
+ - ### Features
+ - Expose ProfilerConfig attribute in SageMaker Search API response.
+
+## __Auto Scaling__
+ - ### Features
+ - This release updates validation for instance types used in the AllowedInstanceTypes and ExcludedInstanceTypes parameters of the InstanceRequirements property of a MixedInstancesPolicy.
+
+# __2.20.112__ __2023-07-26__
+## __AWS Cloud Control API__
+ - ### Features
+ - Updates the documentation for CreateResource.
+
+## __AWS Elemental MediaConvert__
+ - ### Features
+ - This release includes general updates to user documentation.
+
+## __AWS EntityResolution__
+ - ### Features
+ - AWS Entity Resolution can effectively match a source record from a customer relationship management (CRM) system with a source record from a marketing system containing campaign information.
+
+## __AWS Glue__
+ - ### Features
+ - Release Glue Studio Snowflake Connector Node for SDK/CLI
+
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+## __Amazon HealthLake__
+ - ### Features
+ - Updating the HealthLake service documentation.
+
+## __Amazon Managed Blockchain Query__
+ - ### Features
+ - Amazon Managed Blockchain (AMB) Query provides serverless access to standardized, multi-blockchain datasets with developer-friendly APIs.
+
+## __Amazon Omics__
+ - ### Features
+ - The service is renaming as a part of AWS Health.
+
+## __Amazon Polly__
+ - ### Features
+ - Amazon Polly adds 1 new voice - Lisa (nl-BE)
+
+## __Amazon Route 53__
+ - ### Features
+ - Update that corrects the documents for received feedback.
+
+## __OpenSearch Service Serverless__
+ - ### Features
+ - This release adds new collection type VectorSearch.
+
+# __2.20.111__ __2023-07-25__
+## __AWS DataSync__
+ - ### Features
+ - AWS DataSync now supports Microsoft Azure Blob Storage locations.
+
+## __AWS IAM Policy Builder__
+ - ### Bugfixes
+ - Fixed bug where actions were written instead of resources.
+
+## __AWS Lambda__
+ - ### Features
+ - Add Python 3.11 (python3.11) support to AWS Lambda
+
+## __AWS Security Token Service__
+ - ### Features
+ - API updates for the AWS Security Token Service
+
+## __AWS SecurityHub__
+ - ### Features
+ - Add support for CONTAINS and NOT_CONTAINS comparison operators for Automation Rules string filters and map filters
+
+## __AWS Transfer Family__
+ - ### Features
+ - This release adds support for SFTP Connectors.
+
+## __AWSBillingConductor__
+ - ### Features
+ - Added support for Auto-Assocate Billing Groups for CreateBillingGroup, UpdateBillingGroup, and ListBillingGroups.
+
+## __Amazon Connect Customer Profiles__
+ - ### Features
+ - Amazon Connect Customer Profiles now supports rule-based resolution to match and merge similar profiles into unified profiles, helping companies deliver faster and more personalized customer service by providing access to relevant customer information for agents and automated experiences.
+
+## __Amazon Connect Wisdom Service__
+ - ### Features
+ - This release added two new data types: AssistantIntegrationConfiguration, and SessionIntegrationConfiguration to support Wisdom integration with Amazon Connect Chat
+
+## __Amazon DynamoDB__
+ - ### Features
+ - Documentation updates for DynamoDB
+
+## __Amazon Elastic Compute Cloud__
+ - ### Features
+ - This release adds an instance's peak and baseline network bandwidth as well as the memory sizes of an instance's inference accelerators to DescribeInstanceTypes.
+
+## __Amazon Relational Database Service__
+ - ### Features
+ - This release adds support for monitoring storage optimization progress on the DescribeDBInstances API.
+
+## __Amazon SageMaker Service__
+ - ### Features
+ - Mark ContentColumn and TargetLabelColumn as required Targets in TextClassificationJobConfig in CreateAutoMLJobV2API
+
+## __EMR Serverless__
+ - ### Features
+ - This release adds support for publishing application logs to CloudWatch.
+
+# __2.20.110__ __2023-07-24__
+## __AWS CloudFormation__
+ - ### Features
+ - This release supports filtering by DRIFT_STATUS for existing API ListStackInstances and adds support for a new API ListStackInstanceResourceDrifts. Customers can now view resource drift information from their StackSet management accounts.
+
+## __AWS Cost Explorer Service__
+ - ### Features
+ - This release introduces the new API 'GetSavingsPlanPurchaseRecommendationDetails', which retrieves the details for a Savings Plan recommendation. It also updates the existing API 'GetSavingsPlansPurchaseRecommendation' to include the recommendation detail ID.
+
+## __AWS Glue__
+ - ### Features
+ - Added support for Data Preparation Recipe node in Glue Studio jobs
+
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+## __Amazon Chime SDK Media Pipelines__
+ - ### Features
+ - AWS Media Pipeline compositing enhancement and Media Insights Pipeline auto language identification.
+
+## __Amazon Elastic Compute Cloud__
+ - ### Features
+ - Add "disabled" enum value to SpotInstanceState.
+
+## __Amazon QuickSight__
+ - ### Features
+ - This release launches new Snapshot APIs for CSV and PDF exports, adds support for info icon for filters and parameters in Exploration APIs, adds modeled exception to the DeleteAccountCustomization API, and introduces AttributeAggregationFunction's ability to add UNIQUE_VALUE aggregation in tooltips.
+
+## __AmazonApiGatewayV2__
+ - ### Features
+ - Documentation updates for Amazon API Gateway.
+
+# __2.20.109__ __2023-07-21__
+## __AWS Elemental MediaConvert__
+ - ### Features
+ - This release includes improvements to Preserve 444 handling, compatibility of HEVC sources without frame rates, and general improvements to MP4 outputs.
+
+## __AWS Glue__
+ - ### Features
+ - This release adds support for AWS Glue Crawler with Apache Hudi Tables, allowing Crawlers to discover Hudi Tables in S3 and register them in Glue Data Catalog for query engines to query against.
+
+## __AWS SDK for Java v2__
+ - ### Features
+ - Updated endpoint and partition metadata.
+
+## __Amazon Relational Database Service__
+ - ### Features
+ - Adds support for the DBSystemID parameter of CreateDBInstance to RDS Custom for Oracle.
+
+## __Amazon WorkSpaces__
+ - ### Features
+ - Fixed VolumeEncryptionKey descriptions
+
+## __S3 Transfer Manager__
+ - ### Bugfixes
+ - Fix a bug where the SSE-C parameters are not copied to the CompleteMultipartUpload request when transforming to a multipart copy.
+
# __2.20.108__ __2023-07-20__
## __Amazon CodeCatalyst__
- ### Features
@@ -676,7 +1184,7 @@ Special thanks to the following contributors to this release:
## __Contributors__
Special thanks to the following contributors to this release:
-[@bmaizels](https://github.com/bmaizels), [@breader124](https://github.com/breader124)
+[@breader124](https://github.com/breader124), [@bmaizels](https://github.com/bmaizels)
# __2.20.85__ __2023-06-13__
## __AWS CloudTrail__
- ### Features
diff --git a/README.md b/README.md
index 24404a45552..e5589f0ea26 100644
--- a/README.md
+++ b/README.md
@@ -4,7 +4,7 @@
[](https://gitter.im/aws/aws-sdk-java-v2?utm_source=badge&utm_medium=badge&utm_campaign=pr-badge)
[](https://codecov.io/gh/aws/aws-sdk-java-v2)
-[](#contributors-)
+[](#contributors-)
The **AWS SDK for Java 2.0** is a rewrite of 1.0 with some great new features. As with version 1.0,
@@ -52,7 +52,7 @@ To automatically manage module versions (currently all modules have the same ver
software.amazon.awssdkbom
- 2.20.108
+ 2.20.125pomimport
@@ -86,12 +86,12 @@ Alternatively you can add dependencies for the specific services you use only:
software.amazon.awssdkec2
- 2.20.108
+ 2.20.125software.amazon.awssdks3
- 2.20.108
+ 2.20.125
```
@@ -103,7 +103,7 @@ You can import the whole SDK into your project (includes *ALL* services). Please
software.amazon.awssdkaws-sdk-java
- 2.20.108
+ 2.20.125
```
@@ -308,6 +308,10 @@ Thanks goes to these wonderful people ([emoji key](https://allcontributors.org/d
+ * If content length of this {@link AsyncRequestBody} is present, each divided {@link AsyncRequestBody} is delivered to the
+ * subscriber right after it's initialized.
+ *
+ * If content length is null, it is sent after the entire content for that chunk is buffered.
+ * In this case, the configured {@code maxMemoryUsageInBytes} must be larger than or equal to {@code chunkSizeInBytes}.
+ *
+ * @see AsyncRequestBodySplitConfiguration
+ */
+ default SdkPublisher split(AsyncRequestBodySplitConfiguration splitConfiguration) {
+ Validate.notNull(splitConfiguration, "splitConfiguration");
+
+ return SplittingPublisher.builder()
+ .asyncRequestBody(this)
+ .chunkSizeInBytes(splitConfiguration.chunkSizeInBytes())
+ .bufferSizeInBytes(splitConfiguration.bufferSizeInBytes())
+ .build();
+ }
+
+ /**
+ * This is a convenience method that passes an instance of the {@link AsyncRequestBodySplitConfiguration} builder,
+ * avoiding the need to create one manually via {@link AsyncRequestBodySplitConfiguration#builder()}.
+ *
+ * @see #split(AsyncRequestBodySplitConfiguration)
+ */
+ default SdkPublisher split(Consumer splitConfiguration) {
+ Validate.notNull(splitConfiguration, "splitConfiguration");
+ return split(AsyncRequestBodySplitConfiguration.builder().applyMutation(splitConfiguration).build());
+ }
}
diff --git a/core/sdk-core/src/main/java/software/amazon/awssdk/core/async/AsyncRequestBodySplitConfiguration.java b/core/sdk-core/src/main/java/software/amazon/awssdk/core/async/AsyncRequestBodySplitConfiguration.java
new file mode 100644
index 00000000000..fe51f33b4ff
--- /dev/null
+++ b/core/sdk-core/src/main/java/software/amazon/awssdk/core/async/AsyncRequestBodySplitConfiguration.java
@@ -0,0 +1,141 @@
+/*
+ * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License").
+ * You may not use this file except in compliance with the License.
+ * A copy of the License is located at
+ *
+ * http://aws.amazon.com/apache2.0
+ *
+ * or in the "license" file accompanying this file. This file is distributed
+ * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
+ * express or implied. See the License for the specific language governing
+ * permissions and limitations under the License.
+ */
+
+package software.amazon.awssdk.core.async;
+
+import java.util.Objects;
+import software.amazon.awssdk.annotations.SdkPublicApi;
+import software.amazon.awssdk.utils.Validate;
+import software.amazon.awssdk.utils.builder.CopyableBuilder;
+import software.amazon.awssdk.utils.builder.ToCopyableBuilder;
+
+/**
+ * Configuration options for {@link AsyncRequestBody#split} to configure how the SDK
+ * should split an {@link SdkPublisher}.
+ */
+@SdkPublicApi
+public final class AsyncRequestBodySplitConfiguration implements ToCopyableBuilder {
+ private final Long chunkSizeInBytes;
+ private final Long bufferSizeInBytes;
+
+ private AsyncRequestBodySplitConfiguration(DefaultBuilder builder) {
+ this.chunkSizeInBytes = Validate.isPositiveOrNull(builder.chunkSizeInBytes, "chunkSizeInBytes");
+ this.bufferSizeInBytes = Validate.isPositiveOrNull(builder.bufferSizeInBytes, "bufferSizeInBytes");
+ }
+
+ /**
+ * The configured chunk size for each divided {@link AsyncRequestBody}.
+ */
+ public Long chunkSizeInBytes() {
+ return chunkSizeInBytes;
+ }
+
+ /**
+ * The configured maximum buffer size the SDK will use to buffer the content from the source {@link SdkPublisher}.
+ */
+ public Long bufferSizeInBytes() {
+ return bufferSizeInBytes;
+ }
+
+ /**
+ * Create a {@link Builder}, used to create a {@link AsyncRequestBodySplitConfiguration}.
+ */
+ public static Builder builder() {
+ return new DefaultBuilder();
+ }
+
+ @Override
+ public boolean equals(Object o) {
+ if (this == o) {
+ return true;
+ }
+ if (o == null || getClass() != o.getClass()) {
+ return false;
+ }
+
+ AsyncRequestBodySplitConfiguration that = (AsyncRequestBodySplitConfiguration) o;
+
+ if (!Objects.equals(chunkSizeInBytes, that.chunkSizeInBytes)) {
+ return false;
+ }
+ return Objects.equals(bufferSizeInBytes, that.bufferSizeInBytes);
+ }
+
+ @Override
+ public int hashCode() {
+ int result = chunkSizeInBytes != null ? chunkSizeInBytes.hashCode() : 0;
+ result = 31 * result + (bufferSizeInBytes != null ? bufferSizeInBytes.hashCode() : 0);
+ return result;
+ }
+
+ @Override
+ public AsyncRequestBodySplitConfiguration.Builder toBuilder() {
+ return new DefaultBuilder(this);
+ }
+
+ public interface Builder extends CopyableBuilder {
+
+ /**
+ * Configures the size for each divided chunk. The last chunk may be smaller than the configured size. The default value
+ * is 2MB.
+ *
+ * @param chunkSizeInBytes the chunk size in bytes
+ * @return This object for method chaining.
+ */
+ Builder chunkSizeInBytes(Long chunkSizeInBytes);
+
+ /**
+ * The maximum buffer size the SDK will use to buffer the content from the source {@link SdkPublisher}. The default value
+ * is 8MB.
+ *
+ * @param bufferSizeInBytes the buffer size in bytes
+ * @return This object for method chaining.
+ */
+ Builder bufferSizeInBytes(Long bufferSizeInBytes);
+ }
+
+ private static final class DefaultBuilder implements Builder {
+ private Long chunkSizeInBytes;
+ private Long bufferSizeInBytes;
+
+ private DefaultBuilder(AsyncRequestBodySplitConfiguration asyncRequestBodySplitConfiguration) {
+ this.chunkSizeInBytes = asyncRequestBodySplitConfiguration.chunkSizeInBytes;
+ this.bufferSizeInBytes = asyncRequestBodySplitConfiguration.bufferSizeInBytes;
+ }
+
+ private DefaultBuilder() {
+
+ }
+
+ @Override
+ public Builder chunkSizeInBytes(Long chunkSizeInBytes) {
+ this.chunkSizeInBytes = chunkSizeInBytes;
+ return this;
+ }
+
+ @Override
+ public Builder bufferSizeInBytes(Long bufferSizeInBytes) {
+ this.bufferSizeInBytes = bufferSizeInBytes;
+ return this;
+ }
+
+ @Override
+ public AsyncRequestBodySplitConfiguration build() {
+ return new AsyncRequestBodySplitConfiguration(this);
+ }
+ }
+}
diff --git a/core/sdk-core/src/main/java/software/amazon/awssdk/core/internal/async/ChecksumCalculatingAsyncRequestBody.java b/core/sdk-core/src/main/java/software/amazon/awssdk/core/internal/async/ChecksumCalculatingAsyncRequestBody.java
index 101aed7b744..146007927c6 100644
--- a/core/sdk-core/src/main/java/software/amazon/awssdk/core/internal/async/ChecksumCalculatingAsyncRequestBody.java
+++ b/core/sdk-core/src/main/java/software/amazon/awssdk/core/internal/async/ChecksumCalculatingAsyncRequestBody.java
@@ -239,7 +239,7 @@ private static final class SynchronousChunkBuffer {
}
private Iterable buffer(ByteBuffer bytes) {
- return chunkBuffer.bufferAndCreateChunks(bytes);
+ return chunkBuffer.split(bytes);
}
}
diff --git a/core/sdk-core/src/main/java/software/amazon/awssdk/core/internal/async/ChunkBuffer.java b/core/sdk-core/src/main/java/software/amazon/awssdk/core/internal/async/ChunkBuffer.java
index 93d6d09578a..c171b078767 100644
--- a/core/sdk-core/src/main/java/software/amazon/awssdk/core/internal/async/ChunkBuffer.java
+++ b/core/sdk-core/src/main/java/software/amazon/awssdk/core/internal/async/ChunkBuffer.java
@@ -19,10 +19,11 @@
import java.nio.ByteBuffer;
import java.util.ArrayList;
+import java.util.Collections;
import java.util.List;
import java.util.concurrent.atomic.AtomicLong;
import software.amazon.awssdk.annotations.SdkInternalApi;
-import software.amazon.awssdk.utils.BinaryUtils;
+import software.amazon.awssdk.utils.Logger;
import software.amazon.awssdk.utils.Validate;
import software.amazon.awssdk.utils.builder.SdkBuilder;
@@ -31,17 +32,20 @@
*/
@SdkInternalApi
public final class ChunkBuffer {
- private final AtomicLong remainingBytes;
+ private static final Logger log = Logger.loggerFor(ChunkBuffer.class);
+ private final AtomicLong transferredBytes;
private final ByteBuffer currentBuffer;
- private final int bufferSize;
+ private final int chunkSize;
+ private final long totalBytes;
private ChunkBuffer(Long totalBytes, Integer bufferSize) {
Validate.notNull(totalBytes, "The totalBytes must not be null");
int chunkSize = bufferSize != null ? bufferSize : DEFAULT_ASYNC_CHUNK_SIZE;
- this.bufferSize = chunkSize;
+ this.chunkSize = chunkSize;
this.currentBuffer = ByteBuffer.allocate(chunkSize);
- this.remainingBytes = new AtomicLong(totalBytes);
+ this.totalBytes = totalBytes;
+ this.transferredBytes = new AtomicLong(0);
}
public static Builder builder() {
@@ -49,52 +53,97 @@ public static Builder builder() {
}
- // currentBuffer and bufferedList can get over written if concurrent Threads calls this method at the same time.
- public synchronized Iterable bufferAndCreateChunks(ByteBuffer buffer) {
- int startPosition = 0;
- List bufferedList = new ArrayList<>();
- int currentBytesRead = buffer.remaining();
- do {
- int bufferedBytes = currentBuffer.position();
- int availableToRead = bufferSize - bufferedBytes;
- int bytesToMove = Math.min(availableToRead, currentBytesRead - startPosition);
+ /**
+ * Split the input {@link ByteBuffer} into multiple smaller {@link ByteBuffer}s, each of which contains {@link #chunkSize}
+ * worth of bytes. If the last chunk of the input ByteBuffer contains less than {@link #chunkSize} data, the last chunk will
+ * be buffered.
+ */
+ public synchronized Iterable split(ByteBuffer inputByteBuffer) {
- byte[] bytes = BinaryUtils.copyAllBytesFrom(buffer);
- if (bufferedBytes == 0) {
- currentBuffer.put(bytes, startPosition, bytesToMove);
- } else {
- currentBuffer.put(bytes, 0, bytesToMove);
+ if (!inputByteBuffer.hasRemaining()) {
+ return Collections.singletonList(inputByteBuffer);
+ }
+
+ List byteBuffers = new ArrayList<>();
+
+ // If current buffer is not empty, fill the buffer first.
+ if (currentBuffer.position() != 0) {
+ fillCurrentBuffer(inputByteBuffer);
+
+ if (isCurrentBufferFull()) {
+ addCurrentBufferToIterable(byteBuffers, chunkSize);
+ }
+ }
+
+ // If the input buffer is not empty, split the input buffer
+ if (inputByteBuffer.hasRemaining()) {
+ splitRemainingInputByteBuffer(inputByteBuffer, byteBuffers);
+ }
+
+ // If this is the last chunk, add data buffered to the iterable
+ if (isLastChunk()) {
+ int remainingBytesInBuffer = currentBuffer.position();
+ addCurrentBufferToIterable(byteBuffers, remainingBytesInBuffer);
+ }
+ return byteBuffers;
+ }
+
+ private boolean isCurrentBufferFull() {
+ return currentBuffer.position() == chunkSize;
+ }
+
+ /**
+ * Splits the input ByteBuffer to multiple chunks and add them to the iterable.
+ */
+ private void splitRemainingInputByteBuffer(ByteBuffer inputByteBuffer, List byteBuffers) {
+ while (inputByteBuffer.hasRemaining()) {
+ ByteBuffer inputByteBufferCopy = inputByteBuffer.asReadOnlyBuffer();
+ if (inputByteBuffer.remaining() < chunkSize) {
+ currentBuffer.put(inputByteBuffer);
+ break;
}
- startPosition = startPosition + bytesToMove;
-
- // Send the data once the buffer is full
- if (currentBuffer.position() == bufferSize) {
- currentBuffer.position(0);
- ByteBuffer bufferToSend = ByteBuffer.allocate(bufferSize);
- bufferToSend.put(currentBuffer.array(), 0, bufferSize);
- bufferToSend.clear();
- currentBuffer.clear();
- bufferedList.add(bufferToSend);
- remainingBytes.addAndGet(-bufferSize);
+ int newLimit = inputByteBufferCopy.position() + chunkSize;
+ inputByteBufferCopy.limit(newLimit);
+ inputByteBuffer.position(newLimit);
+ byteBuffers.add(inputByteBufferCopy);
+ transferredBytes.addAndGet(chunkSize);
+ }
+ }
+
+ private boolean isLastChunk() {
+ long remainingBytes = totalBytes - transferredBytes.get();
+ return remainingBytes != 0 && remainingBytes == currentBuffer.position();
+ }
+
+ private void addCurrentBufferToIterable(List byteBuffers, int capacity) {
+ ByteBuffer bufferedChunk = ByteBuffer.allocate(capacity);
+ currentBuffer.flip();
+ bufferedChunk.put(currentBuffer);
+ bufferedChunk.flip();
+ byteBuffers.add(bufferedChunk);
+ transferredBytes.addAndGet(bufferedChunk.remaining());
+ currentBuffer.clear();
+ }
+
+ private void fillCurrentBuffer(ByteBuffer inputByteBuffer) {
+ while (currentBuffer.position() < chunkSize) {
+ if (!inputByteBuffer.hasRemaining()) {
+ break;
+ }
+
+ int remainingCapacity = chunkSize - currentBuffer.position();
+
+ if (inputByteBuffer.remaining() < remainingCapacity) {
+ currentBuffer.put(inputByteBuffer);
+ } else {
+ ByteBuffer remainingChunk = inputByteBuffer.asReadOnlyBuffer();
+ int newLimit = inputByteBuffer.position() + remainingCapacity;
+ remainingChunk.limit(newLimit);
+ inputByteBuffer.position(newLimit);
+ currentBuffer.put(remainingChunk);
}
- } while (startPosition < currentBytesRead);
-
- int remainingBytesInBuffer = currentBuffer.position();
-
- // Send the remaining buffer when
- // 1. remainingBytes in buffer are same as the last few bytes to be read.
- // 2. If it is a zero byte and the last byte to be read.
- if (remainingBytes.get() == remainingBytesInBuffer &&
- (buffer.remaining() == 0 || remainingBytesInBuffer > 0)) {
- currentBuffer.clear();
- ByteBuffer trimmedBuffer = ByteBuffer.allocate(remainingBytesInBuffer);
- trimmedBuffer.put(currentBuffer.array(), 0, remainingBytesInBuffer);
- trimmedBuffer.clear();
- bufferedList.add(trimmedBuffer);
- remainingBytes.addAndGet(-remainingBytesInBuffer);
}
- return bufferedList;
}
public interface Builder extends SdkBuilder {
diff --git a/core/sdk-core/src/main/java/software/amazon/awssdk/core/internal/async/SplittingPublisher.java b/core/sdk-core/src/main/java/software/amazon/awssdk/core/internal/async/SplittingPublisher.java
new file mode 100644
index 00000000000..c56d1b6437d
--- /dev/null
+++ b/core/sdk-core/src/main/java/software/amazon/awssdk/core/internal/async/SplittingPublisher.java
@@ -0,0 +1,331 @@
+/*
+ * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License").
+ * You may not use this file except in compliance with the License.
+ * A copy of the License is located at
+ *
+ * http://aws.amazon.com/apache2.0
+ *
+ * or in the "license" file accompanying this file. This file is distributed
+ * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
+ * express or implied. See the License for the specific language governing
+ * permissions and limitations under the License.
+ */
+
+package software.amazon.awssdk.core.internal.async;
+
+import java.nio.ByteBuffer;
+import java.util.Optional;
+import java.util.concurrent.atomic.AtomicBoolean;
+import java.util.concurrent.atomic.AtomicInteger;
+import java.util.concurrent.atomic.AtomicLong;
+import org.reactivestreams.Subscriber;
+import org.reactivestreams.Subscription;
+import software.amazon.awssdk.annotations.SdkInternalApi;
+import software.amazon.awssdk.core.async.AsyncRequestBody;
+import software.amazon.awssdk.core.async.SdkPublisher;
+import software.amazon.awssdk.core.exception.NonRetryableException;
+import software.amazon.awssdk.core.internal.util.NoopSubscription;
+import software.amazon.awssdk.utils.Logger;
+import software.amazon.awssdk.utils.Validate;
+import software.amazon.awssdk.utils.async.SimplePublisher;
+
+/**
+ * Splits an {@link AsyncRequestBody} to multiple smaller {@link AsyncRequestBody}s, each of which publishes a specific portion of
+ * the original data.
+ *
+ *
Alphanumeric string that can be used to distinguish between calls to the IssueCertificate action. Idempotency tokens for IssueCertificate time out after one minute. Therefore, if you call IssueCertificate multiple times with the same idempotency token within one minute, Amazon Web Services Private CA recognizes that you are requesting only one certificate and will issue only one. If you change the idempotency token for each call, Amazon Web Services Private CA recognizes that you are requesting multiple certificates.
"
+ "documentation":"
Alphanumeric string that can be used to distinguish between calls to the IssueCertificate action. Idempotency tokens for IssueCertificate time out after five minutes. Therefore, if you call IssueCertificate multiple times with the same idempotency token within five minutes, Amazon Web Services Private CA recognizes that you are requesting only one certificate and will issue only one. If you change the idempotency token for each call, Amazon Web Services Private CA recognizes that you are requesting multiple certificates.
This is the Amazon Web Services Private Certificate Authority API Reference. It provides descriptions, syntax, and usage examples for each of the actions and data types involved in creating and managing a private certificate authority (CA) for your organization.
The documentation for each action shows the API request parameters and the JSON response. Alternatively, you can use one of the Amazon Web Services SDKs to access an API that is tailored to the programming language or platform that you prefer. For more information, see Amazon Web Services SDKs.
Each Amazon Web Services Private CA API operation has a quota that determines the number of times the operation can be called per second. Amazon Web Services Private CA throttles API requests at different rates depending on the operation. Throttling means that Amazon Web Services Private CA rejects an otherwise valid request because the request exceeds the operation's quota for the number of requests per second. When a request is throttled, Amazon Web Services Private CA returns a ThrottlingException error. Amazon Web Services Private CA does not guarantee a minimum request rate for APIs.
To see an up-to-date list of your Amazon Web Services Private CA quotas, or to request a quota increase, log into your Amazon Web Services account and visit the Service Quotas console.
"
+ "documentation":"
This is the Amazon Web Services Private Certificate Authority API Reference. It provides descriptions, syntax, and usage examples for each of the actions and data types involved in creating and managing a private certificate authority (CA) for your organization.
The documentation for each action shows the API request parameters and the JSON response. Alternatively, you can use one of the Amazon Web Services SDKs to access an API that is tailored to the programming language or platform that you prefer. For more information, see Amazon Web Services SDKs.
Each Amazon Web Services Private CA API operation has a quota that determines the number of times the operation can be called per second. Amazon Web Services Private CA throttles API requests at different rates depending on the operation. Throttling means that Amazon Web Services Private CA rejects an otherwise valid request because the request exceeds the operation's quota for the number of requests per second. When a request is throttled, Amazon Web Services Private CA returns a ThrottlingException error. Amazon Web Services Private CA does not guarantee a minimum request rate for APIs.
To see an up-to-date list of your Amazon Web Services Private CA quotas, or to request a quota increase, log into your Amazon Web Services account and visit the Service Quotas console.
Represents the event action configuration for an element of a Component or ComponentChild. Use for the workflow feature in Amplify Studio that allows you to bind events and actions to components. ActionParameters defines the action that is performed when an event occurs on the component.
For WebSocket APIs, a key-value map specifying request parameters that are passed from the method request to the backend. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the backend. The method request parameter value must match the pattern of method.request.{location}.{name}\n , where \n {location}\n is querystring, path, or header; and \n {name}\n must be a valid and unique method request parameter name.
For HTTP API integrations with a specified integrationSubtype, request parameters are a key-value map specifying parameters that are passed to AWS_PROXY integrations. You can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Working with AWS service integrations for HTTP APIs.
For HTTP API itegrations, without a specified integrationSubtype request parameters are a key-value map specifying how to transform HTTP requests before sending them to backend integrations. The key should follow the pattern <action>:<header|querystring|path>.<location>. The action can be append, overwrite or remove. For values, you can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Transforming API requests and responses.
"
+ "documentation" : "
For WebSocket APIs, a key-value map specifying request parameters that are passed from the method request to the backend. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the backend. The method request parameter value must match the pattern of method.request.{location}.{name}\n , where \n {location}\n is querystring, path, or header; and \n {name}\n must be a valid and unique method request parameter name.
For HTTP API integrations with a specified integrationSubtype, request parameters are a key-value map specifying parameters that are passed to AWS_PROXY integrations. You can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Working with AWS service integrations for HTTP APIs.
For HTTP API integrations, without a specified integrationSubtype request parameters are a key-value map specifying how to transform HTTP requests before sending them to backend integrations. The key should follow the pattern <action>:<header|querystring|path>.<location>. The action can be append, overwrite or remove. For values, you can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Transforming API requests and responses.
For WebSocket APIs, a key-value map specifying request parameters that are passed from the method request to the backend. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the backend. The method request parameter value must match the pattern of method.request.{location}.{name}\n , where \n {location}\n is querystring, path, or header; and \n {name}\n must be a valid and unique method request parameter name.
For HTTP API integrations with a specified integrationSubtype, request parameters are a key-value map specifying parameters that are passed to AWS_PROXY integrations. You can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Working with AWS service integrations for HTTP APIs.
For HTTP API itegrations, without a specified integrationSubtype request parameters are a key-value map specifying how to transform HTTP requests before sending them to backend integrations. The key should follow the pattern <action>:<header|querystring|path>.<location>. The action can be append, overwrite or remove. For values, you can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Transforming API requests and responses.
"
+ "documentation" : "
For WebSocket APIs, a key-value map specifying request parameters that are passed from the method request to the backend. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the backend. The method request parameter value must match the pattern of method.request.{location}.{name}\n , where \n {location}\n is querystring, path, or header; and \n {name}\n must be a valid and unique method request parameter name.
For HTTP API integrations with a specified integrationSubtype, request parameters are a key-value map specifying parameters that are passed to AWS_PROXY integrations. You can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Working with AWS service integrations for HTTP APIs.
For HTTP API integrations, without a specified integrationSubtype request parameters are a key-value map specifying how to transform HTTP requests before sending them to backend integrations. The key should follow the pattern <action>:<header|querystring|path>.<location>. The action can be append, overwrite or remove. For values, you can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Transforming API requests and responses.
For WebSocket APIs, a key-value map specifying request parameters that are passed from the method request to the backend. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the backend. The method request parameter value must match the pattern of method.request.{location}.{name}\n , where \n {location}\n is querystring, path, or header; and \n {name}\n must be a valid and unique method request parameter name.
For HTTP API integrations with a specified integrationSubtype, request parameters are a key-value map specifying parameters that are passed to AWS_PROXY integrations. You can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Working with AWS service integrations for HTTP APIs.
For HTTP API itegrations, without a specified integrationSubtype request parameters are a key-value map specifying how to transform HTTP requests before sending them to backend integrations. The key should follow the pattern <action>:<header|querystring|path>.<location>. The action can be append, overwrite or remove. For values, you can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Transforming API requests and responses.
"
+ "documentation" : "
For WebSocket APIs, a key-value map specifying request parameters that are passed from the method request to the backend. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the backend. The method request parameter value must match the pattern of method.request.{location}.{name}\n , where \n {location}\n is querystring, path, or header; and \n {name}\n must be a valid and unique method request parameter name.
For HTTP API integrations with a specified integrationSubtype, request parameters are a key-value map specifying parameters that are passed to AWS_PROXY integrations. You can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Working with AWS service integrations for HTTP APIs.
For HTTP API integrations, without a specified integrationSubtype request parameters are a key-value map specifying how to transform HTTP requests before sending them to backend integrations. The key should follow the pattern <action>:<header|querystring|path>.<location>. The action can be append, overwrite or remove. For values, you can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Transforming API requests and responses.
For WebSocket APIs, a key-value map specifying request parameters that are passed from the method request to the backend. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the backend. The method request parameter value must match the pattern of method.request.{location}.{name}\n , where \n {location}\n is querystring, path, or header; and \n {name}\n must be a valid and unique method request parameter name.
For HTTP API integrations with a specified integrationSubtype, request parameters are a key-value map specifying parameters that are passed to AWS_PROXY integrations. You can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Working with AWS service integrations for HTTP APIs.
For HTTP API integrations, without a specified integrationSubtype request parameters are a key-value map specifying how to transform HTTP requests before sending them to the backend. The key should follow the pattern <action>:<header|querystring|path>.<location> where action can be append, overwrite or remove. For values, you can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Transforming API requests and responses.
"
+ "documentation" : "
For WebSocket APIs, a key-value map specifying request parameters that are passed from the method request to the backend. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the backend. The method request parameter value must match the pattern of method.request.{location}.{name}\n , where \n {location}\n is querystring, path, or header; and \n {name}\n must be a valid and unique method request parameter name.
For HTTP API integrations with a specified integrationSubtype, request parameters are a key-value map specifying parameters that are passed to AWS_PROXY integrations. You can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Working with AWS service integrations for HTTP APIs.
For HTTP API integrations, without a specified integrationSubtype request parameters are a key-value map specifying how to transform HTTP requests before sending them to the backend. The key should follow the pattern <action>:<header|querystring|path>.<location> where action can be append, overwrite or remove. For values, you can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Transforming API requests and responses.
For WebSocket APIs, a key-value map specifying request parameters that are passed from the method request to the backend. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the backend. The method request parameter value must match the pattern of method.request.{location}.{name}\n , where \n {location}\n is querystring, path, or header; and \n {name}\n must be a valid and unique method request parameter name.
For HTTP API integrations with a specified integrationSubtype, request parameters are a key-value map specifying parameters that are passed to AWS_PROXY integrations. You can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Working with AWS service integrations for HTTP APIs.
For HTTP API integrations, without a specified integrationSubtype request parameters are a key-value map specifying how to transform HTTP requests before sending them to the backend. The key should follow the pattern <action>:<header|querystring|path>.<location> where action can be append, overwrite or remove. For values, you can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Transforming API requests and responses.
"
+ "documentation" : "
For WebSocket APIs, a key-value map specifying request parameters that are passed from the method request to the backend. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the backend. The method request parameter value must match the pattern of method.request.{location}.{name}\n , where \n {location}\n is querystring, path, or header; and \n {name}\n must be a valid and unique method request parameter name.
For HTTP API integrations with a specified integrationSubtype, request parameters are a key-value map specifying parameters that are passed to AWS_PROXY integrations. You can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Working with AWS service integrations for HTTP APIs.
For HTTP API integrations, without a specified integrationSubtype request parameters are a key-value map specifying how to transform HTTP requests before sending them to the backend. The key should follow the pattern <action>:<header|querystring|path>.<location> where action can be append, overwrite or remove. For values, you can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Transforming API requests and responses.
For WebSocket APIs, a key-value map specifying request parameters that are passed from the method request to the backend. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the backend. The method request parameter value must match the pattern of method.request.{location}.{name}\n , where \n {location}\n is querystring, path, or header; and \n {name}\n must be a valid and unique method request parameter name.
For HTTP API integrations with a specified integrationSubtype, request parameters are a key-value map specifying parameters that are passed to AWS_PROXY integrations. You can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Working with AWS service integrations for HTTP APIs.
For HTTP API itegrations, without a specified integrationSubtype request parameters are a key-value map specifying how to transform HTTP requests before sending them to backend integrations. The key should follow the pattern <action>:<header|querystring|path>.<location>. The action can be append, overwrite or remove. For values, you can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Transforming API requests and responses.
"
+ "documentation" : "
For WebSocket APIs, a key-value map specifying request parameters that are passed from the method request to the backend. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the backend. The method request parameter value must match the pattern of method.request.{location}.{name}\n , where \n {location}\n is querystring, path, or header; and \n {name}\n must be a valid and unique method request parameter name.
For HTTP API integrations with a specified integrationSubtype, request parameters are a key-value map specifying parameters that are passed to AWS_PROXY integrations. You can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Working with AWS service integrations for HTTP APIs.
For HTTP API integrations, without a specified integrationSubtype request parameters are a key-value map specifying how to transform HTTP requests before sending them to backend integrations. The key should follow the pattern <action>:<header|querystring|path>.<location>. The action can be append, overwrite or remove. For values, you can provide static values, or map request data, stage variables, or context variables that are evaluated at runtime. To learn more, see Transforming API requests and responses.
"
},
"RequestTemplates" : {
"shape" : "TemplateMap",
diff --git a/services/appconfig/pom.xml b/services/appconfig/pom.xml
index cf727211dfc..05f5db83bc2 100644
--- a/services/appconfig/pom.xml
+++ b/services/appconfig/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTappconfigAWS Java SDK :: Services :: AppConfig
diff --git a/services/appconfigdata/pom.xml b/services/appconfigdata/pom.xml
index 371999a0061..81d7c63e533 100644
--- a/services/appconfigdata/pom.xml
+++ b/services/appconfigdata/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTappconfigdataAWS Java SDK :: Services :: App Config Data
diff --git a/services/appfabric/pom.xml b/services/appfabric/pom.xml
index 30471310556..8238986e198 100644
--- a/services/appfabric/pom.xml
+++ b/services/appfabric/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTappfabricAWS Java SDK :: Services :: App Fabric
diff --git a/services/appflow/pom.xml b/services/appflow/pom.xml
index b79c3d24395..e7969fad583 100644
--- a/services/appflow/pom.xml
+++ b/services/appflow/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTappflowAWS Java SDK :: Services :: Appflow
diff --git a/services/appintegrations/pom.xml b/services/appintegrations/pom.xml
index 2eb0a9ef3e0..feed4b1f497 100644
--- a/services/appintegrations/pom.xml
+++ b/services/appintegrations/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTappintegrationsAWS Java SDK :: Services :: App Integrations
diff --git a/services/applicationautoscaling/pom.xml b/services/applicationautoscaling/pom.xml
index 7772dfeaf5c..902912c3295 100644
--- a/services/applicationautoscaling/pom.xml
+++ b/services/applicationautoscaling/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTapplicationautoscalingAWS Java SDK :: Services :: AWS Application Auto Scaling
diff --git a/services/applicationcostprofiler/pom.xml b/services/applicationcostprofiler/pom.xml
index cd91c777451..c235229a287 100644
--- a/services/applicationcostprofiler/pom.xml
+++ b/services/applicationcostprofiler/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTapplicationcostprofilerAWS Java SDK :: Services :: Application Cost Profiler
diff --git a/services/applicationdiscovery/pom.xml b/services/applicationdiscovery/pom.xml
index 7641d51769d..0f37c85e99b 100644
--- a/services/applicationdiscovery/pom.xml
+++ b/services/applicationdiscovery/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTapplicationdiscoveryAWS Java SDK :: Services :: AWS Application Discovery Service
diff --git a/services/applicationinsights/pom.xml b/services/applicationinsights/pom.xml
index 6de209338a4..95f7a902d9c 100644
--- a/services/applicationinsights/pom.xml
+++ b/services/applicationinsights/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTapplicationinsightsAWS Java SDK :: Services :: Application Insights
diff --git a/services/applicationinsights/src/main/resources/codegen-resources/endpoint-rule-set.json b/services/applicationinsights/src/main/resources/codegen-resources/endpoint-rule-set.json
index 75e13ac5dd1..5aa99ae02e8 100644
--- a/services/applicationinsights/src/main/resources/codegen-resources/endpoint-rule-set.json
+++ b/services/applicationinsights/src/main/resources/codegen-resources/endpoint-rule-set.json
@@ -32,13 +32,12 @@
{
"conditions": [
{
- "fn": "aws.partition",
+ "fn": "isSet",
"argv": [
{
- "ref": "Region"
+ "ref": "Endpoint"
}
- ],
- "assign": "PartitionResult"
+ ]
}
],
"type": "tree",
@@ -46,23 +45,20 @@
{
"conditions": [
{
- "fn": "isSet",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "Endpoint"
- }
+ "ref": "UseFIPS"
+ },
+ true
]
- },
- {
- "fn": "parseURL",
- "argv": [
- {
- "ref": "Endpoint"
- }
- ],
- "assign": "url"
}
],
+ "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "type": "error"
+ },
+ {
+ "conditions": [],
"type": "tree",
"rules": [
{
@@ -71,67 +67,42 @@
"fn": "booleanEquals",
"argv": [
{
- "ref": "UseFIPS"
+ "ref": "UseDualStack"
},
true
]
}
],
- "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
"type": "error"
},
{
"conditions": [],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
- },
- true
- ]
- }
- ],
- "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
- "type": "error"
+ "endpoint": {
+ "url": {
+ "ref": "Endpoint"
},
- {
- "conditions": [],
- "endpoint": {
- "url": {
- "ref": "Endpoint"
- },
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
}
]
- },
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
{
"conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
- },
- true
- ]
- },
- {
- "fn": "booleanEquals",
+ "fn": "isSet",
"argv": [
{
- "ref": "UseDualStack"
- },
- true
+ "ref": "Region"
+ }
]
}
],
@@ -140,90 +111,215 @@
{
"conditions": [
{
- "fn": "booleanEquals",
+ "fn": "aws.partition",
"argv": [
- true,
{
- "fn": "getAttr",
+ "ref": "Region"
+ }
+ ],
+ "assign": "PartitionResult"
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseFIPS"
},
- "supportsFIPS"
+ true
+ ]
+ },
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ {
+ "ref": "UseDualStack"
+ },
+ true
]
}
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsFIPS"
+ ]
+ }
+ ]
+ },
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://applicationinsights-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
+ "type": "error"
+ }
]
},
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseFIPS"
},
- "supportsDualStack"
+ true
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsFIPS"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://applicationinsights-fips.{Region}.{PartitionResult#dnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
]
+ },
+ {
+ "conditions": [],
+ "error": "FIPS is enabled but this partition does not support FIPS",
+ "type": "error"
}
]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
- "endpoint": {
- "url": "https://applicationinsights-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
- },
- {
- "conditions": [],
- "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
},
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseDualStack"
},
- "supportsFIPS"
+ true
]
}
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://applicationinsights.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "error": "DualStack is enabled but this partition does not support DualStack",
+ "type": "error"
+ }
]
- }
- ],
- "type": "tree",
- "rules": [
+ },
{
"conditions": [],
"type": "tree",
@@ -231,7 +327,7 @@
{
"conditions": [],
"endpoint": {
- "url": "https://applicationinsights-fips.{Region}.{PartitionResult#dnsSuffix}",
+ "url": "https://applicationinsights.{Region}.{PartitionResult#dnsSuffix}",
"properties": {},
"headers": {}
},
@@ -240,74 +336,13 @@
]
}
]
- },
- {
- "conditions": [],
- "error": "FIPS is enabled but this partition does not support FIPS",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
- },
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- true,
- {
- "fn": "getAttr",
- "argv": [
- {
- "ref": "PartitionResult"
- },
- "supportsDualStack"
- ]
- }
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
- "endpoint": {
- "url": "https://applicationinsights.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
- },
- {
- "conditions": [],
- "error": "DualStack is enabled but this partition does not support DualStack",
- "type": "error"
}
]
},
{
"conditions": [],
- "endpoint": {
- "url": "https://applicationinsights.{Region}.{PartitionResult#dnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "error": "Invalid Configuration: Missing Region",
+ "type": "error"
}
]
}
diff --git a/services/applicationinsights/src/main/resources/codegen-resources/endpoint-tests.json b/services/applicationinsights/src/main/resources/codegen-resources/endpoint-tests.json
index 7f3fbce2a03..af6b8c47b29 100644
--- a/services/applicationinsights/src/main/resources/codegen-resources/endpoint-tests.json
+++ b/services/applicationinsights/src/main/resources/codegen-resources/endpoint-tests.json
@@ -1,770 +1,29 @@
{
"testCases": [
{
- "documentation": "For region ap-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.ap-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.ap-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.ap-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.ap-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.eu-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.eu-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.eu-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.eu-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.us-gov-east-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-gov-east-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.us-gov-east-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-gov-east-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.us-gov-east-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-gov-east-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.us-gov-east-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-gov-east-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.ca-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ca-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.ca-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ca-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.ca-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ca-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.ca-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ca-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.eu-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.eu-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.eu-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.eu-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.us-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.us-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.us-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.us-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.us-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.us-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.us-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.us-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.af-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "af-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.af-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "af-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.af-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "af-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.af-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "af-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.eu-north-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-north-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.eu-north-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-north-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.eu-north-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-north-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.eu-north-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-north-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.eu-west-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.eu-west-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.eu-west-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.eu-west-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.eu-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.eu-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.eu-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.eu-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.eu-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.eu-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.eu-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.eu-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.ap-northeast-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.ap-northeast-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.ap-northeast-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.ap-northeast-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.ap-northeast-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.ap-northeast-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.ap-northeast-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights.ap-northeast-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.ap-northeast-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.ap-northeast-1.amazonaws.com"
+ "url": "https://applicationinsights.af-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-northeast-1",
- "UseFIPS": true
+ "Region": "af-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.ap-northeast-1.api.aws"
+ "url": "https://applicationinsights.ap-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-northeast-1",
- "UseFIPS": false
+ "Region": "ap-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
@@ -775,542 +34,534 @@
}
},
"params": {
- "UseDualStack": false,
"Region": "ap-northeast-1",
- "UseFIPS": false
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.me-south-1.api.aws"
+ "url": "https://applicationinsights.ap-northeast-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "me-south-1",
- "UseFIPS": true
+ "Region": "ap-northeast-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.me-south-1.amazonaws.com"
+ "url": "https://applicationinsights.ap-northeast-3.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "me-south-1",
- "UseFIPS": true
+ "Region": "ap-northeast-3",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.me-south-1.api.aws"
+ "url": "https://applicationinsights.ap-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "me-south-1",
- "UseFIPS": false
+ "Region": "ap-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.me-south-1.amazonaws.com"
+ "url": "https://applicationinsights.ap-southeast-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "me-south-1",
- "UseFIPS": false
+ "Region": "ap-southeast-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region sa-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.sa-east-1.api.aws"
+ "url": "https://applicationinsights.ap-southeast-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "sa-east-1",
- "UseFIPS": true
+ "Region": "ap-southeast-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region sa-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.sa-east-1.amazonaws.com"
+ "url": "https://applicationinsights.ca-central-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "sa-east-1",
- "UseFIPS": true
+ "Region": "ca-central-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region sa-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.sa-east-1.api.aws"
+ "url": "https://applicationinsights.eu-central-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "sa-east-1",
- "UseFIPS": false
+ "Region": "eu-central-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.sa-east-1.amazonaws.com"
+ "url": "https://applicationinsights.eu-north-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "sa-east-1",
- "UseFIPS": false
+ "Region": "eu-north-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region eu-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.ap-east-1.api.aws"
+ "url": "https://applicationinsights.eu-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-east-1",
- "UseFIPS": true
+ "Region": "eu-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.ap-east-1.amazonaws.com"
+ "url": "https://applicationinsights.eu-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-east-1",
- "UseFIPS": true
+ "Region": "eu-west-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.ap-east-1.api.aws"
+ "url": "https://applicationinsights.eu-west-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-east-1",
- "UseFIPS": false
+ "Region": "eu-west-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.ap-east-1.amazonaws.com"
+ "url": "https://applicationinsights.eu-west-3.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-east-1",
- "UseFIPS": false
+ "Region": "eu-west-3",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.cn-north-1.api.amazonwebservices.com.cn"
+ "url": "https://applicationinsights.me-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "cn-north-1",
- "UseFIPS": true
+ "Region": "me-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.cn-north-1.amazonaws.com.cn"
+ "url": "https://applicationinsights.sa-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "cn-north-1",
- "UseFIPS": true
+ "Region": "sa-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.cn-north-1.api.amazonwebservices.com.cn"
+ "url": "https://applicationinsights.us-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "cn-north-1",
- "UseFIPS": false
+ "Region": "us-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.cn-north-1.amazonaws.com.cn"
+ "url": "https://applicationinsights.us-east-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "cn-north-1",
- "UseFIPS": false
+ "Region": "us-east-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.us-gov-west-1.api.aws"
+ "url": "https://applicationinsights.us-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-gov-west-1",
- "UseFIPS": true
+ "Region": "us-west-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.us-gov-west-1.amazonaws.com"
+ "url": "https://applicationinsights.us-west-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-gov-west-1",
- "UseFIPS": true
+ "Region": "us-west-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.us-gov-west-1.api.aws"
+ "url": "https://applicationinsights-fips.us-east-1.api.aws"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-gov-west-1",
- "UseFIPS": false
+ "Region": "us-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.us-gov-west-1.amazonaws.com"
+ "url": "https://applicationinsights-fips.us-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-gov-west-1",
- "UseFIPS": false
+ "Region": "us-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.ap-southeast-1.api.aws"
+ "url": "https://applicationinsights.us-east-1.api.aws"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-1",
- "UseFIPS": true
+ "Region": "us-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.ap-southeast-1.amazonaws.com"
+ "url": "https://applicationinsights.cn-north-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-1",
- "UseFIPS": true
+ "Region": "cn-north-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.ap-southeast-1.api.aws"
+ "url": "https://applicationinsights.cn-northwest-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-1",
- "UseFIPS": false
+ "Region": "cn-northwest-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.ap-southeast-1.amazonaws.com"
+ "url": "https://applicationinsights-fips.cn-north-1.api.amazonwebservices.com.cn"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-1",
- "UseFIPS": false
+ "Region": "cn-north-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.ap-southeast-2.api.aws"
+ "url": "https://applicationinsights-fips.cn-north-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-2",
- "UseFIPS": true
+ "Region": "cn-north-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.ap-southeast-2.amazonaws.com"
+ "url": "https://applicationinsights.cn-north-1.api.amazonwebservices.com.cn"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-2",
- "UseFIPS": true
+ "Region": "cn-north-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.ap-southeast-2.api.aws"
+ "url": "https://applicationinsights.us-gov-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-2",
- "UseFIPS": false
+ "Region": "us-gov-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.ap-southeast-2.amazonaws.com"
+ "url": "https://applicationinsights.us-gov-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-2",
- "UseFIPS": false
+ "Region": "us-gov-west-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.us-east-1.api.aws"
+ "url": "https://applicationinsights-fips.us-gov-east-1.api.aws"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-1",
- "UseFIPS": true
+ "Region": "us-gov-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.us-east-1.amazonaws.com"
+ "url": "https://applicationinsights-fips.us-gov-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-1",
- "UseFIPS": true
+ "Region": "us-gov-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.us-east-1.api.aws"
+ "url": "https://applicationinsights.us-gov-east-1.api.aws"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-1",
- "UseFIPS": false
+ "Region": "us-gov-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://applicationinsights.us-east-1.amazonaws.com"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-1",
- "UseFIPS": false
+ "Region": "us-iso-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.us-east-2.api.aws"
+ "url": "https://applicationinsights-fips.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-2",
- "UseFIPS": true
+ "Region": "us-iso-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.us-east-2.amazonaws.com"
- }
+ "error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-2",
- "UseFIPS": true
+ "Region": "us-iso-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.us-east-2.api.aws"
+ "url": "https://applicationinsights.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-2",
- "UseFIPS": false
+ "Region": "us-iso-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://applicationinsights.us-east-2.amazonaws.com"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-2",
- "UseFIPS": false
+ "Region": "us-isob-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights-fips.cn-northwest-1.api.amazonwebservices.com.cn"
+ "url": "https://applicationinsights-fips.us-isob-east-1.sc2s.sgov.gov"
}
},
"params": {
- "UseDualStack": true,
- "Region": "cn-northwest-1",
- "UseFIPS": true
+ "Region": "us-isob-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://applicationinsights-fips.cn-northwest-1.amazonaws.com.cn"
- }
+ "error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
- "UseDualStack": false,
- "Region": "cn-northwest-1",
- "UseFIPS": true
+ "Region": "us-isob-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.cn-northwest-1.api.amazonwebservices.com.cn"
+ "url": "https://applicationinsights.us-isob-east-1.sc2s.sgov.gov"
}
},
"params": {
- "UseDualStack": true,
- "Region": "cn-northwest-1",
- "UseFIPS": false
+ "Region": "us-isob-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For custom endpoint with region set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
- "url": "https://applicationinsights.cn-northwest-1.amazonaws.com.cn"
+ "url": "https://example.com"
}
},
"params": {
+ "Region": "us-east-1",
+ "UseFIPS": false,
"UseDualStack": false,
- "Region": "cn-northwest-1",
- "UseFIPS": false
+ "Endpoint": "https://example.com"
}
},
{
- "documentation": "For custom endpoint with fips disabled and dualstack disabled",
+ "documentation": "For custom endpoint with region not set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
"url": "https://example.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-1",
"UseFIPS": false,
+ "UseDualStack": false,
"Endpoint": "https://example.com"
}
},
@@ -1320,9 +571,9 @@
"error": "Invalid Configuration: FIPS and custom endpoint are not supported"
},
"params": {
- "UseDualStack": false,
"Region": "us-east-1",
"UseFIPS": true,
+ "UseDualStack": false,
"Endpoint": "https://example.com"
}
},
@@ -1332,11 +583,17 @@
"error": "Invalid Configuration: Dualstack and custom endpoint are not supported"
},
"params": {
- "UseDualStack": true,
"Region": "us-east-1",
"UseFIPS": false,
+ "UseDualStack": true,
"Endpoint": "https://example.com"
}
+ },
+ {
+ "documentation": "Missing region",
+ "expect": {
+ "error": "Invalid Configuration: Missing Region"
+ }
}
],
"version": "1.0"
diff --git a/services/applicationinsights/src/main/resources/codegen-resources/paginators-1.json b/services/applicationinsights/src/main/resources/codegen-resources/paginators-1.json
index 2f237e4c958..c9e1dab5611 100644
--- a/services/applicationinsights/src/main/resources/codegen-resources/paginators-1.json
+++ b/services/applicationinsights/src/main/resources/codegen-resources/paginators-1.json
@@ -29,6 +29,11 @@
"input_token": "NextToken",
"output_token": "NextToken",
"limit_key": "MaxResults"
+ },
+ "ListWorkloads": {
+ "input_token": "NextToken",
+ "output_token": "NextToken",
+ "limit_key": "MaxResults"
}
}
}
diff --git a/services/applicationinsights/src/main/resources/codegen-resources/service-2.json b/services/applicationinsights/src/main/resources/codegen-resources/service-2.json
index 00880d07607..7d1cbec399d 100644
--- a/services/applicationinsights/src/main/resources/codegen-resources/service-2.json
+++ b/services/applicationinsights/src/main/resources/codegen-resources/service-2.json
@@ -14,6 +14,22 @@
"uid":"application-insights-2018-11-25"
},
"operations":{
+ "AddWorkload":{
+ "name":"AddWorkload",
+ "http":{
+ "method":"POST",
+ "requestUri":"/"
+ },
+ "input":{"shape":"AddWorkloadRequest"},
+ "output":{"shape":"AddWorkloadResponse"},
+ "errors":[
+ {"shape":"ResourceInUseException"},
+ {"shape":"ResourceNotFoundException"},
+ {"shape":"ValidationException"},
+ {"shape":"InternalServerException"}
+ ],
+ "documentation":"
Adds a workload to a component. Each component can have at most five workloads.
Retrieve a list of the tags (keys and values) that are associated with a specified application. A tag is a label that you optionally define and associate with an application. Each tag consists of a required tag key and an optional associated tag value. A tag key is a general label that acts as a category for more specific tag values. A tag value acts as a descriptor within a tag key.
The NextToken value returned from a previous paginated ListConfigurationHistory request where MaxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the NextToken value. This value is null when there are no more results to return.
Specifies how the problem was resolved. If the value is AUTOMATIC, the system resolved the problem. If the value is MANUAL, the user resolved the problem. If the value is UNRESOLVED, then the problem is not resolved.
"
}
},
"documentation":"
Describes a problem that is detected by correlating observations.
The visibility of a problem. When you pass a value of IGNORED, the problem is removed from the default view, and all notifications for the problem are suspended. When VISIBLE is passed, the IGNORED action is reversed.
Completes the lifecycle action for the specified token or instance with the specified result.
This step is a part of the procedure for adding a lifecycle hook to an Auto Scaling group:
(Optional) Create a launch template or launch configuration with a user data script that runs while an instance is in a wait state due to a lifecycle hook.
(Optional) Create a Lambda function and a rule that allows Amazon EventBridge to invoke your Lambda function when an instance is put into a wait state due to a lifecycle hook.
(Optional) Create a notification target and an IAM role. The target can be either an Amazon SQS queue or an Amazon SNS topic. The role allows Amazon EC2 Auto Scaling to publish lifecycle notifications to the target.
Create the lifecycle hook. Specify whether the hook is used when the instances launch or terminate.
If you need more time, record the lifecycle action heartbeat to keep the instance in a wait state.
If you finish before the timeout period ends, send a callback by using the CompleteLifecycleAction API call.
Completes the lifecycle action for the specified token or instance with the specified result.
This step is a part of the procedure for adding a lifecycle hook to an Auto Scaling group:
(Optional) Create a launch template or launch configuration with a user data script that runs while an instance is in a wait state due to a lifecycle hook.
(Optional) Create a Lambda function and a rule that allows Amazon EventBridge to invoke your Lambda function when an instance is put into a wait state due to a lifecycle hook.
(Optional) Create a notification target and an IAM role. The target can be either an Amazon SQS queue or an Amazon SNS topic. The role allows Amazon EC2 Auto Scaling to publish lifecycle notifications to the target.
Create the lifecycle hook. Specify whether the hook is used when the instances launch or terminate.
If you need more time, record the lifecycle action heartbeat to keep the instance in a wait state.
If you finish before the timeout period ends, send a callback by using the CompleteLifecycleAction API call.
Detaches one or more traffic sources from the specified Auto Scaling group.
When you detach a taffic, it enters the Removing state while deregistering the instances in the group. When all instances are deregistered, then you can no longer describe the traffic source using the DescribeTrafficSources API call. The instances continue to run.
"
+ "documentation":"
Detaches one or more traffic sources from the specified Auto Scaling group.
When you detach a traffic source, it enters the Removing state while deregistering the instances in the group. When all instances are deregistered, then you can no longer describe the traffic source using the DescribeTrafficSources API call. The instances continue to run.
The amount by which to scale, based on the specified adjustment type. A positive value adds to the current capacity while a negative number removes from the current capacity. For exact capacity, you must specify a positive value.
Required if the policy type is SimpleScaling. (Not used with any other policy type.)
"
+ "documentation":"
The amount by which to scale, based on the specified adjustment type. A positive value adds to the current capacity while a negative number removes from the current capacity. For exact capacity, you must specify a non-negative value.
Required if the policy type is SimpleScaling. (Not used with any other policy type.)
(Optional) Indicates whether to roll back the Auto Scaling group to its previous configuration if the instance refresh fails. The default is false.
A rollback is not supported in the following situations:
There is no desired configuration specified for the instance refresh.
The Auto Scaling group has a launch template that uses an Amazon Web Services Systems Manager parameter instead of an AMI ID for the ImageId property.
The Auto Scaling group uses the launch template's $Latest or $Default version.
"
+ "documentation":"
(Optional) Indicates whether to roll back the Auto Scaling group to its previous configuration if the instance refresh fails or a CloudWatch alarm threshold is met. The default is false.
A rollback is not supported in the following situations:
There is no desired configuration specified for the instance refresh.
The Auto Scaling group has a launch template that uses an Amazon Web Services Systems Manager parameter instead of an AMI ID for the ImageId property.
The Auto Scaling group uses the launch template's $Latest or $Default version.
(Optional) The CloudWatch alarm specification. CloudWatch alarms can be used to identify any issues and fail the operation if an alarm threshold is met.
"
}
},
"documentation":"
Describes the preferences for an instance refresh.
Sets your preferences for the instance refresh so that it performs as expected when you start it. Includes the instance warmup time, the minimum healthy percentage, and the behaviors that you want Amazon EC2 Auto Scaling to use if instances that are in Standby state or protected from scale in are found. You can also choose to enable additional features, such as the following:
Auto rollback
Checkpoints
Skip matching
"
+ "documentation":"
Sets your preferences for the instance refresh so that it performs as expected when you start it. Includes the instance warmup time, the minimum healthy percentage, and the behaviors that you want Amazon EC2 Auto Scaling to use if instances that are in Standby state or protected from scale in are found. You can also choose to enable additional features, such as the following:
The amount by which to scale, based on the specified adjustment type. A positive value adds to the current capacity while a negative number removes from the current capacity.
The amount by which to scale. The adjustment is based on the value that you specified in the AdjustmentType property (either an absolute number or a percentage). A positive value adds to the current capacity and a negative number subtracts from the current capacity.
"
+ "documentation":"
The amount by which to scale, based on the specified adjustment type. A positive value adds to the current capacity while a negative number removes from the current capacity. For exact capacity, you must specify a non-negative value.
"
}
},
"documentation":"
Describes information used to create a step adjustment for a step scaling policy.
For the following examples, suppose that you have an alarm with a breach threshold of 50:
To trigger the adjustment when the metric is greater than or equal to 50 and less than 60, specify a lower bound of 0 and an upper bound of 10.
To trigger the adjustment when the metric is greater than 40 and less than or equal to 50, specify a lower bound of -10 and an upper bound of 0.
There are a few rules for the step adjustments for your step policy:
The ranges of your step adjustments can't overlap or have a gap.
At most, one step adjustment can have a null lower bound. If one step adjustment has a negative lower bound, then there must be a step adjustment with a null lower bound.
At most, one step adjustment can have a null upper bound. If one step adjustment has a positive upper bound, then there must be a step adjustment with a null upper bound.
The upper and lower bound can't be null in the same step adjustment.
For more information, see Step adjustments in the Amazon EC2 Auto Scaling User Guide.
This action creates a legal hold on a recovery point (backup). A legal hold is a restraint on altering or deleting a backup until an authorized user cancels the legal hold. Any actions to delete or disassociate a recovery point will fail with an error if one or more active legal holds are on the recovery point.
Returns an array of resources successfully backed up by Backup, including the time the resource was saved, an Amazon Resource Name (ARN) of the resource, and a resource type.
A value in minutes after a backup is scheduled before a job will be canceled if it doesn't start successfully. This value is optional. If this value is included, it must be at least 60 minutes to avoid errors.
During the start window, the backup job status remains in CREATED status until it has successfully begun or until the start window time has run out. If within the start window time Backup receives an error that allows the job to be retried, Backup will automatically retry to begin the job at least every 10 minutes until the backup successfully begins (the job status changes to RUNNING) or until the job status changes to EXPIRED (which is expected to occur when the start window time is over).
"
+ "documentation":"
A value in minutes after a backup is scheduled before a job will be canceled if it doesn't start successfully. This value is optional. If this value is included, it must be at least 60 minutes to avoid errors.
This parameter has a maximum value of 100 years (52,560,000 minutes).
During the start window, the backup job status remains in CREATED status until it has successfully begun or until the start window time has run out. If within the start window time Backup receives an error that allows the job to be retried, Backup will automatically retry to begin the job at least every 10 minutes until the backup successfully begins (the job status changes to RUNNING) or until the job status changes to EXPIRED (which is expected to occur when the start window time is over).
The lifecycle defines when a protected resource is transitioned to cold storage and when it expires. Backup will transition and expire backups automatically according to the lifecycle that you define.
Backups transitioned to cold storage must be stored in cold storage for a minimum of 90 days. Therefore, the “retention” setting must be 90 days greater than the “transition to cold after days” setting. The “transition to cold after days” setting cannot be changed after a backup has been transitioned to cold.
Resource types that are able to be transitioned to cold storage are listed in the \"Lifecycle to cold storage\" section of the Feature availability by resource table. Backup ignores this expression for other resource types.
"
+ "documentation":"
The lifecycle defines when a protected resource is transitioned to cold storage and when it expires. Backup will transition and expire backups automatically according to the lifecycle that you define.
Backups transitioned to cold storage must be stored in cold storage for a minimum of 90 days. Therefore, the “retention” setting must be 90 days greater than the “transition to cold after days” setting. The “transition to cold after days” setting cannot be changed after a backup has been transitioned to cold.
Resource types that are able to be transitioned to cold storage are listed in the \"Lifecycle to cold storage\" section of the Feature availability by resource table. Backup ignores this expression for other resource types.
This parameter has a maximum value of 100 years (36,500 days).
This setting specifies the minimum retention period that the vault retains its recovery points. If this parameter is not specified, no minimum retention period is enforced.
If specified, any backup or copy job to the vault must have a lifecycle policy with a retention period equal to or longer than the minimum retention period. If a job retention period is shorter than that minimum retention period, then the vault fails the backup or copy job, and you should either modify your lifecycle settings or use a different vault.
This is the setting that specifies the maximum retention period that the vault retains its recovery points. If this parameter is not specified, Backup does not enforce a maximum retention period on the recovery points in the vault (allowing indefinite storage).
If specified, any backup or copy job to the vault must have a lifecycle policy with a retention period equal to or shorter than the maximum retention period. If the job retention period is longer than that maximum retention period, then the vault fails the backup or copy job, and you should either modify your lifecycle settings or use a different vault.
The name of a logical container where backups are stored. Logically air-gapped backup vaults are identified by names that are unique to the account used to create them and the Region where they are created. They consist of lowercase letters, numbers, and hyphens.
This value is in Unix format, Coordinated Universal Time (UTC), and accurate to milliseconds. For example, the value 1516925490.087 represents Friday, January 26, 2018 12:11:30.087 AM.
The name of a logical container where backups are stored. Backup vaults are identified by names that are unique to the account used to create them and the Amazon Web Services Region where they are created. They consist of lowercase letters, numbers, and hyphens.
The server-side encryption key that is used to protect your backups; for example, arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab.
"
@@ -2836,6 +2933,12 @@
"documentation":"
An Amazon Resource Name (ARN) that uniquely identifies a recovery point; for example, arn:aws:backup:us-east-1:123456789012:recovery-point:1EB3B5E7-9EB0-435A-A80B-108B488B0D45.
An Amazon Resource Name (ARN) that uniquely identifies a recovery point; for example, arn:aws:backup:us-east-1:123456789012:recovery-point:1EB3B5E7-9EB0-435A-A80B-108B488B0D45.
The next item following a partial list of returned items. For example, if a request is made to return maxResults number of items, NextToken allows you to return more items in your list starting at the location pointed to by the next token.
The next item following a partial list of returned items. For example, if a request is made to return maxResults number of items, NextToken allows you to return more items in your list starting at the location pointed to by the next token.
The next item following a partial list of returned items. For example, if a request is made to return maxResults number of items, NextToken allows you to return more items in your list starting at the location pointed to by the next token.
The next item following a partial list of returned items. For example, if a request is made to return maxResults number of items, NextToken allows you to return more items in your list starting at the location pointed to by the next token.
A value in minutes after a backup is scheduled before a job will be canceled if it doesn't start successfully. This value is optional, and the default is 8 hours. If this value is included, it must be at least 60 minutes to avoid errors.
During the start window, the backup job status remains in CREATED status until it has successfully begun or until the start window time has run out. If within the start window time Backup receives an error that allows the job to be retried, Backup will automatically retry to begin the job at least every 10 minutes until the backup successfully begins (the job status changes to RUNNING) or until the job status changes to EXPIRED (which is expected to occur when the start window time is over).
"
+ "documentation":"
A value in minutes after a backup is scheduled before a job will be canceled if it doesn't start successfully. This value is optional, and the default is 8 hours. If this value is included, it must be at least 60 minutes to avoid errors.
This parameter has a maximum value of 100 years (52,560,000 minutes).
During the start window, the backup job status remains in CREATED status until it has successfully begun or until the start window time has run out. If within the start window time Backup receives an error that allows the job to be retried, Backup will automatically retry to begin the job at least every 10 minutes until the backup successfully begins (the job status changes to RUNNING) or until the job status changes to EXPIRED (which is expected to occur when the start window time is over).
A value in minutes during which a successfully started backup must complete, or else Backup will cancel the job. This value is optional. This value begins counting down from when the backup was scheduled. It does not add additional time for StartWindowMinutes, or if the backup started later than scheduled.
"
+ "documentation":"
A value in minutes during which a successfully started backup must complete, or else Backup will cancel the job. This value is optional. This value begins counting down from when the backup was scheduled. It does not add additional time for StartWindowMinutes, or if the backup started later than scheduled.
Like StartWindowMinutes, this parameter has a maximum value of 100 years (52,560,000 minutes).
The lifecycle defines when a protected resource is transitioned to cold storage and when it expires. Backup will transition and expire backups automatically according to the lifecycle that you define.
Backups transitioned to cold storage must be stored in cold storage for a minimum of 90 days. Therefore, the “retention” setting must be 90 days greater than the “transition to cold after days” setting. The “transition to cold after days” setting cannot be changed after a backup has been transitioned to cold.
Resource types that are able to be transitioned to cold storage are listed in the \"Lifecycle to cold storage\" section of the Feature availability by resource table. Backup ignores this expression for other resource types.
"
+ "documentation":"
The lifecycle defines when a protected resource is transitioned to cold storage and when it expires. Backup will transition and expire backups automatically according to the lifecycle that you define.
Backups transitioned to cold storage must be stored in cold storage for a minimum of 90 days. Therefore, the “retention” setting must be 90 days greater than the “transition to cold after days” setting. The “transition to cold after days” setting cannot be changed after a backup has been transitioned to cold.
Resource types that are able to be transitioned to cold storage are listed in the \"Lifecycle to cold storage\" section of the Feature availability by resource table. Backup ignores this expression for other resource types.
This parameter has a maximum value of 100 years (36,500 days).
Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources.
In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price.
Multi-node parallel jobs aren't supported on Spot Instances.
In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIs in the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironments operation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instance in the Amazon Elastic Container Service Developer Guide.
To create a compute environment that uses EKS resources, the caller must have permissions to call eks:DescribeCluster.
Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps:
Create a new compute environment with the new AMI.
Add the compute environment to an existing job queue.
Remove the earlier compute environment from your job queue.
Delete the earlier compute environment.
In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules:
Either don't set the service role (serviceRole) parameter or set it to the AWSBatchServiceRole service-linked role.
Set the allocation strategy (allocationStrategy) parameter to BEST_FIT_PROGRESSIVE or SPOT_CAPACITY_OPTIMIZED.
Set the update to latest image version (updateToLatestImageVersion) parameter to true. The updateToLatestImageVersion parameter is used when you update a compute environment. This parameter is ignored when you create a compute environment.
Don't specify an AMI ID in imageId, imageIdOverride (in ec2Configuration), or in the launch template (launchTemplate). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest, by setting either a new default version for the launch template (if $Default) or by adding a new version to the launch template (if $Latest).
If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template (launchTemplate) is set to $Latest or $Default, the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
"
+ "documentation":"
Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources.
In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch template that you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price.
Multi-node parallel jobs aren't supported on Spot Instances.
In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIs in the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironments operation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instance in the Amazon Elastic Container Service Developer Guide.
To create a compute environment that uses EKS resources, the caller must have permissions to call eks:DescribeCluster.
Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps:
Create a new compute environment with the new AMI.
Add the compute environment to an existing job queue.
Remove the earlier compute environment from your job queue.
Delete the earlier compute environment.
In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules:
Either don't set the service role (serviceRole) parameter or set it to the AWSBatchServiceRole service-linked role.
Set the allocation strategy (allocationStrategy) parameter to BEST_FIT_PROGRESSIVE, SPOT_CAPACITY_OPTIMIZED, or SPOT_PRICE_CAPACITY_OPTIMIZED.
Set the update to latest image version (updateToLatestImageVersion) parameter to true. The updateToLatestImageVersion parameter is used when you update a compute environment. This parameter is ignored when you create a compute environment.
Don't specify an AMI ID in imageId, imageIdOverride (in ec2Configuration), or in the launch template (launchTemplate). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest, by setting either a new default version for the launch template (if $Default) or by adding a new version to the launch template (if $Latest).
If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template (launchTemplate) is set to $Latest or $Default, the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
The allocation strategy to use for the compute resource if not enough instances of the best fitting instance type can be allocated. This might be because of availability of the instance type in the Region or Amazon EC2 service limits. For more information, see Allocation strategies in the Batch User Guide.
This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
BEST_FIT (default)
Batch selects an instance type that best fits the needs of the jobs with a preference for the lowest-cost instance type. If additional instances of the selected instance type aren't available, Batch waits for the additional instances to be available. If there aren't enough instances available or the user is reaching Amazon EC2 service limits, additional jobs aren't run until the currently running jobs are completed. This allocation strategy keeps costs lower but can limit scaling. If you're using Spot Fleets with BEST_FIT, the Spot Fleet IAM Role must be specified. Compute resources that use a BEST_FIT allocation strategy don't support infrastructure updates and can't update some parameters. For more information, see Updating compute environments in the Batch User Guide.
BEST_FIT_PROGRESSIVE
Batch selects additional instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types with lower cost vCPUs. If additional instances of the previously selected instance types aren't available, Batch selects new instance types.
SPOT_CAPACITY_OPTIMIZED
Batch selects one or more instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types that are less likely to be interrupted. This allocation strategy is only available for Spot Instance compute resources.
With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance.
"
+ "documentation":"
The allocation strategy to use for the compute resource if not enough instances of the best fitting instance type can be allocated. This might be because of availability of the instance type in the Region or Amazon EC2 service limits. For more information, see Allocation strategies in the Batch User Guide.
This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
BEST_FIT (default)
Batch selects an instance type that best fits the needs of the jobs with a preference for the lowest-cost instance type. If additional instances of the selected instance type aren't available, Batch waits for the additional instances to be available. If there aren't enough instances available or the user is reaching Amazon EC2 service limits, additional jobs aren't run until the currently running jobs are completed. This allocation strategy keeps costs lower but can limit scaling. If you're using Spot Fleets with BEST_FIT, the Spot Fleet IAM Role must be specified. Compute resources that use a BEST_FIT allocation strategy don't support infrastructure updates and can't update some parameters. For more information, see Updating compute environments in the Batch User Guide.
BEST_FIT_PROGRESSIVE
Batch selects additional instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types with lower cost vCPUs. If additional instances of the previously selected instance types aren't available, Batch selects new instance types.
SPOT_CAPACITY_OPTIMIZED
Batch selects one or more instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types that are less likely to be interrupted. This allocation strategy is only available for Spot Instance compute resources.
SPOT_PRICE_CAPACITY_OPTIMIZED
The price and capacity optimized allocation strategy looks at both price and capacity to select the Spot Instance pools that are the least likely to be interrupted and have the lowest possible price. This allocation strategy is only available for Spot Instance compute resources.
With BEST_FIT_PROGRESSIVE,SPOT_CAPACITY_OPTIMIZED and SPOT_PRICE_CAPACITY_OPTIMIZED strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance.
The maximum number of vCPUs that a compute environment can support.
With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED allocation strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance. For example, no more than a single instance from among those specified in your compute environment is allocated.
"
+ "documentation":"
The maximum number of vCPUs that a compute environment can support.
With BEST_FIT_PROGRESSIVE, SPOT_CAPACITY_OPTIMIZED and SPOT_PRICE_CAPACITY_OPTIMIZED allocation strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance. For example, no more than a single instance from among those specified in your compute environment is allocated.
The maximum number of Amazon EC2 vCPUs that an environment can reach.
With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED allocation strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance. That is, no more than a single instance from among those specified in your compute environment.
"
+ "documentation":"
The maximum number of Amazon EC2 vCPUs that an environment can reach.
With BEST_FIT_PROGRESSIVE, SPOT_CAPACITY_OPTIMIZED, and SPOT_PRICE_CAPACITY_OPTIMIZED allocation strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance. That is, no more than a single instance from among those specified in your compute environment.
The allocation strategy to use for the compute resource if there's not enough instances of the best fitting instance type that can be allocated. This might be because of availability of the instance type in the Region or Amazon EC2 service limits. For more information, see Allocation strategies in the Batch User Guide.
When updating a compute environment, changing the allocation strategy requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. BEST_FIT isn't supported when updating a compute environment.
This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
BEST_FIT_PROGRESSIVE
Batch selects additional instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types with lower cost vCPUs. If additional instances of the previously selected instance types aren't available, Batch selects new instance types.
SPOT_CAPACITY_OPTIMIZED
Batch selects one or more instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types that are less likely to be interrupted. This allocation strategy is only available for Spot Instance compute resources.
With both BEST_FIT_PROGRESSIVE and SPOT_CAPACITY_OPTIMIZED strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance.
"
+ "documentation":"
The allocation strategy to use for the compute resource if there's not enough instances of the best fitting instance type that can be allocated. This might be because of availability of the instance type in the Region or Amazon EC2 service limits. For more information, see Allocation strategies in the Batch User Guide.
When updating a compute environment, changing the allocation strategy requires an infrastructure update of the compute environment. For more information, see Updating compute environments in the Batch User Guide. BEST_FIT isn't supported when updating a compute environment.
This parameter isn't applicable to jobs that are running on Fargate resources. Don't specify it.
BEST_FIT_PROGRESSIVE
Batch selects additional instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types with lower cost vCPUs. If additional instances of the previously selected instance types aren't available, Batch selects new instance types.
SPOT_CAPACITY_OPTIMIZED
Batch selects one or more instance types that are large enough to meet the requirements of the jobs in the queue. Its preference is for instance types that are less likely to be interrupted. This allocation strategy is only available for Spot Instance compute resources.
SPOT_PRICE_CAPACITY_OPTIMIZED
The price and capacity optimized allocation strategy looks at both price and capacity to select the Spot Instance pools that are the least likely to be interrupted and have the lowest possible price. This allocation strategy is only available for Spot Instance compute resources.
With both BEST_FIT_PROGRESSIVE, SPOT_CAPACITY_OPTIMIZED, and SPOT_PRICE_CAPACITY_OPTIMIZED strategies using On-Demand or Spot Instances, and the BEST_FIT strategy using Spot Instances, Batch might need to exceed maxvCpus to meet your capacity requirements. In this event, Batch never exceeds maxvCpus by more than a single instance.
The account IDs that make up the billing group. Account IDs must be a part of the consolidated billing family, and not associated with another billing group.
"
+ "documentation":"
The account IDs that make up the billing group. Account IDs must be a part of the consolidated billing family, and not associated with another billing group.
A summary report of actual Amazon Web Services charges and calculated Amazon Web Services charges, based on the associated pricing plan of a billing group.
"
+ "documentation":"
A summary report of actual Amazon Web Services charges and calculated Amazon Web Services charges, based on the associated pricing plan of a billing group.
Operation is the specific Amazon Web Services action covered by this line item. This describes the specific usage of the line item.
If the Scope attribute is set to SKU, this attribute indicates which operation the PricingRule is modifying. For example, a value of RunInstances:0202 indicates the operation of running an Amazon EC2 instance.
The name of a budget. The name must be unique within an account. The : and \\ characters aren't allowed in BudgetName.
"
+ "documentation":"
The name of a budget. The name must be unique within an account. The : and \\ characters, and the \"/action/\" substring, aren't allowed in BudgetName.
The runtime configuration for the S3 recording sink.
"
+ "documentation":"
The runtime configuration for the S3 recording sink. If specified, the settings in this structure override any settings in S3RecordingSinkConfiguration.
Error reasons for collaboration analysis templates that could not be retrieved. One error is returned for every collaboration analysis template that could not be retrieved.
The types of analysis rules associated with this configured table. Valid values are `AGGREGATION` and `LIST`. Currently, only one analysis rule may be associated with a configured table.
"
+ "documentation":"
The types of analysis rules associated with this configured table. Currently, only one analysis rule may be associated with a configured table.
An optional label that you can assign to a resource when you create it. Each tag consists of a key and an optional value, both of which you define. When you use tagging, you can also use tag-based access control in IAM policies to control access to this resource.
"
+ }
+ }
+ },
"UpdateCollaborationInput":{
"type":"structure",
"required":["collaborationIdentifier"],
@@ -3695,7 +4603,8 @@
"enum":[
"FIELD_VALIDATION_FAILED",
"INVALID_CONFIGURATION",
- "INVALID_QUERY"
+ "INVALID_QUERY",
+ "IAM_SYNCHRONIZATION_DELAY"
]
}
},
diff --git a/services/cloud9/pom.xml b/services/cloud9/pom.xml
index 71733ad3f0c..a777b374e60 100644
--- a/services/cloud9/pom.xml
+++ b/services/cloud9/pom.xml
@@ -20,7 +20,7 @@
servicessoftware.amazon.awssdk
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOT4.0.0cloud9
diff --git a/services/cloud9/src/main/resources/codegen-resources/endpoint-rule-set.json b/services/cloud9/src/main/resources/codegen-resources/endpoint-rule-set.json
index 7f41719e6ef..fbaea21794b 100644
--- a/services/cloud9/src/main/resources/codegen-resources/endpoint-rule-set.json
+++ b/services/cloud9/src/main/resources/codegen-resources/endpoint-rule-set.json
@@ -3,7 +3,7 @@
"parameters": {
"Region": {
"builtIn": "AWS::Region",
- "required": true,
+ "required": false,
"documentation": "The AWS region used to dispatch the request.",
"type": "String"
},
@@ -32,13 +32,12 @@
{
"conditions": [
{
- "fn": "aws.partition",
+ "fn": "isSet",
"argv": [
{
- "ref": "Region"
+ "ref": "Endpoint"
}
- ],
- "assign": "PartitionResult"
+ ]
}
],
"type": "tree",
@@ -46,14 +45,20 @@
{
"conditions": [
{
- "fn": "isSet",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "Endpoint"
- }
+ "ref": "UseFIPS"
+ },
+ true
]
}
],
+ "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "type": "error"
+ },
+ {
+ "conditions": [],
"type": "tree",
"rules": [
{
@@ -62,67 +67,42 @@
"fn": "booleanEquals",
"argv": [
{
- "ref": "UseFIPS"
+ "ref": "UseDualStack"
},
true
]
}
],
- "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
"type": "error"
},
{
"conditions": [],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
- },
- true
- ]
- }
- ],
- "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
- "type": "error"
+ "endpoint": {
+ "url": {
+ "ref": "Endpoint"
},
- {
- "conditions": [],
- "endpoint": {
- "url": {
- "ref": "Endpoint"
- },
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
}
]
- },
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
{
"conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
- },
- true
- ]
- },
- {
- "fn": "booleanEquals",
+ "fn": "isSet",
"argv": [
{
- "ref": "UseDualStack"
- },
- true
+ "ref": "Region"
+ }
]
}
],
@@ -131,90 +111,215 @@
{
"conditions": [
{
- "fn": "booleanEquals",
+ "fn": "aws.partition",
"argv": [
- true,
{
- "fn": "getAttr",
+ "ref": "Region"
+ }
+ ],
+ "assign": "PartitionResult"
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ {
+ "ref": "UseFIPS"
+ },
+ true
+ ]
+ },
+ {
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseDualStack"
+ },
+ true
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsFIPS"
+ ]
+ }
+ ]
},
- "supportsFIPS"
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://cloud9-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
]
+ },
+ {
+ "conditions": [],
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
+ "type": "error"
}
]
},
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseFIPS"
},
- "supportsDualStack"
+ true
]
}
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsFIPS"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://cloud9-fips.{Region}.{PartitionResult#dnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "error": "FIPS is enabled but this partition does not support FIPS",
+ "type": "error"
+ }
]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
- "endpoint": {
- "url": "https://cloud9-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
- },
- {
- "conditions": [],
- "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
},
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseDualStack"
},
- "supportsFIPS"
+ true
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://cloud9.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
]
+ },
+ {
+ "conditions": [],
+ "error": "DualStack is enabled but this partition does not support DualStack",
+ "type": "error"
}
]
- }
- ],
- "type": "tree",
- "rules": [
+ },
{
"conditions": [],
"type": "tree",
@@ -222,7 +327,7 @@
{
"conditions": [],
"endpoint": {
- "url": "https://cloud9-fips.{Region}.{PartitionResult#dnsSuffix}",
+ "url": "https://cloud9.{Region}.{PartitionResult#dnsSuffix}",
"properties": {},
"headers": {}
},
@@ -231,74 +336,13 @@
]
}
]
- },
- {
- "conditions": [],
- "error": "FIPS is enabled but this partition does not support FIPS",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
- },
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- true,
- {
- "fn": "getAttr",
- "argv": [
- {
- "ref": "PartitionResult"
- },
- "supportsDualStack"
- ]
- }
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
- "endpoint": {
- "url": "https://cloud9.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
- },
- {
- "conditions": [],
- "error": "DualStack is enabled but this partition does not support DualStack",
- "type": "error"
}
]
},
{
"conditions": [],
- "endpoint": {
- "url": "https://cloud9.{Region}.{PartitionResult#dnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "error": "Invalid Configuration: Missing Region",
+ "type": "error"
}
]
}
diff --git a/services/cloud9/src/main/resources/codegen-resources/endpoint-tests.json b/services/cloud9/src/main/resources/codegen-resources/endpoint-tests.json
index 247e617ca0f..b28ec45a62f 100644
--- a/services/cloud9/src/main/resources/codegen-resources/endpoint-tests.json
+++ b/services/cloud9/src/main/resources/codegen-resources/endpoint-tests.json
@@ -1,1099 +1,533 @@
{
"testCases": [
{
- "documentation": "For region ap-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.ap-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.ap-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.ap-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ap-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.ap-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ap-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.eu-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.eu-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.eu-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.eu-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.ca-central-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ca-central-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.ca-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ca-central-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.ca-central-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ca-central-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.ca-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ca-central-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.eu-central-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-central-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.eu-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-central-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.eu-central-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-central-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.eu-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-central-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.us-west-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-west-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.us-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-west-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.us-west-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "us-west-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.us-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "us-west-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.us-west-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-west-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.us-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-west-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.us-west-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "us-west-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.us-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "us-west-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.af-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "af-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.af-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "af-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.af-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "af-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.af-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "af-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.eu-north-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-north-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.eu-north-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-north-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.eu-north-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-north-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.eu-north-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-north-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.eu-west-3.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-west-3",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.eu-west-3.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-west-3",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.eu-west-3.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-west-3",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.eu-west-3.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-west-3",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.eu-west-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-west-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.eu-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-west-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.eu-west-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-west-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.eu-west-2.amazonaws.com"
+ "url": "https://cloud9.af-south-1.amazonaws.com"
}
},
"params": {
+ "Region": "af-south-1",
"UseFIPS": false,
- "Region": "eu-west-2",
"UseDualStack": false
}
},
{
- "documentation": "For region eu-west-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.eu-west-1.api.aws"
+ "url": "https://cloud9.ap-east-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "eu-west-1",
- "UseDualStack": true
+ "Region": "ap-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region eu-west-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.eu-west-1.amazonaws.com"
+ "url": "https://cloud9.ap-northeast-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "eu-west-1",
+ "Region": "ap-northeast-1",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.eu-west-1.api.aws"
+ "url": "https://cloud9.ap-northeast-2.amazonaws.com"
}
},
"params": {
+ "Region": "ap-northeast-2",
"UseFIPS": false,
- "Region": "eu-west-1",
- "UseDualStack": true
+ "UseDualStack": false
}
},
{
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.eu-west-1.amazonaws.com"
+ "url": "https://cloud9.ap-northeast-3.amazonaws.com"
}
},
"params": {
+ "Region": "ap-northeast-3",
"UseFIPS": false,
- "Region": "eu-west-1",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-3 with FIPS enabled and DualStack enabled",
+ "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.ap-northeast-3.api.aws"
+ "url": "https://cloud9.ap-south-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-northeast-3",
- "UseDualStack": true
+ "Region": "ap-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-3 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.ap-northeast-3.amazonaws.com"
+ "url": "https://cloud9.ap-southeast-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-northeast-3",
+ "Region": "ap-southeast-1",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack enabled",
+ "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.ap-northeast-3.api.aws"
+ "url": "https://cloud9.ap-southeast-2.amazonaws.com"
}
},
"params": {
+ "Region": "ap-southeast-2",
"UseFIPS": false,
- "Region": "ap-northeast-3",
- "UseDualStack": true
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.ap-northeast-3.amazonaws.com"
+ "url": "https://cloud9.ca-central-1.amazonaws.com"
}
},
"params": {
+ "Region": "ca-central-1",
"UseFIPS": false,
- "Region": "ap-northeast-3",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.ap-northeast-2.api.aws"
+ "url": "https://cloud9.eu-central-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-northeast-2",
- "UseDualStack": true
+ "Region": "eu-central-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.ap-northeast-2.amazonaws.com"
+ "url": "https://cloud9.eu-north-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-northeast-2",
+ "Region": "eu-north-1",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region eu-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.ap-northeast-2.api.aws"
+ "url": "https://cloud9.eu-south-1.amazonaws.com"
}
},
"params": {
+ "Region": "eu-south-1",
"UseFIPS": false,
- "Region": "ap-northeast-2",
- "UseDualStack": true
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.ap-northeast-2.amazonaws.com"
+ "url": "https://cloud9.eu-west-1.amazonaws.com"
}
},
"params": {
+ "Region": "eu-west-1",
"UseFIPS": false,
- "Region": "ap-northeast-2",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.ap-northeast-1.api.aws"
+ "url": "https://cloud9.eu-west-2.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-northeast-1",
- "UseDualStack": true
+ "Region": "eu-west-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.ap-northeast-1.amazonaws.com"
+ "url": "https://cloud9.eu-west-3.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-northeast-1",
+ "Region": "eu-west-3",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.ap-northeast-1.api.aws"
+ "url": "https://cloud9.me-south-1.amazonaws.com"
}
},
"params": {
+ "Region": "me-south-1",
"UseFIPS": false,
- "Region": "ap-northeast-1",
- "UseDualStack": true
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.ap-northeast-1.amazonaws.com"
+ "url": "https://cloud9.sa-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "sa-east-1",
"UseFIPS": false,
- "Region": "ap-northeast-1",
"UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.me-south-1.api.aws"
+ "url": "https://cloud9.us-east-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "me-south-1",
- "UseDualStack": true
+ "Region": "us-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.me-south-1.amazonaws.com"
+ "url": "https://cloud9.us-east-2.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "me-south-1",
+ "Region": "us-east-2",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.me-south-1.api.aws"
+ "url": "https://cloud9.us-west-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-west-1",
"UseFIPS": false,
- "Region": "me-south-1",
- "UseDualStack": true
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.me-south-1.amazonaws.com"
+ "url": "https://cloud9.us-west-2.amazonaws.com"
}
},
"params": {
+ "Region": "us-west-2",
"UseFIPS": false,
- "Region": "me-south-1",
"UseDualStack": false
}
},
{
- "documentation": "For region sa-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.sa-east-1.api.aws"
+ "url": "https://cloud9-fips.us-east-1.api.aws"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": true,
- "Region": "sa-east-1",
"UseDualStack": true
}
},
{
- "documentation": "For region sa-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.sa-east-1.amazonaws.com"
+ "url": "https://cloud9-fips.us-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": true,
- "Region": "sa-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region sa-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.sa-east-1.api.aws"
+ "url": "https://cloud9.us-east-1.api.aws"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": false,
- "Region": "sa-east-1",
"UseDualStack": true
}
},
{
- "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.sa-east-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "sa-east-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ap-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.ap-east-1.api.aws"
+ "url": "https://cloud9-fips.cn-north-1.api.amazonwebservices.com.cn"
}
},
"params": {
+ "Region": "cn-north-1",
"UseFIPS": true,
- "Region": "ap-east-1",
"UseDualStack": true
}
},
{
- "documentation": "For region ap-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.ap-east-1.amazonaws.com"
+ "url": "https://cloud9-fips.cn-north-1.amazonaws.com.cn"
}
},
"params": {
+ "Region": "cn-north-1",
"UseFIPS": true,
- "Region": "ap-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.ap-east-1.api.aws"
+ "url": "https://cloud9.cn-north-1.api.amazonwebservices.com.cn"
}
},
"params": {
+ "Region": "cn-north-1",
"UseFIPS": false,
- "Region": "ap-east-1",
"UseDualStack": true
}
},
{
- "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.ap-east-1.amazonaws.com"
+ "url": "https://cloud9.cn-north-1.amazonaws.com.cn"
}
},
"params": {
+ "Region": "cn-north-1",
"UseFIPS": false,
- "Region": "ap-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.ap-southeast-1.api.aws"
+ "url": "https://cloud9-fips.us-gov-east-1.api.aws"
}
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": true,
- "Region": "ap-southeast-1",
"UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.ap-southeast-1.amazonaws.com"
+ "url": "https://cloud9-fips.us-gov-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": true,
- "Region": "ap-southeast-1",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.ap-southeast-1.api.aws"
+ "url": "https://cloud9.us-gov-east-1.api.aws"
}
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": false,
- "Region": "ap-southeast-1",
"UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.ap-southeast-1.amazonaws.com"
+ "url": "https://cloud9.us-gov-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": false,
- "Region": "ap-southeast-1",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://cloud9-fips.ap-southeast-2.api.aws"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": true,
- "Region": "ap-southeast-2",
"UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.ap-southeast-2.amazonaws.com"
+ "url": "https://cloud9-fips.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": true,
- "Region": "ap-southeast-2",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://cloud9.ap-southeast-2.api.aws"
- }
+ "error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": false,
- "Region": "ap-southeast-2",
"UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.ap-southeast-2.amazonaws.com"
+ "url": "https://cloud9.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": false,
- "Region": "ap-southeast-2",
"UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://cloud9-fips.us-east-1.api.aws"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
+ "Region": "us-isob-east-1",
"UseFIPS": true,
- "Region": "us-east-1",
"UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9-fips.us-east-1.amazonaws.com"
+ "url": "https://cloud9-fips.us-isob-east-1.sc2s.sgov.gov"
}
},
"params": {
+ "Region": "us-isob-east-1",
"UseFIPS": true,
- "Region": "us-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://cloud9.us-east-1.api.aws"
- }
+ "error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
+ "Region": "us-isob-east-1",
"UseFIPS": false,
- "Region": "us-east-1",
"UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.us-east-1.amazonaws.com"
+ "url": "https://cloud9.us-isob-east-1.sc2s.sgov.gov"
}
},
"params": {
+ "Region": "us-isob-east-1",
"UseFIPS": false,
- "Region": "us-east-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region us-east-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.us-east-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-east-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9-fips.us-east-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-east-2",
"UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloud9.us-east-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "us-east-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For custom endpoint with region set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
- "url": "https://cloud9.us-east-2.amazonaws.com"
+ "url": "https://example.com"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": false,
- "Region": "us-east-2",
- "UseDualStack": false
+ "UseDualStack": false,
+ "Endpoint": "https://example.com"
}
},
{
- "documentation": "For custom endpoint with fips disabled and dualstack disabled",
+ "documentation": "For custom endpoint with region not set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
"url": "https://example.com"
@@ -1101,7 +535,6 @@
},
"params": {
"UseFIPS": false,
- "Region": "us-east-1",
"UseDualStack": false,
"Endpoint": "https://example.com"
}
@@ -1112,8 +545,8 @@
"error": "Invalid Configuration: FIPS and custom endpoint are not supported"
},
"params": {
- "UseFIPS": true,
"Region": "us-east-1",
+ "UseFIPS": true,
"UseDualStack": false,
"Endpoint": "https://example.com"
}
@@ -1124,11 +557,17 @@
"error": "Invalid Configuration: Dualstack and custom endpoint are not supported"
},
"params": {
- "UseFIPS": false,
"Region": "us-east-1",
+ "UseFIPS": false,
"UseDualStack": true,
"Endpoint": "https://example.com"
}
+ },
+ {
+ "documentation": "Missing region",
+ "expect": {
+ "error": "Invalid Configuration: Missing Region"
+ }
}
],
"version": "1.0"
diff --git a/services/cloud9/src/main/resources/codegen-resources/service-2.json b/services/cloud9/src/main/resources/codegen-resources/service-2.json
index 7e9cca54ffb..2d2f489783d 100644
--- a/services/cloud9/src/main/resources/codegen-resources/service-2.json
+++ b/services/cloud9/src/main/resources/codegen-resources/service-2.json
@@ -330,7 +330,7 @@
},
"imageId":{
"shape":"ImageId",
- "documentation":"
The identifier for the Amazon Machine Image (AMI) that's used to create the EC2 instance. To choose an AMI for the instance, you must specify a valid AMI alias or a valid Amazon EC2 Systems Manager (SSM) path.
The default Amazon Linux AMI is currently used if the parameter isn't explicitly assigned a value in the request.
In the future the parameter for Amazon Linux will no longer be available when you specify an AMI for your instance. Amazon Linux 2 will then become the default AMI, which is used to launch your instance if no parameter is explicitly defined.
AMI aliases
Amazon Linux (default): amazonlinux-1-x86_64
Amazon Linux 2: amazonlinux-2-x86_64
Ubuntu 18.04: ubuntu-18.04-x86_64
SSM paths
Amazon Linux (default): resolve:ssm:/aws/service/cloud9/amis/amazonlinux-1-x86_64
Amazon Linux 2: resolve:ssm:/aws/service/cloud9/amis/amazonlinux-2-x86_64
The identifier for the Amazon Machine Image (AMI) that's used to create the EC2 instance. To choose an AMI for the instance, you must specify a valid AMI alias or a valid Amazon EC2 Systems Manager (SSM) path.
The default Amazon Linux AMI is currently used if the parameter isn't explicitly assigned a value in the request. Because Amazon Linux AMI has ended standard support as of December 31, 2020, we recommend you choose Amazon Linux 2, which includes long term support through 2023.
From December 31, 2023, the parameter for Amazon Linux will no longer be available when you specify an AMI for your instance. Amazon Linux 2 will then become the default AMI, which is used to launch your instance if no parameter is explicitly defined.
AMI aliases
Amazon Linux (default): amazonlinux-1-x86_64
Amazon Linux 2: amazonlinux-2-x86_64
Ubuntu 18.04: ubuntu-18.04-x86_64
SSM paths
Amazon Linux (default): resolve:ssm:/aws/service/cloud9/amis/amazonlinux-1-x86_64
Amazon Linux 2: resolve:ssm:/aws/service/cloud9/amis/amazonlinux-2-x86_64
"
},
"automaticStopTimeMinutes":{
"shape":"AutomaticStopTimeMinutes",
diff --git a/services/cloudcontrol/pom.xml b/services/cloudcontrol/pom.xml
index b81b90ff714..c6f585ad8ac 100644
--- a/services/cloudcontrol/pom.xml
+++ b/services/cloudcontrol/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTcloudcontrolAWS Java SDK :: Services :: Cloud Control
diff --git a/services/cloudcontrol/src/main/resources/codegen-resources/endpoint-rule-set.json b/services/cloudcontrol/src/main/resources/codegen-resources/endpoint-rule-set.json
index e215956866a..1812986e28a 100644
--- a/services/cloudcontrol/src/main/resources/codegen-resources/endpoint-rule-set.json
+++ b/services/cloudcontrol/src/main/resources/codegen-resources/endpoint-rule-set.json
@@ -32,13 +32,12 @@
{
"conditions": [
{
- "fn": "aws.partition",
+ "fn": "isSet",
"argv": [
{
- "ref": "Region"
+ "ref": "Endpoint"
}
- ],
- "assign": "PartitionResult"
+ ]
}
],
"type": "tree",
@@ -46,23 +45,20 @@
{
"conditions": [
{
- "fn": "isSet",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "Endpoint"
- }
+ "ref": "UseFIPS"
+ },
+ true
]
- },
- {
- "fn": "parseURL",
- "argv": [
- {
- "ref": "Endpoint"
- }
- ],
- "assign": "url"
}
],
+ "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "type": "error"
+ },
+ {
+ "conditions": [],
"type": "tree",
"rules": [
{
@@ -71,67 +67,42 @@
"fn": "booleanEquals",
"argv": [
{
- "ref": "UseFIPS"
+ "ref": "UseDualStack"
},
true
]
}
],
- "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
"type": "error"
},
{
"conditions": [],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
- },
- true
- ]
- }
- ],
- "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
- "type": "error"
+ "endpoint": {
+ "url": {
+ "ref": "Endpoint"
},
- {
- "conditions": [],
- "endpoint": {
- "url": {
- "ref": "Endpoint"
- },
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
}
]
- },
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
{
"conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
- },
- true
- ]
- },
- {
- "fn": "booleanEquals",
+ "fn": "isSet",
"argv": [
{
- "ref": "UseDualStack"
- },
- true
+ "ref": "Region"
+ }
]
}
],
@@ -140,90 +111,215 @@
{
"conditions": [
{
- "fn": "booleanEquals",
+ "fn": "aws.partition",
"argv": [
- true,
{
- "fn": "getAttr",
+ "ref": "Region"
+ }
+ ],
+ "assign": "PartitionResult"
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseFIPS"
},
- "supportsFIPS"
+ true
+ ]
+ },
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ {
+ "ref": "UseDualStack"
+ },
+ true
]
}
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsFIPS"
+ ]
+ }
+ ]
+ },
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://cloudcontrolapi-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
+ "type": "error"
+ }
]
},
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseFIPS"
},
- "supportsDualStack"
+ true
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsFIPS"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://cloudcontrolapi-fips.{Region}.{PartitionResult#dnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
]
+ },
+ {
+ "conditions": [],
+ "error": "FIPS is enabled but this partition does not support FIPS",
+ "type": "error"
}
]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
- },
- {
- "conditions": [],
- "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
},
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseDualStack"
},
- "supportsFIPS"
+ true
]
}
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://cloudcontrolapi.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "error": "DualStack is enabled but this partition does not support DualStack",
+ "type": "error"
+ }
]
- }
- ],
- "type": "tree",
- "rules": [
+ },
{
"conditions": [],
"type": "tree",
@@ -231,7 +327,7 @@
{
"conditions": [],
"endpoint": {
- "url": "https://cloudcontrolapi-fips.{Region}.{PartitionResult#dnsSuffix}",
+ "url": "https://cloudcontrolapi.{Region}.{PartitionResult#dnsSuffix}",
"properties": {},
"headers": {}
},
@@ -240,74 +336,13 @@
]
}
]
- },
- {
- "conditions": [],
- "error": "FIPS is enabled but this partition does not support FIPS",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
- },
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- true,
- {
- "fn": "getAttr",
- "argv": [
- {
- "ref": "PartitionResult"
- },
- "supportsDualStack"
- ]
- }
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
- "endpoint": {
- "url": "https://cloudcontrolapi.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
- },
- {
- "conditions": [],
- "error": "DualStack is enabled but this partition does not support DualStack",
- "type": "error"
}
]
},
{
"conditions": [],
- "endpoint": {
- "url": "https://cloudcontrolapi.{Region}.{PartitionResult#dnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "error": "Invalid Configuration: Missing Region",
+ "type": "error"
}
]
}
diff --git a/services/cloudcontrol/src/main/resources/codegen-resources/endpoint-tests.json b/services/cloudcontrol/src/main/resources/codegen-resources/endpoint-tests.json
index ab39a6b29bf..d21e7072890 100644
--- a/services/cloudcontrol/src/main/resources/codegen-resources/endpoint-tests.json
+++ b/services/cloudcontrol/src/main/resources/codegen-resources/endpoint-tests.json
@@ -1,276 +1,133 @@
{
"testCases": [
{
- "documentation": "For region ap-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.ap-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.ap-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.eu-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.eu-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-south-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-south-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-south-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-south-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-south-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-south-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-south-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.eu-south-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-south-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-south-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.eu-south-2.amazonaws.com"
+ "url": "https://cloudcontrolapi.af-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "eu-south-2",
- "UseFIPS": false
+ "Region": "af-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.us-gov-east-1.api.aws"
+ "url": "https://cloudcontrolapi.ap-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-gov-east-1",
- "UseFIPS": true
+ "Region": "ap-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.us-gov-east-1.amazonaws.com"
+ "url": "https://cloudcontrolapi.ap-northeast-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-gov-east-1",
- "UseFIPS": true
+ "Region": "ap-northeast-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.us-gov-east-1.api.aws"
+ "url": "https://cloudcontrolapi.ap-northeast-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-gov-east-1",
- "UseFIPS": false
+ "Region": "ap-northeast-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.us-gov-east-1.amazonaws.com"
+ "url": "https://cloudcontrolapi.ap-northeast-3.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-gov-east-1",
- "UseFIPS": false
+ "Region": "ap-northeast-3",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-central-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.me-central-1.api.aws"
+ "url": "https://cloudcontrolapi.ap-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "me-central-1",
- "UseFIPS": true
+ "Region": "ap-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-central-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.me-central-1.amazonaws.com"
+ "url": "https://cloudcontrolapi.ap-southeast-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "me-central-1",
- "UseFIPS": true
+ "Region": "ap-southeast-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-central-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.me-central-1.api.aws"
+ "url": "https://cloudcontrolapi.ap-southeast-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "me-central-1",
- "UseFIPS": false
+ "Region": "ap-southeast-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-central-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-southeast-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.me-central-1.amazonaws.com"
+ "url": "https://cloudcontrolapi.ap-southeast-3.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "me-central-1",
- "UseFIPS": false
+ "Region": "ap-southeast-3",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ca-central-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.ca-central-1.api.aws"
+ "url": "https://cloudcontrolapi.ca-central-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
"Region": "ca-central-1",
- "UseFIPS": true
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
@@ -281,74 +138,9 @@
}
},
"params": {
- "UseDualStack": false,
- "Region": "ca-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.ca-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ca-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.ca-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
"Region": "ca-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.eu-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-central-1",
- "UseFIPS": false
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
@@ -359,1114 +151,495 @@
}
},
"params": {
- "UseDualStack": false,
- "Region": "eu-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.us-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.us-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.us-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.us-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.us-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.us-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.us-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.us-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.af-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "af-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.af-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "af-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.af-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "af-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.af-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "af-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-north-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-north-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-north-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-north-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.eu-north-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-north-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.eu-north-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-north-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-west-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-west-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.eu-west-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.eu-west-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.eu-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.eu-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.eu-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.eu-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.eu-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-northeast-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-northeast-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.ap-northeast-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.ap-northeast-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-northeast-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-northeast-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.ap-northeast-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.ap-northeast-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-northeast-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-northeast-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.ap-northeast-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.ap-northeast-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region me-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.me-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "me-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region me-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.me-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "me-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region me-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.me-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "me-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.me-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "me-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region sa-east-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.sa-east-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "sa-east-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region sa-east-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.sa-east-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "sa-east-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region sa-east-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.sa-east-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "sa-east-1",
- "UseFIPS": false
+ "Region": "eu-central-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.sa-east-1.amazonaws.com"
+ "url": "https://cloudcontrolapi.eu-north-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "sa-east-1",
- "UseFIPS": false
+ "Region": "eu-north-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region eu-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-east-1.api.aws"
+ "url": "https://cloudcontrolapi.eu-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-east-1",
- "UseFIPS": true
+ "Region": "eu-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-east-1.amazonaws.com"
+ "url": "https://cloudcontrolapi.eu-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-east-1",
- "UseFIPS": true
+ "Region": "eu-west-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.ap-east-1.api.aws"
+ "url": "https://cloudcontrolapi.eu-west-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-east-1",
- "UseFIPS": false
+ "Region": "eu-west-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.ap-east-1.amazonaws.com"
+ "url": "https://cloudcontrolapi.eu-west-3.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-east-1",
- "UseFIPS": false
+ "Region": "eu-west-3",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.cn-north-1.api.amazonwebservices.com.cn"
+ "url": "https://cloudcontrolapi.me-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "cn-north-1",
- "UseFIPS": true
+ "Region": "me-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.cn-north-1.amazonaws.com.cn"
+ "url": "https://cloudcontrolapi.sa-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "cn-north-1",
- "UseFIPS": true
+ "Region": "sa-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.cn-north-1.api.amazonwebservices.com.cn"
+ "url": "https://cloudcontrolapi.us-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "cn-north-1",
- "UseFIPS": false
+ "Region": "us-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.cn-north-1.amazonaws.com.cn"
+ "url": "https://cloudcontrolapi-fips.us-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "cn-north-1",
- "UseFIPS": false
+ "Region": "us-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.us-gov-west-1.api.aws"
+ "url": "https://cloudcontrolapi.us-east-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-gov-west-1",
- "UseFIPS": true
+ "Region": "us-east-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.us-gov-west-1.amazonaws.com"
+ "url": "https://cloudcontrolapi-fips.us-east-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-gov-west-1",
- "UseFIPS": true
+ "Region": "us-east-2",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.us-gov-west-1.api.aws"
+ "url": "https://cloudcontrolapi.us-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-gov-west-1",
- "UseFIPS": false
+ "Region": "us-west-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-west-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.us-gov-west-1.amazonaws.com"
+ "url": "https://cloudcontrolapi-fips.us-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-gov-west-1",
- "UseFIPS": false
+ "Region": "us-west-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-southeast-1.api.aws"
+ "url": "https://cloudcontrolapi.us-west-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-1",
- "UseFIPS": true
+ "Region": "us-west-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-west-2 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-southeast-1.amazonaws.com"
+ "url": "https://cloudcontrolapi-fips.us-west-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-1",
- "UseFIPS": true
+ "Region": "us-west-2",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.ap-southeast-1.api.aws"
+ "url": "https://cloudcontrolapi-fips.us-east-1.api.aws"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-1",
- "UseFIPS": false
+ "Region": "us-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.ap-southeast-1.amazonaws.com"
+ "url": "https://cloudcontrolapi.us-east-1.api.aws"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-1",
- "UseFIPS": false
+ "Region": "us-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-southeast-2.api.aws"
+ "url": "https://cloudcontrolapi.cn-north-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-2",
- "UseFIPS": true
+ "Region": "cn-north-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-southeast-2.amazonaws.com"
+ "url": "https://cloudcontrolapi.cn-northwest-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-2",
- "UseFIPS": true
+ "Region": "cn-northwest-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.ap-southeast-2.api.aws"
+ "url": "https://cloudcontrolapi-fips.cn-north-1.api.amazonwebservices.com.cn"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-2",
- "UseFIPS": false
+ "Region": "cn-north-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.ap-southeast-2.amazonaws.com"
+ "url": "https://cloudcontrolapi-fips.cn-north-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-2",
- "UseFIPS": false
+ "Region": "cn-north-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS enabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-southeast-3.api.aws"
+ "url": "https://cloudcontrolapi.cn-north-1.api.amazonwebservices.com.cn"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-3",
- "UseFIPS": true
+ "Region": "cn-north-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.ap-southeast-3.amazonaws.com"
+ "url": "https://cloudcontrolapi.us-gov-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-3",
- "UseFIPS": true
+ "Region": "us-gov-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.ap-southeast-3.api.aws"
+ "url": "https://cloudcontrolapi-fips.us-gov-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-3",
- "UseFIPS": false
+ "Region": "us-gov-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.ap-southeast-3.amazonaws.com"
+ "url": "https://cloudcontrolapi.us-gov-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-3",
- "UseFIPS": false
+ "Region": "us-gov-west-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.us-east-1.api.aws"
+ "url": "https://cloudcontrolapi-fips.us-gov-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-1",
- "UseFIPS": true
+ "Region": "us-gov-west-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.us-east-1.amazonaws.com"
+ "url": "https://cloudcontrolapi-fips.us-gov-east-1.api.aws"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-1",
- "UseFIPS": true
+ "Region": "us-gov-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.us-east-1.api.aws"
+ "url": "https://cloudcontrolapi.us-gov-east-1.api.aws"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-1",
- "UseFIPS": false
+ "Region": "us-gov-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.us-east-1.amazonaws.com"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-1",
- "UseFIPS": false
+ "Region": "us-iso-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.us-east-2.api.aws"
+ "url": "https://cloudcontrolapi-fips.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-2",
- "UseFIPS": true
+ "Region": "us-iso-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.us-east-2.amazonaws.com"
- }
+ "error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-2",
- "UseFIPS": true
+ "Region": "us-iso-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.us-east-2.api.aws"
+ "url": "https://cloudcontrolapi.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-2",
- "UseFIPS": false
+ "Region": "us-iso-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi.us-east-2.amazonaws.com"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-2",
- "UseFIPS": false
+ "Region": "us-isob-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi-fips.cn-northwest-1.api.amazonwebservices.com.cn"
+ "url": "https://cloudcontrolapi-fips.us-isob-east-1.sc2s.sgov.gov"
}
},
"params": {
- "UseDualStack": true,
- "Region": "cn-northwest-1",
- "UseFIPS": true
+ "Region": "us-isob-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://cloudcontrolapi-fips.cn-northwest-1.amazonaws.com.cn"
- }
+ "error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
- "UseDualStack": false,
- "Region": "cn-northwest-1",
- "UseFIPS": true
+ "Region": "us-isob-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.cn-northwest-1.api.amazonwebservices.com.cn"
+ "url": "https://cloudcontrolapi.us-isob-east-1.sc2s.sgov.gov"
}
},
"params": {
- "UseDualStack": true,
- "Region": "cn-northwest-1",
- "UseFIPS": false
+ "Region": "us-isob-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For custom endpoint with region set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
- "url": "https://cloudcontrolapi.cn-northwest-1.amazonaws.com.cn"
+ "url": "https://example.com"
}
},
"params": {
+ "Region": "us-east-1",
+ "UseFIPS": false,
"UseDualStack": false,
- "Region": "cn-northwest-1",
- "UseFIPS": false
+ "Endpoint": "https://example.com"
}
},
{
- "documentation": "For custom endpoint with fips disabled and dualstack disabled",
+ "documentation": "For custom endpoint with region not set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
"url": "https://example.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-1",
"UseFIPS": false,
+ "UseDualStack": false,
"Endpoint": "https://example.com"
}
},
@@ -1476,9 +649,9 @@
"error": "Invalid Configuration: FIPS and custom endpoint are not supported"
},
"params": {
- "UseDualStack": false,
"Region": "us-east-1",
"UseFIPS": true,
+ "UseDualStack": false,
"Endpoint": "https://example.com"
}
},
@@ -1488,11 +661,17 @@
"error": "Invalid Configuration: Dualstack and custom endpoint are not supported"
},
"params": {
- "UseDualStack": true,
"Region": "us-east-1",
"UseFIPS": false,
+ "UseDualStack": true,
"Endpoint": "https://example.com"
}
+ },
+ {
+ "documentation": "Missing region",
+ "expect": {
+ "error": "Invalid Configuration: Missing Region"
+ }
}
],
"version": "1.0"
diff --git a/services/cloudcontrol/src/main/resources/codegen-resources/service-2.json b/services/cloudcontrol/src/main/resources/codegen-resources/service-2.json
index b59888f2c05..9e34fe76313 100644
--- a/services/cloudcontrol/src/main/resources/codegen-resources/service-2.json
+++ b/services/cloudcontrol/src/main/resources/codegen-resources/service-2.json
@@ -286,7 +286,7 @@
},
"DesiredState":{
"shape":"Properties",
- "documentation":"
Structured data format representing the desired state of the resource, consisting of that resource's properties and their desired values.
Cloud Control API currently supports JSON as a structured data format.
<p>Specify the desired state as one of the following:</p> <ul> <li> <p>A JSON blob</p> </li> <li> <p>A local path containing the desired state in JSON data format</p> </li> </ul> <p>For more information, see <a href="https://docs.aws.amazon.com/cloudcontrolapi/latest/userguide/resource-operations-create.html#resource-operations-create-desiredstate">Composing the desired state of the resource</a> in the <i>Amazon Web Services Cloud Control API User Guide</i>.</p> <p>For more information about the properties of a specific resource, refer to the related topic for the resource in the <a href="https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-template-resource-type-ref.html">Resource and property types reference</a> in the <i>CloudFormation Users Guide</i>.</p>
"
+ "documentation":"
Structured data format representing the desired state of the resource, consisting of that resource's properties and their desired values.
Cloud Control API currently supports JSON as a structured data format.
Specify the desired state as one of the following:
A JSON blob
A local path containing the desired state in JSON data format
For more information about the properties of a specific resource, refer to the related topic for the resource in the Resource and property types reference in the CloudFormation Users Guide.
Lists all stacks that are importing an exported output value. To modify or remove an exported output value, first use this action to see which stacks are using it. To see the exported output values in your account, see ListExports.
For more information about importing an exported output value, see the Fn::ImportValue function.
Returns drift information for resources in a stack instance.
ListStackInstanceResourceDrifts returns drift information for the most recent drift detection operation. If an operation is in progress, it may only return partial results.
In some cases, you must explicitly acknowledge that your stack template contains certain capabilities in order for CloudFormation to create the stack.
CAPABILITY_IAM and CAPABILITY_NAMED_IAM
Some stack templates might include resources that can affect permissions in your Amazon Web Services account; for example, by creating new Identity and Access Management (IAM) users. For those stacks, you must explicitly acknowledge this by specifying one of these capabilities.
The following IAM resources require you to specify either the CAPABILITY_IAM or CAPABILITY_NAMED_IAM capability.
If you have IAM resources, you can specify either capability.
If you have IAM resources with custom names, you must specify CAPABILITY_NAMED_IAM.
If you don't specify either of these capabilities, CloudFormation returns an InsufficientCapabilities error.
If your stack template contains these resources, we suggest that you review all permissions associated with them and edit their permissions if necessary.
Some template contain macros. Macros perform custom processing on templates; this can include simple actions like find-and-replace operations, all the way to extensive transformations of entire templates. Because of this, users typically create a change set from the processed template, so that they can review the changes resulting from the macros before actually creating the stack. If your stack template contains one or more macros, and you choose to create a stack directly from the processed template, without first reviewing the resulting changes in a change set, you must acknowledge this capability. This includes the AWS::Include and AWS::Serverless transforms, which are macros hosted by CloudFormation.
This capacity doesn't apply to creating change sets, and specifying it when creating change sets has no effect.
If you want to create a stack from a stack template that contains macros and nested stacks, you must create or update the stack directly from the template using the CreateStack or UpdateStack action, and specifying this capability.
In some cases, you must explicitly acknowledge that your stack template contains certain capabilities in order for CloudFormation to create the stack.
CAPABILITY_IAM and CAPABILITY_NAMED_IAM
Some stack templates might include resources that can affect permissions in your Amazon Web Services account; for example, by creating new Identity and Access Management (IAM) users. For those stacks, you must explicitly acknowledge this by specifying one of these capabilities.
The following IAM resources require you to specify either the CAPABILITY_IAM or CAPABILITY_NAMED_IAM capability.
If you have IAM resources, you can specify either capability.
If you have IAM resources with custom names, you must specify CAPABILITY_NAMED_IAM.
If you don't specify either of these capabilities, CloudFormation returns an InsufficientCapabilities error.
If your stack template contains these resources, we suggest that you review all permissions associated with them and edit their permissions if necessary.
Some template contain macros. Macros perform custom processing on templates; this can include simple actions like find-and-replace operations, all the way to extensive transformations of entire templates. Because of this, users typically create a change set from the processed template, so that they can review the changes resulting from the macros before actually creating the stack. If your stack template contains one or more macros, and you choose to create a stack directly from the processed template, without first reviewing the resulting changes in a change set, you must acknowledge this capability. This includes the AWS::Include and AWS::Serverless transforms, which are macros hosted by CloudFormation.
This capacity doesn't apply to creating change sets, and specifying it when creating change sets has no effect.
If you want to create a stack from a stack template that contains macros and nested stacks, you must create or update the stack directly from the template using the CreateStack or UpdateStack action, and specifying this capability.
In some cases, you must explicitly acknowledge that your stack template contains certain capabilities in order for CloudFormation to create the stack.
CAPABILITY_IAM and CAPABILITY_NAMED_IAM
Some stack templates might include resources that can affect permissions in your Amazon Web Services account; for example, by creating new Identity and Access Management (IAM) users. For those stacks, you must explicitly acknowledge this by specifying one of these capabilities.
The following IAM resources require you to specify either the CAPABILITY_IAM or CAPABILITY_NAMED_IAM capability.
If you have IAM resources, you can specify either capability.
If you have IAM resources with custom names, you must specify CAPABILITY_NAMED_IAM.
If you don't specify either of these capabilities, CloudFormation returns an InsufficientCapabilities error.
If your stack template contains these resources, we recommend that you review all permissions associated with them and edit their permissions if necessary.
Some template contain macros. Macros perform custom processing on templates; this can include simple actions like find-and-replace operations, all the way to extensive transformations of entire templates. Because of this, users typically create a change set from the processed template, so that they can review the changes resulting from the macros before actually creating the stack. If your stack template contains one or more macros, and you choose to create a stack directly from the processed template, without first reviewing the resulting changes in a change set, you must acknowledge this capability. This includes the AWS::Include and AWS::Serverless transforms, which are macros hosted by CloudFormation.
If you want to create a stack from a stack template that contains macros and nested stacks, you must create the stack directly from the template using this capability.
You should only create stacks directly from a stack template that contains macros if you know what processing the macro performs.
Each macro relies on an underlying Lambda service function for processing stack templates. Be aware that the Lambda function owner can update the function operation without CloudFormation being notified.
In some cases, you must explicitly acknowledge that your stack template contains certain capabilities in order for CloudFormation to create the stack.
CAPABILITY_IAM and CAPABILITY_NAMED_IAM
Some stack templates might include resources that can affect permissions in your Amazon Web Services account; for example, by creating new Identity and Access Management (IAM) users. For those stacks, you must explicitly acknowledge this by specifying one of these capabilities.
The following IAM resources require you to specify either the CAPABILITY_IAM or CAPABILITY_NAMED_IAM capability.
If you have IAM resources, you can specify either capability.
If you have IAM resources with custom names, you must specify CAPABILITY_NAMED_IAM.
If you don't specify either of these capabilities, CloudFormation returns an InsufficientCapabilities error.
If your stack template contains these resources, we recommend that you review all permissions associated with them and edit their permissions if necessary.
Some template contain macros. Macros perform custom processing on templates; this can include simple actions like find-and-replace operations, all the way to extensive transformations of entire templates. Because of this, users typically create a change set from the processed template, so that they can review the changes resulting from the macros before actually creating the stack. If your stack template contains one or more macros, and you choose to create a stack directly from the processed template, without first reviewing the resulting changes in a change set, you must acknowledge this capability. This includes the AWS::Include and AWS::Serverless transforms, which are macros hosted by CloudFormation.
If you want to create a stack from a stack template that contains macros and nested stacks, you must create the stack directly from the template using this capability.
You should only create stacks directly from a stack template that contains macros if you know what processing the macro performs.
Each macro relies on an underlying Lambda service function for processing stack templates. Be aware that the Lambda function owner can update the function operation without CloudFormation being notified.
Whether to enable termination protection on the specified stack. If a user attempts to delete a stack with termination protection enabled, the operation fails and the stack remains unchanged. For more information, see Protecting a Stack From Being Deleted in the CloudFormation User Guide. Termination protection is deactivated on stacks by default.
For nested stacks, termination protection is set on the root stack and can't be changed directly on the nested stack.
"
+ "documentation":"
Whether to enable termination protection on the specified stack. If a user attempts to delete a stack with termination protection enabled, the operation fails and the stack remains unchanged. For more information, see Protecting a Stack From Being Deleted in the CloudFormation User Guide. Termination protection is deactivated on stacks by default.
For nested stacks, termination protection is set on the root stack and can't be changed directly on the nested stack.
This deletion policy deletes newly created resources, but retains existing resources, when a stack operation is rolled back. This ensures new, empty, and unused resources are deleted, while critical resources and their data are retained. RetainExceptOnCreate can be specified for any resource that supports the DeletionPolicy attribute.
In some cases, you must explicitly acknowledge that your stack set template contains certain capabilities in order for CloudFormation to create the stack set and related stack instances.
CAPABILITY_IAM and CAPABILITY_NAMED_IAM
Some stack templates might include resources that can affect permissions in your Amazon Web Services account; for example, by creating new Identity and Access Management (IAM) users. For those stack sets, you must explicitly acknowledge this by specifying one of these capabilities.
The following IAM resources require you to specify either the CAPABILITY_IAM or CAPABILITY_NAMED_IAM capability.
If you have IAM resources, you can specify either capability.
If you have IAM resources with custom names, you must specify CAPABILITY_NAMED_IAM.
If you don't specify either of these capabilities, CloudFormation returns an InsufficientCapabilities error.
If your stack template contains these resources, we recommend that you review all permissions associated with them and edit their permissions if necessary.
Some templates reference macros. If your stack set template references one or more macros, you must create the stack set directly from the processed template, without first reviewing the resulting changes in a change set. To create the stack set directly, you must acknowledge this capability. For more information, see Using CloudFormation Macros to Perform Custom Processing on Templates.
Stack sets with service-managed permissions don't currently support the use of macros in templates. (This includes the AWS::Include and AWS::Serverless transforms, which are macros hosted by CloudFormation.) Even if you specify this capability for a stack set with service-managed permissions, if you reference a macro in your template the stack set operation will fail.
"
+ "documentation":"
In some cases, you must explicitly acknowledge that your stack set template contains certain capabilities in order for CloudFormation to create the stack set and related stack instances.
CAPABILITY_IAM and CAPABILITY_NAMED_IAM
Some stack templates might include resources that can affect permissions in your Amazon Web Services account; for example, by creating new Identity and Access Management (IAM) users. For those stack sets, you must explicitly acknowledge this by specifying one of these capabilities.
The following IAM resources require you to specify either the CAPABILITY_IAM or CAPABILITY_NAMED_IAM capability.
If you have IAM resources, you can specify either capability.
If you have IAM resources with custom names, you must specify CAPABILITY_NAMED_IAM.
If you don't specify either of these capabilities, CloudFormation returns an InsufficientCapabilities error.
If your stack template contains these resources, we recommend that you review all permissions associated with them and edit their permissions if necessary.
Some templates reference macros. If your stack set template references one or more macros, you must create the stack set directly from the processed template, without first reviewing the resulting changes in a change set. To create the stack set directly, you must acknowledge this capability. For more information, see Using CloudFormation Macros to Perform Custom Processing on Templates.
Stack sets with service-managed permissions don't currently support the use of macros in templates. (This includes the AWS::Include and AWS::Serverless transforms, which are macros hosted by CloudFormation.) Even if you specify this capability for a stack set with service-managed permissions, if you reference a macro in your template the stack set operation will fail.
The Amazon Resource Name (ARN) of the IAM role to use to create this stack set.
Specify an IAM role only if you are using customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Prerequisites: Granting Permissions for Stack Set Operations in the CloudFormation User Guide.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the IAM role to use to create this stack set.
Specify an IAM role only if you are using customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Prerequisites: Granting Permissions for Stack Set Operations in the CloudFormation User Guide.
Preserves the state of previously provisioned resources when an operation fails. This parameter can't be specified when the OnStackFailure parameter to the CreateChangeSet API operation was specified.
True - if the stack creation fails, do nothing. This is equivalent to specifying DO_NOTHING for the OnStackFailure parameter to the CreateChangeSet API operation.
False - if the stack creation fails, roll back the stack. This is equivalent to specifying ROLLBACK for the OnStackFailure parameter to the CreateChangeSet API operation.
This deletion policy deletes newly created resources, but retains existing resources, when a stack operation is rolled back. This ensures new, empty, and unused resources are deleted, while critical resources and their data are retained. RetainExceptOnCreate can be specified for any resource that supports the DeletionPolicy attribute.
If the previous paginated request didn't return all of the remaining results, the response object's NextToken parameter value is set to a token. To retrieve the next set of results, call this action again and assign that token to the request object's NextToken parameter. If there are no remaining results, the previous response object's NextToken parameter is set to null.
The maximum number of results to be returned with a single call. If the number of available results exceeds this maximum, the response includes a NextToken value that you can assign to the NextToken request parameter to get the next set of results.
[Service-managed permissions] Specifies whether you are acting as an account administrator in the organization's management account or as a delegated administrator in a member account.
By default, SELF is specified. Use SELF for stack sets with self-managed permissions.
If you are signed in to the management account, specify SELF.
If you are signed in to a delegated administrator account, specify DELEGATED_ADMIN.
Your Amazon Web Services account must be registered as a delegated administrator in the management account. For more information, see Register a delegated administrator in the CloudFormation User Guide.
If the previous paginated request didn't return all of the remaining results, the response object's NextToken parameter value is set to a token. To retrieve the next set of results, call this action again and assign that token to the request object's NextToken parameter. If there are no remaining results, the previous response object's NextToken parameter is set to null.
The amount of time, in minutes, during which CloudFormation should monitor all the rollback triggers after the stack creation or update operation deploys all necessary resources.
The default is 0 minutes.
If you specify a monitoring period but don't specify any rollback triggers, CloudFormation still waits the specified period of time before cleaning up old resources after update operations. You can use this monitoring period to perform any manual stack validation desired, and manually cancel the stack creation or update (using CancelUpdateStack, for example) as necessary.
If you specify 0 for this parameter, CloudFormation still monitors the specified rollback triggers during stack creation and update operations. Then, for update operations, it begins disposing of old resources immediately once the operation completes.
"
}
},
- "documentation":"
Structure containing the rollback triggers for CloudFormation to monitor during stack creation and updating operations, and for the specified monitoring period afterwards.
Rollback triggers enable you to have CloudFormation monitor the state of your application during stack creation and updating, and to roll back that operation if the application breaches the threshold of any of the alarms you've specified. For more information, see Monitor and Roll Back Stack Operations.
"
+ "documentation":"
Structure containing the rollback triggers for CloudFormation to monitor during stack creation and updating operations, and for the specified monitoring period afterwards.
Rollback triggers enable you to have CloudFormation monitor the state of your application during stack creation and updating, and to roll back that operation if the application breaches the threshold of any of the alarms you've specified. For more information, see Monitor and Roll Back Stack Operations.
This deletion policy deletes newly created resources, but retains existing resources, when a stack operation is rolled back. This ensures new, empty, and unused resources are deleted, while critical resources and their data are retained. RetainExceptOnCreate can be specified for any resource that supports the DeletionPolicy attribute.
Whether termination protection is enabled for the stack.
For nested stacks, termination protection is set on the root stack and can't be changed directly on the nested stack. For more information, see Protecting a Stack From Being Deleted in the CloudFormation User Guide.
"
+ "documentation":"
Whether termination protection is enabled for the stack.
For nested stacks, termination protection is set on the root stack and can't be changed directly on the nested stack. For more information, see Protecting a Stack From Being Deleted in the CloudFormation User Guide.
For nested stacks--stacks created as resources for another stack--the stack ID of the direct parent of this stack. For the first level of nested stacks, the root stack is also the parent stack.
For nested stacks--stacks created as resources for another stack--the stack ID of the direct parent of this stack. For the first level of nested stacks, the root stack is also the parent stack.
Information about whether a stack's actual configuration differs, or has drifted, from its expected configuration, as defined in the stack template and any values specified as template parameters. For more information, see Detecting Unregulated Configuration Changes to Stacks and Resources.
"
+ "documentation":"
Information about whether a stack's actual configuration differs, or has drifted, from its expected configuration, as defined in the stack template and any values specified as template parameters. For more information, see Detecting Unregulated Configuration Changes to Stacks and Resources.
This deletion policy deletes newly created resources, but retains existing resources, when a stack operation is rolled back. This ensures new, empty, and unused resources are deleted, while critical resources and their data are retained. RetainExceptOnCreate can be specified for any resource that supports the DeletionPolicy attribute.
Context information that enables CloudFormation to uniquely identify a resource. CloudFormation uses context key-value pairs in cases where a resource's logical and physical IDs aren't enough to uniquely identify that resource. Each context key-value pair specifies a unique resource that contains the targeted resource.
Status of the actual configuration of the resource compared to its expected configuration. These will be present only for resources whose StackInstanceResourceDriftStatus is MODIFIED.
Information about whether the resource's actual configuration differs, or has drifted, from its expected configuration, as defined in the stack template and any values specified as template parameters. For more information, see Detecting Unregulated Configuration Changes to Stacks and Resources.
"
+ "documentation":"
Information about whether the resource's actual configuration differs, or has drifted, from its expected configuration, as defined in the stack template and any values specified as template parameters. For more information, see Detecting Unregulated Configuration Changes to Stacks and Resources.
Information about whether the resource's actual configuration differs, or has drifted, from its expected configuration, as defined in the stack template and any values specified as template parameters. For more information, see Detecting Unregulated Configuration Changes to Stacks and Resources.
"
+ "documentation":"
Information about whether the resource's actual configuration differs, or has drifted, from its expected configuration, as defined in the stack template and any values specified as template parameters. For more information, see Detecting Unregulated Configuration Changes to Stacks and Resources.
Contains information about the module from which the resource was created, if the resource was created from a module included in the stack template.
"
}
},
- "documentation":"
Contains the drift information for a resource that has been checked for drift. This includes actual and expected property values for resources in which CloudFormation has detected drift. Only resource properties explicitly defined in the stack template are checked for drift. For more information, see Detecting Unregulated Configuration Changes to Stacks and Resources.
Resources that don't currently support drift detection can't be checked. For a list of resources that support drift detection, see Resources that Support Drift Detection.
Use DetectStackResourceDrift to detect drift on individual resources, or DetectStackDrift to detect drift on all resources in a given stack that support drift detection.
"
+ "documentation":"
Contains the drift information for a resource that has been checked for drift. This includes actual and expected property values for resources in which CloudFormation has detected drift. Only resource properties explicitly defined in the stack template are checked for drift. For more information, see Detecting Unregulated Configuration Changes to Stacks and Resources.
Resources that don't currently support drift detection can't be checked. For a list of resources that support drift detection, see Resources that Support Drift Detection.
Use DetectStackResourceDrift to detect drift on individual resources, or DetectStackDrift to detect drift on all resources in a given stack that support drift detection.
Information about whether the resource's actual configuration differs, or has drifted, from its expected configuration, as defined in the stack template and any values specified as template parameters. For more information, see Detecting Unregulated Configuration Changes to Stacks and Resources.
"
+ "documentation":"
Information about whether the resource's actual configuration differs, or has drifted, from its expected configuration, as defined in the stack template and any values specified as template parameters. For more information, see Detecting Unregulated Configuration Changes to Stacks and Resources.
The Amazon Resource Name (ARN) of the IAM role used to create or update the stack set.
Use customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Prerequisites: Granting Permissions for Stack Set Operations in the CloudFormation User Guide.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the IAM role used to create or update the stack set.
Use customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Prerequisites: Granting Permissions for Stack Set Operations in the CloudFormation User Guide.
The Amazon Resource Name (ARN) of the IAM role used to perform this stack set operation.
Use customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Define Permissions for Multiple Administrators in the CloudFormation User Guide.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the IAM role used to perform this stack set operation.
Use customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Define Permissions for Multiple Administrators in the CloudFormation User Guide.
For nested stacks--stacks created as resources for another stack--the stack ID of the direct parent of this stack. For the first level of nested stacks, the root stack is also the parent stack.
For nested stacks--stacks created as resources for another stack--the stack ID of the direct parent of this stack. For the first level of nested stacks, the root stack is also the parent stack.
Summarizes information about whether a stack's actual configuration differs, or has drifted, from its expected configuration, as defined in the stack template and any values specified as template parameters. For more information, see Detecting Unregulated Configuration Changes to Stacks and Resources.
"
+ "documentation":"
Summarizes information about whether a stack's actual configuration differs, or has drifted, from its expected configuration, as defined in the stack template and any values specified as template parameters. For more information, see Detecting Unregulated Configuration Changes to Stacks and Resources.
In some cases, you must explicitly acknowledge that your stack template contains certain capabilities in order for CloudFormation to update the stack.
CAPABILITY_IAM and CAPABILITY_NAMED_IAM
Some stack templates might include resources that can affect permissions in your Amazon Web Services account; for example, by creating new Identity and Access Management (IAM) users. For those stacks, you must explicitly acknowledge this by specifying one of these capabilities.
The following IAM resources require you to specify either the CAPABILITY_IAM or CAPABILITY_NAMED_IAM capability.
If you have IAM resources, you can specify either capability.
If you have IAM resources with custom names, you must specify CAPABILITY_NAMED_IAM.
If you don't specify either of these capabilities, CloudFormation returns an InsufficientCapabilities error.
If your stack template contains these resources, we suggest that you review all permissions associated with them and edit their permissions if necessary.
Some template contain macros. Macros perform custom processing on templates; this can include simple actions like find-and-replace operations, all the way to extensive transformations of entire templates. Because of this, users typically create a change set from the processed template, so that they can review the changes resulting from the macros before actually updating the stack. If your stack template contains one or more macros, and you choose to update a stack directly from the processed template, without first reviewing the resulting changes in a change set, you must acknowledge this capability. This includes the AWS::Include and AWS::Serverless transforms, which are macros hosted by CloudFormation.
If you want to update a stack from a stack template that contains macros and nested stacks, you must update the stack directly from the template using this capability.
You should only update stacks directly from a stack template that contains macros if you know what processing the macro performs.
Each macro relies on an underlying Lambda service function for processing stack templates. Be aware that the Lambda function owner can update the function operation without CloudFormation being notified.
In some cases, you must explicitly acknowledge that your stack template contains certain capabilities in order for CloudFormation to update the stack.
CAPABILITY_IAM and CAPABILITY_NAMED_IAM
Some stack templates might include resources that can affect permissions in your Amazon Web Services account; for example, by creating new Identity and Access Management (IAM) users. For those stacks, you must explicitly acknowledge this by specifying one of these capabilities.
The following IAM resources require you to specify either the CAPABILITY_IAM or CAPABILITY_NAMED_IAM capability.
If you have IAM resources, you can specify either capability.
If you have IAM resources with custom names, you must specify CAPABILITY_NAMED_IAM.
If you don't specify either of these capabilities, CloudFormation returns an InsufficientCapabilities error.
If your stack template contains these resources, we suggest that you review all permissions associated with them and edit their permissions if necessary.
Some template contain macros. Macros perform custom processing on templates; this can include simple actions like find-and-replace operations, all the way to extensive transformations of entire templates. Because of this, users typically create a change set from the processed template, so that they can review the changes resulting from the macros before actually updating the stack. If your stack template contains one or more macros, and you choose to update a stack directly from the processed template, without first reviewing the resulting changes in a change set, you must acknowledge this capability. This includes the AWS::Include and AWS::Serverless transforms, which are macros hosted by CloudFormation.
If you want to update a stack from a stack template that contains macros and nested stacks, you must update the stack directly from the template using this capability.
You should only update stacks directly from a stack template that contains macros if you know what processing the macro performs.
Each macro relies on an underlying Lambda service function for processing stack templates. Be aware that the Lambda function owner can update the function operation without CloudFormation being notified.
A unique identifier for this UpdateStack request. Specify this token if you plan to retry requests so that CloudFormation knows that you're not attempting to update a stack with the same name. You might retry UpdateStack requests to ensure that CloudFormation successfully received them.
All events triggered by a given stack operation are assigned the same client request token, which you can use to track operations. For example, if you execute a CreateStack operation with the token token1, then all the StackEvents generated by that operation will have ClientRequestToken set as token1.
In the console, stack operations display the client request token on the Events tab. Stack operations that are initiated from the console use the token format Console-StackOperation-ID, which helps you easily identify the stack operation . For example, if you create a stack using the console, each stack event would be assigned the same token in the following format: Console-CreateStack-7f59c3cf-00d2-40c7-b2ff-e75db0987002.
This deletion policy deletes newly created resources, but retains existing resources, when a stack operation is rolled back. This ensures new, empty, and unused resources are deleted, while critical resources and their data are retained. RetainExceptOnCreate can be specified for any resource that supports the DeletionPolicy attribute.
In some cases, you must explicitly acknowledge that your stack template contains certain capabilities in order for CloudFormation to update the stack set and its associated stack instances.
CAPABILITY_IAM and CAPABILITY_NAMED_IAM
Some stack templates might include resources that can affect permissions in your Amazon Web Services account; for example, by creating new Identity and Access Management (IAM) users. For those stacks sets, you must explicitly acknowledge this by specifying one of these capabilities.
The following IAM resources require you to specify either the CAPABILITY_IAM or CAPABILITY_NAMED_IAM capability.
If you have IAM resources, you can specify either capability.
If you have IAM resources with custom names, you must specify CAPABILITY_NAMED_IAM.
If you don't specify either of these capabilities, CloudFormation returns an InsufficientCapabilities error.
If your stack template contains these resources, we recommend that you review all permissions associated with them and edit their permissions if necessary.
Some templates reference macros. If your stack set template references one or more macros, you must update the stack set directly from the processed template, without first reviewing the resulting changes in a change set. To update the stack set directly, you must acknowledge this capability. For more information, see Using CloudFormation Macros to Perform Custom Processing on Templates.
Stack sets with service-managed permissions do not currently support the use of macros in templates. (This includes the AWS::Include and AWS::Serverless transforms, which are macros hosted by CloudFormation.) Even if you specify this capability for a stack set with service-managed permissions, if you reference a macro in your template the stack set operation will fail.
"
+ "documentation":"
In some cases, you must explicitly acknowledge that your stack template contains certain capabilities in order for CloudFormation to update the stack set and its associated stack instances.
CAPABILITY_IAM and CAPABILITY_NAMED_IAM
Some stack templates might include resources that can affect permissions in your Amazon Web Services account; for example, by creating new Identity and Access Management (IAM) users. For those stacks sets, you must explicitly acknowledge this by specifying one of these capabilities.
The following IAM resources require you to specify either the CAPABILITY_IAM or CAPABILITY_NAMED_IAM capability.
If you have IAM resources, you can specify either capability.
If you have IAM resources with custom names, you must specify CAPABILITY_NAMED_IAM.
If you don't specify either of these capabilities, CloudFormation returns an InsufficientCapabilities error.
If your stack template contains these resources, we recommend that you review all permissions associated with them and edit their permissions if necessary.
Some templates reference macros. If your stack set template references one or more macros, you must update the stack set directly from the processed template, without first reviewing the resulting changes in a change set. To update the stack set directly, you must acknowledge this capability. For more information, see Using CloudFormation Macros to Perform Custom Processing on Templates.
Stack sets with service-managed permissions do not currently support the use of macros in templates. (This includes the AWS::Include and AWS::Serverless transforms, which are macros hosted by CloudFormation.) Even if you specify this capability for a stack set with service-managed permissions, if you reference a macro in your template the stack set operation will fail.
The Amazon Resource Name (ARN) of the IAM role to use to update this stack set.
Specify an IAM role only if you are using customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Granting Permissions for Stack Set Operations in the CloudFormation User Guide.
If you specified a customized administrator role when you created the stack set, you must specify a customized administrator role, even if it is the same customized administrator role used with this stack set previously.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the IAM role to use to update this stack set.
Specify an IAM role only if you are using customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Granting Permissions for Stack Set Operations in the CloudFormation User Guide.
If you specified a customized administrator role when you created the stack set, you must specify a customized administrator role, even if it is the same customized administrator role used with this stack set previously.
Creates a staging distribution using the configuration of the provided primary distribution. A staging distribution is a copy of an existing distribution (called the primary distribution) that you can use in a continuous deployment workflow.
After you create a staging distribution, you can use UpdateDistribution to modify the staging distribution's configuration. Then you can use CreateContinuousDeploymentPolicy to incrementally move traffic to the staging distribution.
"
+ "documentation":"
Creates a staging distribution using the configuration of the provided primary distribution. A staging distribution is a copy of an existing distribution (called the primary distribution) that you can use in a continuous deployment workflow.
After you create a staging distribution, you can use UpdateDistribution to modify the staging distribution's configuration. Then you can use CreateContinuousDeploymentPolicy to incrementally move traffic to the staging distribution.
This API operation requires the following IAM permissions:
Copies the staging distribution's configuration to its corresponding primary distribution. The primary distribution retains its Aliases (also known as alternate domain names or CNAMEs) and ContinuousDeploymentPolicyId value, but otherwise its configuration is overwritten to match the staging distribution.
You can use this operation in a continuous deployment workflow after you have tested configuration changes on the staging distribution. After using a continuous deployment policy to move a portion of your domain name's traffic to the staging distribution and verifying that it works as intended, you can use this operation to copy the staging distribution's configuration to the primary distribution. This action will disable the continuous deployment policy and move your domain's traffic back to the primary distribution.
"
+ "documentation":"
Copies the staging distribution's configuration to its corresponding primary distribution. The primary distribution retains its Aliases (also known as alternate domain names or CNAMEs) and ContinuousDeploymentPolicyId value, but otherwise its configuration is overwritten to match the staging distribution.
You can use this operation in a continuous deployment workflow after you have tested configuration changes on the staging distribution. After using a continuous deployment policy to move a portion of your domain name's traffic to the staging distribution and verifying that it works as intended, you can use this operation to copy the staging distribution's configuration to the primary distribution. This action will disable the continuous deployment policy and move your domain's traffic back to the primary distribution.
This API operation requires the following IAM permissions:
This field is true if any of the Amazon Web Services accounts in the list have active CloudFront key pairs that CloudFront can use to verify the signatures of signed URLs and signed cookies. If not, this field is false.
"
+ "documentation":"
This field is true if any of the Amazon Web Services accounts in the list are configured as trusted signers. If not, this field is false.
A value that uniquely identifies a request to create a resource. This helps to prevent CloudFront from creating a duplicate resource if you accidentally resubmit an identical request.
A Boolean flag to specify the state of the staging distribution when it's created. When you set this value to True, the staging distribution is enabled. When you set this value to False, the staging distribution is disabled.
If you omit this field, the default value is True.
This field is true if any of the Amazon Web Services accounts have public keys that CloudFront can use to verify the signatures of signed URLs and signed cookies. If not, this field is false.
"
+ "documentation":"
This field is true if any of the Amazon Web Services accounts in the list are configured as trusted signers. If not, this field is false.
A field in a CloudTrail event record on which to filter events to be logged. For event data stores for Config configuration items, Audit Manager evidence, or non-Amazon Web Services events, the field is used only for selecting events as filtering is not supported.
For CloudTrail event records, supported fields include readOnly, eventCategory, eventSource (for management events), eventName, resources.type, and resources.ARN.
For event data stores for Config configuration items, Audit Manager evidence, or non-Amazon Web Services events, the only supported field is eventCategory.
readOnly - Optional. Can be set to Equals a value of true or false. If you do not add this field, CloudTrail logs both read and write events. A value of true logs only read events. A value of false logs only write events.
eventSource - For filtering management events only. This can be set only to NotEqualskms.amazonaws.com.
eventName - Can use any operator. You can use it to filter in or filter out any data event logged to CloudTrail, such as PutBucket or GetSnapshotBlock. You can have multiple values for this field, separated by commas.
eventCategory - This is required and must be set to Equals.
For CloudTrail event records, the value must be Management or Data.
For Config configuration items, the value must be ConfigurationItem.
For Audit Manager evidence, the value must be Evidence.
For non-Amazon Web Services events, the value must be ActivityAuditLog.
resources.type - This field is required for CloudTrail data events. resources.type can only use the Equals operator, and the value can be one of the following:
AWS::DynamoDB::Table
AWS::Lambda::Function
AWS::S3::Object
AWS::CloudTrail::Channel
AWS::CodeWhisperer::Profile
AWS::Cognito::IdentityPool
AWS::DynamoDB::Stream
AWS::EC2::Snapshot
AWS::EMRWAL::Workspace
AWS::FinSpace::Environment
AWS::Glue::Table
AWS::GuardDuty::Detector
AWS::KendraRanking::ExecutionPlan
AWS::ManagedBlockchain::Node
AWS::SageMaker::ExperimentTrialComponent
AWS::SageMaker::FeatureGroup
AWS::S3::AccessPoint
AWS::S3ObjectLambda::AccessPoint
AWS::S3Outposts::Object
You can have only one resources.type field per selector. To log data events on more than one resource type, add another selector.
resources.ARN - You can use any operator with resources.ARN, but if you use Equals or NotEquals, the value must exactly match the ARN of a valid resource of the type you've specified in the template as the value of resources.type. For example, if resources.type equals AWS::S3::Object, the ARN must be in one of the following formats. To log all data events for all objects in a specific S3 bucket, use the StartsWith operator, and include only the bucket ARN as the matching value.
The trailing slash is intentional; do not exclude it. Replace the text between less than and greater than symbols (<>) with resource-specific information.
arn:<partition>:s3:::<bucket_name>/
arn:<partition>:s3:::<bucket_name>/<object_path>/
When resources.type equals AWS::DynamoDB::Table, and the operator is set to Equals or NotEquals, the ARN must be in the following format:
When resources.type equals AWS::SageMaker::ExperimentTrialComponent, and the operator is set to Equals or NotEquals, the ARN must be in the following format:
When resources.type equals AWS::S3::AccessPoint, and the operator is set to Equals or NotEquals, the ARN must be in one of the following formats. To log events on all objects in an S3 access point, we recommend that you use only the access point ARN, don’t include the object path, and use the StartsWith or NotStartsWith operators.
A field in a CloudTrail event record on which to filter events to be logged. For event data stores for Config configuration items, Audit Manager evidence, or non-Amazon Web Services events, the field is used only for selecting events as filtering is not supported.
For CloudTrail event records, supported fields include readOnly, eventCategory, eventSource (for management events), eventName, resources.type, and resources.ARN.
For event data stores for Config configuration items, Audit Manager evidence, or non-Amazon Web Services events, the only supported field is eventCategory.
readOnly - Optional. Can be set to Equals a value of true or false. If you do not add this field, CloudTrail logs both read and write events. A value of true logs only read events. A value of false logs only write events.
eventSource - For filtering management events only. This can be set only to NotEqualskms.amazonaws.com.
eventName - Can use any operator. You can use it to filter in or filter out any data event logged to CloudTrail, such as PutBucket or GetSnapshotBlock. You can have multiple values for this field, separated by commas.
eventCategory - This is required and must be set to Equals.
For CloudTrail event records, the value must be Management or Data.
For Config configuration items, the value must be ConfigurationItem.
For Audit Manager evidence, the value must be Evidence.
For non-Amazon Web Services events, the value must be ActivityAuditLog.
resources.type - This field is required for CloudTrail data events. resources.type can only use the Equals operator, and the value can be one of the following:
AWS::DynamoDB::Table
AWS::Lambda::Function
AWS::S3::Object
AWS::CloudTrail::Channel
AWS::CodeWhisperer::Profile
AWS::Cognito::IdentityPool
AWS::DynamoDB::Stream
AWS::EC2::Snapshot
AWS::EMRWAL::Workspace
AWS::FinSpace::Environment
AWS::Glue::Table
AWS::GuardDuty::Detector
AWS::KendraRanking::ExecutionPlan
AWS::ManagedBlockchain::Network
AWS::ManagedBlockchain::Node
AWS::MedicalImaging::Datastore
AWS::SageMaker::ExperimentTrialComponent
AWS::SageMaker::FeatureGroup
AWS::S3::AccessPoint
AWS::S3ObjectLambda::AccessPoint
AWS::S3Outposts::Object
AWS::SSMMessages::ControlChannel
AWS::VerifiedPermissions::PolicyStore
You can have only one resources.type field per selector. To log data events on more than one resource type, add another selector.
resources.ARN - You can use any operator with resources.ARN, but if you use Equals or NotEquals, the value must exactly match the ARN of a valid resource of the type you've specified in the template as the value of resources.type. For example, if resources.type equals AWS::S3::Object, the ARN must be in one of the following formats. To log all data events for all objects in a specific S3 bucket, use the StartsWith operator, and include only the bucket ARN as the matching value.
The trailing slash is intentional; do not exclude it. Replace the text between less than and greater than symbols (<>) with resource-specific information.
arn:<partition>:s3:::<bucket_name>/
arn:<partition>:s3:::<bucket_name>/<object_path>/
When resources.type equals AWS::DynamoDB::Table, and the operator is set to Equals or NotEquals, the ARN must be in the following format:
When resources.type equals AWS::SageMaker::ExperimentTrialComponent, and the operator is set to Equals or NotEquals, the ARN must be in the following format:
When resources.type equals AWS::S3::AccessPoint, and the operator is set to Equals or NotEquals, the ARN must be in one of the following formats. To log events on all objects in an S3 access point, we recommend that you use only the access point ARN, don’t include the object path, and use the StartsWith or NotStartsWith operators.
When resources.type equals AWS::VerifiedPermissions::PolicyStore, and the operator is set to Equals or NotEquals, the ARN must be in the following format:
The retention period of the event data store, in days. You can set a retention period of up to 2557 days, the equivalent of seven years.
"
+ "documentation":"
The retention period of the event data store, in days. You can set a retention period of up to 2557 days, the equivalent of seven years. CloudTrail Lake determines whether to retain an event by checking if the eventTime of the event is within the specified retention period. For example, if you set a retention period of 90 days, CloudTrail will remove events when the eventTime is older than 90 days.
If you plan to copy trail events to this event data store, we recommend that you consider both the age of the events that you want to copy as well as how long you want to keep the copied events in your event data store. For example, if you copy trail events that are 5 years old and specify a retention period of 7 years, the event data store will retain those events for two years.
The resource type in which you want to log data events. You can specify the following basic event selector resource types:
AWS::DynamoDB::Table
AWS::Lambda::Function
AWS::S3::Object
The following resource types are also available through advanced event selectors. Basic event selector resource types are valid in advanced event selectors, but advanced event selector resource types are not valid in basic event selectors. For more information, see AdvancedFieldSelector.
AWS::CloudTrail::Channel
AWS::CodeWhisperer::Profile
AWS::Cognito::IdentityPool
AWS::DynamoDB::Stream
AWS::EC2::Snapshot
AWS::EMRWAL::Workspace
AWS::FinSpace::Environment
AWS::Glue::Table
AWS::GuardDuty::Detector
AWS::KendraRanking::ExecutionPlan
AWS::ManagedBlockchain::Node
AWS::SageMaker::ExperimentTrialComponent
AWS::SageMaker::FeatureGroup
AWS::S3::AccessPoint
AWS::S3ObjectLambda::AccessPoint
AWS::S3Outposts::Object
"
+ "documentation":"
The resource type in which you want to log data events. You can specify the following basic event selector resource types:
AWS::DynamoDB::Table
AWS::Lambda::Function
AWS::S3::Object
The following resource types are also available through advanced event selectors. Basic event selector resource types are valid in advanced event selectors, but advanced event selector resource types are not valid in basic event selectors. For more information, see AdvancedFieldSelector.
You are already running the maximum number of concurrent queries. Wait a minute for some queries to finish, and then run the query again.
",
+ "documentation":"
You are already running the maximum number of concurrent queries. The maximum number of concurrent queries is 10. Wait a minute for some queries to finish, and then run the query again.
The retention period of the event data store, in days. You can set a retention period of up to 2557 days, the equivalent of seven years. CloudTrail Lake determines whether to retain an event by checking if the eventTime of the event is within the specified retention period. For example, if you set a retention period of 90 days, CloudTrail will remove events when the eventTime is older than 90 days.
If you decrease the retention period of an event data store, CloudTrail will remove any events with an eventTime older than the new retention period. For example, if the previous retention period was 365 days and you decrease it to 100 days, CloudTrail will remove events with an eventTime older than 100 days.
Creates a connection that can then be given to other AWS services like CodePipeline so that it can access third-party code repositories. The connection is in pending status until the third-party connection handshake is completed from the console.
"
+ "documentation":"
Creates a connection that can then be given to other Amazon Web Services services like CodePipeline so that it can access third-party code repositories. The connection is in pending status until the third-party connection handshake is completed from the console.
The Amazon Resource Name (ARN) of the connection. The ARN is used as the connection reference when the connection is shared between AWS services.
The ARN is never reused if the connection is deleted.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the connection. The ARN is used as the connection reference when the connection is shared between Amazon Web Services.
The ARN is never reused if the connection is deleted.
The Amazon Resource Name (ARN) of the host associated with the connection.
"
}
},
- "documentation":"
A resource that is used to connect third-party source providers with services like AWS CodePipeline.
Note: A connection created through CloudFormation, the CLI, or the SDK is in `PENDING` status by default. You can make its status `AVAILABLE` by updating the connection in the console.
"
+ "documentation":"
A resource that is used to connect third-party source providers with services like CodePipeline.
Note: A connection created through CloudFormation, the CLI, or the SDK is in `PENDING` status by default. You can make its status `AVAILABLE` by updating the connection in the console.
The Amazon Resource Name (ARN) of the connection to be created. The ARN is used as the connection reference when the connection is shared between AWS services.
The ARN is never reused if the connection is deleted.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the connection to be created. The ARN is used as the connection reference when the connection is shared between Amazon Web Services services.
The ARN is never reused if the connection is deleted.
This AWS CodeStar Connections API Reference provides descriptions and usage examples of the operations and data types for the AWS CodeStar Connections API. You can use the connections API to work with connections and installations.
Connections are configurations that you use to connect AWS resources to external code repositories. Each connection is a resource that can be given to services such as CodePipeline to connect to a third-party repository such as Bitbucket. For example, you can add the connection in CodePipeline so that it triggers your pipeline when a code change is made to your third-party code repository. Each connection is named and associated with a unique ARN that is used to reference the connection.
When you create a connection, the console initiates a third-party connection handshake. Installations are the apps that are used to conduct this handshake. For example, the installation for the Bitbucket provider type is the Bitbucket app. When you create a connection, you can choose an existing installation or create one.
When you want to create a connection to an installed provider type such as GitHub Enterprise Server, you create a host for your connections.
You can work with connections by calling:
CreateConnection, which creates a uniquely named connection that can be referenced by services such as CodePipeline.
This Amazon Web Services CodeStar Connections API Reference provides descriptions and usage examples of the operations and data types for the Amazon Web Services CodeStar Connections API. You can use the connections API to work with connections and installations.
Connections are configurations that you use to connect Amazon Web Services resources to external code repositories. Each connection is a resource that can be given to services such as CodePipeline to connect to a third-party repository such as Bitbucket. For example, you can add the connection in CodePipeline so that it triggers your pipeline when a code change is made to your third-party code repository. Each connection is named and associated with a unique ARN that is used to reference the connection.
When you create a connection, the console initiates a third-party connection handshake. Installations are the apps that are used to conduct this handshake. For example, the installation for the Bitbucket provider type is the Bitbucket app. When you create a connection, you can choose an existing installation or create one.
When you want to create a connection to an installed provider type such as GitHub Enterprise Server, you create a host for your connections.
You can work with connections by calling:
CreateConnection, which creates a uniquely named connection that can be referenced by services such as CodePipeline.
GetHost, which returns information about the host, including the setup status.
ListHosts, which lists the hosts associated with your account.
You can work with tags in Amazon Web Services CodeStar Connections by calling the following:
ListTagsForResource, which gets information about Amazon Web Services tags for a specified Amazon Resource Name (ARN) in Amazon Web Services CodeStar Connections.
TagResource, which adds or updates tags for a resource in Amazon Web Services CodeStar Connections.
UntagResource, which removes tags for a resource in Amazon Web Services CodeStar Connections.
For information about how to use Amazon Web Services CodeStar Connections, see the Developer Tools User Guide.
Adds additional user attributes to the user pool schema.
"
+ "documentation":"
Adds additional user attributes to the user pool schema.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Calling this action requires developer credentials.
"
+ "documentation":"
Adds the specified user to the specified group.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Confirms user registration as an admin without using a confirmation code. Works on any user.
Calling this action requires developer credentials.
"
+ "documentation":"
Confirms user registration as an admin without using a confirmation code. Works on any user.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
If MessageAction isn't set, the default is to send a welcome message via email or phone (SMS).
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
This message is based on a template that you configured in your call to create or update a user pool. This template includes your custom sign-up instructions and placeholders for user name and temporary password.
Alternatively, you can call AdminCreateUser with SUPPRESS for the MessageAction parameter, and Amazon Cognito won't send any email.
In either case, the user will be in the FORCE_CHANGE_PASSWORD state until they sign in and change their password.
AdminCreateUser requires developer credentials.
"
+ "documentation":"
Creates a new user in the specified user pool.
If MessageAction isn't set, the default is to send a welcome message via email or phone (SMS).
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
This message is based on a template that you configured in your call to create or update a user pool. This template includes your custom sign-up instructions and placeholders for user name and temporary password.
Alternatively, you can call AdminCreateUser with SUPPRESS for the MessageAction parameter, and Amazon Cognito won't send any email.
In either case, the user will be in the FORCE_CHANGE_PASSWORD state until they sign in and change their password.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Deletes a user as an administrator. Works on any user.
Calling this action requires developer credentials.
"
+ "documentation":"
Deletes a user as an administrator. Works on any user.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Deletes the user attributes in a user pool as an administrator. Works on any user.
Calling this action requires developer credentials.
"
+ "documentation":"
Deletes the user attributes in a user pool as an administrator. Works on any user.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Prevents the user from signing in with the specified external (SAML or social) identity provider (IdP). If the user that you want to deactivate is a Amazon Cognito user pools native username + password user, they can't use their password to sign in. If the user to deactivate is a linked external IdP user, any link between that user and an existing user is removed. When the external user signs in again, and the user is no longer attached to the previously linked DestinationUser, the user must create a new user account. See AdminLinkProviderForUser.
This action is enabled only for admin access and requires developer credentials.
The ProviderName must match the value specified when creating an IdP for the pool.
To deactivate a native username + password user, the ProviderName value must be Cognito and the ProviderAttributeName must be Cognito_Subject. The ProviderAttributeValue must be the name that is used in the user pool for the user.
The ProviderAttributeName must always be Cognito_Subject for social IdPs. The ProviderAttributeValue must always be the exact subject that was used when the user was originally linked as a source user.
For de-linking a SAML identity, there are two scenarios. If the linked identity has not yet been used to sign in, the ProviderAttributeName and ProviderAttributeValue must be the same values that were used for the SourceUser when the identities were originally linked using AdminLinkProviderForUser call. (If the linking was done with ProviderAttributeName set to Cognito_Subject, the same applies here). However, if the user has already signed in, the ProviderAttributeName must be Cognito_Subject and ProviderAttributeValue must be the subject of the SAML assertion.
"
+ "documentation":"
Prevents the user from signing in with the specified external (SAML or social) identity provider (IdP). If the user that you want to deactivate is a Amazon Cognito user pools native username + password user, they can't use their password to sign in. If the user to deactivate is a linked external IdP user, any link between that user and an existing user is removed. When the external user signs in again, and the user is no longer attached to the previously linked DestinationUser, the user must create a new user account. See AdminLinkProviderForUser.
The ProviderName must match the value specified when creating an IdP for the pool.
To deactivate a native username + password user, the ProviderName value must be Cognito and the ProviderAttributeName must be Cognito_Subject. The ProviderAttributeValue must be the name that is used in the user pool for the user.
The ProviderAttributeName must always be Cognito_Subject for social IdPs. The ProviderAttributeValue must always be the exact subject that was used when the user was originally linked as a source user.
For de-linking a SAML identity, there are two scenarios. If the linked identity has not yet been used to sign in, the ProviderAttributeName and ProviderAttributeValue must be the same values that were used for the SourceUser when the identities were originally linked using AdminLinkProviderForUser call. (If the linking was done with ProviderAttributeName set to Cognito_Subject, the same applies here). However, if the user has already signed in, the ProviderAttributeName must be Cognito_Subject and ProviderAttributeValue must be the subject of the SAML assertion.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Deactivates a user and revokes all access tokens for the user. A deactivated user can't sign in, but still appears in the responses to GetUser and ListUsers API requests.
You must make this API request with Amazon Web Services credentials that have cognito-idp:AdminDisableUser permissions.
"
+ "documentation":"
Deactivates a user and revokes all access tokens for the user. A deactivated user can't sign in, but still appears in the responses to GetUser and ListUsers API requests.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Enables the specified user as an administrator. Works on any user.
Calling this action requires developer credentials.
"
+ "documentation":"
Enables the specified user as an administrator. Works on any user.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Calling this action requires developer credentials.
"
+ "documentation":"
Forgets the device, as an administrator.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Calling this action requires developer credentials.
"
+ "documentation":"
Gets the device, as an administrator.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Gets the specified user by user name in a user pool as an administrator. Works on any user.
Calling this action requires developer credentials.
"
+ "documentation":"
Gets the specified user by user name in a user pool as an administrator. Works on any user.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Initiates the authentication flow, as an administrator.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Calling this action requires developer credentials.
"
+ "documentation":"
Initiates the authentication flow, as an administrator.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Links an existing user account in a user pool (DestinationUser) to an identity from an external IdP (SourceUser) based on a specified attribute name and value from the external IdP. This allows you to create a link from the existing user account to an external federated user identity that has not yet been used to sign in. You can then use the federated user identity to sign in as the existing user account.
For example, if there is an existing user with a username and password, this API links that user to a federated user identity. When the user signs in with a federated user identity, they sign in as the existing user account.
The maximum number of federated identities linked to a user is five.
Because this API allows a user with an external federated identity to sign in as an existing user in the user pool, it is critical that it only be used with external IdPs and provider attributes that have been trusted by the application owner.
This action is administrative and requires developer credentials.
"
+ "documentation":"
Links an existing user account in a user pool (DestinationUser) to an identity from an external IdP (SourceUser) based on a specified attribute name and value from the external IdP. This allows you to create a link from the existing user account to an external federated user identity that has not yet been used to sign in. You can then use the federated user identity to sign in as the existing user account.
For example, if there is an existing user with a username and password, this API links that user to a federated user identity. When the user signs in with a federated user identity, they sign in as the existing user account.
The maximum number of federated identities linked to a user is five.
Because this API allows a user with an external federated identity to sign in as an existing user in the user pool, it is critical that it only be used with external IdPs and provider attributes that have been trusted by the application owner.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Calling this action requires developer credentials.
"
+ "documentation":"
Lists devices, as an administrator.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Calling this action requires developer credentials.
"
+ "documentation":"
Lists the groups that the user belongs to.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
A history of user activity and any risks detected as part of Amazon Cognito advanced security.
"
+ "documentation":"
A history of user activity and any risks detected as part of Amazon Cognito advanced security.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Removes the specified user from the specified group.
Calling this action requires developer credentials.
"
+ "documentation":"
Removes the specified user from the specified group.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Resets the specified user's password in a user pool as an administrator. Works on any user.
When a developer calls this API, the current password is invalidated, so it must be changed. If a user tries to sign in after the API is called, the app will get a PasswordResetRequiredException exception back and should direct the user down the flow to reset the password, which is the same as the forgot password flow. In addition, if the user pool has phone verification selected and a verified phone number exists for the user, or if email verification is selected and a verified email exists for the user, calling this API will also result in sending a message to the end user with the code to change their password.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Calling this action requires developer credentials.
"
+ "documentation":"
Resets the specified user's password in a user pool as an administrator. Works on any user.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Deactivates a user's password, requiring them to change it. If a user tries to sign in after the API is called, Amazon Cognito responds with a PasswordResetRequiredException error. Your app must then perform the actions that reset your user's password: the forgot-password flow. In addition, if the user pool has phone verification selected and a verified phone number exists for the user, or if email verification is selected and a verified email exists for the user, calling this API will also result in sending a message to the end user with the code to change their password.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Responds to an authentication challenge, as an administrator.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Calling this action requires developer credentials.
"
+ "documentation":"
Responds to an authentication challenge, as an administrator.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
The user's multi-factor authentication (MFA) preference, including which MFA options are activated, and if any are preferred. Only one factor can be set as preferred. The preferred MFA factor will be used to authenticate a user if multiple factors are activated. If multiple options are activated and no preference is set, a challenge to choose an MFA option will be returned during sign-in.
"
+ "documentation":"
The user's multi-factor authentication (MFA) preference, including which MFA options are activated, and if any are preferred. Only one factor can be set as preferred. The preferred MFA factor will be used to authenticate a user if multiple factors are activated. If multiple options are activated and no preference is set, a challenge to choose an MFA option will be returned during sign-in.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Sets the specified user's password in a user pool as an administrator. Works on any user.
The password can be temporary or permanent. If it is temporary, the user status enters the FORCE_CHANGE_PASSWORD state. When the user next tries to sign in, the InitiateAuth/AdminInitiateAuth response will contain the NEW_PASSWORD_REQUIRED challenge. If the user doesn't sign in before it expires, the user won't be able to sign in, and an administrator must reset their password.
Once the user has set a new password, or the password is permanent, the user status is set to Confirmed.
"
+ "documentation":"
Sets the specified user's password in a user pool as an administrator. Works on any user.
The password can be temporary or permanent. If it is temporary, the user status enters the FORCE_CHANGE_PASSWORD state. When the user next tries to sign in, the InitiateAuth/AdminInitiateAuth response will contain the NEW_PASSWORD_REQUIRED challenge. If the user doesn't sign in before it expires, the user won't be able to sign in, and an administrator must reset their password.
Once the user has set a new password, or the password is permanent, the user status is set to Confirmed.
AdminSetUserPassword can set a password for the user profile that Amazon Cognito creates for third-party federated users. When you set a password, the federated user's status changes from EXTERNAL_PROVIDER to CONFIRMED. A user in this state can sign in as a federated user, and initiate authentication flows in the API like a linked native user. They can also modify their password and attributes in token-authenticated API requests like ChangePassword and UpdateUserAttributes. As a best security practice and to keep users in sync with your external IdP, don't set passwords on federated user profiles. To set up a federated user for native sign-in with a linked native user, refer to Linking federated users to an existing user profile.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
This action is no longer supported. You can use it to configure only SMS MFA. You can't use it to configure time-based one-time password (TOTP) software token MFA. To configure either type of MFA, use AdminSetUserMFAPreference instead.
"
+ "documentation":"
This action is no longer supported. You can use it to configure only SMS MFA. You can't use it to configure time-based one-time password (TOTP) software token MFA. To configure either type of MFA, use AdminSetUserMFAPreference instead.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Provides feedback for an authentication event indicating if it was from a valid user. This feedback is used for improving the risk evaluation decision for the user pool as part of Amazon Cognito advanced security.
"
+ "documentation":"
Provides feedback for an authentication event indicating if it was from a valid user. This feedback is used for improving the risk evaluation decision for the user pool as part of Amazon Cognito advanced security.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Calling this action requires developer credentials.
"
+ "documentation":"
Updates the device status as an administrator.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Updates the specified user's attributes, including developer attributes, as an administrator. Works on any user.
For custom attributes, you must prepend the custom: prefix to the attribute name.
In addition to updating user attributes, this API can also be used to mark phone and email as verified.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Calling this action requires developer credentials.
"
+ "documentation":"
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Updates the specified user's attributes, including developer attributes, as an administrator. Works on any user.
For custom attributes, you must prepend the custom: prefix to the attribute name.
In addition to updating user attributes, this API can also be used to mark phone and email as verified.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Signs out a user from all devices. You must sign AdminUserGlobalSignOut requests with Amazon Web Services credentials. It also invalidates all refresh tokens that Amazon Cognito has issued to a user. The user's current access and ID tokens remain valid until they expire. By default, access and ID tokens expire one hour after they're issued. A user can still use a hosted UI cookie to retrieve new tokens for the duration of the cookie validity period of 1 hour.
Calling this action requires developer credentials.
"
+ "documentation":"
Signs out a user from all devices. AdminUserGlobalSignOut invalidates all identity, access and refresh tokens that Amazon Cognito has issued to a user. A user can still use a hosted UI cookie to retrieve new tokens for the duration of the 1-hour cookie validity period.
Your app isn't aware that a user's access token is revoked unless it attempts to authorize a user pools API request with an access token that contains the scope aws.cognito.signin.user.admin. Your app might otherwise accept access tokens until they expire.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Begins setup of time-based one-time password (TOTP) multi-factor authentication (MFA) for a user, with a unique private key that Amazon Cognito generates and returns in the API response. You can authorize an AssociateSoftwareToken request with either the user's access token, or a session string from a challenge response that you received from Amazon Cognito.
Amazon Cognito disassociates an existing software token when you verify the new token in a VerifySoftwareToken API request. If you don't verify the software token and your user pool doesn't require MFA, the user can then authenticate with user name and password credentials alone. If your user pool requires TOTP MFA, Amazon Cognito generates an MFA_SETUP or SOFTWARE_TOKEN_SETUP challenge each time your user signs. Complete setup with AssociateSoftwareToken and VerifySoftwareToken.
After you set up software token MFA for your user, Amazon Cognito generates a SOFTWARE_TOKEN_MFA challenge when they authenticate. Respond to this challenge with your user's TOTP.
"
+ "documentation":"
Begins setup of time-based one-time password (TOTP) multi-factor authentication (MFA) for a user, with a unique private key that Amazon Cognito generates and returns in the API response. You can authorize an AssociateSoftwareToken request with either the user's access token, or a session string from a challenge response that you received from Amazon Cognito.
Amazon Cognito disassociates an existing software token when you verify the new token in a VerifySoftwareToken API request. If you don't verify the software token and your user pool doesn't require MFA, the user can then authenticate with user name and password credentials alone. If your user pool requires TOTP MFA, Amazon Cognito generates an MFA_SETUP or SOFTWARE_TOKEN_SETUP challenge each time your user signs. Complete setup with AssociateSoftwareToken and VerifySoftwareToken.
After you set up software token MFA for your user, Amazon Cognito generates a SOFTWARE_TOKEN_MFA challenge when they authenticate. Respond to this challenge with your user's TOTP.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Changes the password for a specified user in a user pool.
",
+ "documentation":"
Changes the password for a specified user in a user pool.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Confirms tracking of the device. This API call is the call that begins device tracking.
"
+ "documentation":"
Confirms tracking of the device. This API call is the call that begins device tracking.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Allows a user to enter a confirmation code to reset a forgotten password.
",
+ "documentation":"
Allows a user to enter a confirmation code to reset a forgotten password.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Calling this action requires developer credentials.
"
+ "documentation":"
Creates a new group in the specified user pool.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Creates a new OAuth2.0 resource server and defines custom scopes within it.
"
+ "documentation":"
Creates a new OAuth2.0 resource server and defines custom scopes within it.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Creates a new Amazon Cognito user pool and sets the password policy for the pool.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
"
+ "documentation":"
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Creates a new Amazon Cognito user pool and sets the password policy for the pool.
If you don't provide a value for an attribute, Amazon Cognito sets it to its default value.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
When you create a new user pool client, token revocation is automatically activated. For more information about revoking tokens, see RevokeToken.
"
+ "documentation":"
Creates the user pool client.
When you create a new user pool client, token revocation is automatically activated. For more information about revoking tokens, see RevokeToken.
If you don't provide a value for an attribute, Amazon Cognito sets it to its default value.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Returns the configuration information and metadata of the specified user pool.
"
+ "documentation":"
Returns the configuration information and metadata of the specified user pool.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Client method for returning the configuration information and metadata of the specified user pool app client.
"
+ "documentation":"
Client method for returning the configuration information and metadata of the specified user pool app client.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Calling this API causes a message to be sent to the end user with a confirmation code that is required to change the user's password. For the Username parameter, you can use the username or user alias. The method used to send the confirmation code is sent according to the specified AccountRecoverySetting. For more information, see Recovering User Accounts in the Amazon Cognito Developer Guide. If neither a verified phone number nor a verified email exists, an InvalidParameterException is thrown. To use the confirmation code for resetting the password, call ConfirmForgotPassword.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
",
+ "documentation":"
Calling this API causes a message to be sent to the end user with a confirmation code that is required to change the user's password. For the Username parameter, you can use the username or user alias. The method used to send the confirmation code is sent according to the specified AccountRecoverySetting. For more information, see Recovering User Accounts in the Amazon Cognito Developer Guide. To use the confirmation code for resetting the password, call ConfirmForgotPassword.
If neither a verified phone number nor a verified email exists, this API returns InvalidParameterException. If your app client has a client secret and you don't provide a SECRET_HASH parameter, this API returns NotAuthorizedException.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Generates a user attribute verification code for the specified attribute name. Sends a message to a user with a code that they must return in a VerifyUserAttribute request.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
",
+ "documentation":"
Generates a user attribute verification code for the specified attribute name. Sends a message to a user with a code that they must return in a VerifyUserAttribute request.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Signs out users from all devices. It also invalidates all refresh tokens that Amazon Cognito has issued to a user. A user can still use a hosted UI cookie to retrieve new tokens for the duration of the 1-hour cookie validity period.
"
+ "documentation":"
Signs out a user from all devices. GlobalSignOut invalidates all identity, access and refresh tokens that Amazon Cognito has issued to a user. A user can still use a hosted UI cookie to retrieve new tokens for the duration of the 1-hour cookie validity period.
Your app isn't aware that a user's access token is revoked unless it attempts to authorize a user pools API request with an access token that contains the scope aws.cognito.signin.user.admin. Your app might otherwise accept access tokens until they expire.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Initiates sign-in for a user in the Amazon Cognito user directory. You can't sign in a user with a federated IdP with InitiateAuth. For more information, see Adding user pool sign-in through a third party.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
",
+ "documentation":"
Initiates sign-in for a user in the Amazon Cognito user directory. You can't sign in a user with a federated IdP with InitiateAuth. For more information, see Adding user pool sign-in through a third party.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Lists the sign-in devices that Amazon Cognito has registered to the current user.
"
+ "documentation":"
Lists the sign-in devices that Amazon Cognito has registered to the current user.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Calling this action requires developer credentials.
"
+ "documentation":"
Lists the groups associated with a user pool.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Lists the clients that have been created for the specified user pool.
"
+ "documentation":"
Lists the clients that have been created for the specified user pool.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Lists the user pools associated with an Amazon Web Services account.
"
+ "documentation":"
Lists the user pools associated with an Amazon Web Services account.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Lists users and their basic details in a user pool.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Calling this action requires developer credentials.
"
+ "documentation":"
Lists the users in the specified group.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Resends the confirmation (for confirmation of registration) to a specific user in the user pool.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
",
+ "documentation":"
Resends the confirmation (for confirmation of registration) to a specific user in the user pool.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
",
+ "documentation":"
Responds to the authentication challenge.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Revokes all of the access tokens generated by, and at the same time as, the specified refresh token. After a token is revoked, you can't use the revoked token to access Amazon Cognito user APIs, or to authorize access to your resource server.
"
+ "documentation":"
Revokes all of the access tokens generated by, and at the same time as, the specified refresh token. After a token is revoked, you can't use the revoked token to access Amazon Cognito user APIs, or to authorize access to your resource server.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Set the user's multi-factor authentication (MFA) method preference, including which MFA factors are activated and if any are preferred. Only one factor can be set as preferred. The preferred MFA factor will be used to authenticate a user if multiple factors are activated. If multiple options are activated and no preference is set, a challenge to choose an MFA option will be returned during sign-in. If an MFA type is activated for a user, the user will be prompted for MFA during all sign-in attempts unless device tracking is turned on and the device has been trusted. If you want MFA to be applied selectively based on the assessed risk level of sign-in attempts, deactivate MFA for users and turn on Adaptive Authentication for the user pool.
"
+ "documentation":"
Set the user's multi-factor authentication (MFA) method preference, including which MFA factors are activated and if any are preferred. Only one factor can be set as preferred. The preferred MFA factor will be used to authenticate a user if multiple factors are activated. If multiple options are activated and no preference is set, a challenge to choose an MFA option will be returned during sign-in. If an MFA type is activated for a user, the user will be prompted for MFA during all sign-in attempts unless device tracking is turned on and the device has been trusted. If you want MFA to be applied selectively based on the assessed risk level of sign-in attempts, deactivate MFA for users and turn on Adaptive Authentication for the user pool.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Sets the user pool multi-factor authentication (MFA) configuration.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
"
+ "documentation":"
Sets the user pool multi-factor authentication (MFA) configuration.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
This action is no longer supported. You can use it to configure only SMS MFA. You can't use it to configure time-based one-time password (TOTP) software token MFA. To configure either type of MFA, use SetUserMFAPreference instead.
",
+ "documentation":"
This action is no longer supported. You can use it to configure only SMS MFA. You can't use it to configure time-based one-time password (TOTP) software token MFA. To configure either type of MFA, use SetUserMFAPreference instead.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Registers the user in the specified user pool and creates a user name, password, and user attributes.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
",
+ "documentation":"
Registers the user in the specified user pool and creates a user name, password, and user attributes.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Provides the feedback for an authentication event, whether it was from a valid user or not. This feedback is used for improving the risk evaluation decision for the user pool as part of Amazon Cognito advanced security.
"
+ "documentation":"
Provides the feedback for an authentication event, whether it was from a valid user or not. This feedback is used for improving the risk evaluation decision for the user pool as part of Amazon Cognito advanced security.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Updates the specified group with the specified attributes.
Calling this action requires developer credentials.
"
+ "documentation":"
Updates the specified group with the specified attributes.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Updates the name and scopes of resource server. All other fields are read-only.
If you don't provide a value for an attribute, it is set to the default value.
"
+ "documentation":"
Updates the name and scopes of resource server. All other fields are read-only.
If you don't provide a value for an attribute, it is set to the default value.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Allows a user to update a specific attribute (one at a time).
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
",
+ "documentation":"
Allows a user to update a specific attribute (one at a time).
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Updates the specified user pool with the specified attributes. You can get a list of the current user pool settings using DescribeUserPool. If you don't provide a value for an attribute, it will be set to the default value.
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
"
+ "documentation":"
This action might generate an SMS text message. Starting June 1, 2021, US telecom carriers require you to register an origination phone number before you can send SMS messages to US phone numbers. If you use SMS text messages in Amazon Cognito, you must register a phone number with Amazon Pinpoint. Amazon Cognito uses the registered number automatically. Otherwise, Amazon Cognito users who must receive SMS messages might not be able to sign up, activate their accounts, or sign in.
If you have never used SMS text messages with Amazon Cognito or any other Amazon Web Service, Amazon Simple Notification Service might place your account in the SMS sandbox. In sandbox mode, you can send messages only to verified phone numbers. After you test your app while in the sandbox environment, you can move out of the sandbox and into production. For more information, see SMS message settings for Amazon Cognito user pools in the Amazon Cognito Developer Guide.
Updates the specified user pool with the specified attributes. You can get a list of the current user pool settings using DescribeUserPool.
If you don't provide a value for an attribute, Amazon Cognito sets it to its default value.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Updates the specified user pool app client with the specified attributes. You can get a list of the current user pool app client settings using DescribeUserPoolClient.
If you don't provide a value for an attribute, it will be set to the default value.
You can also use this operation to enable token revocation for user pool clients. For more information about revoking tokens, see RevokeToken.
"
+ "documentation":"
Updates the specified user pool app client with the specified attributes. You can get a list of the current user pool app client settings using DescribeUserPoolClient.
If you don't provide a value for an attribute, Amazon Cognito sets it to its default value.
You can also use this operation to enable token revocation for user pool clients. For more information about revoking tokens, see RevokeToken.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Updates the Secure Sockets Layer (SSL) certificate for the custom domain for your user pool.
You can use this operation to provide the Amazon Resource Name (ARN) of a new certificate to Amazon Cognito. You can't use it to change the domain for a user pool.
A custom domain is used to host the Amazon Cognito hosted UI, which provides sign-up and sign-in pages for your application. When you set up a custom domain, you provide a certificate that you manage with Certificate Manager (ACM). When necessary, you can use this operation to change the certificate that you applied to your custom domain.
Usually, this is unnecessary following routine certificate renewal with ACM. When you renew your existing certificate in ACM, the ARN for your certificate remains the same, and your custom domain uses the new certificate automatically.
However, if you replace your existing certificate with a new one, ACM gives the new certificate a new ARN. To apply the new certificate to your custom domain, you must provide this ARN to Amazon Cognito.
When you add your new certificate in ACM, you must choose US East (N. Virginia) as the Amazon Web Services Region.
After you submit your request, Amazon Cognito requires up to 1 hour to distribute your new certificate to your custom domain.
Updates the Secure Sockets Layer (SSL) certificate for the custom domain for your user pool.
You can use this operation to provide the Amazon Resource Name (ARN) of a new certificate to Amazon Cognito. You can't use it to change the domain for a user pool.
A custom domain is used to host the Amazon Cognito hosted UI, which provides sign-up and sign-in pages for your application. When you set up a custom domain, you provide a certificate that you manage with Certificate Manager (ACM). When necessary, you can use this operation to change the certificate that you applied to your custom domain.
Usually, this is unnecessary following routine certificate renewal with ACM. When you renew your existing certificate in ACM, the ARN for your certificate remains the same, and your custom domain uses the new certificate automatically.
However, if you replace your existing certificate with a new one, ACM gives the new certificate a new ARN. To apply the new certificate to your custom domain, you must provide this ARN to Amazon Cognito.
When you add your new certificate in ACM, you must choose US East (N. Virginia) as the Amazon Web Services Region.
After you submit your request, Amazon Cognito requires up to 1 hour to distribute your new certificate to your custom domain.
Amazon Cognito evaluates Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you must use IAM credentials to authorize requests, and you must grant yourself the corresponding IAM permission in a policy.
Use this API to register a user's entered time-based one-time password (TOTP) code and mark the user's software token MFA status as \"verified\" if successful. The request takes an access token or a session string, but not both.
"
+ "documentation":"
Use this API to register a user's entered time-based one-time password (TOTP) code and mark the user's software token MFA status as \"verified\" if successful. The request takes an access token or a session string, but not both.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
Verifies the specified user attributes in the user pool.
If your user pool requires verification before Amazon Cognito updates the attribute value, VerifyUserAttribute updates the affected attribute to its pending value. For more information, see UserAttributeUpdateSettingsType.
",
+ "documentation":"
Verifies the specified user attributes in the user pool.
If your user pool requires verification before Amazon Cognito updates the attribute value, VerifyUserAttribute updates the affected attribute to its pending value. For more information, see UserAttributeUpdateSettingsType.
Amazon Cognito doesn't evaluate Identity and Access Management (IAM) policies in requests for this API operation. For this operation, you can't use IAM credentials to authorize requests, and you can't grant IAM permissions in policies. For more information about authorization models in Amazon Cognito, see Using the Amazon Cognito native and OIDC APIs.
A map of custom key-value pairs that you can provide as input for any custom workflows that this action triggers.
If your user pool configuration includes triggers, the AdminConfirmSignUp API action invokes the Lambda function that is specified for the post confirmation trigger. When Amazon Cognito invokes this function, it passes a JSON payload, which the function receives as input. In this payload, the clientMetadata attribute provides the data that you assigned to the ClientMetadata parameter in your AdminConfirmSignUp request. In your function code in Lambda, you can process the ClientMetadata value to enhance your workflow for your specific needs.
When you use the ClientMetadata parameter, remember that Amazon Cognito won't do the following:
Store the ClientMetadata value. This data is available only to Lambda triggers that are assigned to a user pool to support custom workflows. If your user pool configuration doesn't include triggers, the ClientMetadata parameter serves no purpose.
Validate the ClientMetadata value.
Encrypt the ClientMetadata value. Don't use Amazon Cognito to provide sensitive information.
"
}
},
- "documentation":"
Represents the request to confirm user registration.
"
+ "documentation":"
Confirm a user's registration as a user pool administrator.
The user's temporary password. This password must conform to the password policy that you specified when you created the user pool.
The temporary password is valid only once. To complete the Admin Create User flow, the user must enter the temporary password in the sign-in page, along with a new password to be used in all future sign-ins.
This parameter isn't required. If you don't specify a value, Amazon Cognito generates one for you.
The temporary password can only be used until the user account expiration limit that you specified when you created the user pool. To reset the account after that time limit, you must call AdminCreateUser again, specifying \"RESEND\" for the MessageAction parameter.
"
+ "documentation":"
The user's temporary password. This password must conform to the password policy that you specified when you created the user pool.
The temporary password is valid only once. To complete the Admin Create User flow, the user must enter the temporary password in the sign-in page, along with a new password to be used in all future sign-ins.
This parameter isn't required. If you don't specify a value, Amazon Cognito generates one for you.
The temporary password can only be used until the user account expiration limit that you set for your user pool. To reset the account after that time limit, you must call AdminCreateUser again and specify RESEND for the MessageAction parameter.
UNCONFIRMED - User has been created but not confirmed.
CONFIRMED - User has been confirmed.
ARCHIVED - User is no longer active.
UNKNOWN - User status isn't known.
RESET_REQUIRED - User is confirmed, but the user must request a code and reset their password before they can sign in.
FORCE_CHANGE_PASSWORD - The user is confirmed and the user can sign in using a temporary password, but on first sign-in, the user must change their password to a new value before doing anything else.
"
+ "documentation":"
The user status. Can be one of the following:
UNCONFIRMED - User has been created but not confirmed.
CONFIRMED - User has been confirmed.
UNKNOWN - User status isn't known.
RESET_REQUIRED - User is confirmed, but the user must request a code and reset their password before they can sign in.
FORCE_CHANGE_PASSWORD - The user is confirmed and the user can sign in using a temporary password, but on first sign-in, the user must change their password to a new value before doing anything else.
The authentication parameters. These are inputs corresponding to the AuthFlow that you're invoking. The required values depend on the value of AuthFlow:
For USER_SRP_AUTH: USERNAME (required), SRP_A (required), SECRET_HASH (required if the app client is configured with a client secret), DEVICE_KEY.
For REFRESH_TOKEN_AUTH/REFRESH_TOKEN: REFRESH_TOKEN (required), SECRET_HASH (required if the app client is configured with a client secret), DEVICE_KEY.
For ADMIN_NO_SRP_AUTH: USERNAME (required), SECRET_HASH (if app client is configured with client secret), PASSWORD (required), DEVICE_KEY.
For CUSTOM_AUTH: USERNAME (required), SECRET_HASH (if app client is configured with client secret), DEVICE_KEY. To start the authentication flow with password verification, include ChallengeName: SRP_A and SRP_A: (The SRP_A Value).
"
+ "documentation":"
The authentication parameters. These are inputs corresponding to the AuthFlow that you're invoking. The required values depend on the value of AuthFlow:
For USER_SRP_AUTH: USERNAME (required), SRP_A (required), SECRET_HASH (required if the app client is configured with a client secret), DEVICE_KEY.
For ADMIN_USER_PASSWORD_AUTH: USERNAME (required), PASSWORD (required), SECRET_HASH (required if the app client is configured with a client secret), DEVICE_KEY.
For REFRESH_TOKEN_AUTH/REFRESH_TOKEN: REFRESH_TOKEN (required), SECRET_HASH (required if the app client is configured with a client secret), DEVICE_KEY.
For CUSTOM_AUTH: USERNAME (required), SECRET_HASH (if app client is configured with client secret), DEVICE_KEY. To start the authentication flow with password verification, include ChallengeName: SRP_A and SRP_A: (The SRP_A Value).
The existing user in the user pool that you want to assign to the external IdP user account. This user can be a native (Username + Password) Amazon Cognito user pools user or a federated user (for example, a SAML or Facebook user). If the user doesn't exist, Amazon Cognito generates an exception. Amazon Cognito returns this user when the new user (with the linked IdP attribute) signs in.
For a native username + password user, the ProviderAttributeValue for the DestinationUser should be the username in the user pool. For a federated user, it should be the provider-specific user_id.
The ProviderAttributeName of the DestinationUser is ignored.
The ProviderName should be set to Cognito for users in Cognito user pools.
All attributes in the DestinationUser profile must be mutable. If you have assigned the user any immutable custom attributes, the operation won't succeed.
"
+ "documentation":"
The existing user in the user pool that you want to assign to the external IdP user account. This user can be a local (Username + Password) Amazon Cognito user pools user or a federated user (for example, a SAML or Facebook user). If the user doesn't exist, Amazon Cognito generates an exception. Amazon Cognito returns this user when the new user (with the linked IdP attribute) signs in.
For a native username + password user, the ProviderAttributeValue for the DestinationUser should be the username in the user pool. For a federated user, it should be the provider-specific user_id.
The ProviderAttributeName of the DestinationUser is ignored.
The ProviderName should be set to Cognito for users in Cognito user pools.
All attributes in the DestinationUser profile must be mutable. If you have assigned the user any immutable custom attributes, the operation won't succeed.
An external IdP account for a user who doesn't exist yet in the user pool. This user must be a federated user (for example, a SAML or Facebook user), not another native user.
If the SourceUser is using a federated social IdP, such as Facebook, Google, or Login with Amazon, you must set the ProviderAttributeName to Cognito_Subject. For social IdPs, the ProviderName will be Facebook, Google, or LoginWithAmazon, and Amazon Cognito will automatically parse the Facebook, Google, and Login with Amazon tokens for id, sub, and user_id, respectively. The ProviderAttributeValue for the user must be the same value as the id, sub, or user_id value found in the social IdP token.
For SAML, the ProviderAttributeName can be any value that matches a claim in the SAML assertion. If you want to link SAML users based on the subject of the SAML assertion, you should map the subject to a claim through the SAML IdP and submit that claim name as the ProviderAttributeName. If you set ProviderAttributeName to Cognito_Subject, Amazon Cognito will automatically parse the default unique identifier found in the subject from the SAML token.
"
+ "documentation":"
An external IdP account for a user who doesn't exist yet in the user pool. This user must be a federated user (for example, a SAML or Facebook user), not another native user.
If the SourceUser is using a federated social IdP, such as Facebook, Google, or Login with Amazon, you must set the ProviderAttributeName to Cognito_Subject. For social IdPs, the ProviderName will be Facebook, Google, or LoginWithAmazon, and Amazon Cognito will automatically parse the Facebook, Google, and Login with Amazon tokens for id, sub, and user_id, respectively. The ProviderAttributeValue for the user must be the same value as the id, sub, or user_id value found in the social IdP token.
For OIDC, the ProviderAttributeName can be any value that matches a claim in the ID token, or that your app retrieves from the userInfo endpoint. You must map the claim to a user pool attribute in your IdP configuration, and set the user pool attribute name as the value of ProviderAttributeName in your AdminLinkProviderForUser request.
For SAML, the ProviderAttributeName can be any value that matches a claim in the SAML assertion. To link SAML users based on the subject of the SAML assertion, map the subject to a claim through the SAML IdP and set that claim name as the value of ProviderAttributeName in your AdminLinkProviderForUser request.
For both OIDC and SAML users, when you set ProviderAttributeName to Cognito_Subject, Amazon Cognito will automatically parse the default unique identifier found in the subject from the IdP token.
The challenge responses. These are inputs corresponding to the value of ChallengeName, for example:
SMS_MFA: SMS_MFA_CODE, USERNAME, SECRET_HASH (if app client is configured with client secret).
PASSWORD_VERIFIER: PASSWORD_CLAIM_SIGNATURE, PASSWORD_CLAIM_SECRET_BLOCK, TIMESTAMP, USERNAME, SECRET_HASH (if app client is configured with client secret).
PASSWORD_VERIFIER requires DEVICE_KEY when signing in with a remembered device.
ADMIN_NO_SRP_AUTH: PASSWORD, USERNAME, SECRET_HASH (if app client is configured with client secret).
NEW_PASSWORD_REQUIRED: NEW_PASSWORD, USERNAME, SECRET_HASH (if app client is configured with client secret). To set any required attributes that Amazon Cognito returned as requiredAttributes in the AdminInitiateAuth response, add a userAttributes.attributename parameter. This parameter can also set values for writable attributes that aren't required by your user pool.
In a NEW_PASSWORD_REQUIRED challenge response, you can't modify a required attribute that already has a value. In AdminRespondToAuthChallenge, set a value for any keys that Amazon Cognito returned in the requiredAttributes parameter, then use the AdminUpdateUserAttributes API operation to modify the value of any additional attributes.
MFA_SETUP requires USERNAME, plus you must use the session value returned by VerifySoftwareToken in the Session parameter.
The value of the USERNAME attribute must be the user's actual username, not an alias (such as an email address or phone number). To make this simpler, the AdminInitiateAuth response includes the actual username value in the USERNAMEUSER_ID_FOR_SRP attribute. This happens even if you specified an alias in your call to AdminInitiateAuth.
"
+ "documentation":"
The challenge responses. These are inputs corresponding to the value of ChallengeName, for example:
SMS_MFA: SMS_MFA_CODE, USERNAME, SECRET_HASH (if app client is configured with client secret).
PASSWORD_VERIFIER: PASSWORD_CLAIM_SIGNATURE, PASSWORD_CLAIM_SECRET_BLOCK, TIMESTAMP, USERNAME, SECRET_HASH (if app client is configured with client secret).
PASSWORD_VERIFIER requires DEVICE_KEY when signing in with a remembered device.
ADMIN_NO_SRP_AUTH: PASSWORD, USERNAME, SECRET_HASH (if app client is configured with client secret).
NEW_PASSWORD_REQUIRED: NEW_PASSWORD, USERNAME, SECRET_HASH (if app client is configured with client secret). To set any required attributes that Amazon Cognito returned as requiredAttributes in the AdminInitiateAuth response, add a userAttributes.attributename parameter. This parameter can also set values for writable attributes that aren't required by your user pool.
In a NEW_PASSWORD_REQUIRED challenge response, you can't modify a required attribute that already has a value. In AdminRespondToAuthChallenge, set a value for any keys that Amazon Cognito returned in the requiredAttributes parameter, then use the AdminUpdateUserAttributes API operation to modify the value of any additional attributes.
MFA_SETUP requires USERNAME, plus you must use the session value returned by VerifySoftwareToken in the Session parameter.
The value of the USERNAME attribute must be the user's actual username, not an alias (such as an email address or phone number). To make this simpler, the AdminInitiateAuth response includes the actual username value in the USERNAMEUSER_ID_FOR_SRP attribute. This happens even if you specified an alias in your call to AdminInitiateAuth.
The authentication event feedback value. When you provide a FeedbackValue value of valid, you tell Amazon Cognito that you trust a user session where Amazon Cognito has evaluated some level of risk. When you provide a FeedbackValue value of invalid, you tell Amazon Cognito that you don't trust a user session, or you don't believe that Amazon Cognito evaluated a high-enough risk level.
"
}
}
},
@@ -3087,7 +3125,7 @@
"documentation":"
If UserDataShared is true, Amazon Cognito includes user data in the events that it publishes to Amazon Pinpoint analytics.
"
}
},
- "documentation":"
The Amazon Pinpoint analytics configuration necessary to collect metrics for a user pool.
In Regions where Amazon Pinpointisn't available, user pools only support sending events to Amazon Pinpoint projects in us-east-1. In Regions where Amazon Pinpoint is available, user pools support sending events to Amazon Pinpoint projects within that same Region.
"
+ "documentation":"
The Amazon Pinpoint analytics configuration necessary to collect metrics for a user pool.
In Regions where Amazon Pinpoint isn't available, user pools only support sending events to Amazon Pinpoint projects in us-east-1. In Regions where Amazon Pinpoint is available, user pools support sending events to Amazon Pinpoint projects within that same Region.
The Amazon Resource Name (arn) of a CloudWatch Logs log group where your user pool sends logs. The log group must not be encrypted with Key Management Service and must be in the same Amazon Web Services account as your user pool.
"
+ }
+ },
+ "documentation":"
The CloudWatch logging destination of a user pool detailed activity logging configuration.
A keyed-hash message authentication code (HMAC) calculated using the secret key of a user pool client and username plus the client ID in the message.
"
+ "documentation":"
A keyed-hash message authentication code (HMAC) calculated using the secret key of a user pool client and username plus the client ID in the message. For more information about SecretHash, see Computing secret hash values.
The ID token time limit. After this limit expires, your user can't use their ID token. To specify the time unit for IdTokenValidity as seconds, minutes, hours, or days, set a TokenValidityUnits value in your API request.
For example, when you set IdTokenValidity as 10 and TokenValidityUnits as hours, your user can authenticate their session with their ID token for 10 hours.
The default time unit for AccessTokenValidity in an API request is hours. Valid range is displayed below in seconds.
If you don't specify otherwise in the configuration of your app client, your ID tokens are valid for one hour.
"
+ "documentation":"
The ID token time limit. After this limit expires, your user can't use their ID token. To specify the time unit for IdTokenValidity as seconds, minutes, hours, or days, set a TokenValidityUnits value in your API request.
For example, when you set IdTokenValidity as 10 and TokenValidityUnits as hours, your user can authenticate their session with their ID token for 10 hours.
The default time unit for IdTokenValidity in an API request is hours. Valid range is displayed below in seconds.
If you don't specify otherwise in the configuration of your app client, your ID tokens are valid for one hour.
Set to true if the client is allowed to follow the OAuth protocol when interacting with Amazon Cognito user pools.
"
+ "documentation":"
Set to true to use OAuth 2.0 features in your user pool app client.
AllowedOAuthFlowsUserPoolClient must be true before you can configure the following features in your app client.
CallBackURLs: Callback URLs.
LogoutURLs: Sign-out redirect URLs.
AllowedOAuthScopes: OAuth 2.0 scopes.
AllowedOAuthFlows: Support for authorization code, implicit, and client credentials OAuth 2.0 grants.
To use OAuth 2.0 features, configure one of these features in the Amazon Cognito console or set AllowedOAuthFlowsUserPoolClient to true in a CreateUserPoolClient or UpdateUserPoolClient API request. If you don't set a value for AllowedOAuthFlowsUserPoolClient in a request with the CLI or SDKs, it defaults to false.
Enables advanced security risk detection. Set the key AdvancedSecurityMode to the value \"AUDIT\".
"
+ "documentation":"
User pool add-ons. Contains settings for activation of advanced security features. To log user security information but take no action, set to AUDIT. To configure automatic security responses to risky traffic to your user pool, set to ENFORCED.
Case sensitivity on the username input for the selected sign-in option. For example, when case sensitivity is set to False, users can sign in using either \"username\" or \"Username\". This configuration is immutable once it has been set. For more information, see UsernameConfigurationType.
"
+ "documentation":"
Case sensitivity on the username input for the selected sign-in option. When case sensitivity is set to False (case insensitive), users can sign in with any combination of capital and lowercase letters. For example, username, USERNAME, or UserName, or for email, email@example.com or EMaiL@eXamplE.Com. For most use cases, set case sensitivity to False (case insensitive) as a best practice. When usernames and email addresses are case insensitive, Amazon Cognito treats any variation in case as the same user, and prevents a case variation from being assigned to the same attribute for a different user.
This configuration is immutable after you set it. For more information, see UsernameConfigurationType.
The ARN of a verified email address in Amazon SES. Amazon Cognito uses this email address in one of the following ways, depending on the value that you specify for the EmailSendingAccount parameter:
If you specify COGNITO_DEFAULT, Amazon Cognito uses this address as the custom FROM address when it emails your users using its built-in email account.
If you specify DEVELOPER, Amazon Cognito emails your users with this address by calling Amazon SES on your behalf.
The Region value of the SourceArn parameter must indicate a supported Amazon Web Services Region of your user pool. Typically, the Region in the SourceArn and the user pool Region are the same. For more information, see Amazon SES email configuration regions in the Amazon Cognito Developer Guide.
"
+ "documentation":"
The ARN of a verified email address or an address from a verified domain in Amazon SES. You can set a SourceArn email from a verified domain only with an API request. You can set a verified email address, but not an address in a verified domain, in the Amazon Cognito console. Amazon Cognito uses the email address that you provide in one of the following ways, depending on the value that you specify for the EmailSendingAccount parameter:
If you specify COGNITO_DEFAULT, Amazon Cognito uses this address as the custom FROM address when it emails your users using its built-in email account.
If you specify DEVELOPER, Amazon Cognito emails your users with this address by calling Amazon SES on your behalf.
The Region value of the SourceArn parameter must indicate a supported Amazon Web Services Region of your user pool. Typically, the Region in the SourceArn and the user pool Region are the same. For more information, see Amazon SES email configuration regions in the Amazon Cognito Developer Guide.
The authentication event feedback value. When you provide a FeedbackValue value of valid, you tell Amazon Cognito that you trust a user session where Amazon Cognito has evaluated some level of risk. When you provide a FeedbackValue value of invalid, you tell Amazon Cognito that you don't trust a user session, or you don't believe that Amazon Cognito evaluated a high-enough risk level.
The authentication parameters. These are inputs corresponding to the AuthFlow that you're invoking. The required values depend on the value of AuthFlow:
For USER_SRP_AUTH: USERNAME (required), SRP_A (required), SECRET_HASH (required if the app client is configured with a client secret), DEVICE_KEY.
For REFRESH_TOKEN_AUTH/REFRESH_TOKEN: REFRESH_TOKEN (required), SECRET_HASH (required if the app client is configured with a client secret), DEVICE_KEY.
For CUSTOM_AUTH: USERNAME (required), SECRET_HASH (if app client is configured with client secret), DEVICE_KEY. To start the authentication flow with password verification, include ChallengeName: SRP_A and SRP_A: (The SRP_A Value).
"
+ "documentation":"
The authentication parameters. These are inputs corresponding to the AuthFlow that you're invoking. The required values depend on the value of AuthFlow:
For USER_SRP_AUTH: USERNAME (required), SRP_A (required), SECRET_HASH (required if the app client is configured with a client secret), DEVICE_KEY.
For USER_PASSWORD_AUTH: USERNAME (required), PASSWORD (required), SECRET_HASH (required if the app client is configured with a client secret), DEVICE_KEY.
For REFRESH_TOKEN_AUTH/REFRESH_TOKEN: REFRESH_TOKEN (required), SECRET_HASH (required if the app client is configured with a client secret), DEVICE_KEY.
For CUSTOM_AUTH: USERNAME (required), SECRET_HASH (if app client is configured with client secret), DEVICE_KEY. To start the authentication flow with password verification, include ChallengeName: SRP_A and SRP_A: (The SRP_A Value).
An array of strings, where each string is the name of a user attribute to be returned for each user in the search results. If the array is null, all attributes are returned.
"
+ "documentation":"
A JSON array of user attribute names, for example given_name, that you want Amazon Cognito to include in the response for each user. When you don't provide an AttributesToGet parameter, Amazon Cognito returns all attributes for each user.
A list of the user pool users, and their attributes, that match your query.
Amazon Cognito creates a profile in your user pool for each native user in your user pool, and each unique user ID from your third-party identity providers (IdPs). When you link users with the AdminLinkProviderForUser API operation, the output of ListUsers displays both the IdP user and the native user that you linked. You can identify IdP users in the Users object of this API response by the IdP prefix that Amazon Cognito appends to Username.
PASSWORD_VERIFIER requires DEVICE_KEY when you sign in with a remembered device.
NEW_PASSWORD_REQUIRED: NEW_PASSWORD, USERNAME, SECRET_HASH (if app client is configured with client secret). To set any required attributes that Amazon Cognito returned as requiredAttributes in the InitiateAuth response, add a userAttributes.attributename parameter. This parameter can also set values for writable attributes that aren't required by your user pool.
In a NEW_PASSWORD_REQUIRED challenge response, you can't modify a required attribute that already has a value. In RespondToAuthChallenge, set a value for any keys that Amazon Cognito returned in the requiredAttributes parameter, then use the UpdateUserAttributes API operation to modify the value of any additional attributes.
SOFTWARE_TOKEN_MFA: USERNAME and SOFTWARE_TOKEN_MFA_CODE are required attributes.
PASSWORD_VERIFIER requires DEVICE_KEY when you sign in with a remembered device.
NEW_PASSWORD_REQUIRED: NEW_PASSWORD, USERNAME, SECRET_HASH (if app client is configured with client secret). To set any required attributes that Amazon Cognito returned as requiredAttributes in the InitiateAuth response, add a userAttributes.attributename parameter. This parameter can also set values for writable attributes that aren't required by your user pool.
In a NEW_PASSWORD_REQUIRED challenge response, you can't modify a required attribute that already has a value. In RespondToAuthChallenge, set a value for any keys that Amazon Cognito returned in the requiredAttributes parameter, then use the UpdateUserAttributes API operation to modify the value of any additional attributes.
SOFTWARE_TOKEN_MFA: USERNAME and SOFTWARE_TOKEN_MFA_CODE are required attributes.
Specifies whether the value of the attribute can be changed.
For any user pool attribute that is mapped to an IdP attribute, you must set this parameter to true. Amazon Cognito updates mapped attributes when users sign in to your application through an IdP. If an attribute is immutable, Amazon Cognito throws an error when it attempts to update the attribute. For more information, see Specifying Identity Provider Attribute Mappings for Your User Pool.
",
+ "documentation":"
Specifies whether the value of the attribute can be changed.
Any user pool attribute whose value you map from an IdP attribute must be mutable, with a parameter value of true. Amazon Cognito updates mapped attributes when users sign in to your application through an IdP. If an attribute is immutable, Amazon Cognito throws an error when it attempts to update the attribute. For more information, see Specifying Identity Provider Attribute Mappings for Your User Pool.
Specifies the constraints for an attribute of the string type.
"
}
},
- "documentation":"
Contains information about the schema attribute.
"
+ "documentation":"
A list of the user attributes and their properties in your user pool. The attribute schema contains standard attributes, custom attributes with a custom: prefix, and developer attributes with a dev: prefix. For more information, see User pool attributes.
Developer-only attributes are a legacy feature of user pools, are read-only to all app clients. You can create and update developer-only attributes only with IAM-authenticated API operations. Use app client read/write permissions instead.
A time unit of seconds, minutes, hours, or days for the value that you set in the AccessTokenValidity parameter. The default AccessTokenValidity time unit is hours.
"
+ "documentation":"
A time unit of seconds, minutes, hours, or days for the value that you set in the AccessTokenValidity parameter. The default AccessTokenValidity time unit is hours. AccessTokenValidity duration can range from five minutes to one day.
A time unit of seconds, minutes, hours, or days for the value that you set in the IdTokenValidity parameter. The default IdTokenValidity time unit is hours.
"
+ "documentation":"
A time unit of seconds, minutes, hours, or days for the value that you set in the IdTokenValidity parameter. The default IdTokenValidity time unit is hours. IdTokenValidity duration can range from five minutes to one day.
A time unit of seconds, minutes, hours, or days for the value that you set in the RefreshTokenValidity parameter. The default RefreshTokenValidity time unit is days.
"
+ "documentation":"
A time unit of seconds, minutes, hours, or days for the value that you set in the RefreshTokenValidity parameter. The default RefreshTokenValidity time unit is days. RefreshTokenValidity duration can range from 60 minutes to 10 years.
"
}
},
"documentation":"
The data type TokenValidityUnits specifies the time units you use when you set the duration of ID, access, and refresh tokens.
The authentication event feedback value. When you provide a FeedbackValue value of valid, you tell Amazon Cognito that you trust a user session where Amazon Cognito has evaluated some level of risk. When you provide a FeedbackValue value of invalid, you tell Amazon Cognito that you don't trust a user session, or you don't believe that Amazon Cognito evaluated a high-enough risk level.
The ID token time limit. After this limit expires, your user can't use their ID token. To specify the time unit for IdTokenValidity as seconds, minutes, hours, or days, set a TokenValidityUnits value in your API request.
For example, when you set IdTokenValidity as 10 and TokenValidityUnits as hours, your user can authenticate their session with their ID token for 10 hours.
The default time unit for AccessTokenValidity in an API request is hours. Valid range is displayed below in seconds.
If you don't specify otherwise in the configuration of your app client, your ID tokens are valid for one hour.
"
+ "documentation":"
The ID token time limit. After this limit expires, your user can't use their ID token. To specify the time unit for IdTokenValidity as seconds, minutes, hours, or days, set a TokenValidityUnits value in your API request.
For example, when you set IdTokenValidity as 10 and TokenValidityUnits as hours, your user can authenticate their session with their ID token for 10 hours.
The default time unit for IdTokenValidity in an API request is hours. Valid range is displayed below in seconds.
If you don't specify otherwise in the configuration of your app client, your ID tokens are valid for one hour.
The units in which the validity times are represented. The default unit for RefreshToken is days, and the default for ID and access tokens is hours.
"
+ "documentation":"
The time units you use when you set the duration of ID, access, and refresh tokens. The default unit for RefreshToken is days, and the default for ID and access tokens is hours.
Set to true if the client is allowed to follow the OAuth protocol when interacting with Amazon Cognito user pools.
"
+ "documentation":"
Set to true to use OAuth 2.0 features in your user pool app client.
AllowedOAuthFlowsUserPoolClient must be true before you can configure the following features in your app client.
CallBackURLs: Callback URLs.
LogoutURLs: Sign-out redirect URLs.
AllowedOAuthScopes: OAuth 2.0 scopes.
AllowedOAuthFlows: Support for authorization code, implicit, and client credentials OAuth 2.0 grants.
To use OAuth 2.0 features, configure one of these features in the Amazon Cognito console or set AllowedOAuthFlowsUserPoolClient to true in a CreateUserPoolClient or UpdateUserPoolClient API request. If you don't set a value for AllowedOAuthFlowsUserPoolClient in a request with the CLI or SDKs, it defaults to false.
Enables advanced security risk detection. Set the key AdvancedSecurityMode to the value \"AUDIT\".
"
+ "documentation":"
User pool add-ons. Contains settings for activation of advanced security features. To log user security information but take no action, set to AUDIT. To configure automatic security responses to risky traffic to your user pool, set to ENFORCED.
Requires that your user verifies their email address, phone number, or both before Amazon Cognito updates the value of that attribute. When you update a user attribute that has this option activated, Amazon Cognito sends a verification message to the new phone number or email address. Amazon Cognito doesn’t change the value of the attribute until your user responds to the verification message and confirms the new value.
When AttributesRequireVerificationBeforeUpdate is false, your user pool doesn't require that your users verify attribute changes before Amazon Cognito updates them. In a user pool where AttributesRequireVerificationBeforeUpdate is false, API operations that change attribute values can immediately update a user’s email or phone_number attribute.
"
+ "documentation":"
Requires that your user verifies their email address, phone number, or both before Amazon Cognito updates the value of that attribute. When you update a user attribute that has this option activated, Amazon Cognito sends a verification message to the new phone number or email address. Amazon Cognito doesn’t change the value of the attribute until your user responds to the verification message and confirms the new value.
You can verify an updated email address or phone number with a VerifyUserAttribute API request. You can also call the AdminUpdateUserAttributes API and set email_verified or phone_number_verified to true.
When AttributesRequireVerificationBeforeUpdate is false, your user pool doesn't require that your users verify attribute changes before Amazon Cognito updates them. In a user pool where AttributesRequireVerificationBeforeUpdate is false, API operations that change attribute values can immediately update a user’s email or phone_number attribute.
"
}
},
"documentation":"
The settings for updates to user attributes. These settings include the property AttributesRequireVerificationBeforeUpdate, a user-pool setting that tells Amazon Cognito how to handle changes to the value of your users' email address and phone number attributes. For more information, see Verifying updates to email addresses and phone numbers.
Contextual data, such as the user's device fingerprint, IP address, or location, used for evaluating the risk of an unexpected event by Amazon Cognito advanced security.
",
- "sensitive":true
+ "documentation":"
Contextual data, such as the user's device fingerprint, IP address, or location, used for evaluating the risk of an unexpected event by Amazon Cognito advanced security.
The operating mode of advanced security features in your user pool.
"
}
},
- "documentation":"
The user pool add-ons type.
"
+ "documentation":"
User pool add-ons. Contains settings for activation of advanced security features. To log user security information but take no action, set to AUDIT. To configure automatic security responses to risky traffic to your user pool, set to ENFORCED.
The ID token time limit. After this limit expires, your user can't use their ID token. To specify the time unit for IdTokenValidity as seconds, minutes, hours, or days, set a TokenValidityUnits value in your API request.
For example, when you set IdTokenValidity as 10 and TokenValidityUnits as hours, your user can authenticate their session with their ID token for 10 hours.
The default time unit for AccessTokenValidity in an API request is hours. Valid range is displayed below in seconds.
If you don't specify otherwise in the configuration of your app client, your ID tokens are valid for one hour.
"
+ "documentation":"
The ID token time limit. After this limit expires, your user can't use their ID token. To specify the time unit for IdTokenValidity as seconds, minutes, hours, or days, set a TokenValidityUnits value in your API request.
For example, when you set IdTokenValidity as 10 and TokenValidityUnits as hours, your user can authenticate their session with their ID token for 10 hours.
The default time unit for IdTokenValidity in an API request is hours. Valid range is displayed below in seconds.
If you don't specify otherwise in the configuration of your app client, your ID tokens are valid for one hour.
Set to true if the client is allowed to follow the OAuth protocol when interacting with Amazon Cognito user pools.
",
+ "documentation":"
Set to true to use OAuth 2.0 features in your user pool app client.
AllowedOAuthFlowsUserPoolClient must be true before you can configure the following features in your app client.
CallBackURLs: Callback URLs.
LogoutURLs: Sign-out redirect URLs.
AllowedOAuthScopes: OAuth 2.0 scopes.
AllowedOAuthFlows: Support for authorization code, implicit, and client credentials OAuth 2.0 grants.
To use OAuth 2.0 features, configure one of these features in the Amazon Cognito console or set AllowedOAuthFlowsUserPoolClient to true in a CreateUserPoolClient or UpdateUserPoolClient API request. If you don't set a value for AllowedOAuthFlowsUserPoolClient in a request with the CLI or SDKs, it defaults to false.
A container with the schema attributes of a user pool.
"
+ "documentation":"
A list of the user attributes and their properties in your user pool. The attribute schema contains standard attributes, custom attributes with a custom: prefix, and developer attributes with a dev: prefix. For more information, see User pool attributes.
Developer-only attributes are a legacy feature of user pools, are read-only to all app clients. You can create and update developer-only attributes only with IAM-authenticated API operations. Use app client read/write permissions instead.
The email configuration of your user pool. The email configuration type sets your preferred sending method, Amazon Web Services Region, and sender for messages tfrom your user pool.
"
+ "documentation":"
The email configuration of your user pool. The email configuration type sets your preferred sending method, Amazon Web Services Region, and sender for messages from your user pool.
The reason why the SMS configuration can't send the messages to your users.
This message might include comma-separated values to describe why your SMS configuration can't send messages to user pool end users.
InvalidSmsRoleAccessPolicyException
The Identity and Access Management role that Amazon Cognito uses to send SMS messages isn't properly configured. For more information, see SmsConfigurationType.
SNSSandbox
The Amazon Web Services account is in the SNS SMS Sandbox and messages will only reach verified end users. This parameter won’t get populated with SNSSandbox if the IAM user creating the user pool doesn’t have SNS permissions. To learn how to move your Amazon Web Services account out of the sandbox, see Moving out of the SMS sandbox.
"
+ "documentation":"
The reason why the SMS configuration can't send the messages to your users.
This message might include comma-separated values to describe why your SMS configuration can't send messages to user pool end users.
InvalidSmsRoleAccessPolicyException
The Identity and Access Management role that Amazon Cognito uses to send SMS messages isn't properly configured. For more information, see SmsConfigurationType.
SNSSandbox
The Amazon Web Services account is in the SNS SMS Sandbox and messages will only reach verified end users. This parameter won’t get populated with SNSSandbox if the user creating the user pool doesn’t have SNS permissions. To learn how to move your Amazon Web Services account out of the sandbox, see Moving out of the SMS sandbox.
User pool add-ons. Contains settings for activation of advanced security features. To log user security information but take no action, set to AUDIT. To configure automatic security responses to risky traffic to your user pool, set to ENFORCED.
The user status. This can be one of the following:
UNCONFIRMED - User has been created but not confirmed.
CONFIRMED - User has been confirmed.
EXTERNAL_PROVIDER - User signed in with a third-party IdP.
ARCHIVED - User is no longer active.
UNKNOWN - User status isn't known.
RESET_REQUIRED - User is confirmed, but the user must request a code and reset their password before they can sign in.
FORCE_CHANGE_PASSWORD - The user is confirmed and the user can sign in using a temporary password, but on first sign-in, the user must change their password to a new value before doing anything else.
"
+ "documentation":"
The user status. This can be one of the following:
UNCONFIRMED - User has been created but not confirmed.
CONFIRMED - User has been confirmed.
EXTERNAL_PROVIDER - User signed in with a third-party IdP.
UNKNOWN - User status isn't known.
RESET_REQUIRED - User is confirmed, but the user must request a code and reset their password before they can sign in.
FORCE_CHANGE_PASSWORD - The user is confirmed and the user can sign in using a temporary password, but on first sign-in, the user must change their password to a new value before doing anything else.
Specifies whether user name case sensitivity will be applied for all users in the user pool through Amazon Cognito APIs.
Valid values include:
True
Enables case sensitivity for all username input. When this option is set to True, users must sign in using the exact capitalization of their given username, such as “UserName”. This is the default value.
False
Enables case insensitivity for all username input. For example, when this option is set to False, users can sign in using either \"username\" or \"Username\". This option also enables both preferred_username and email alias to be case insensitive, in addition to the username attribute.
"
+ "documentation":"
Specifies whether user name case sensitivity will be applied for all users in the user pool through Amazon Cognito APIs. For most use cases, set case sensitivity to False (case insensitive) as a best practice. When usernames and email addresses are case insensitive, users can sign in as the same user when they enter a different capitalization of their user name.
Valid values include:
True
Enables case sensitivity for all username input. When this option is set to True, users must sign in using the exact capitalization of their given username, such as “UserName”. This is the default value.
False
Enables case insensitivity for all username input. For example, when this option is set to False, users can sign in using username, USERNAME, or UserName. This option also enables both preferred_username and email alias to be case insensitive, in addition to the username attribute.
Using the Amazon Cognito user pools API, you can create a user pool to manage directories and users. You can authenticate a user to obtain tokens related to user identity and access policies.
This API reference provides information about user pools in Amazon Cognito user pools.
This API reference provides detailed information about API operations and object types in Amazon Cognito. At the bottom of the page for each API operation and object, under See Also, you can learn how to use it in an Amazon Web Services SDK in the language of your choice.
Along with resource management operations, the Amazon Cognito user pools API includes classes of operations and authorization models for client-side and server-side user operations. For more information, see Using the Amazon Cognito native and OIDC APIs in the Amazon Cognito Developer Guide.
You can also start reading about the CognitoIdentityProvider client in the following SDK guides.
Lists phone numbers claimed to your Amazon Connect instance or traffic distribution group. If the provided TargetArn is a traffic distribution group, you can call this API in both Amazon Web Services Regions associated with traffic distribution group.
Lists phone numbers claimed to your Amazon Connect instance or traffic distribution group. If the provided TargetArn is a traffic distribution group, you can call this API in both Amazon Web Services Regions associated with traffic distribution group.
Updates the name and description of a quick connect. The request accepts the following data in JSON format. At least Name or Description must be provided.
Updates the traffic distribution for a given traffic distribution group.
You can change the SignInConfig only for a default TrafficDistributionGroup. If you call UpdateTrafficDistribution with a modified SignInConfig and a non-default TrafficDistributionGroup, an InvalidRequestException is returned.
The identifier of the traffic distribution group. This can be the ID or the ARN if the API is being called in the Region where the traffic distribution group was created. The ARN must be provided if the call is from the replicated Region.
The identifier of the traffic distribution group. This can be the ID or the ARN if the API is being called in the Region where the traffic distribution group was created. The ARN must be provided if the call is from the replicated Region.
The identifier of the traffic distribution group. This can be the ID or the ARN if the API is being called in the Region where the traffic distribution group was created. The ARN must be provided if the call is from the replicated Region.
Whether this is the default traffic distribution group created during instance replication. The default traffic distribution group cannot be deleted by the DeleteTrafficDistributionGroup API. The default traffic distribution group is deleted as part of the process for deleting a replica.
You can change the SignInConfig only for a default TrafficDistributionGroup. If you call UpdateTrafficDistribution with a modified SignInConfig and a non-default TrafficDistributionGroup, an InvalidRequestException is returned.
Whether this is the default traffic distribution group created during instance replication. The default traffic distribution group cannot be deleted by the DeleteTrafficDistributionGroup API. The default traffic distribution group is deleted as part of the process for deleting a replica.
Creates recommendations that help you save cost by identifying idle and underutilized Amazon EC2 instances.
Recommendations are generated to either downsize or terminate instances, along with providing savings detail and metrics. For more information about calculation and function, see Optimizing Your Cost with Rightsizing Recommendations in the Billing and Cost Management User Guide.
Retrieves the details for a Savings Plan recommendation. These details include the hourly data-points that construct the new cost, coverage, and utilization charts.
Requests a Savings Plans recommendation generation. This enables you to calculate a fresh set of Savings Plans recommendations that takes your latest usage data and current Savings Plans inventory into account. You can refresh Savings Plans recommendations up to three times daily for a consolidated billing family.
StartSavingsPlansPurchaseRecommendationGeneration has no request syntax because no input parameters are needed to support this operation.
The dollar value that triggers a notification if the threshold is exceeded.
This field has been deprecated. To specify a threshold, use ThresholdExpression. Continued use of Threshold will be treated as shorthand syntax for a ThresholdExpression.
One of Threshold or ThresholdExpression is required for this resource.
",
+ "documentation":"
(deprecated)
An absolute dollar value that must be exceeded by the anomaly's total impact (see Impact for more details) for an anomaly notification to be generated.
This field has been deprecated. To specify a threshold, use ThresholdExpression. Continued use of Threshold will be treated as shorthand syntax for a ThresholdExpression.
One of Threshold or ThresholdExpression is required for this resource. You cannot specify both.
",
"deprecated":true,
"deprecatedMessage":"Threshold has been deprecated in favor of ThresholdExpression"
},
"Frequency":{
"shape":"AnomalySubscriptionFrequency",
- "documentation":"
The frequency that anomaly reports are sent over email.
"
+ "documentation":"
The frequency that anomaly notifications are sent. Notifications are sent either over email (for DAILY and WEEKLY frequencies) or SNS (for IMMEDIATE frequency). For more information, see Creating an Amazon SNS topic for anomaly notifications.
An Expression object used to specify the anomalies that you want to generate alerts for. This supports dimensions and nested expressions. The supported dimensions are ANOMALY_TOTAL_IMPACT_ABSOLUTE and ANOMALY_TOTAL_IMPACT_PERCENTAGE. The supported nested expression types are AND and OR. The match option GREATER_THAN_OR_EQUAL is required. Values must be numbers between 0 and 10,000,000,000.
One of Threshold or ThresholdExpression is required for this resource.
The following are examples of valid ThresholdExpressions:
An Expression object used to specify the anomalies that you want to generate alerts for. This supports dimensions and nested expressions. The supported dimensions are ANOMALY_TOTAL_IMPACT_ABSOLUTE and ANOMALY_TOTAL_IMPACT_PERCENTAGE, corresponding to an anomaly’s TotalImpact and TotalImpactPercentage, respectively (see Impact for more details). The supported nested expression types are AND and OR. The match option GREATER_THAN_OR_EQUAL is required. Values must be numbers between 0 and 10,000,000,000 in string format.
One of Threshold or ThresholdExpression is required for this resource. You cannot specify both.
The following are examples of valid ThresholdExpressions:
The association between a monitor, threshold, and list of subscribers used to deliver notifications about anomalies detected by a monitor that exceeds a threshold. The content consists of the detailed metadata and the current status of the AnomalySubscription object.
"
+ "documentation":"
An AnomalySubscription resource (also referred to as an alert subscription) sends notifications about specific anomalies that meet an alerting criteria defined by you.
You can specify the frequency of the alerts and the subscribers to notify.
Anomaly subscriptions can be associated with one or more AnomalyMonitor resources, and they only send notifications about anomalies detected by those associated monitors. You can also configure a threshold to further control which anomalies are included in the notifications.
Anomalies that don’t exceed the chosen threshold and therefore don’t trigger notifications from an anomaly subscription will still be available on the console and from the GetAnomalies API.
An Expression object used to categorize costs. This supports dimensions, tags, and nested expressions. Currently the only dimensions supported are LINKED_ACCOUNT, SERVICE_CODE, RECORD_TYPE, and LINKED_ACCOUNT_NAME.
Root level OR isn't supported. We recommend that you create a separate rule instead.
RECORD_TYPE is a dimension used for Cost Explorer APIs, and is also supported for Cost Category expressions. This dimension uses different terms, depending on whether you're using the console or API/JSON editor. For a detailed comparison, see Term Comparisons in the Billing and Cost Management User Guide.
"
+ "documentation":"
An Expression object used to categorize costs. This supports dimensions, tags, and nested expressions. Currently the only dimensions supported are LINKED_ACCOUNT, SERVICE_CODE, RECORD_TYPE, LINKED_ACCOUNT_NAME, REGION, and USAGE_TYPE.
RECORD_TYPE is a dimension used for Cost Explorer APIs, and is also supported for Cost Category expressions. This dimension uses different terms, depending on whether you're using the console or API/JSON editor. For a detailed comparison, see Term Comparisons in the Billing and Cost Management User Guide.
Use Expression to filter in various Cost Explorer APIs.
Not all Expression types are supported in each API. Refer to the documentation for each specific API to see what is supported.
There are two patterns:
Simple dimension values.
There are three types of simple dimension values: CostCategories, Tags, and Dimensions.
Specify the CostCategories field to define a filter that acts on Cost Categories.
Specify the Tags field to define a filter that acts on Cost Allocation Tags.
Specify the Dimensions field to define a filter that acts on the DimensionValues.
For each filter type, you can set the dimension name and values for the filters that you plan to use.
For example, you can filter for REGION==us-east-1 OR REGION==us-west-1. For GetRightsizingRecommendation, the Region is a full name (for example, REGION==US East (N. Virginia).
The corresponding Expression for this example is as follows: { \"Dimensions\": { \"Key\": \"REGION\", \"Values\": [ \"us-east-1\", “us-west-1” ] } }
As shown in the previous example, lists of dimension values are combined with OR when applying the filter.
You can also set different match options to further control how the filter behaves. Not all APIs support match options. Refer to the documentation for each specific API to see what is supported.
For example, you can filter for linked account names that start with “a”.
The corresponding Expression for this example is as follows: { \"Dimensions\": { \"Key\": \"LINKED_ACCOUNT_NAME\", \"MatchOptions\": [ \"STARTS_WITH\" ], \"Values\": [ \"a\" ] } }
Compound Expression types with logical operations.
You can use multiple Expression types and the logical operators AND/OR/NOT to create a list of one or more Expression objects. By doing this, you can filter by more advanced options.
For example, you can filter by ((REGION == us-east-1 OR REGION == us-west-1) OR (TAG.Type == Type1)) AND (USAGE_TYPE != DataTransfer).
The corresponding Expression for this example is as follows: { \"And\": [ {\"Or\": [ {\"Dimensions\": { \"Key\": \"REGION\", \"Values\": [ \"us-east-1\", \"us-west-1\" ] }}, {\"Tags\": { \"Key\": \"TagName\", \"Values\": [\"Value1\"] } } ]}, {\"Not\": {\"Dimensions\": { \"Key\": \"USAGE_TYPE\", \"Values\": [\"DataTransfer\"] }}} ] }
Because each Expression can have only one operator, the service returns an error if more than one is specified. The following example shows an Expression object that creates an error: { \"And\": [ ... ], \"Dimensions\": { \"Key\": \"USAGE_TYPE\", \"Values\": [ \"DataTransfer\" ] } }
The following is an example of the corresponding error message: \"Expression has more than one roots. Only one root operator is allowed for each expression: And, Or, Not, Dimensions, Tags, CostCategories\"
For the GetRightsizingRecommendation action, a combination of OR and NOT isn't supported. OR isn't supported between different dimensions, or dimensions and tags. NOT operators aren't supported. Dimensions are also limited to LINKED_ACCOUNT, REGION, or RIGHTSIZING_TYPE.
For the GetReservationPurchaseRecommendation action, only NOT is supported. AND and OR aren't supported. Dimensions are limited to LINKED_ACCOUNT.
"
+ "documentation":"
Use Expression to filter in various Cost Explorer APIs.
Not all Expression types are supported in each API. Refer to the documentation for each specific API to see what is supported.
There are two patterns:
Simple dimension values.
There are three types of simple dimension values: CostCategories, Tags, and Dimensions.
Specify the CostCategories field to define a filter that acts on Cost Categories.
Specify the Tags field to define a filter that acts on Cost Allocation Tags.
Specify the Dimensions field to define a filter that acts on the DimensionValues.
For each filter type, you can set the dimension name and values for the filters that you plan to use.
For example, you can filter for REGION==us-east-1 OR REGION==us-west-1. For GetRightsizingRecommendation, the Region is a full name (for example, REGION==US East (N. Virginia).
The corresponding Expression for this example is as follows: { \"Dimensions\": { \"Key\": \"REGION\", \"Values\": [ \"us-east-1\", \"us-west-1\" ] } }
As shown in the previous example, lists of dimension values are combined with OR when applying the filter.
You can also set different match options to further control how the filter behaves. Not all APIs support match options. Refer to the documentation for each specific API to see what is supported.
For example, you can filter for linked account names that start with \"a\".
The corresponding Expression for this example is as follows: { \"Dimensions\": { \"Key\": \"LINKED_ACCOUNT_NAME\", \"MatchOptions\": [ \"STARTS_WITH\" ], \"Values\": [ \"a\" ] } }
Compound Expression types with logical operations.
You can use multiple Expression types and the logical operators AND/OR/NOT to create a list of one or more Expression objects. By doing this, you can filter by more advanced options.
For example, you can filter by ((REGION == us-east-1 OR REGION == us-west-1) OR (TAG.Type == Type1)) AND (USAGE_TYPE != DataTransfer).
The corresponding Expression for this example is as follows: { \"And\": [ {\"Or\": [ {\"Dimensions\": { \"Key\": \"REGION\", \"Values\": [ \"us-east-1\", \"us-west-1\" ] }}, {\"Tags\": { \"Key\": \"TagName\", \"Values\": [\"Value1\"] } } ]}, {\"Not\": {\"Dimensions\": { \"Key\": \"USAGE_TYPE\", \"Values\": [\"DataTransfer\"] }}} ] }
Because each Expression can have only one operator, the service returns an error if more than one is specified. The following example shows an Expression object that creates an error: { \"And\": [ ... ], \"Dimensions\": { \"Key\": \"USAGE_TYPE\", \"Values\": [ \"DataTransfer\" ] } }
The following is an example of the corresponding error message: \"Expression has more than one roots. Only one root operator is allowed for each expression: And, Or, Not, Dimensions, Tags, CostCategories\"
For the GetRightsizingRecommendation action, a combination of OR and NOT isn't supported. OR isn't supported between different dimensions, or dimensions and tags. NOT operators aren't supported. Dimensions are also limited to LINKED_ACCOUNT, REGION, or RIGHTSIZING_TYPE.
For the GetReservationPurchaseRecommendation action, only NOT is supported. AND and OR aren't supported. Dimensions are limited to LINKED_ACCOUNT.
The account scope that you want your recommendations for. Amazon Web Services calculates recommendations including the management account and member accounts if the value is set to PAYER. If the value is LINKED, recommendations are calculated for individual member accounts only.
The estimated return on investment that's based on the recommended Savings Plan that you purchased. This is calculated as estimatedSavingsAmount/estimatedSPCost*100.
The update to the threshold value for receiving notifications.
This field has been deprecated. To update a threshold, use ThresholdExpression. Continued use of Threshold will be treated as shorthand syntax for a ThresholdExpression.
",
+ "documentation":"
(deprecated)
The update to the threshold value for receiving notifications.
This field has been deprecated. To update a threshold, use ThresholdExpression. Continued use of Threshold will be treated as shorthand syntax for a ThresholdExpression.
You can specify either Threshold or ThresholdExpression, but not both.
",
"deprecated":true,
"deprecatedMessage":"Threshold has been deprecated in favor of ThresholdExpression"
},
@@ -4667,7 +4859,7 @@
},
"ThresholdExpression":{
"shape":"Expression",
- "documentation":"
The update to the Expression object used to specify the anomalies that you want to generate alerts for. This supports dimensions and nested expressions. The supported dimensions are ANOMALY_TOTAL_IMPACT_ABSOLUTE and ANOMALY_TOTAL_IMPACT_PERCENTAGE. The supported nested expression types are AND and OR. The match option GREATER_THAN_OR_EQUAL is required. Values must be numbers between 0 and 10,000,000,000.
The following are examples of valid ThresholdExpressions:
The update to the Expression object used to specify the anomalies that you want to generate alerts for. This supports dimensions and nested expressions. The supported dimensions are ANOMALY_TOTAL_IMPACT_ABSOLUTE and ANOMALY_TOTAL_IMPACT_PERCENTAGE, corresponding to an anomaly’s TotalImpact and TotalImpactPercentage, respectively (see Impact for more details). The supported nested expression types are AND and OR. The match option GREATER_THAN_OR_EQUAL is required. Values must be numbers between 0 and 10,000,000,000 in string format.
You can specify either Threshold or ThresholdExpression, but not both.
The following are examples of valid ThresholdExpressions:
Creates a domain, which is a container for all customer data, such as customer profile attributes, object types, profile keys, and encryption keys. You can create multiple domains, and each domain can have multiple third-party integrations.
Each Amazon Connect instance can be associated with only one domain. Multiple Amazon Connect instances can be associated with one domain.
Creates a domain, which is a container for all customer data, such as customer profile attributes, object types, profile keys, and encryption keys. You can create multiple domains, and each domain can have multiple third-party integrations.
Each Amazon Connect instance can be associated with only one domain. Multiple Amazon Connect instances can be associated with one domain.
Returns the template information for a specific object type.
A template is a predefined ProfileObjectType, such as “Salesforce-Account” or “Salesforce-Contact.” When a user sends a ProfileObject, using the PutProfileObject API, with an ObjectTypeName that matches one of the TemplateIds, it uses the mappings from the template.
Returns a set of profiles that belong to the same matching group using the matchId or profileId. You can also specify the type of matching that you want for finding similar profiles using either RULE_BASED_MATCHING or ML_BASED_MATCHING.
The Address type. You can choose from Address, BusinessAddress, MaillingAddress, and ShippingAddress.
You only can use the Address type in the MatchingRule. For example, if you want to match profile based on BusinessAddress.City or MaillingAddress.City, you need to choose the BusinessAddress and the MaillingAddress to represent the Address type and specify the Address.City on the matching rule.
The PhoneNumber type. You can choose from PhoneNumber, HomePhoneNumber, and MobilePhoneNumber.
You only can use the PhoneNumber type in the MatchingRule. For example, if you want to match a profile based on Phone or HomePhone, you need to choose the Phone and the HomePhone to represent the PhoneNumber type and only specify the PhoneNumber on the matching rule.
The Email type. You can choose from EmailAddress, BusinessEmailAddress and PersonalEmailAddress.
You only can use the EmailAddress type in the MatchingRule. For example, if you want to match profile based on PersonalEmailAddress or BusinessEmailAddress, you need to choose the PersonalEmailAddress and the BusinessEmailAddress to represent the EmailAddress type and only specify the EmailAddress on the matching rule.
"
+ }
+ },
+ "documentation":"
Configuration information about the AttributeTypesSelector where the rule-based identity resolution uses to match profiles. You can choose how profiles are compared across attribute types and which attribute to use for matching from each type. There are three attribute types you can configure:
Email type
You can choose from Email, BusinessEmail, and PersonalEmail
Phone number type
You can choose from Phone, HomePhone, and MobilePhone
Address type
You can choose from Address, BusinessAddress, MaillingAddress, and ShippingAddress
You can either choose ONE_TO_ONE or MANY_TO_MANY as the AttributeMatchingModel. When choosing MANY_TO_MANY, the system can match attribute across the sub-types of an attribute type. For example, if the value of the Email field of Profile A and the value of BusinessEmail field of Profile B matches, the two profiles are matched on the Email type. When choosing ONE_TO_ONE the system can only match if the sub-types are exact matches. For example, only when the value of the Email field of Profile A and the value of the Email field of Profile B matches, the two profiles are matched on the Email type.
The process of matching duplicate profiles. If Matching = true, Amazon Connect Customer Profiles starts a weekly batch process called Identity Resolution Job. If you do not specify a date and time for Identity Resolution Job to run, by default it runs every Saturday at 12AM UTC to detect duplicate profiles in your domains.
After the Identity Resolution Job completes, use the GetMatches API to return and review the results. Or, if you have configured ExportingConfig in the MatchingRequest, you can download the results from S3.
The process of matching duplicate profiles using the Rule-Based matching. If RuleBasedMatching = true, Amazon Connect Customer Profiles will start to match and merge your profiles according to your configuration in the RuleBasedMatchingRequest. You can use the ListRuleBasedMatches and GetSimilarProfiles API to return and review the results. Also, if you have configured ExportingConfig in the RuleBasedMatchingRequest, you can download the results from S3.
The process of matching duplicate profiles. If Matching = true, Amazon Connect Customer Profiles starts a weekly batch process called Identity Resolution Job. If you do not specify a date and time for Identity Resolution Job to run, by default it runs every Saturday at 12AM UTC to detect duplicate profiles in your domains.
After the Identity Resolution Job completes, use the GetMatches API to return and review the results. Or, if you have configured ExportingConfig in the MatchingRequest, you can download the results from S3.
The process of matching duplicate profiles using the Rule-Based matching. If RuleBasedMatching = true, Amazon Connect Customer Profiles will start to match and merge your profiles according to your configuration in the RuleBasedMatchingRequest. You can use the ListRuleBasedMatches and GetSimilarProfiles API to return and review the results. Also, if you have configured ExportingConfig in the RuleBasedMatchingRequest, you can download the results from S3.
The process of matching duplicate profiles. If Matching = true, Amazon Connect Customer Profiles starts a weekly batch process called Identity Resolution Job. If you do not specify a date and time for Identity Resolution Job to run, by default it runs every Saturday at 12AM UTC to detect duplicate profiles in your domains.
After the Identity Resolution Job completes, use the GetMatches API to return and review the results. Or, if you have configured ExportingConfig in the MatchingRequest, you can download the results from S3.
The process of matching duplicate profiles using the Rule-Based matching. If RuleBasedMatching = true, Amazon Connect Customer Profiles will start to match and merge your profiles according to your configuration in the RuleBasedMatchingRequest. You can use the ListRuleBasedMatches and GetSimilarProfiles API to return and review the results. Also, if you have configured ExportingConfig in the RuleBasedMatchingRequest, you can download the results from S3.
It only has value when the MatchType is ML_BASED_MATCHING.A number between 0 and 1, where a higher score means higher similarity. Examining match confidence scores lets you distinguish between groups of similar records in which the system is highly confident (which you may decide to merge), groups of similar records about which the system is uncertain (which you may decide to have reviewed by a human), and groups of similar records that the system deems to be unlikely (which you may decide to reject). Given confidence scores vary as per the data input, it should not be used as an absolute measure of matching quality.
The first status after configuration a rule-based matching rule. If it is an existing domain, the rule-based Identity Resolution waits one hour before creating the matching rule. If it is a new domain, the system will skip the PENDING stage.
IN_PROGRESS
The system is creating the rule-based matching rule. Under this status, the system is evaluating the existing data and you can no longer change the Rule-based matching configuration.
ACTIVE
The rule is ready to use. You can change the rule a day after the status is in ACTIVE.
The process of matching duplicate profiles. If Matching = true, Amazon Connect Customer Profiles starts a weekly batch process called Identity Resolution Job. If you do not specify a date and time for Identity Resolution Job to run, by default it runs every Saturday at 12AM UTC to detect duplicate profiles in your domains.
After the Identity Resolution Job completes, use the GetMatches API to return and review the results. Or, if you have configured ExportingConfig in the MatchingRequest, you can download the results from S3.
The process of matching duplicate profiles using the rule-Based matching. If RuleBasedMatching = true, Amazon Connect Customer Profiles will start to match and merge your profiles according to your configuration in the RuleBasedMatchingRequest. You can use the ListRuleBasedMatches and GetSimilarProfiles API to return and review the results. Also, if you have configured ExportingConfig in the RuleBasedMatchingRequest, you can download the results from S3.
The process of matching duplicate profiles. If Matching = true, Amazon Connect Customer Profiles starts a weekly batch process called Identity Resolution Job. If you do not specify a date and time for Identity Resolution Job to run, by default it runs every Saturday at 12AM UTC to detect duplicate profiles in your domains.
After the Identity Resolution Job completes, use the GetMatches API to return and review the results. Or, if you have configured ExportingConfig in the MatchingRequest, you can download the results from S3.
The process of matching duplicate profiles using the rule-Based matching. If RuleBasedMatching = true, Amazon Connect Customer Profiles will start to match and merge your profiles according to your configuration in the RuleBasedMatchingRequest. You can use the ListRuleBasedMatches and GetSimilarProfiles API to return and review the results. Also, if you have configured ExportingConfig in the RuleBasedMatchingRequest, you can download the results from S3.
This operation prevents any individual assessments from running if they haven't started running. It also attempts to cancel any individual assessments that are currently running.
Describes the status of the connections that have been made between the replication instance and an endpoint. Connections are created when you test an endpoint.
Lists events for a given source identifier and source type. You can also specify a start and end time. For more information on DMS events, see Working with Events and Notifications in the Database Migration Service User Guide.
Returns a paginated list of extension pack associations for the specified migration project. An extension pack is an add-on module that emulates functions present in a source database that are required when converting objects to the target database.
Returns table statistics on the database migration task, including table name, rows inserted, rows updated, and rows deleted.
Note that the \"last updated\" column the DMS console only indicates the time that DMS last updated the table statistics record for a table. It does not indicate the time of the last update to the table.
Saves a copy of a database migration assessment report to your Amazon S3 bucket. DMS can save your assessment report as a comma-separated value (CSV) or a PDF file.
Lists all metadata tags attached to an DMS resource, including replication instance, endpoint, subnet group, and migration task. For more information, see Tag data type description.
Applies the extension pack to your target database. An extension pack is an add-on module that emulates functions present in a source database that are required when converting objects to the target database.
For a given DMS Serverless replication configuration, DMS connects to the source endpoint and collects the metadata to analyze the replication workload. Using this metadata, DMS then computes and provisions the required capacity and starts replicating to the target endpoint using the server resources that DMS has provisioned for the DMS Serverless replication.
"
+ "documentation":"
Creates a database migration assessment report by assessing the migration complexity for your source database. A database migration assessment report summarizes all of the schema conversion tasks. It also details the action items for database objects that can't be converted to the database engine of your target database instance.
For a given DMS Serverless replication configuration, DMS connects to the source endpoint and collects the metadata to analyze the replication workload. Using this metadata, DMS then computes and provisions the required capacity and starts replicating to the target endpoint using the server resources that DMS has provisioned for the DMS Serverless replication.
The type of database engine for the data provider. Valid values include \"aurora\", \"aurora_postgresql\", \"mysql\", \"oracle\", \"postgres\", and \"sqlserver\". A value of \"aurora\" represents Amazon Aurora MySQL-Compatible Edition.
The Availability Zone where the instance profile will be created. The default value is a random, system-chosen Availability Zone in the Amazon Web Services Region where your data provider is created, for examplem us-east-1d.
The Amazon Resource Name (ARN) of the KMS key that is used to encrypt the connection parameters for the instance profile.
If you don't specify a value for the KmsKeyArn parameter, then DMS uses your default encryption key.
KMS creates the default encryption key for your Amazon Web Services account. Your Amazon Web Services account has a different default encryption key for each Amazon Web Services Region.
Specifies the accessibility options for the instance profile. A value of true represents an instance profile with a public IP address. A value of false represents an instance profile with a private IP address. The default value is true.
Specifies the network type for the instance profile. A value of IPV4 represents an instance profile with IPv4 network type and only supports IPv4 addressing. A value of IPV6 represents an instance profile with IPv6 network type and only supports IPv6 addressing. A value of DUAL represents an instance profile with dual network type that supports IPv4 and IPv6 addressing.
Specifies the VPC security group names to be used with the instance profile. The VPC security group must work with the VPC containing the instance profile.
The identifier of the associated instance profile. Identifiers must begin with a letter and must contain only ASCII letters, digits, and hyphens. They can't end with a hyphen, or contain two consecutive hyphens.
The settings in JSON format for migration rules. Migration rules make it possible for you to change the object names according to the rules that you specify. For example, you can change an object name to lowercase or uppercase, add or remove a prefix or suffix, or rename objects.
A description of the data provider. Descriptions can have up to 31 characters. A description can contain only ASCII letters, digits, and hyphens ('-'). Also, it can't end with a hyphen or contain two consecutive hyphens, and can only begin with a letter.
The type of database engine for the data provider. Valid values include \"aurora\", \"aurora_postgresql\", \"mysql\", \"oracle\", \"postgres\", and \"sqlserver\". A value of \"aurora\" represents Amazon Aurora MySQL-Compatible Edition.
The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, DMS includes a pagination token in the response so that you can retrieve the remaining results.
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.
An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, DMS includes a pagination token in the response so that you can retrieve the remaining results.
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
If NextToken is returned, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
"
+ "documentation":"
If NextToken is returned, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
If NextToken is returned by a previous response, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
If NextToken is returned, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
If NextToken is returned by a previous response, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
If NextToken is returned, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
If NextToken is returned by a previous response, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
If NextToken is returned, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
If NextToken is returned by a previous response, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
If NextToken is returned, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, DMS includes a pagination token in the response so that you can retrieve the remaining results.
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, DMS includes a pagination token in the response so that you can retrieve the remaining results.
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, DMS includes a pagination token in the response so that you can retrieve the remaining results.
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
If NextToken is returned by a previous response, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
"
+ "documentation":"
The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, DMS includes a pagination token in the response so that you can retrieve the remaining results.
If NextToken is returned, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
"
+ "documentation":"
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
If NextToken is returned by a previous response, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
"
+ "documentation":"
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, DMS includes a pagination token in the response so that you can retrieve the remaining results.
If NextToken is returned, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
"
+ "documentation":"
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
If NextToken is returned by a previous response, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
If NextToken is returned, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
"
+ "documentation":"
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
Sets the maximum number of records returned in the response.
"
+ "documentation":"
The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, DMS includes a pagination token in the response so that you can retrieve the remaining results.
If NextToken is returned by a previous response, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
"
+ "documentation":"
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
If NextToken is returned, there are more results available. The value of NextToken is a unique pagination token for each page. Make the call again using the returned token to retrieve the next page. Keep all other arguments unchanged.
"
+ "documentation":"
Specifies the unique pagination token that makes it possible to display the next page of results. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
If Marker is returned by a previous response, there are more results available. The value of Marker is a unique pagination token for each page. To retrieve the next page, make the call again using the returned token and keeping all other arguments unchanged.
The Amazon Resource Name (ARN) of the KMS key that is used to encrypt the connection parameters for the instance profile.
If you don't specify a value for the KmsKeyArn parameter, then DMS uses your default encryption key.
KMS creates the default encryption key for your Amazon Web Services account. Your Amazon Web Services account has a different default encryption key for each Amazon Web Services Region.
Specifies the accessibility options for the instance profile. A value of true represents an instance profile with a public IP address. A value of false represents an instance profile with a private IP address. The default value is true.
Specifies the network type for the instance profile. A value of IPV4 represents an instance profile with IPv4 network type and only supports IPv4 addressing. A value of IPV6 represents an instance profile with IPv6 network type and only supports IPv6 addressing. A value of DUAL represents an instance profile with dual network type that supports IPv4 and IPv6 addressing.
A description of the instance profile. Descriptions can have up to 31 characters. A description can contain only ASCII letters, digits, and hyphens ('-'). Also, it can't end with a hyphen or contain two consecutive hyphens, and can only begin with a letter.
The settings in JSON format for migration rules. Migration rules make it possible for you to change the object names according to the rules that you specify. For example, you can change an object name to lowercase or uppercase, add or remove a prefix or suffix, or rename objects.
The identifier of the data provider. Identifiers must begin with a letter and must contain only ASCII letters, digits, and hyphens. They can't end with a hyphen, or contain two consecutive hyphens.
The type of database engine for the data provider. Valid values include \"aurora\", \"aurora_postgresql\", \"mysql\", \"oracle\", \"postgres\", and \"sqlserver\". A value of \"aurora\" represents Amazon Aurora MySQL-Compatible Edition.
If this attribute is Y, the current call to ModifyDataProvider replaces all existing data provider settings with the exact settings that you specify in this call. If this attribute is N, the current call to ModifyDataProvider does two things:
It replaces any data provider settings that already exist with new values, for settings with the same names.
It creates new data provider settings that you specify in the call, for settings with different names.
The identifier of the instance profile. Identifiers must begin with a letter and must contain only ASCII letters, digits, and hyphens. They can't end with a hyphen, or contain two consecutive hyphens.
The Amazon Resource Name (ARN) of the KMS key that is used to encrypt the connection parameters for the instance profile.
If you don't specify a value for the KmsKeyArn parameter, then DMS uses your default encryption key.
KMS creates the default encryption key for your Amazon Web Services account. Your Amazon Web Services account has a different default encryption key for each Amazon Web Services Region.
Specifies the accessibility options for the instance profile. A value of true represents an instance profile with a public IP address. A value of false represents an instance profile with a private IP address. The default value is true.
Specifies the network type for the instance profile. A value of IPV4 represents an instance profile with IPv4 network type and only supports IPv4 addressing. A value of IPV6 represents an instance profile with IPv6 network type and only supports IPv6 addressing. A value of DUAL represents an instance profile with dual network type that supports IPv4 and IPv6 addressing.
The identifier of the migration project. Identifiers must begin with a letter and must contain only ASCII letters, digits, and hyphens. They can't end with a hyphen, or contain two consecutive hyphens.
The settings in JSON format for migration rules. Migration rules make it possible for you to change the object names according to the rules that you specify. For example, you can change an object name to lowercase or uppercase, add or remove a prefix or suffix, or rename objects.
The address of your Oracle Automatic Storage Management (ASM) server. You can set this value from the asm_server value. You set asm_server as part of the extra connection attribute string to access an Oracle server with Binary Reader that uses ASM. For more information, see Configuration for change data capture (CDC) on an Oracle source database.
The identifier of the secret in Secrets Manager that contains the transparent data encryption (TDE) password. DMS requires this password to access Oracle redo logs encrypted by TDE using Binary Reader.
Specifies whether to use default or custom replication behavior for PostgreSQL-compatible endpoints. You can use this setting to specify replication behavior for endpoints that require additional configuration, such as Babelfish endpoints.
"
+ "documentation":"
Specifies the default behavior of the replication's handling of PostgreSQL- compatible endpoints that require some additional configuration, such as Babelfish endpoints.
Whether to overwrite the migration project extension pack. An extension pack is an add-on module that emulates functions present in a source database that are required when converting objects to the target database.
Activates an DataSync agent that you've deployed in your storage environment. The activation process associates the agent with your Amazon Web Services account.
If you haven't deployed an agent yet, see the following topics to learn more:
If you're copying data to or from an Snowcone device, you can also use CreateLocationNfs to create your transfer location. For more information, see Configuring transfers with Snowcone.
Creates an endpoint for a Server Message Block (SMB) file server that DataSync can access for a transfer. For more information, see Creating an SMB location.
"
+ "documentation":"
Creates an endpoint for a Server Message Block (SMB) file server that DataSync can use for a data transfer.
Updates some of the parameters of a previously created location for Network File System (NFS) access. For information about creating an NFS location, see Creating a location for NFS.
"
+ "documentation":"
Modifies some configurations of the Network File System (NFS) transfer location that you're using with DataSync.
Specifies the Amazon Resource Name (ARN) of the DataSync agent that connects to and reads from your on-premises storage system's management interface.
"
+ "documentation":"
Specifies the Amazon Resource Name (ARN) of the DataSync agent that connects to and reads from your on-premises storage system's management interface. You can only specify one ARN.
Specifies the ARN of the subnet where you want to run your DataSync task when using a VPC endpoint. This is the subnet where DataSync creates and manages the network interfaces for your transfer.
"
+ "documentation":"
Specifies the ARN of the subnet where you want to run your DataSync task when using a VPC endpoint. This is the subnet where DataSync creates and manages the network interfaces for your transfer. You can only specify one ARN.
Specifies the authentication method DataSync uses to access your Azure Blob Storage. DataSync can access blob storage using a shared access signature (SAS).
Specifies the type of blob that you want your objects or files to be when transferring them into Azure Blob Storage. Currently, DataSync only supports moving data into Azure Blob Storage as block blobs. For more information on blob types, see the Azure Blob Storage documentation.
Specifies the access tier that you want your objects or files transferred into. This only applies when using the location as a transfer destination. For more information, see Access tiers.
Specifies labels that help you categorize, filter, and search for your Amazon Web Services resources. We recommend creating at least a name tag for your transfer location.
The subdirectory in the NFS file system that is used to read data from the NFS source location or write data to the NFS destination. The NFS path should be a path that's exported by the NFS server, or a subdirectory of that path. The path should be such that it can be mounted by other NFS clients in your network.
To see all the paths exported by your NFS server, run \"showmount -e nfs-server-name\" from an NFS client that has access to your server. You can specify any directory that appears in the results, and any subdirectory of that directory. Ensure that the NFS export is accessible without Kerberos authentication.
To transfer all the data in the folder you specified, DataSync needs to have permissions to read all the data. To ensure this, either configure the NFS export with no_root_squash, or ensure that the permissions for all of the files that you want DataSync allow read access for all users. Doing either enables the agent to read the files. For the agent to access directories, you must additionally enable all execute access.
If you are copying data to or from your Snowcone device, see NFS Server on Snowcone for more information.
For information about NFS export configuration, see 18.7. The /etc/exports Configuration File in the Red Hat Enterprise Linux documentation.
"
+ "documentation":"
Specifies the export path in your NFS file server that you want DataSync to mount.
This path (or a subdirectory of the path) is where DataSync transfers data to or from. For information on configuring an export for DataSync, see Accessing NFS file servers.
The name of the NFS server. This value is the IP address or Domain Name Service (DNS) name of the NFS server. An agent that is installed on-premises uses this hostname to mount the NFS server in a network.
If you are copying data to or from your Snowcone device, see NFS Server on Snowcone for more information.
This name must either be DNS-compliant or must be an IP version 4 (IPv4) address.
"
+ "documentation":"
Specifies the Domain Name System (DNS) name or IP version 4 address of the NFS file server that your DataSync agent connects to.
The key-value pair that represents the tag that you want to add to the location. The value can be an empty string. We recommend using tags to name your resources.
"
+ "documentation":"
Specifies labels that help you categorize, filter, and search for your Amazon Web Services resources. We recommend creating at least a name tag for your location.
The type of blob that you want your objects or files to be when transferring them into Azure Blob Storage. Currently, DataSync only supports moving data into Azure Blob Storage as block blobs. For more information on blob types, see the Azure Blob Storage documentation.
The access tier that you want your objects or files transferred into. This only applies when using the location as a transfer destination. For more information, see Access tiers.
The ID of the VPC endpoint that is configured for an agent. An agent that is configured with a VPC endpoint will not be accessible over the public internet.
"
+ "documentation":"
Specifies the ID of the VPC endpoint that your agent connects to.
The private endpoint that is configured for an agent that has access to IP addresses in a PrivateLink. An agent that is configured with this endpoint will not be accessible over the public internet.
The Amazon Resource Names (ARNs) of the security groups that are configured for the EC2 resource that hosts an agent activated in a VPC or an agent that has access to a VPC endpoint.
"
+ "documentation":"
Specifies the Amazon Resource Names (ARN) of the security group that provides DataSync access to your VPC endpoint. You can only specify one ARN.
"
}
},
- "documentation":"
The VPC endpoint, subnet, and security group that an agent uses to access IP addresses in a VPC (Virtual Private Cloud).
"
+ "documentation":"
Specifies how your DataSync agent connects to Amazon Web Services using a virtual private cloud (VPC) service endpoint. An agent that uses a VPC endpoint isn't accessible over the public internet.
Specifies the authentication method DataSync uses to access your Azure Blob Storage. DataSync can access blob storage using a shared access signature (SAS).
Specifies the type of blob that you want your objects or files to be when transferring them into Azure Blob Storage. Currently, DataSync only supports moving data into Azure Blob Storage as block blobs. For more information on blob types, see the Azure Blob Storage documentation.
Specifies the access tier that you want your objects or files transferred into. This only applies when using the location as a transfer destination. For more information, see Access tiers.
The subdirectory in the NFS file system that is used to read data from the NFS source location or write data to the NFS destination. The NFS path should be a path that's exported by the NFS server, or a subdirectory of that path. The path should be such that it can be mounted by other NFS clients in your network.
To see all the paths exported by your NFS server, run \"showmount -e nfs-server-name\" from an NFS client that has access to your server. You can specify any directory that appears in the results, and any subdirectory of that directory. Ensure that the NFS export is accessible without Kerberos authentication.
To transfer all the data in the folder that you specified, DataSync must have permissions to read all the data. To ensure this, either configure the NFS export with no_root_squash, or ensure that the files you want DataSync to access have permissions that allow read access for all users. Doing either option enables the agent to read the files. For the agent to access directories, you must additionally enable all execute access.
If you are copying data to or from your Snowcone device, see NFS Server on Snowcone for more information.
For information about NFS export configuration, see 18.7. The /etc/exports Configuration File in the Red Hat Enterprise Linux documentation.
"
+ "documentation":"
Specifies the export path in your NFS file server that you want DataSync to mount.
This path (or a subdirectory of the path) is where DataSync transfers data to or from. For information on configuring an export for DataSync, see Accessing NFS file servers.
Specifies the Amazon Resource Name (ARN) of the DataSync agent that connects to and reads your on-premises storage system.
"
+ "documentation":"
Specifies the Amazon Resource Name (ARN) of the DataSync agent that connects to and reads your on-premises storage system. You can only specify one ARN.
The Scan operation returns one or more items and item attributes by accessing every item in a table or a secondary index. To have DynamoDB return fewer items, you can provide a FilterExpression operation.
If the total number of scanned items exceeds the maximum dataset size limit of 1 MB, the scan stops and results are returned to the user as a LastEvaluatedKey value to continue the scan in a subsequent operation. The results also include the number of items exceeding the limit. A scan can result in no table data meeting the filter criteria.
A single Scan operation reads up to the maximum number of items set (if using the Limit parameter) or a maximum of 1 MB of data and then apply any filtering to the results using FilterExpression. If LastEvaluatedKey is present in the response, you need to paginate the result set. For more information, see Paginating the Results in the Amazon DynamoDB Developer Guide.
Scan operations proceed sequentially; however, for faster performance on a large table or secondary index, applications can request a parallel Scan operation by providing the Segment and TotalSegments parameters. For more information, see Parallel Scan in the Amazon DynamoDB Developer Guide.
Scan uses eventually consistent reads when accessing the data in a table; therefore, the result set might not include the changes to data in the table immediately before the operation began. If you need a consistent copy of the data, as of the time that the Scan begins, you can set the ConsistentRead parameter to true.
",
+ "documentation":"
The Scan operation returns one or more items and item attributes by accessing every item in a table or a secondary index. To have DynamoDB return fewer items, you can provide a FilterExpression operation.
If the total size of scanned items exceeds the maximum dataset size limit of 1 MB, the scan completes and results are returned to the user. The LastEvaluatedKey value is also returned and the requestor can use the LastEvaluatedKey to continue the scan in a subsequent operation. Each scan response also includes number of items that were scanned (ScannedCount) as part of the request. If using a FilterExpression, a scan result can result in no items meeting the criteria and the Count will result in zero. If you did not use a FilterExpression in the scan request, then Count is the same as ScannedCount.
Count and ScannedCount only return the count of items specific to a single scan request and, unless the table is less than 1MB, do not represent the total number of items in the table.
A single Scan operation first reads up to the maximum number of items set (if using the Limit parameter) or a maximum of 1 MB of data and then applies any filtering to the results if a FilterExpression is provided. If LastEvaluatedKey is present in the response, pagination is required to complete the full table scan. For more information, see Paginating the Results in the Amazon DynamoDB Developer Guide.
Scan operations proceed sequentially; however, for faster performance on a large table or secondary index, applications can request a parallel Scan operation by providing the Segment and TotalSegments parameters. For more information, see Parallel Scan in the Amazon DynamoDB Developer Guide.
By default, a Scan uses eventually consistent reads when accessing the items in a table. Therefore, the results from an eventually consistent Scan may not include the latest item changes at the time the scan iterates through each item in the table. If you require a strongly consistent read of each item as the scan iterates through the items in the table, you can set the ConsistentRead parameter to true. Strong consistency only relates to the consistency of the read at the item level.
DynamoDB does not provide snapshot isolation for a scan operation when the ConsistentRead parameter is set to true. Thus, a DynamoDB scan operation does not guarantee that all reads in a scan see a consistent snapshot of the table when the scan operation was requested.
Use ReturnValuesOnConditionCheckFailure to get the item attributes if the Update condition fails. For ReturnValuesOnConditionCheckFailure, the valid values are: NONE, ALL_OLD, UPDATED_OLD, ALL_NEW, UPDATED_NEW.
"
+ "documentation":"
Use ReturnValuesOnConditionCheckFailure to get the item attributes if the Update condition fails. For ReturnValuesOnConditionCheckFailure, the valid values are: NONE and ALL_OLD.
"
}
},
"documentation":"
Represents a request to perform an UpdateItem operation.
Seals and completes the snapshot after all of the required blocks of data have been written to it. Completing the snapshot changes the status to completed. You cannot write new blocks to a snapshot after it has been completed.
"
+ "documentation":"
Seals and completes the snapshot after all of the required blocks of data have been written to it. Completing the snapshot changes the status to completed. You cannot write new blocks to a snapshot after it has been completed.
You should always retry requests that receive server (5xx) error responses, and ThrottlingException and RequestThrottledException client error responses. For more information see Error retries in the Amazon Elastic Compute Cloud User Guide.
Returns the data in a block in an Amazon Elastic Block Store snapshot.
"
+ "documentation":"
Returns the data in a block in an Amazon Elastic Block Store snapshot.
You should always retry requests that receive server (5xx) error responses, and ThrottlingException and RequestThrottledException client error responses. For more information see Error retries in the Amazon Elastic Compute Cloud User Guide.
Returns information about the blocks that are different between two Amazon Elastic Block Store snapshots of the same volume/snapshot lineage.
"
+ "documentation":"
Returns information about the blocks that are different between two Amazon Elastic Block Store snapshots of the same volume/snapshot lineage.
You should always retry requests that receive server (5xx) error responses, and ThrottlingException and RequestThrottledException client error responses. For more information see Error retries in the Amazon Elastic Compute Cloud User Guide.
Returns information about the blocks in an Amazon Elastic Block Store snapshot.
"
+ "documentation":"
Returns information about the blocks in an Amazon Elastic Block Store snapshot.
You should always retry requests that receive server (5xx) error responses, and ThrottlingException and RequestThrottledException client error responses. For more information see Error retries in the Amazon Elastic Compute Cloud User Guide.
Writes a block of data to a snapshot. If the specified block contains data, the existing data is overwritten. The target snapshot must be in the pending state.
Data written to a snapshot must be aligned with 512-KiB sectors.
",
+ "documentation":"
Writes a block of data to a snapshot. If the specified block contains data, the existing data is overwritten. The target snapshot must be in the pending state.
Data written to a snapshot must be aligned with 512-KiB sectors.
You should always retry requests that receive server (5xx) error responses, and ThrottlingException and RequestThrottledException client error responses. For more information see Error retries in the Amazon Elastic Compute Cloud User Guide.
Creates a new Amazon EBS snapshot. The new snapshot enters the pending state after the request completes.
After creating the snapshot, use PutSnapshotBlock to write blocks of data to the snapshot.
"
+ "documentation":"
Creates a new Amazon EBS snapshot. The new snapshot enters the pending state after the request completes.
After creating the snapshot, use PutSnapshotBlock to write blocks of data to the snapshot.
You should always retry requests that receive server (5xx) error responses, and ThrottlingException and RequestThrottledException client error responses. For more information see Error retries in the Amazon Elastic Compute Cloud User Guide.
If you’re creating a network interface in a dual-stack or IPv6-only subnet, you have the option to assign a primary IPv6 IP address. A primary IPv6 address is an IPv6 GUA address associated with an ENI that you have enabled to use a primary IPv6 address. Use this option if the instance that this ENI will be attached to relies on its IPv6 address not changing. Amazon Web Services will automatically assign an IPv6 address associated with the ENI attached to your instance to be the primary IPv6 address. Once you enable an IPv6 GUA address to be a primary IPv6, you cannot disable it. When you enable an IPv6 GUA address to be a primary IPv6, the first IPv6 GUA will be made the primary IPv6 address until the instance is terminated or the network interface is detached. If you have multiple IPv6 addresses associated with an ENI attached to your instance and you enable a primary IPv6 address, the first IPv6 GUA address associated with the ENI becomes the primary IPv6 address.
Determines if an IPv6 address associated with a network interface is the primary IPv6 address. When you enable an IPv6 GUA address to be a primary IPv6, the first IPv6 GUA will be made the primary IPv6 address until the instance is terminated or the network interface is detached. For more information, see RunInstances.
The number of IPv6 delegated prefixes to be automatically assigned to the network interface. You cannot use this option if you use the Ipv6Prefix option.
The primary IPv6 address of the network interface. When you enable an IPv6 GUA address to be a primary IPv6, the first IPv6 GUA will be made the primary IPv6 address until the instance is terminated or the network interface is detached. For more information about primary IPv6 addresses, see RunInstances.
The primary IPv6 address of the network interface. When you enable an IPv6 GUA address to be a primary IPv6, the first IPv6 GUA will be made the primary IPv6 address until the instance is terminated or the network interface is detached. For more information about primary IPv6 addresses, see RunInstances.
The primary IPv6 address of the network interface. When you enable an IPv6 GUA address to be a primary IPv6, the first IPv6 GUA will be made the primary IPv6 address until the instance is terminated or the network interface is detached. For more information about primary IPv6 addresses, see RunInstances.
If you’re modifying a network interface in a dual-stack or IPv6-only subnet, you have the option to assign a primary IPv6 IP address. A primary IPv6 address is an IPv6 GUA address associated with an ENI that you have enabled to use a primary IPv6 address. Use this option if the instance that this ENI will be attached to relies on its IPv6 address not changing. Amazon Web Services will automatically assign an IPv6 address associated with the ENI attached to your instance to be the primary IPv6 address. Once you enable an IPv6 GUA address to be a primary IPv6, you cannot disable it. When you enable an IPv6 GUA address to be a primary IPv6, the first IPv6 GUA will be made the primary IPv6 address until the instance is terminated or the network interface is detached. If you have multiple IPv6 addresses associated with an ENI attached to your instance and you enable a primary IPv6 address, the first IPv6 GUA address associated with the ENI becomes the primary IPv6 address.
"
}
},
"documentation":"
Contains the parameters for ModifyNetworkInterfaceAttribute.
Determines if an IPv6 address associated with a network interface is the primary IPv6 address. When you enable an IPv6 GUA address to be a primary IPv6, the first IPv6 GUA will be made the primary IPv6 address until the instance is terminated or the network interface is detached. For more information, see ModifyNetworkInterfaceAttribute.
If you’re launching an instance into a dual-stack or IPv6-only subnet, you can enable assigning a primary IPv6 address. A primary IPv6 address is an IPv6 GUA address associated with an ENI that you have enabled to use a primary IPv6 address. Use this option if an instance relies on its IPv6 address not changing. When you launch the instance, Amazon Web Services will automatically assign an IPv6 address associated with the ENI attached to your instance to be the primary IPv6 address. Once you enable an IPv6 GUA address to be a primary IPv6, you cannot disable it. When you enable an IPv6 GUA address to be a primary IPv6, the first IPv6 GUA will be made the primary IPv6 address until the instance is terminated or the network interface is detached. If you have multiple IPv6 addresses associated with an ENI attached to your instance and you enable a primary IPv6 address, the first IPv6 GUA address associated with the ENI becomes the primary IPv6 address.
Only for archived snapshots that are temporarily restored. Indicates the date and time when a temporarily restored snapshot will be automatically re-archived.
The ARN of the Outpost on which the snapshot is stored. For more information, see Amazon EBS local snapshots on Outposts in the Amazon Elastic Compute Cloud User Guide.
Amazon Elastic Compute Cloud (Amazon EC2) provides secure and resizable computing capacity in the Amazon Web Services Cloud. Using Amazon EC2 eliminates the need to invest in hardware up front, so you can develop and deploy applications faster. Amazon Virtual Private Cloud (Amazon VPC) enables you to provision a logically isolated section of the Amazon Web Services Cloud where you can launch Amazon Web Services resources in a virtual network that you've defined. Amazon Elastic Block Store (Amazon EBS) provides block level storage volumes for use with EC2 instances. EBS volumes are highly available and reliable storage volumes that can be attached to any running instance and used like a hard drive.
Currently, stopped tasks appear in the returned results for at least one hour.
"
+ "documentation":"
Describes a specified task or tasks.
Currently, stopped tasks appear in the returned results for at least one hour.
If you have tasks with tags, and then delete the cluster, the tagged tasks are returned in the response. If you create a new cluster with the same name as the deleted cluster, the tagged tasks are not included in the response.
Returns a list of tasks. You can filter the results by cluster, task definition family, container instance, launch type, what IAM principal started the task, or by the desired status of the task.
Recently stopped tasks might appear in the returned results. Currently, stopped tasks appear in the returned results for at least one hour.
"
+ "documentation":"
Returns a list of tasks. You can filter the results by cluster, task definition family, container instance, launch type, what IAM principal started the task, or by the desired status of the task.
Recently stopped tasks might appear in the returned results.
For services using the rolling update (ECS) you can update the desired count, deployment configuration, network configuration, load balancers, service registries, enable ECS managed tags option, propagate tags option, task placement constraints and strategies, and task definition. When you update any of these parameters, Amazon ECS starts new tasks with the new configuration.
For services using the blue/green (CODE_DEPLOY) deployment controller, only the desired count, deployment configuration, health check grace period, task placement constraints and strategies, enable ECS managed tags option, and propagate tags can be updated using this API. If the network configuration, platform version, task definition, or load balancer need to be updated, create a new CodeDeploy deployment. For more information, see CreateDeployment in the CodeDeploy API Reference.
For services using an external deployment controller, you can update only the desired count, task placement constraints and strategies, health check grace period, enable ECS managed tags option, and propagate tags option, using this API. If the launch type, load balancer, network configuration, platform version, or task definition need to be updated, create a new task set For more information, see CreateTaskSet.
You can add to or subtract from the number of instantiations of a task definition in a service by specifying the cluster that the service is running in and a new desiredCount parameter.
If you have updated the Docker image of your application, you can create a new task definition with that image and deploy it to your service. The service scheduler uses the minimum healthy percent and maximum percent parameters (in the service's deployment configuration) to determine the deployment strategy.
If your updated Docker image uses the same tag as what is in the existing task definition for your service (for example, my_image:latest), you don't need to create a new revision of your task definition. You can update the service using the forceNewDeployment option. The new tasks launched by the deployment pull the current image/tag combination from your repository when they start.
You can also update the deployment configuration of a service. When a deployment is triggered by updating the task definition of a service, the service scheduler uses the deployment configuration parameters, minimumHealthyPercent and maximumPercent, to determine the deployment strategy.
If minimumHealthyPercent is below 100%, the scheduler can ignore desiredCount temporarily during a deployment. For example, if desiredCount is four tasks, a minimum of 50% allows the scheduler to stop two existing tasks before starting two new tasks. Tasks for services that don't use a load balancer are considered healthy if they're in the RUNNING state. Tasks for services that use a load balancer are considered healthy if they're in the RUNNING state and are reported as healthy by the load balancer.
The maximumPercent parameter represents an upper limit on the number of running tasks during a deployment. You can use it to define the deployment batch size. For example, if desiredCount is four tasks, a maximum of 200% starts four new tasks before stopping the four older tasks (provided that the cluster resources required to do this are available).
When UpdateService stops a task during a deployment, the equivalent of docker stop is issued to the containers running in the task. This results in a SIGTERM and a 30-second timeout. After this, SIGKILL is sent and the containers are forcibly stopped. If the container handles the SIGTERM gracefully and exits within 30 seconds from receiving it, no SIGKILL is sent.
When the service scheduler launches new tasks, it determines task placement in your cluster with the following logic.
Determine which of the container instances in your cluster can support your service's task definition. For example, they have the required CPU, memory, ports, and container instance attributes.
By default, the service scheduler attempts to balance tasks across Availability Zones in this manner even though you can choose a different placement strategy.
Sort the valid container instances by the fewest number of running tasks for this service in the same Availability Zone as the instance. For example, if zone A has one running service task and zones B and C each have zero, valid container instances in either zone B or C are considered optimal for placement.
Place the new service task on a valid container instance in an optimal Availability Zone (based on the previous steps), favoring container instances with the fewest number of running tasks for this service.
When the service scheduler stops running tasks, it attempts to maintain balance across the Availability Zones in your cluster using the following logic:
Sort the container instances by the largest number of running tasks for this service in the same Availability Zone as the instance. For example, if zone A has one running service task and zones B and C each have two, container instances in either zone B or C are considered optimal for termination.
Stop the task on a container instance in an optimal Availability Zone (based on the previous steps), favoring container instances with the largest number of running tasks for this service.
You must have a service-linked role when you update any of the following service properties. If you specified a custom role when you created the service, Amazon ECS automatically replaces the roleARN associated with the service with the ARN of your service-linked role. For more information, see Service-linked roles in the Amazon Elastic Container Service Developer Guide.
loadBalancers,
serviceRegistries
"
+ "documentation":"
Modifies the parameters of a service.
For services using the rolling update (ECS) you can update the desired count, deployment configuration, network configuration, load balancers, service registries, enable ECS managed tags option, propagate tags option, task placement constraints and strategies, and task definition. When you update any of these parameters, Amazon ECS starts new tasks with the new configuration.
For services using the blue/green (CODE_DEPLOY) deployment controller, only the desired count, deployment configuration, health check grace period, task placement constraints and strategies, enable ECS managed tags option, and propagate tags can be updated using this API. If the network configuration, platform version, task definition, or load balancer need to be updated, create a new CodeDeploy deployment. For more information, see CreateDeployment in the CodeDeploy API Reference.
For services using an external deployment controller, you can update only the desired count, task placement constraints and strategies, health check grace period, enable ECS managed tags option, and propagate tags option, using this API. If the launch type, load balancer, network configuration, platform version, or task definition need to be updated, create a new task set For more information, see CreateTaskSet.
You can add to or subtract from the number of instantiations of a task definition in a service by specifying the cluster that the service is running in and a new desiredCount parameter.
If you have updated the Docker image of your application, you can create a new task definition with that image and deploy it to your service. The service scheduler uses the minimum healthy percent and maximum percent parameters (in the service's deployment configuration) to determine the deployment strategy.
If your updated Docker image uses the same tag as what is in the existing task definition for your service (for example, my_image:latest), you don't need to create a new revision of your task definition. You can update the service using the forceNewDeployment option. The new tasks launched by the deployment pull the current image/tag combination from your repository when they start.
You can also update the deployment configuration of a service. When a deployment is triggered by updating the task definition of a service, the service scheduler uses the deployment configuration parameters, minimumHealthyPercent and maximumPercent, to determine the deployment strategy.
If minimumHealthyPercent is below 100%, the scheduler can ignore desiredCount temporarily during a deployment. For example, if desiredCount is four tasks, a minimum of 50% allows the scheduler to stop two existing tasks before starting two new tasks. Tasks for services that don't use a load balancer are considered healthy if they're in the RUNNING state. Tasks for services that use a load balancer are considered healthy if they're in the RUNNING state and are reported as healthy by the load balancer.
The maximumPercent parameter represents an upper limit on the number of running tasks during a deployment. You can use it to define the deployment batch size. For example, if desiredCount is four tasks, a maximum of 200% starts four new tasks before stopping the four older tasks (provided that the cluster resources required to do this are available).
When UpdateService stops a task during a deployment, the equivalent of docker stop is issued to the containers running in the task. This results in a SIGTERM and a 30-second timeout. After this, SIGKILL is sent and the containers are forcibly stopped. If the container handles the SIGTERM gracefully and exits within 30 seconds from receiving it, no SIGKILL is sent.
When the service scheduler launches new tasks, it determines task placement in your cluster with the following logic.
Determine which of the container instances in your cluster can support your service's task definition. For example, they have the required CPU, memory, ports, and container instance attributes.
By default, the service scheduler attempts to balance tasks across Availability Zones in this manner even though you can choose a different placement strategy.
Sort the valid container instances by the fewest number of running tasks for this service in the same Availability Zone as the instance. For example, if zone A has one running service task and zones B and C each have zero, valid container instances in either zone B or C are considered optimal for placement.
Place the new service task on a valid container instance in an optimal Availability Zone (based on the previous steps), favoring container instances with the fewest number of running tasks for this service.
When the service scheduler stops running tasks, it attempts to maintain balance across the Availability Zones in your cluster using the following logic:
Sort the container instances by the largest number of running tasks for this service in the same Availability Zone as the instance. For example, if zone A has one running service task and zones B and C each have two, container instances in either zone B or C are considered optimal for termination.
Stop the task on a container instance in an optimal Availability Zone (based on the previous steps), favoring container instances with the largest number of running tasks for this service.
You must have a service-linked role when you update any of the following service properties:
loadBalancers,
serviceRegistries
For more information about the role see the CreateService request parameter role.
Time duration (in seconds) to wait before giving up on resolving dependencies for a container. For example, you specify two containers in a task definition with containerA having a dependency on containerB reaching a COMPLETE, SUCCESS, or HEALTHY status. If a startTimeout value is specified for containerB and it doesn't reach the desired status within that time then containerA gives up and not start. This results in the task transitioning to a STOPPED state.
When the ECS_CONTAINER_START_TIMEOUT container agent configuration variable is used, it's enforced independently from this start timeout value.
For tasks using the Fargate launch type, the task or service requires the following platforms:
Linux platform version 1.3.0 or later.
Windows platform version 1.0.0 or later.
For tasks using the EC2 launch type, your container instances require at least version 1.26.0 of the container agent to use a container start timeout value. However, we recommend using the latest container agent version. For information about checking your agent version and updating to the latest version, see Updating the Amazon ECS Container Agent in the Amazon Elastic Container Service Developer Guide. If you're using an Amazon ECS-optimized Linux AMI, your instance needs at least version 1.26.0-1 of the ecs-init package. If your container instances are launched from version 20190301 or later, then they contain the required versions of the container agent and ecs-init. For more information, see Amazon ECS-optimized Linux AMI in the Amazon Elastic Container Service Developer Guide.
"
+ "documentation":"
Time duration (in seconds) to wait before giving up on resolving dependencies for a container. For example, you specify two containers in a task definition with containerA having a dependency on containerB reaching a COMPLETE, SUCCESS, or HEALTHY status. If a startTimeout value is specified for containerB and it doesn't reach the desired status within that time then containerA gives up and not start. This results in the task transitioning to a STOPPED state.
When the ECS_CONTAINER_START_TIMEOUT container agent configuration variable is used, it's enforced independently from this start timeout value.
For tasks using the Fargate launch type, the task or service requires the following platforms:
Linux platform version 1.3.0 or later.
Windows platform version 1.0.0 or later.
For tasks using the EC2 launch type, your container instances require at least version 1.26.0 of the container agent to use a container start timeout value. However, we recommend using the latest container agent version. For information about checking your agent version and updating to the latest version, see Updating the Amazon ECS Container Agent in the Amazon Elastic Container Service Developer Guide. If you're using an Amazon ECS-optimized Linux AMI, your instance needs at least version 1.26.0-1 of the ecs-init package. If your container instances are launched from version 20190301 or later, then they contain the required versions of the container agent and ecs-init. For more information, see Amazon ECS-optimized Linux AMI in the Amazon Elastic Container Service Developer Guide.
Time duration (in seconds) to wait before the container is forcefully killed if it doesn't exit normally on its own.
For tasks using the Fargate launch type, the task or service requires the following platforms:
Linux platform version 1.3.0 or later.
Windows platform version 1.0.0 or later.
The max stop timeout value is 120 seconds and if the parameter is not specified, the default value of 30 seconds is used.
For tasks that use the EC2 launch type, if the stopTimeout parameter isn't specified, the value set for the Amazon ECS container agent configuration variable ECS_CONTAINER_STOP_TIMEOUT is used. If neither the stopTimeout parameter or the ECS_CONTAINER_STOP_TIMEOUT agent configuration variable are set, then the default values of 30 seconds for Linux containers and 30 seconds on Windows containers are used. Your container instances require at least version 1.26.0 of the container agent to use a container stop timeout value. However, we recommend using the latest container agent version. For information about checking your agent version and updating to the latest version, see Updating the Amazon ECS Container Agent in the Amazon Elastic Container Service Developer Guide. If you're using an Amazon ECS-optimized Linux AMI, your instance needs at least version 1.26.0-1 of the ecs-init package. If your container instances are launched from version 20190301 or later, then they contain the required versions of the container agent and ecs-init. For more information, see Amazon ECS-optimized Linux AMI in the Amazon Elastic Container Service Developer Guide.
"
+ "documentation":"
Time duration (in seconds) to wait before the container is forcefully killed if it doesn't exit normally on its own.
For tasks using the Fargate launch type, the task or service requires the following platforms:
Linux platform version 1.3.0 or later.
Windows platform version 1.0.0 or later.
The max stop timeout value is 120 seconds and if the parameter is not specified, the default value of 30 seconds is used.
For tasks that use the EC2 launch type, if the stopTimeout parameter isn't specified, the value set for the Amazon ECS container agent configuration variable ECS_CONTAINER_STOP_TIMEOUT is used. If neither the stopTimeout parameter or the ECS_CONTAINER_STOP_TIMEOUT agent configuration variable are set, then the default values of 30 seconds for Linux containers and 30 seconds on Windows containers are used. Your container instances require at least version 1.26.0 of the container agent to use a container stop timeout value. However, we recommend using the latest container agent version. For information about checking your agent version and updating to the latest version, see Updating the Amazon ECS Container Agent in the Amazon Elastic Container Service Developer Guide. If you're using an Amazon ECS-optimized Linux AMI, your instance needs at least version 1.26.0-1 of the ecs-init package. If your container instances are launched from version 20190301 or later, then they contain the required versions of the container agent and ecs-init. For more information, see Amazon ECS-optimized Linux AMI in the Amazon Elastic Container Service Developer Guide.
A list of ARNs in SSM or Amazon S3 to a credential spec (credspeccode>) file that configures a container for Active Directory authentication. This parameter is only used with domainless authentication.
The format for each ARN is credentialspecdomainless:MyARN. Replace MyARN with the ARN in SSM or Amazon S3.
The credspec must provide a ARN in Secrets Manager for a secret containing the username, password, and the domain to connect to. For better security, the instance isn't joined to the domain for domainless authentication. Other applications on the instance can't use the domainless credentials. You can use this parameter to run tasks on the same instance, even it the tasks need to join different domains. For more information, see Using gMSAs for Windows Containers and Using gMSAs for Linux Containers.
"
+ "documentation":"
A list of ARNs in SSM or Amazon S3 to a credential spec (CredSpec) file that configures the container for Active Directory authentication. We recommend that you use this parameter instead of the dockerSecurityOptions. The maximum number of ARNs is 1.
There are two formats for each ARN.
credentialspecdomainless:MyARN
You use credentialspecdomainless:MyARN to provide a CredSpec with an additional section for a secret in Secrets Manager. You provide the login credentials to the domain in the secret.
Each task that runs on any container instance can join different domains.
You can use this format without joining the container instance to a domain.
credentialspec:MyARN
You use credentialspec:MyARN to provide a CredSpec for a single domain.
You must join the container instance to the domain before you start any tasks that use this task definition.
In both formats, replace MyARN with the ARN in SSM or Amazon S3.
If you provide a credentialspecdomainless:MyARN, the credspec must provide a ARN in Secrets Manager for a secret containing the username, password, and the domain to connect to. For better security, the instance isn't joined to the domain for domainless authentication. Other applications on the instance can't use the domainless credentials. You can use this parameter to run tasks on the same instance, even it the tasks need to join different domains. For more information, see Using gMSAs for Windows Containers and Using gMSAs for Linux Containers.
"
}
},
"documentation":"
Container definitions are used in task definitions to describe the different containers that are launched as part of a task.
"
@@ -1870,7 +1870,7 @@
"documentation":"
The type and amount of a resource to assign to a container, instead of the default value from the task definition. The only supported resource is a GPU.
"
}
},
- "documentation":"
The overrides that are sent to a container. An empty container override can be passed in. An example of an empty container override is {\"containerOverrides\": [ ] }. If a non-empty container override is specified, the name parameter must be included.
"
+ "documentation":"
The overrides that are sent to a container. An empty container override can be passed in. An example of an empty container override is {\"containerOverrides\": [ ] }. If a non-empty container override is specified, the name parameter must be included.
You can use Secrets Manager or Amazon Web Services Systems Manager Parameter Store to store the sensitive data. For more information, see Retrieve secrets through environment variables in the Amazon ECS Developer Guide.
Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is on, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
"
}
},
- "documentation":"
The deployment circuit breaker can only be used for services using the rolling update (ECS) deployment type.
The deployment circuit breaker determines whether a service deployment will fail if the service can't reach a steady state. If it is turned on, a service deployment will transition to a failed state and stop launching new tasks. You can also configure Amazon ECS to roll back your service to the last completed deployment after a failure. For more information, see Rolling update in the Amazon Elastic Container Service Developer Guide.
"
+ "documentation":"
The deployment circuit breaker can only be used for services using the rolling update (ECS) deployment type.
The deployment circuit breaker determines whether a service deployment will fail if the service can't reach a steady state. If it is turned on, a service deployment will transition to a failed state and stop launching new tasks. You can also configure Amazon ECS to roll back your service to the last completed deployment after a failure. For more information, see Rolling update in the Amazon Elastic Container Service Developer Guide.
For more information about API failure reasons, see API failure reasons in the Amazon Elastic Container Service Developer Guide.
The file type to use. The only supported value is s3.
"
}
},
- "documentation":"
A list of files containing the environment variables to pass to a container. You can specify up to ten environment files. The file must have a .env file extension. Each line in an environment file should contain an environment variable in VARIABLE=VALUE format. Lines beginning with # are treated as comments and are ignored. For more information about the environment variable file syntax, see Declare default environment variables in file.
If there are environment variables specified using the environment parameter in a container definition, they take precedence over the variables contained within an environment file. If multiple environment files are specified that contain the same variable, they're processed from the top down. We recommend that you use unique variable names. For more information, see Specifying environment variables in the Amazon Elastic Container Service Developer Guide.
This parameter is only supported for tasks hosted on Fargate using the following platform versions:
Linux platform version 1.4.0 or later.
Windows platform version 1.0.0 or later.
"
+ "documentation":"
A list of files containing the environment variables to pass to a container. You can specify up to ten environment files. The file must have a .env file extension. Each line in an environment file should contain an environment variable in VARIABLE=VALUE format. Lines beginning with # are treated as comments and are ignored. For more information about the environment variable file syntax, see Declare default environment variables in file.
If there are environment variables specified using the environment parameter in a container definition, they take precedence over the variables contained within an environment file. If multiple environment files are specified that contain the same variable, they're processed from the top down. We recommend that you use unique variable names. For more information, see Specifying environment variables in the Amazon Elastic Container Service Developer Guide.
You must use the following platforms for the Fargate launch type:
The full Amazon Resource Name (ARN) of the Elastic Load Balancing target group or groups associated with a service or task set.
A target group ARN is only specified when using an Application Load Balancer or Network Load Balancer. If you're using a Classic Load Balancer, omit the target group ARN.
For services using the ECS deployment controller, you can specify one or multiple target groups. For more information, see Registering multiple target groups with a service in the Amazon Elastic Container Service Developer Guide.
For services using the CODE_DEPLOY deployment controller, you're required to define two target groups for the load balancer. For more information, see Blue/green deployment with CodeDeploy in the Amazon Elastic Container Service Developer Guide.
If your service's task definition uses the awsvpc network mode, you must choose ip as the target type, not instance. Do this when creating your target groups because tasks that use the awsvpc network mode are associated with an elastic network interface, not an Amazon EC2 instance. This network mode is required for the Fargate launch type.
"
+ "documentation":"
The full Amazon Resource Name (ARN) of the Elastic Load Balancing target group or groups associated with a service or task set.
A target group ARN is only specified when using an Application Load Balancer or Network Load Balancer.
For services using the ECS deployment controller, you can specify one or multiple target groups. For more information, see Registering multiple target groups with a service in the Amazon Elastic Container Service Developer Guide.
For services using the CODE_DEPLOY deployment controller, you're required to define two target groups for the load balancer. For more information, see Blue/green deployment with CodeDeploy in the Amazon Elastic Container Service Developer Guide.
If your service's task definition uses the awsvpc network mode, you must choose ip as the target type, not instance. Do this when creating your target groups because tasks that use the awsvpc network mode are associated with an elastic network interface, not an Amazon EC2 instance. This network mode is required for the Fargate launch type.
The name of the load balancer to associate with the Amazon ECS service or task set.
A load balancer name is only specified when using a Classic Load Balancer. If you are using an Application Load Balancer or a Network Load Balancer the load balancer name parameter should be omitted.
"
+ "documentation":"
The name of the load balancer to associate with the Amazon ECS service or task set.
If you are using an Application Load Balancer or a Network Load Balancer the load balancer name parameter should be omitted.
The secrets to pass to the log configuration. For more information, see Specifying sensitive data in the Amazon Elastic Container Service Developer Guide.
By default, containers use the same logging driver that the Docker daemon uses. However, the container might use a different logging driver than the Docker daemon by specifying a log driver configuration in the container definition. For more information about the options for different supported log drivers, see Configure logging drivers in the Docker documentation.
Understand the following when specifying a log configuration for your containers.
Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon (shown in the valid values below). Additional log drivers may be available in future releases of the Amazon ECS container agent.
This parameter requires version 1.18 of the Docker Remote API or greater on your container instance.
For tasks that are hosted on Amazon EC2 instances, the Amazon ECS container agent must register the available logging drivers with the ECS_AVAILABLE_LOGGING_DRIVERS environment variable before containers placed on that instance can use these log configuration options. For more information, see Amazon ECS container agent configuration in the Amazon Elastic Container Service Developer Guide.
For tasks that are on Fargate, because you don't have access to the underlying infrastructure your tasks are hosted on, any additional software needed must be installed outside of the task. For example, the Fluentd output aggregators or a remote host running Logstash to send Gelf logs to.
By default, containers use the same logging driver that the Docker daemon uses. However, the container might use a different logging driver than the Docker daemon by specifying a log driver configuration in the container definition. For more information about the options for different supported log drivers, see Configure logging drivers in the Docker documentation.
Understand the following when specifying a log configuration for your containers.
Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on Fargate, the supported log drivers are awslogs, splunk, and awsfirelens.
For tasks hosted on Amazon EC2 instances, the supported log drivers are awslogs, fluentd, gelf, json-file, journald, logentries,syslog, splunk, and awsfirelens.
This parameter requires version 1.18 of the Docker Remote API or greater on your container instance.
For tasks that are hosted on Amazon EC2 instances, the Amazon ECS container agent must register the available logging drivers with the ECS_AVAILABLE_LOGGING_DRIVERS environment variable before containers placed on that instance can use these log configuration options. For more information, see Amazon ECS container agent configuration in the Amazon Elastic Container Service Developer Guide.
For tasks that are on Fargate, because you don't have access to the underlying infrastructure your tasks are hosted on, any additional software needed must be installed outside of the task. For example, the Fluentd output aggregators or a remote host running Logstash to send Gelf logs to.
The maximum number of Amazon EC2 instances that Amazon ECS will scale out at one time. The scale in process is not affected by this parameter. If this parameter is omitted, the default value of 1 is used.
"
+ "documentation":"
The maximum number of Amazon EC2 instances that Amazon ECS will scale out at one time. The scale in process is not affected by this parameter. If this parameter is omitted, the default value of 10000 is used.
The port number on the container instance to reserve for your container.
If you specify a containerPortRange, leave this field empty and the value of the hostPort is set as follows:
For containers in a task with the awsvpc network mode, the hostPort is set to the same value as the containerPort. This is a static mapping strategy.
For containers in a task with the bridge network mode, the Amazon ECS agent finds open ports on the host and automatically binds them to the container ports. This is a dynamic mapping strategy.
If you use containers in a task with the awsvpc or host network mode, the hostPort can either be left blank or set to the same value as the containerPort.
If you use containers in a task with the bridge network mode, you can specify a non-reserved host port for your container port mapping, or you can omit the hostPort (or set it to 0) while specifying a containerPort and your container automatically receives a port in the ephemeral port range for your container instance operating system and Docker version.
The default ephemeral port range for Docker version 1.6.0 and later is listed on the instance under /proc/sys/net/ipv4/ip_local_port_range. If this kernel parameter is unavailable, the default ephemeral port range from 49153 through 65535 is used. Do not attempt to specify a host port in the ephemeral port range as these are reserved for automatic assignment. In general, ports below 32768 are outside of the ephemeral port range.
The default reserved ports are 22 for SSH, the Docker ports 2375 and 2376, and the Amazon ECS container agent ports 51678-51680. Any host port that was previously specified in a running task is also reserved while the task is running. That is, after a task stops, the host port is released. The current reserved ports are displayed in the remainingResources of DescribeContainerInstances output. A container instance can have up to 100 reserved ports at a time. This number includes the default reserved ports. Automatically assigned ports aren't included in the 100 reserved ports quota.
"
+ "documentation":"
The port number on the container instance to reserve for your container.
If you specify a containerPortRange, leave this field empty and the value of the hostPort is set as follows:
For containers in a task with the awsvpc network mode, the hostPort is set to the same value as the containerPort. This is a static mapping strategy.
For containers in a task with the bridge network mode, the Amazon ECS agent finds open ports on the host and automatically binds them to the container ports. This is a dynamic mapping strategy.
If you use containers in a task with the awsvpc or host network mode, the hostPort can either be left blank or set to the same value as the containerPort.
If you use containers in a task with the bridge network mode, you can specify a non-reserved host port for your container port mapping, or you can omit the hostPort (or set it to 0) while specifying a containerPort and your container automatically receives a port in the ephemeral port range for your container instance operating system and Docker version.
The default ephemeral port range for Docker version 1.6.0 and later is listed on the instance under /proc/sys/net/ipv4/ip_local_port_range. If this kernel parameter is unavailable, the default ephemeral port range from 49153 through 65535 (Linux) or 49152 through 65535 (Windows) is used. Do not attempt to specify a host port in the ephemeral port range as these are reserved for automatic assignment. In general, ports below 32768 are outside of the ephemeral port range.
The default reserved ports are 22 for SSH, the Docker ports 2375 and 2376, and the Amazon ECS container agent ports 51678-51680. Any host port that was previously specified in a running task is also reserved while the task is running. That is, after a task stops, the host port is released. The current reserved ports are displayed in the remainingResources of DescribeContainerInstances output. A container instance can have up to 100 reserved ports at a time. This number includes the default reserved ports. Automatically assigned ports aren't included in the 100 reserved ports quota.
The Unix timestamp for the time when the task stops. More specifically, it's for the time when the task transitions from the RUNNING state to STOPPED.
"
+ "documentation":"
The Unix timestamp for the time when the task stops. More specifically, it's for the time when the task transitions from the RUNNING state to STOPPING.
The task launch types the task definition was validated against. For more information, see Amazon ECS launch types in the Amazon Elastic Container Service Developer Guide.
"
+ "documentation":"
The task launch types the task definition was validated against. The valid values are EC2, FARGATE, and EXTERNAL. For more information, see Amazon ECS launch types in the Amazon Elastic Container Service Developer Guide.
The name of the volume. Up to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed. This name is referenced in the sourceVolume parameter of container definition mountPoints.
"
+ "documentation":"
The name of the volume. Up to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed. This name is referenced in the sourceVolume parameter of container definition mountPoints.
This is required wwhen you use an Amazon EFS volume.
"
},
"host":{
"shape":"HostVolumeProperties",
diff --git a/services/efs/pom.xml b/services/efs/pom.xml
index de91d9cbcb0..f55f3d96d77 100644
--- a/services/efs/pom.xml
+++ b/services/efs/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTefsAWS Java SDK :: Services :: Amazon Elastic File System
diff --git a/services/eks/pom.xml b/services/eks/pom.xml
index 1d7e19bafd7..ad9730b620f 100644
--- a/services/eks/pom.xml
+++ b/services/eks/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTeksAWS Java SDK :: Services :: EKS
diff --git a/services/eks/src/main/resources/codegen-resources/endpoint-rule-set.json b/services/eks/src/main/resources/codegen-resources/endpoint-rule-set.json
index 4b665084edc..4d1971250c5 100644
--- a/services/eks/src/main/resources/codegen-resources/endpoint-rule-set.json
+++ b/services/eks/src/main/resources/codegen-resources/endpoint-rule-set.json
@@ -3,7 +3,7 @@
"parameters": {
"Region": {
"builtIn": "AWS::Region",
- "required": true,
+ "required": false,
"documentation": "The AWS region used to dispatch the request.",
"type": "String"
},
@@ -32,13 +32,12 @@
{
"conditions": [
{
- "fn": "aws.partition",
+ "fn": "isSet",
"argv": [
{
- "ref": "Region"
+ "ref": "Endpoint"
}
- ],
- "assign": "PartitionResult"
+ ]
}
],
"type": "tree",
@@ -46,14 +45,20 @@
{
"conditions": [
{
- "fn": "isSet",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "Endpoint"
- }
+ "ref": "UseFIPS"
+ },
+ true
]
}
],
+ "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "type": "error"
+ },
+ {
+ "conditions": [],
"type": "tree",
"rules": [
{
@@ -62,67 +67,42 @@
"fn": "booleanEquals",
"argv": [
{
- "ref": "UseFIPS"
+ "ref": "UseDualStack"
},
true
]
}
],
- "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
"type": "error"
},
{
"conditions": [],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
- },
- true
- ]
- }
- ],
- "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
- "type": "error"
+ "endpoint": {
+ "url": {
+ "ref": "Endpoint"
},
- {
- "conditions": [],
- "endpoint": {
- "url": {
- "ref": "Endpoint"
- },
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
}
]
- },
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
{
"conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
- },
- true
- ]
- },
- {
- "fn": "booleanEquals",
+ "fn": "isSet",
"argv": [
{
- "ref": "UseDualStack"
- },
- true
+ "ref": "Region"
+ }
]
}
],
@@ -131,224 +111,288 @@
{
"conditions": [
{
- "fn": "booleanEquals",
+ "fn": "aws.partition",
"argv": [
- true,
{
- "fn": "getAttr",
- "argv": [
- {
- "ref": "PartitionResult"
- },
- "supportsFIPS"
- ]
+ "ref": "Region"
}
- ]
- },
+ ],
+ "assign": "PartitionResult"
+ }
+ ],
+ "type": "tree",
+ "rules": [
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseFIPS"
},
- "supportsDualStack"
+ true
]
- }
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
- "endpoint": {
- "url": "https://eks-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
- },
- {
- "conditions": [],
- "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
- },
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- true,
+ },
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseDualStack"
},
- "supportsFIPS"
+ true
]
}
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
+ ],
"type": "tree",
"rules": [
{
"conditions": [
{
- "fn": "stringEquals",
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsFIPS"
+ ]
+ }
+ ]
+ },
+ {
+ "fn": "booleanEquals",
"argv": [
- "aws",
+ true,
{
"fn": "getAttr",
"argv": [
{
"ref": "PartitionResult"
},
- "name"
+ "supportsDualStack"
]
}
]
}
],
- "endpoint": {
- "url": "https://fips.eks.{Region}.{PartitionResult#dnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://eks-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
},
+ {
+ "conditions": [],
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
+ "type": "error"
+ }
+ ]
+ },
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ {
+ "ref": "UseFIPS"
+ },
+ true
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
{
"conditions": [
{
- "fn": "stringEquals",
+ "fn": "booleanEquals",
"argv": [
- "aws-us-gov",
+ true,
{
"fn": "getAttr",
"argv": [
{
"ref": "PartitionResult"
},
- "name"
+ "supportsFIPS"
]
}
]
}
],
- "endpoint": {
- "url": "https://eks.{Region}.{PartitionResult#dnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "stringEquals",
+ "argv": [
+ "aws",
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "name"
+ ]
+ }
+ ]
+ }
+ ],
+ "endpoint": {
+ "url": "https://fips.eks.{Region}.amazonaws.com",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ },
+ {
+ "conditions": [
+ {
+ "fn": "stringEquals",
+ "argv": [
+ "aws-us-gov",
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "name"
+ ]
+ }
+ ]
+ }
+ ],
+ "endpoint": {
+ "url": "https://eks.{Region}.amazonaws.com",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ },
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://eks-fips.{Region}.{PartitionResult#dnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
},
{
"conditions": [],
- "endpoint": {
- "url": "https://eks-fips.{Region}.{PartitionResult#dnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "error": "FIPS is enabled but this partition does not support FIPS",
+ "type": "error"
}
]
- }
- ]
- },
- {
- "conditions": [],
- "error": "FIPS is enabled but this partition does not support FIPS",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
},
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseDualStack"
},
- "supportsDualStack"
+ true
]
}
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://eks.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "error": "DualStack is enabled but this partition does not support DualStack",
+ "type": "error"
+ }
]
- }
- ],
- "type": "tree",
- "rules": [
+ },
{
"conditions": [],
- "endpoint": {
- "url": "https://eks.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://eks.{Region}.{PartitionResult#dnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
}
]
- },
- {
- "conditions": [],
- "error": "DualStack is enabled but this partition does not support DualStack",
- "type": "error"
}
]
},
{
"conditions": [],
- "endpoint": {
- "url": "https://eks.{Region}.{PartitionResult#dnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "error": "Invalid Configuration: Missing Region",
+ "type": "error"
}
]
}
diff --git a/services/eks/src/main/resources/codegen-resources/endpoint-tests.json b/services/eks/src/main/resources/codegen-resources/endpoint-tests.json
index 8cc05949df5..0355c7dcc07 100644
--- a/services/eks/src/main/resources/codegen-resources/endpoint-tests.json
+++ b/services/eks/src/main/resources/codegen-resources/endpoint-tests.json
@@ -1,980 +1,31 @@
{
"testCases": [
{
- "documentation": "For region ap-south-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.ap-south-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-south-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ap-south-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.ap-south-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-south-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ap-south-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.ap-south-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ap-south-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ap-south-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.ap-south-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ap-south-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.ap-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.ap-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.ap-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ap-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.ap-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ap-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.eu-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.eu-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-south-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.eu-south-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-south-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-south-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.eu-south-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-south-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-south-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-south-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-south-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-south-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-south-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-south-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.us-gov-east-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-gov-east-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.us-gov-east-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-gov-east-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.us-gov-east-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "us-gov-east-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.us-gov-east-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "us-gov-east-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region me-central-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.me-central-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "me-central-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region me-central-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.me-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "me-central-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region me-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.me-central-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "me-central-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region me-central-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.me-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "me-central-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.ca-central-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ca-central-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.ca-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ca-central-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.ca-central-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ca-central-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.ca-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ca-central-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.eu-central-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-central-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.eu-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-central-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-central-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-central-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-central-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-central-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.eu-central-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-central-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-central-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.eu-central-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-central-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-central-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-central-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-central-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-central-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-central-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-central-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.us-west-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-west-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.us-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-west-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.us-west-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "us-west-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.us-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "us-west-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.us-west-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-west-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.us-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-west-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.us-west-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "us-west-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.us-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "us-west-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.af-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "af-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.af-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "af-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.af-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "af-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.af-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "af-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.eu-north-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-north-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.eu-north-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-north-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-north-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-north-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-north-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-north-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.eu-west-3.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-west-3",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.eu-west-3.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-west-3",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-west-3.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-west-3",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-west-3.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-west-3",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.eu-west-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-west-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.eu-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-west-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-west-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-west-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-west-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.eu-west-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-west-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.eu-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "eu-west-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-west-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-west-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.eu-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "eu-west-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.ap-northeast-3.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-northeast-3",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.ap-northeast-3.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-northeast-3",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.ap-northeast-3.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ap-northeast-3",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack disabled",
+ "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.ap-northeast-3.amazonaws.com"
+ "url": "https://eks.af-south-1.amazonaws.com"
}
},
"params": {
+ "Region": "af-south-1",
"UseFIPS": false,
- "Region": "ap-northeast-3",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.ap-northeast-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-northeast-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.ap-northeast-2.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-northeast-2",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.ap-northeast-2.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ap-northeast-2",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.ap-northeast-2.amazonaws.com"
+ "url": "https://eks.ap-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "ap-east-1",
"UseFIPS": false,
- "Region": "ap-northeast-2",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.ap-northeast-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-northeast-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.ap-northeast-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-northeast-1",
"UseDualStack": false
}
},
- {
- "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.ap-northeast-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "ap-northeast-1",
- "UseDualStack": true
- }
- },
{
"documentation": "For region ap-northeast-1 with FIPS disabled and DualStack disabled",
"expect": {
@@ -983,680 +34,537 @@
}
},
"params": {
- "UseFIPS": false,
"Region": "ap-northeast-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region me-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.me-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "me-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region me-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.me-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "me-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region me-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.me-south-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "me-south-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.me-south-1.amazonaws.com"
- }
- },
- "params": {
"UseFIPS": false,
- "Region": "me-south-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region sa-east-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.sa-east-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "sa-east-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region sa-east-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.sa-east-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "sa-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region sa-east-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks.sa-east-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": false,
- "Region": "sa-east-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.sa-east-1.amazonaws.com"
+ "url": "https://eks.ap-northeast-2.amazonaws.com"
}
},
"params": {
+ "Region": "ap-northeast-2",
"UseFIPS": false,
- "Region": "sa-east-1",
- "UseDualStack": false
- }
- },
- {
- "documentation": "For region ap-east-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.ap-east-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-east-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region ap-east-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://fips.eks.ap-east-1.amazonaws.com"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "ap-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.ap-east-1.api.aws"
+ "url": "https://eks.ap-northeast-3.amazonaws.com"
}
},
"params": {
+ "Region": "ap-northeast-3",
"UseFIPS": false,
- "Region": "ap-east-1",
- "UseDualStack": true
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.ap-east-1.amazonaws.com"
+ "url": "https://eks.ap-south-1.amazonaws.com"
}
},
"params": {
+ "Region": "ap-south-1",
"UseFIPS": false,
- "Region": "ap-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.cn-north-1.api.amazonwebservices.com.cn"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "cn-north-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks-fips.cn-north-1.amazonaws.com.cn"
+ "url": "https://eks.ap-southeast-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "cn-north-1",
+ "Region": "ap-southeast-1",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.cn-north-1.api.amazonwebservices.com.cn"
+ "url": "https://eks.ap-southeast-2.amazonaws.com"
}
},
"params": {
+ "Region": "ap-southeast-2",
"UseFIPS": false,
- "Region": "cn-north-1",
- "UseDualStack": true
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-southeast-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.cn-north-1.amazonaws.com.cn"
+ "url": "https://eks.ap-southeast-3.amazonaws.com"
}
},
"params": {
+ "Region": "ap-southeast-3",
"UseFIPS": false,
- "Region": "cn-north-1",
"UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://eks-fips.us-gov-west-1.api.aws"
- }
- },
- "params": {
- "UseFIPS": true,
- "Region": "us-gov-west-1",
- "UseDualStack": true
- }
- },
- {
- "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.us-gov-west-1.amazonaws.com"
+ "url": "https://eks.ca-central-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "us-gov-west-1",
+ "Region": "ca-central-1",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.us-gov-west-1.api.aws"
+ "url": "https://eks.eu-central-1.amazonaws.com"
}
},
"params": {
+ "Region": "eu-central-1",
"UseFIPS": false,
- "Region": "us-gov-west-1",
- "UseDualStack": true
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.us-gov-west-1.amazonaws.com"
+ "url": "https://eks.eu-north-1.amazonaws.com"
}
},
"params": {
+ "Region": "eu-north-1",
"UseFIPS": false,
- "Region": "us-gov-west-1",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region eu-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks-fips.ap-southeast-1.api.aws"
+ "url": "https://eks.eu-south-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-southeast-1",
- "UseDualStack": true
+ "Region": "eu-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://fips.eks.ap-southeast-1.amazonaws.com"
+ "url": "https://eks.eu-west-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-southeast-1",
+ "Region": "eu-west-1",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.ap-southeast-1.api.aws"
+ "url": "https://eks.eu-west-2.amazonaws.com"
}
},
"params": {
+ "Region": "eu-west-2",
"UseFIPS": false,
- "Region": "ap-southeast-1",
- "UseDualStack": true
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.ap-southeast-1.amazonaws.com"
+ "url": "https://eks.eu-west-3.amazonaws.com"
}
},
"params": {
+ "Region": "eu-west-3",
"UseFIPS": false,
- "Region": "ap-southeast-1",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks-fips.ap-southeast-2.api.aws"
+ "url": "https://eks.me-south-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-southeast-2",
- "UseDualStack": true
+ "Region": "me-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://fips.eks.ap-southeast-2.amazonaws.com"
+ "url": "https://eks.sa-east-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-southeast-2",
+ "Region": "sa-east-1",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.ap-southeast-2.api.aws"
+ "url": "https://eks.us-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": false,
- "Region": "ap-southeast-2",
- "UseDualStack": true
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.ap-southeast-2.amazonaws.com"
+ "url": "https://fips.eks.us-east-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": false,
- "Region": "ap-southeast-2",
+ "Region": "us-east-1",
+ "UseFIPS": true,
"UseDualStack": false
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
"expect": {
- "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
+ "endpoint": {
+ "url": "https://eks.us-east-2.amazonaws.com"
+ }
},
"params": {
- "UseFIPS": true,
- "Region": "us-iso-east-1",
- "UseDualStack": true
+ "Region": "us-east-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks-fips.us-iso-east-1.c2s.ic.gov"
+ "url": "https://fips.eks.us-east-2.amazonaws.com"
}
},
"params": {
+ "Region": "us-east-2",
"UseFIPS": true,
- "Region": "us-iso-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-west-1 with FIPS disabled and DualStack disabled",
"expect": {
- "error": "DualStack is enabled but this partition does not support DualStack"
+ "endpoint": {
+ "url": "https://eks.us-west-1.amazonaws.com"
+ }
},
"params": {
+ "Region": "us-west-1",
"UseFIPS": false,
- "Region": "us-iso-east-1",
- "UseDualStack": true
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-west-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.us-iso-east-1.c2s.ic.gov"
+ "url": "https://fips.eks.us-west-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": false,
- "Region": "us-iso-east-1",
+ "Region": "us-west-1",
+ "UseFIPS": true,
"UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks-fips.ap-southeast-3.api.aws"
+ "url": "https://eks.us-west-2.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-southeast-3",
- "UseDualStack": true
+ "Region": "us-west-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-west-2 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://fips.eks.ap-southeast-3.amazonaws.com"
+ "url": "https://fips.eks.us-west-2.amazonaws.com"
}
},
"params": {
+ "Region": "us-west-2",
"UseFIPS": true,
- "Region": "ap-southeast-3",
"UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://eks.ap-southeast-3.api.aws"
+ "url": "https://eks-fips.us-east-1.api.aws"
}
},
"params": {
- "UseFIPS": false,
- "Region": "ap-southeast-3",
+ "Region": "us-east-1",
+ "UseFIPS": true,
"UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://eks.ap-southeast-3.amazonaws.com"
+ "url": "https://eks.us-east-1.api.aws"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": false,
- "Region": "ap-southeast-3",
- "UseDualStack": false
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-4 with FIPS enabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks-fips.ap-southeast-4.api.aws"
+ "url": "https://eks.cn-north-1.amazonaws.com.cn"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-southeast-4",
- "UseDualStack": true
+ "Region": "cn-north-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-4 with FIPS enabled and DualStack disabled",
+ "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://fips.eks.ap-southeast-4.amazonaws.com"
+ "url": "https://eks.cn-northwest-1.amazonaws.com.cn"
}
},
"params": {
- "UseFIPS": true,
- "Region": "ap-southeast-4",
+ "Region": "cn-northwest-1",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-4 with FIPS disabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://eks.ap-southeast-4.api.aws"
+ "url": "https://eks-fips.cn-north-1.api.amazonwebservices.com.cn"
}
},
"params": {
- "UseFIPS": false,
- "Region": "ap-southeast-4",
+ "Region": "cn-north-1",
+ "UseFIPS": true,
"UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-4 with FIPS disabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.ap-southeast-4.amazonaws.com"
+ "url": "https://eks-fips.cn-north-1.amazonaws.com.cn"
}
},
"params": {
- "UseFIPS": false,
- "Region": "ap-southeast-4",
+ "Region": "cn-north-1",
+ "UseFIPS": true,
"UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://eks-fips.us-east-1.api.aws"
+ "url": "https://eks.cn-north-1.api.amazonwebservices.com.cn"
}
},
"params": {
- "UseFIPS": true,
- "Region": "us-east-1",
+ "Region": "cn-north-1",
+ "UseFIPS": false,
"UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://fips.eks.us-east-1.amazonaws.com"
+ "url": "https://eks.us-gov-east-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": true,
- "Region": "us-east-1",
+ "Region": "us-gov-east-1",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.us-east-1.api.aws"
+ "url": "https://eks.us-gov-east-1.amazonaws.com"
}
},
"params": {
- "UseFIPS": false,
- "Region": "us-east-1",
- "UseDualStack": true
+ "Region": "us-gov-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.us-east-1.amazonaws.com"
+ "url": "https://eks.us-gov-west-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-gov-west-1",
"UseFIPS": false,
- "Region": "us-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks-fips.us-east-2.api.aws"
+ "url": "https://eks.us-gov-west-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-gov-west-1",
"UseFIPS": true,
- "Region": "us-east-2",
- "UseDualStack": true
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://fips.eks.us-east-2.amazonaws.com"
+ "url": "https://eks-fips.us-gov-east-1.api.aws"
}
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": true,
- "Region": "us-east-2",
- "UseDualStack": false
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://eks.us-east-2.api.aws"
+ "url": "https://eks.us-gov-east-1.api.aws"
}
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": false,
- "Region": "us-east-2",
"UseDualStack": true
}
},
{
- "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.us-east-2.amazonaws.com"
+ "url": "https://eks.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": false,
- "Region": "us-east-2",
"UseDualStack": false
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://eks-fips.cn-northwest-1.api.amazonwebservices.com.cn"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": true,
- "Region": "cn-northwest-1",
"UseDualStack": true
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks-fips.cn-northwest-1.amazonaws.com.cn"
+ "url": "https://eks-fips.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": true,
- "Region": "cn-northwest-1",
"UseDualStack": false
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://eks.cn-northwest-1.api.amazonwebservices.com.cn"
- }
+ "error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": false,
- "Region": "cn-northwest-1",
"UseDualStack": true
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.cn-northwest-1.amazonaws.com.cn"
+ "url": "https://eks.us-isob-east-1.sc2s.sgov.gov"
}
},
"params": {
+ "Region": "us-isob-east-1",
"UseFIPS": false,
- "Region": "cn-northwest-1",
"UseDualStack": false
}
},
@@ -1666,8 +574,8 @@
"error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
- "UseFIPS": true,
"Region": "us-isob-east-1",
+ "UseFIPS": true,
"UseDualStack": true
}
},
@@ -1679,8 +587,8 @@
}
},
"params": {
- "UseFIPS": true,
"Region": "us-isob-east-1",
+ "UseFIPS": true,
"UseDualStack": false
}
},
@@ -1690,26 +598,27 @@
"error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
- "UseFIPS": false,
"Region": "us-isob-east-1",
+ "UseFIPS": false,
"UseDualStack": true
}
},
{
- "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For custom endpoint with region set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
- "url": "https://eks.us-isob-east-1.sc2s.sgov.gov"
+ "url": "https://example.com"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": false,
- "Region": "us-isob-east-1",
- "UseDualStack": false
+ "UseDualStack": false,
+ "Endpoint": "https://example.com"
}
},
{
- "documentation": "For custom endpoint with fips disabled and dualstack disabled",
+ "documentation": "For custom endpoint with region not set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
"url": "https://example.com"
@@ -1717,7 +626,6 @@
},
"params": {
"UseFIPS": false,
- "Region": "us-east-1",
"UseDualStack": false,
"Endpoint": "https://example.com"
}
@@ -1728,8 +636,8 @@
"error": "Invalid Configuration: FIPS and custom endpoint are not supported"
},
"params": {
- "UseFIPS": true,
"Region": "us-east-1",
+ "UseFIPS": true,
"UseDualStack": false,
"Endpoint": "https://example.com"
}
@@ -1740,11 +648,17 @@
"error": "Invalid Configuration: Dualstack and custom endpoint are not supported"
},
"params": {
- "UseFIPS": false,
"Region": "us-east-1",
+ "UseFIPS": false,
"UseDualStack": true,
"Endpoint": "https://example.com"
}
+ },
+ {
+ "documentation": "Missing region",
+ "expect": {
+ "error": "Invalid Configuration: Missing Region"
+ }
}
],
"version": "1.0"
diff --git a/services/eks/src/main/resources/codegen-resources/service-2.json b/services/eks/src/main/resources/codegen-resources/service-2.json
index 1c034287282..4452bcc29fe 100644
--- a/services/eks/src/main/resources/codegen-resources/service-2.json
+++ b/services/eks/src/main/resources/codegen-resources/service-2.json
@@ -2669,6 +2669,7 @@
"Ec2SubnetNotFound",
"Ec2SubnetInvalidConfiguration",
"IamInstanceProfileNotFound",
+ "Ec2SubnetMissingIpv6Assignment",
"IamLimitExceeded",
"IamNodeRoleNotFound",
"NodeCreationFailure",
@@ -2678,7 +2679,20 @@
"AccessDenied",
"InternalFailure",
"ClusterUnreachable",
- "Ec2SubnetMissingIpv6Assignment"
+ "AmiIdNotFound",
+ "AutoScalingGroupOptInRequired",
+ "AutoScalingGroupRateLimitExceeded",
+ "Ec2LaunchTemplateDeletionFailure",
+ "Ec2LaunchTemplateInvalidConfiguration",
+ "Ec2LaunchTemplateMaxLimitExceeded",
+ "Ec2SubnetListTooLong",
+ "IamThrottling",
+ "NodeTerminationFailure",
+ "PodEvictionFailure",
+ "SourceEc2LaunchTemplateNotFound",
+ "LimitExceeded",
+ "Unknown",
+ "AutoScalingGroupInstanceRefreshActive"
]
},
"NodegroupResources":{
diff --git a/services/elasticache/pom.xml b/services/elasticache/pom.xml
index 615d184e3eb..dcb5c9d4b09 100644
--- a/services/elasticache/pom.xml
+++ b/services/elasticache/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTelasticacheAWS Java SDK :: Services :: Amazon ElastiCache
diff --git a/services/elasticache/src/main/resources/codegen-resources/service-2.json b/services/elasticache/src/main/resources/codegen-resources/service-2.json
index 7721d619e4d..902bdcd4f48 100644
--- a/services/elasticache/src/main/resources/codegen-resources/service-2.json
+++ b/services/elasticache/src/main/resources/codegen-resources/service-2.json
@@ -1329,6 +1329,25 @@
{"shape":"InvalidParameterCombinationException"}
],
"documentation":"
Represents the input of a TestFailover operation which test automatic failover on a specified node group (called shard in the console) in a replication group (called cluster in the console).
This API is designed for testing the behavior of your application in case of ElastiCache failover. It is not designed to be an operational tool for initiating a failover to overcome a problem you may have with the cluster. Moreover, in certain conditions such as large-scale operational events, Amazon may block this API.
Note the following
A customer can use this operation to test automatic failover on up to 5 shards (called node groups in the ElastiCache API and Amazon CLI) in any rolling 24-hour period.
If calling this operation on shards in different clusters (called replication groups in the API and CLI), the calls can be made concurrently.
If calling this operation multiple times on different shards in the same Redis (cluster mode enabled) replication group, the first node replacement must complete before a subsequent call can be made.
To determine whether the node replacement is complete you can check Events using the Amazon ElastiCache console, the Amazon CLI, or the ElastiCache API. Look for the following automatic failover related events, listed here in order of occurrance:
Replication group message: Test Failover API called for node group <node-group-id>
Cache cluster message: Failover from primary node <primary-node-id> to replica node <node-id> completed
Replication group message: Failover from primary node <primary-node-id> to replica node <node-id> completed
List of endpoints from which data should be migrated. List should have only one element.
"
+ }
+ }
+ },
+ "TestMigrationResponse":{
+ "type":"structure",
+ "members":{
+ "ReplicationGroup":{"shape":"ReplicationGroup"}
+ }
+ },
"TimeRangeFilter":{
"type":"structure",
"members":{
diff --git a/services/elasticbeanstalk/pom.xml b/services/elasticbeanstalk/pom.xml
index 4014a8391a8..aba0b7f911e 100644
--- a/services/elasticbeanstalk/pom.xml
+++ b/services/elasticbeanstalk/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTelasticbeanstalkAWS Java SDK :: Services :: AWS Elastic Beanstalk
diff --git a/services/elasticinference/pom.xml b/services/elasticinference/pom.xml
index f11c052f1f1..a04cda02a18 100644
--- a/services/elasticinference/pom.xml
+++ b/services/elasticinference/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTelasticinferenceAWS Java SDK :: Services :: Elastic Inference
diff --git a/services/elasticloadbalancing/pom.xml b/services/elasticloadbalancing/pom.xml
index 66fee6bfd61..1a5e796e3d5 100644
--- a/services/elasticloadbalancing/pom.xml
+++ b/services/elasticloadbalancing/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTelasticloadbalancingAWS Java SDK :: Services :: Elastic Load Balancing
diff --git a/services/elasticloadbalancingv2/pom.xml b/services/elasticloadbalancingv2/pom.xml
index 37f64408806..eb38fefa86e 100644
--- a/services/elasticloadbalancingv2/pom.xml
+++ b/services/elasticloadbalancingv2/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTelasticloadbalancingv2AWS Java SDK :: Services :: Elastic Load Balancing V2
diff --git a/services/elasticloadbalancingv2/src/main/resources/codegen-resources/endpoint-rule-set.json b/services/elasticloadbalancingv2/src/main/resources/codegen-resources/endpoint-rule-set.json
index 79fee5593dd..78128a590f9 100644
--- a/services/elasticloadbalancingv2/src/main/resources/codegen-resources/endpoint-rule-set.json
+++ b/services/elasticloadbalancingv2/src/main/resources/codegen-resources/endpoint-rule-set.json
@@ -3,7 +3,7 @@
"parameters": {
"Region": {
"builtIn": "AWS::Region",
- "required": true,
+ "required": false,
"documentation": "The AWS region used to dispatch the request.",
"type": "String"
},
@@ -32,13 +32,12 @@
{
"conditions": [
{
- "fn": "aws.partition",
+ "fn": "isSet",
"argv": [
{
- "ref": "Region"
+ "ref": "Endpoint"
}
- ],
- "assign": "PartitionResult"
+ ]
}
],
"type": "tree",
@@ -46,14 +45,20 @@
{
"conditions": [
{
- "fn": "isSet",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "Endpoint"
- }
+ "ref": "UseFIPS"
+ },
+ true
]
}
],
+ "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "type": "error"
+ },
+ {
+ "conditions": [],
"type": "tree",
"rules": [
{
@@ -62,67 +67,42 @@
"fn": "booleanEquals",
"argv": [
{
- "ref": "UseFIPS"
+ "ref": "UseDualStack"
},
true
]
}
],
- "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
"type": "error"
},
{
"conditions": [],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
- },
- true
- ]
- }
- ],
- "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
- "type": "error"
+ "endpoint": {
+ "url": {
+ "ref": "Endpoint"
},
- {
- "conditions": [],
- "endpoint": {
- "url": {
- "ref": "Endpoint"
- },
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
}
]
- },
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
{
"conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
- },
- true
- ]
- },
- {
- "fn": "booleanEquals",
+ "fn": "isSet",
"argv": [
{
- "ref": "UseDualStack"
- },
- true
+ "ref": "Region"
+ }
]
}
],
@@ -131,179 +111,240 @@
{
"conditions": [
{
- "fn": "booleanEquals",
+ "fn": "aws.partition",
"argv": [
- true,
{
- "fn": "getAttr",
+ "ref": "Region"
+ }
+ ],
+ "assign": "PartitionResult"
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseFIPS"
},
- "supportsFIPS"
+ true
]
- }
- ]
- },
- {
- "fn": "booleanEquals",
- "argv": [
- true,
+ },
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseDualStack"
},
- "supportsDualStack"
+ true
]
}
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
+ ],
"type": "tree",
"rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsFIPS"
+ ]
+ }
+ ]
+ },
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://elasticloadbalancing-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
+ },
{
"conditions": [],
- "endpoint": {
- "url": "https://elasticloadbalancing-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
+ "type": "error"
}
]
- }
- ]
- },
- {
- "conditions": [],
- "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
},
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseFIPS"
},
- "supportsFIPS"
+ true
]
}
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
+ ],
"type": "tree",
"rules": [
{
"conditions": [
{
- "fn": "stringEquals",
+ "fn": "booleanEquals",
"argv": [
- "aws-us-gov",
+ true,
{
"fn": "getAttr",
"argv": [
{
"ref": "PartitionResult"
},
- "name"
+ "supportsFIPS"
]
}
]
}
],
- "endpoint": {
- "url": "https://elasticloadbalancing.{Region}.amazonaws.com",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "stringEquals",
+ "argv": [
+ "aws-us-gov",
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "name"
+ ]
+ }
+ ]
+ }
+ ],
+ "endpoint": {
+ "url": "https://elasticloadbalancing.{Region}.amazonaws.com",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ },
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://elasticloadbalancing-fips.{Region}.{PartitionResult#dnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
},
{
"conditions": [],
- "endpoint": {
- "url": "https://elasticloadbalancing-fips.{Region}.{PartitionResult#dnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "error": "FIPS is enabled but this partition does not support FIPS",
+ "type": "error"
}
]
- }
- ]
- },
- {
- "conditions": [],
- "error": "FIPS is enabled but this partition does not support FIPS",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
},
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseDualStack"
},
- "supportsDualStack"
+ true
]
}
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://elasticloadbalancing.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "error": "DualStack is enabled but this partition does not support DualStack",
+ "type": "error"
+ }
]
- }
- ],
- "type": "tree",
- "rules": [
+ },
{
"conditions": [],
"type": "tree",
@@ -311,7 +352,7 @@
{
"conditions": [],
"endpoint": {
- "url": "https://elasticloadbalancing.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "url": "https://elasticloadbalancing.{Region}.{PartitionResult#dnsSuffix}",
"properties": {},
"headers": {}
},
@@ -320,28 +361,13 @@
]
}
]
- },
- {
- "conditions": [],
- "error": "DualStack is enabled but this partition does not support DualStack",
- "type": "error"
}
]
},
{
"conditions": [],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
- "endpoint": {
- "url": "https://elasticloadbalancing.{Region}.{PartitionResult#dnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
+ "error": "Invalid Configuration: Missing Region",
+ "type": "error"
}
]
}
diff --git a/services/elasticloadbalancingv2/src/main/resources/codegen-resources/endpoint-tests.json b/services/elasticloadbalancingv2/src/main/resources/codegen-resources/endpoint-tests.json
index 30ae4512f69..a5e023b96c1 100644
--- a/services/elasticloadbalancingv2/src/main/resources/codegen-resources/endpoint-tests.json
+++ b/services/elasticloadbalancingv2/src/main/resources/codegen-resources/endpoint-tests.json
@@ -1,42 +1,55 @@
{
"testCases": [
{
- "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.us-iso-east-1.c2s.ic.gov"
+ "url": "https://elasticloadbalancing.af-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "af-south-1",
"UseFIPS": false,
- "Region": "us-iso-east-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-iso-west-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.us-iso-west-1.c2s.ic.gov"
+ "url": "https://elasticloadbalancing.ap-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "ap-east-1",
"UseFIPS": false,
- "Region": "us-iso-west-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing-fips.us-iso-east-1.c2s.ic.gov"
+ "url": "https://elasticloadbalancing.ap-northeast-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "UseFIPS": true,
- "Region": "us-iso-east-1"
+ "Region": "ap-northeast-1",
+ "UseFIPS": false,
+ "UseDualStack": false
+ }
+ },
+ {
+ "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
+ "expect": {
+ "endpoint": {
+ "url": "https://elasticloadbalancing.ap-northeast-2.amazonaws.com"
+ }
+ },
+ "params": {
+ "Region": "ap-northeast-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
@@ -47,87 +60,87 @@
}
},
"params": {
- "UseDualStack": false,
+ "Region": "ap-northeast-3",
"UseFIPS": false,
- "Region": "ap-northeast-3"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.us-east-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.ap-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "ap-south-1",
"UseFIPS": false,
- "Region": "us-east-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing-fips.us-east-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.ap-southeast-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "UseFIPS": true,
- "Region": "us-east-1"
+ "Region": "ap-southeast-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.eu-west-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.ap-southeast-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "ap-southeast-2",
"UseFIPS": false,
- "Region": "eu-west-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-southeast-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.eu-west-2.amazonaws.com"
+ "url": "https://elasticloadbalancing.ap-southeast-3.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "ap-southeast-3",
"UseFIPS": false,
- "Region": "eu-west-2"
+ "UseDualStack": false
}
},
{
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.eu-west-3.amazonaws.com"
+ "url": "https://elasticloadbalancing.ca-central-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "ca-central-1",
"UseFIPS": false,
- "Region": "eu-west-3"
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.me-south-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.eu-central-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "eu-central-1",
"UseFIPS": false,
- "Region": "me-south-1"
+ "UseDualStack": false
}
},
{
@@ -138,139 +151,139 @@
}
},
"params": {
- "UseDualStack": false,
+ "Region": "eu-north-1",
"UseFIPS": false,
- "Region": "eu-north-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.us-east-2.amazonaws.com"
+ "url": "https://elasticloadbalancing.eu-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "eu-south-1",
"UseFIPS": false,
- "Region": "us-east-2"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing-fips.us-east-2.amazonaws.com"
+ "url": "https://elasticloadbalancing.eu-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "UseFIPS": true,
- "Region": "us-east-2"
+ "Region": "eu-west-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.sa-east-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.eu-west-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "eu-west-2",
"UseFIPS": false,
- "Region": "sa-east-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.ap-east-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.eu-west-3.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "eu-west-3",
"UseFIPS": false,
- "Region": "ap-east-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region eu-south-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.eu-south-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.me-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "me-south-1",
"UseFIPS": false,
- "Region": "eu-south-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.eu-central-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.sa-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "sa-east-1",
"UseFIPS": false,
- "Region": "eu-central-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.ap-southeast-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.us-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-east-1",
"UseFIPS": false,
- "Region": "ap-southeast-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.ap-southeast-2.amazonaws.com"
+ "url": "https://elasticloadbalancing-fips.us-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "UseFIPS": false,
- "Region": "ap-southeast-2"
+ "Region": "us-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.ap-southeast-3.amazonaws.com"
+ "url": "https://elasticloadbalancing.us-east-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-east-2",
"UseFIPS": false,
- "Region": "ap-southeast-3"
+ "UseDualStack": false
}
},
{
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.ca-central-1.amazonaws.com"
+ "url": "https://elasticloadbalancing-fips.us-east-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "UseFIPS": false,
- "Region": "ca-central-1"
+ "Region": "us-east-2",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
@@ -281,9 +294,9 @@
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-west-1",
"UseFIPS": false,
- "Region": "us-west-1"
+ "UseDualStack": false
}
},
{
@@ -294,9 +307,9 @@
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-west-1",
"UseFIPS": true,
- "Region": "us-west-1"
+ "UseDualStack": false
}
},
{
@@ -307,9 +320,9 @@
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-west-2",
"UseFIPS": false,
- "Region": "us-west-2"
+ "UseDualStack": false
}
},
{
@@ -320,139 +333,152 @@
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-west-2",
"UseFIPS": true,
- "Region": "us-west-2"
+ "UseDualStack": false
}
},
{
- "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.af-south-1.amazonaws.com"
+ "url": "https://elasticloadbalancing-fips.us-east-1.api.aws"
}
},
"params": {
- "UseDualStack": false,
- "UseFIPS": false,
- "Region": "af-south-1"
+ "Region": "us-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.ap-south-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.us-east-1.api.aws"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-east-1",
"UseFIPS": false,
- "Region": "ap-south-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.ap-northeast-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.cn-north-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "cn-north-1",
"UseFIPS": false,
- "Region": "ap-northeast-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.ap-northeast-2.amazonaws.com"
+ "url": "https://elasticloadbalancing.cn-northwest-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "cn-northwest-1",
"UseFIPS": false,
- "Region": "ap-northeast-2"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing-fips.us-east-1.api.aws"
+ "url": "https://elasticloadbalancing-fips.cn-north-1.api.amazonwebservices.com.cn"
}
},
"params": {
- "UseDualStack": true,
+ "Region": "cn-north-1",
"UseFIPS": true,
- "Region": "us-east-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.us-east-1.api.aws"
+ "url": "https://elasticloadbalancing-fips.cn-north-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": true,
+ "Region": "cn-north-1",
+ "UseFIPS": true,
+ "UseDualStack": false
+ }
+ },
+ {
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled",
+ "expect": {
+ "endpoint": {
+ "url": "https://elasticloadbalancing.cn-north-1.api.amazonwebservices.com.cn"
+ }
+ },
+ "params": {
+ "Region": "cn-north-1",
"UseFIPS": false,
- "Region": "us-east-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.us-gov-west-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.us-gov-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-gov-east-1",
"UseFIPS": false,
- "Region": "us-gov-west-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.us-gov-west-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.us-gov-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-gov-east-1",
"UseFIPS": true,
- "Region": "us-gov-west-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.us-gov-east-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.us-gov-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-gov-west-1",
"UseFIPS": false,
- "Region": "us-gov-east-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.us-gov-east-1.amazonaws.com"
+ "url": "https://elasticloadbalancing.us-gov-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-gov-west-1",
"UseFIPS": true,
- "Region": "us-gov-east-1"
+ "UseDualStack": false
}
},
{
@@ -463,9 +489,9 @@
}
},
"params": {
- "UseDualStack": true,
+ "Region": "us-gov-east-1",
"UseFIPS": true,
- "Region": "us-gov-east-1"
+ "UseDualStack": true
}
},
{
@@ -476,113 +502,144 @@
}
},
"params": {
- "UseDualStack": true,
+ "Region": "us-gov-east-1",
"UseFIPS": false,
- "Region": "us-gov-east-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.us-isob-east-1.sc2s.sgov.gov"
+ "url": "https://elasticloadbalancing.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-iso-east-1",
"UseFIPS": false,
- "Region": "us-isob-east-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-iso-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing-fips.us-isob-east-1.sc2s.sgov.gov"
+ "url": "https://elasticloadbalancing.us-iso-west-1.c2s.ic.gov"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-iso-west-1",
+ "UseFIPS": false,
+ "UseDualStack": false
+ }
+ },
+ {
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
+ "expect": {
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
+ },
+ "params": {
+ "Region": "us-iso-east-1",
"UseFIPS": true,
- "Region": "us-isob-east-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.cn-northwest-1.amazonaws.com.cn"
+ "url": "https://elasticloadbalancing-fips.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-iso-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
+ }
+ },
+ {
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
+ "expect": {
+ "error": "DualStack is enabled but this partition does not support DualStack"
+ },
+ "params": {
+ "Region": "us-iso-east-1",
"UseFIPS": false,
- "Region": "cn-northwest-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.cn-north-1.amazonaws.com.cn"
+ "url": "https://elasticloadbalancing.us-isob-east-1.sc2s.sgov.gov"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-isob-east-1",
"UseFIPS": false,
- "Region": "cn-north-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://elasticloadbalancing-fips.cn-north-1.api.amazonwebservices.com.cn"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
- "UseDualStack": true,
+ "Region": "us-isob-east-1",
"UseFIPS": true,
- "Region": "cn-north-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing-fips.cn-north-1.amazonaws.com.cn"
+ "url": "https://elasticloadbalancing-fips.us-isob-east-1.sc2s.sgov.gov"
}
},
"params": {
- "UseDualStack": false,
+ "Region": "us-isob-east-1",
"UseFIPS": true,
- "Region": "cn-north-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack enabled",
+ "expect": {
+ "error": "DualStack is enabled but this partition does not support DualStack"
+ },
+ "params": {
+ "Region": "us-isob-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
+ }
+ },
+ {
+ "documentation": "For custom endpoint with region set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
- "url": "https://elasticloadbalancing.cn-north-1.api.amazonwebservices.com.cn"
+ "url": "https://example.com"
}
},
"params": {
- "UseDualStack": true,
+ "Region": "us-east-1",
"UseFIPS": false,
- "Region": "cn-north-1"
+ "UseDualStack": false,
+ "Endpoint": "https://example.com"
}
},
{
- "documentation": "For custom endpoint with fips disabled and dualstack disabled",
+ "documentation": "For custom endpoint with region not set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
"url": "https://example.com"
}
},
"params": {
- "UseDualStack": false,
"UseFIPS": false,
- "Region": "us-east-1",
+ "UseDualStack": false,
"Endpoint": "https://example.com"
}
},
@@ -592,9 +649,9 @@
"error": "Invalid Configuration: FIPS and custom endpoint are not supported"
},
"params": {
- "UseDualStack": false,
- "UseFIPS": true,
"Region": "us-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false,
"Endpoint": "https://example.com"
}
},
@@ -604,11 +661,17 @@
"error": "Invalid Configuration: Dualstack and custom endpoint are not supported"
},
"params": {
- "UseDualStack": true,
- "UseFIPS": false,
"Region": "us-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true,
"Endpoint": "https://example.com"
}
+ },
+ {
+ "documentation": "Missing region",
+ "expect": {
+ "error": "Invalid Configuration: Missing Region"
+ }
}
],
"version": "1.0"
diff --git a/services/elasticloadbalancingv2/src/main/resources/codegen-resources/service-2.json b/services/elasticloadbalancingv2/src/main/resources/codegen-resources/service-2.json
index d2f143bcdec..df77c9e4543 100644
--- a/services/elasticloadbalancingv2/src/main/resources/codegen-resources/service-2.json
+++ b/services/elasticloadbalancingv2/src/main/resources/codegen-resources/service-2.json
@@ -244,7 +244,7 @@
{"shape":"TargetGroupNotFoundException"},
{"shape":"InvalidTargetException"}
],
- "documentation":"
Deregisters the specified targets from the specified target group. After the targets are deregistered, they no longer receive traffic from the load balancer.
"
+ "documentation":"
Deregisters the specified targets from the specified target group. After the targets are deregistered, they no longer receive traffic from the load balancer.
Note: If the specified target does not exist, the action returns successfully.
Associates the specified security groups with the specified Application Load Balancer. The specified security groups override the previously associated security groups.
You can't specify a security group for a Network Load Balancer or Gateway Load Balancer.
"
+ "documentation":"
Associates the specified security groups with the specified Application Load Balancer or Network Load Balancer. The specified security groups override the previously associated security groups.
You can't perform this operation on a Network Load Balancer unless you specified a security group for the load balancer when you created it.
You can't associate a security group with a Gateway Load Balancer.
Indicates whether to evaluate inbound security group rules for traffic sent to a Network Load Balancer through Amazon Web Services PrivateLink.
"
}
},
"documentation":"
Information about a load balancer.
"
@@ -2682,7 +2694,7 @@
"documentation":"
Information for a source IP condition. Specify only when Field is source-ip.
"
}
},
- "documentation":"
Information about a condition for a rule.
Each rule can optionally include up to one of each of the following conditions: http-request-method, host-header, path-pattern, and source-ip. Each rule can also optionally include one or more of each of the following conditions: http-header and query-string. Note that the value for a condition cannot be empty.
"
+ "documentation":"
Information about a condition for a rule.
Each rule can optionally include up to one of each of the following conditions: http-request-method, host-header, path-pattern, and source-ip. Each rule can also optionally include one or more of each of the following conditions: http-header and query-string. Note that the value for a condition cannot be empty.
Indicates whether to evaluate inbound security group rules for traffic sent to a Network Load Balancer through Amazon Web Services PrivateLink. The default is on.
[Network Load Balancers] The type of IP addresses used by the subnets for your load balancer. The possible values are ipv4 (for IPv4 addresses) and dualstack (for IPv4 and IPv6 addresses). You can’t specify dualstack for a load balancer with a UDP or TCP_UDP listener. .
"
+ "documentation":"
[Network Load Balancers] The type of IP addresses used by the subnets for your load balancer. The possible values are ipv4 (for IPv4 addresses) and dualstack (for IPv4 and IPv6 addresses). You can’t specify dualstack for a load balancer with a UDP or TCP_UDP listener.
The port on which the target is listening. If the target group protocol is GENEVE, the supported port is 6081. If the target type is alb, the targeted Application Load Balancer must have at least one listener whose port matches the target group port. Not used if the target is a Lambda function.
"
+ "documentation":"
The port on which the target is listening. If the target group protocol is GENEVE, the supported port is 6081. If the target type is alb, the targeted Application Load Balancer must have at least one listener whose port matches the target group port. This parameter is not used if the target is a Lambda function.
The Amazon Resource Names (ARN) of the load balancers that route traffic to this target group.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the load balancer that routes traffic to this target group. You can use each target group with only one load balancer.
Returns a URL to access the job run dashboard. The generated URL is valid for one hour, after which you must invoke the API again to generate a new URL.
"
+ "documentation":"
Creates and returns a URL that you can use to access the application UIs for a job run.
For jobs in a running state, the application UI is a live user interface such as the Spark or Tez web UI. For completed jobs, the application UI is a persistent application user interface such as the Spark History Server or persistent Tez UI.
The URL is valid for one hour after you generate it. To access the application UI after that hour elapses, you must invoke the API again to generate a new URL.
The types of logs that you want to publish to CloudWatch. If you don't specify any log types, driver STDOUT and STDERR logs will be published to CloudWatch Logs by default. For more information including the supported worker types for Hive and Spark, see Logging for EMR Serverless with CloudWatch.
Creates a MatchingWorkflow object which stores the configuration of the data processing job to be run. It is important to note that there should not be a pre-existing MatchingWorkflow with the same name. To modify an existing workflow, utilize the UpdateMatchingWorkflow API.
Creates a schema mapping, which defines the schema of the input customer records table. The SchemaMapping also provides Entity Resolution with some metadata about the table, such as the attribute types of the columns and which columns to match on.
Deletes the SchemaMapping with a given name. This operation will succeed even if a schema with the given name does not exist. This operation will fail if there is a DataIntegrationWorkflow object that references the SchemaMapping in the workflow's InputSourceConfig.
Assigns one or more tags (key-value pairs) to the specified AWS Entity Resolution resource. Tags can help you organize and categorize your resources. You can also use them to scope user permissions by granting a user permission to access or change only resources with certain tag values. In Entity Resolution, SchemaMapping, and MatchingWorkflow can be tagged. Tags don't have any semantic meaning to AWS and are interpreted strictly as strings of characters. You can use the TagResource action with a resource that already has tags. If you specify a new tag key, this tag is appended to the list of tags associated with the resource. If you specify a tag key that is already associated with the resource, the new tag value that you specify replaces the previous value for that tag.
Updates an existing MatchingWorkflow. This method is identical to CreateMatchingWorkflow, except it uses an HTTP PUT request instead of a POST request, and the MatchingWorkflow must already exist for the method to succeed.
The request could not be processed because of conflict in the current state of the resource. Example: Workflow already exists, Schema already exists, Workflow is currently running, etc. HTTP Status Code: 400
The Amazon Resource Name (ARN) of the IAM role. AWS Entity Resolution assumes this role to create resources on your behalf as part of workflow execution.
The Amazon Resource Name (ARN) of the IAM role. AWS Entity Resolution assumes this role to create resources on your behalf as part of workflow execution.
A list of MappedInputFields. Each MappedInputField corresponds to a column the source data table, and contains column name plus additional information that Entity Resolution uses for matching.
A list of MappedInputFields. Each MappedInputField corresponds to a column the source data table, and contains column name plus additional information that Entity Resolution uses for matching.
The request was rejected because it attempted to create resources beyond the current AWS Entity Resolution account limits. The error message describes the limit exceeded. HTTP Status Code: 402
A list of MappedInputFields. Each MappedInputField corresponds to a column the source data table, and contains column name plus additional information Venice uses for matching.
Normalizes the attributes defined in the schema in the input data. For example, if an attribute has an AttributeType of PHONE_NUMBER, and the data in the input table is in a format of 1234567890, Entity Resolution will normalize this field in the output to (123)-456-7890.
A name of a column to be written to the output. This must be an InputField name in the schema mapping.
"
+ }
+ },
+ "documentation":"
A list of OutputAttribute objects, each of which have the fields Name and Hashed. Each of these objects selects a column to be included in the output table, and whether the values of the column should be hashed.
Normalizes the attributes defined in the schema in the input data. For example, if an attribute has an AttributeType of PHONE_NUMBER, and the data in the input table is in a format of 1234567890, Entity Resolution will normalize this field in the output to (123)-456-7890.
A list of OutputAttribute objects, each of which have the fields Name and Hashed. Each of these objects selects a column to be included in the output table, and whether the values of the column should be hashed.
The S3 path to which Entity Resolution will write the output table.
"
+ }
+ },
+ "documentation":"
A list of OutputAttribute objects, each of which have the fields Name and Hashed. Each of these objects selects a column to be included in the output table, and whether the values of the column should be hashed.
A list of MatchingKeys. The MatchingKeys must have been defined in the SchemaMapping. Two records are considered to match according to this rule if all of the MatchingKeys match.
You can either choose ONE_TO_ONE or MANY_TO_MANY as the AttributeMatchingModel. When choosing MANY_TO_MANY, the system can match attribute across the sub-types of an attribute type. For example, if the value of the Email field of Profile A and the value of BusinessEmail field of Profile B matches, the two profiles are matched on the Email type. When choosing ONE_TO_ONE the system can only match if the sub-types are exact matches. For example, only when the value of the Email field of Profile A and the value of the Email field of Profile B matches, the two profiles are matched on the Email type.
Instruct Entity Resolution to combine several columns into a unified column with the identical attribute type. For example, when working with columns such as first_name, middle_name, and last_name, assigning them a common GroupName will prompt Entity Resolution to concatenate them into a single value.
A key that allows grouping of multiple input attributes into a unified matching group. For example, let's consider a scenario where the source table contains various addresses, such as business_address and shipping_address. By assigning the MatchKeyAddress' to both attributes, Entity Resolution will match records across these fields to create a consolidated matching group. If no MatchKey is specified for a column, it won't be utilized for matching purposes but will still be included in the output table.
The Amazon Resource Name (ARN) of the IAM role. AWS Entity Resolution assumes this role to create resources on your behalf as part of workflow execution.
The Amazon Resource Name (ARN) of the IAM role. AWS Entity Resolution assumes this role to create resources on your behalf as part of workflow execution.
Welcome to the AWS Entity Resolution API Reference.
AWS Entity Resolution is an AWS service that provides pre-configured entity resolution capabilities that enable developers and analysts at advertising and marketing companies to build an accurate and complete view of their consumers.
With AWS Entity Resolution, you have the ability to match source records containing consumer identifiers, such as name, email address, and phone number. This holds true even when these records have incomplete or conflicting identifiers. For example, AWS Entity Resolution can effectively match a source record from a customer relationship management (CRM) system, which includes account information like first name, last name, postal address, phone number, and email address, with a source record from a marketing system containing campaign information, such as username and email address.
Cancels an existing Amazon FSx for Lustre data repository task if that task is in either the PENDING or EXECUTING state. When you cancel a task, Amazon FSx does the following.
Any files that FSx has already exported are not reverted.
FSx continues to export any files that are \"in-flight\" when the cancel operation is received.
FSx does not export any files that have not yet been exported.
",
+ "documentation":"
Cancels an existing Amazon FSx for Lustre data repository task if that task is in either the PENDING or EXECUTING state. When you cancel am export task, Amazon FSx does the following.
Any files that FSx has already exported are not reverted.
FSx continues to export any files that are in-flight when the cancel operation is received.
FSx does not export any files that have not yet been exported.
For a release task, Amazon FSx will stop releasing files upon cancellation. Any files that have already been released will remain in the released state.
Creates an Amazon FSx for Lustre data repository task. You use data repository tasks to perform bulk operations between your Amazon FSx file system and its linked data repositories. An example of a data repository task is exporting any data and metadata changes, including POSIX metadata, to files, directories, and symbolic links (symlinks) from your FSx file system to a linked data repository. A CreateDataRepositoryTask operation will fail if a data repository is not linked to the FSx file system. To learn more about data repository tasks, see Data Repository Tasks. To learn more about linking a data repository to your file system, see Linking your file system to an S3 bucket.
",
+ "documentation":"
Creates an Amazon FSx for Lustre data repository task. A CreateDataRepositoryTask operation will fail if a data repository is not linked to the FSx file system.
You use import and export data repository tasks to perform bulk operations between your FSx for Lustre file system and its linked data repositories. An example of a data repository task is exporting any data and metadata changes, including POSIX metadata, to files, directories, and symbolic links (symlinks) from your FSx file system to a linked data repository.
You use release data repository tasks to release data from your file system for files that are archived to S3. The metadata of released files remains on the file system so users or applications can still access released files by reading the files again, which will restore data from Amazon S3 to the FSx for Lustre file system.
Describes the type of administrative action, as follows:
FILE_SYSTEM_UPDATE - A file system update administrative action initiated from the Amazon FSx console, API (UpdateFileSystem), or CLI (update-file-system).
STORAGE_OPTIMIZATION - After the FILE_SYSTEM_UPDATE task to increase a file system's storage capacity has been completed successfully, a STORAGE_OPTIMIZATION task starts.
For Windows and ONTAP, storage optimization is the process of migrating the file system data to newer larger disks.
For Lustre, storage optimization consists of rebalancing the data across the existing and newly added file servers.
You can track the storage-optimization progress using the ProgressPercent property. When STORAGE_OPTIMIZATION has been completed successfully, the parent FILE_SYSTEM_UPDATE action status changes to COMPLETED. For more information, see Managing storage capacity in the Amazon FSx for Windows File Server User Guide, Managing storage and throughput capacity in the Amazon FSx for Lustre User Guide, and Managing storage capacity and provisioned IOPS in the Amazon FSx for NetApp ONTAP User Guide.
FILE_SYSTEM_ALIAS_ASSOCIATION - A file system update to associate a new Domain Name System (DNS) alias with the file system. For more information, see AssociateFileSystemAliases.
FILE_SYSTEM_ALIAS_DISASSOCIATION - A file system update to disassociate a DNS alias from the file system. For more information, see DisassociateFileSystemAliases.
VOLUME_UPDATE - A volume update to an Amazon FSx for NetApp ONTAP or Amazon FSx for OpenZFS volume initiated from the Amazon FSx console, API (UpdateVolume), or CLI (update-volume).
VOLUME_RESTORE - An Amazon FSx for OpenZFS volume is returned to the state saved by the specified snapshot, initiated from an API (RestoreVolumeFromSnapshot) or CLI (restore-volume-from-snapshot).
SNAPSHOT_UPDATE - A snapshot update to an Amazon FSx for OpenZFS volume initiated from the Amazon FSx console, API (UpdateSnapshot), or CLI (update-snapshot).
RELEASE_NFS_V3_LOCKS - Tracks the release of Network File System (NFS) V3 locks on an Amazon FSx for OpenZFS file system.
",
+ "documentation":"
Describes the type of administrative action, as follows:
FILE_SYSTEM_UPDATE - A file system update administrative action initiated from the Amazon FSx console, API (UpdateFileSystem), or CLI (update-file-system).
THROUGHPUT_OPTIMIZATION - After the FILE_SYSTEM_UPDATE task to increase a file system's throughput capacity has been completed successfully, a THROUGHPUT_OPTIMIZATION task starts.
You can track the storage-optimization progress using the ProgressPercent property. When THROUGHPUT_OPTIMIZATION has been completed successfully, the parent FILE_SYSTEM_UPDATE action status changes to COMPLETED. For more information, see Managing throughput capacity in the Amazon FSx for Windows File Server User Guide.
STORAGE_OPTIMIZATION - After the FILE_SYSTEM_UPDATE task to increase a file system's storage capacity has been completed successfully, a STORAGE_OPTIMIZATION task starts.
For Windows and ONTAP, storage optimization is the process of migrating the file system data to newer larger disks.
For Lustre, storage optimization consists of rebalancing the data across the existing and newly added file servers.
You can track the storage-optimization progress using the ProgressPercent property. When STORAGE_OPTIMIZATION has been completed successfully, the parent FILE_SYSTEM_UPDATE action status changes to COMPLETED. For more information, see Managing storage capacity in the Amazon FSx for Windows File Server User Guide, Managing storage and throughput capacity in the Amazon FSx for Lustre User Guide, and Managing storage capacity and provisioned IOPS in the Amazon FSx for NetApp ONTAP User Guide.
FILE_SYSTEM_ALIAS_ASSOCIATION - A file system update to associate a new Domain Name System (DNS) alias with the file system. For more information, see AssociateFileSystemAliases.
FILE_SYSTEM_ALIAS_DISASSOCIATION - A file system update to disassociate a DNS alias from the file system. For more information, see DisassociateFileSystemAliases.
IOPS_OPTIMIZATION - After the FILE_SYSTEM_UPDATE task to increase a file system's throughput capacity has been completed successfully, a IOPS_OPTIMIZATION task starts.
You can track the storage-optimization progress using the ProgressPercent property. When IOPS_OPTIMIZATION has been completed successfully, the parent FILE_SYSTEM_UPDATE action status changes to COMPLETED. For more information, see Managing provisioned SSD IOPS in the Amazon FSx for Windows File Server User Guide.
STORAGE_TYPE_OPTIMIZATION - After the FILE_SYSTEM_UPDATE task to increase a file system's throughput capacity has been completed successfully, a STORAGE_TYPE_OPTIMIZATION task starts.
You can track the storage-optimization progress using the ProgressPercent property. When STORAGE_TYPE_OPTIMIZATION has been completed successfully, the parent FILE_SYSTEM_UPDATE action status changes to COMPLETED.
VOLUME_UPDATE - A volume update to an Amazon FSx for NetApp ONTAP or Amazon FSx for OpenZFS volume initiated from the Amazon FSx console, API (UpdateVolume), or CLI (update-volume).
VOLUME_RESTORE - An Amazon FSx for OpenZFS volume is returned to the state saved by the specified snapshot, initiated from an API (RestoreVolumeFromSnapshot) or CLI (restore-volume-from-snapshot).
SNAPSHOT_UPDATE - A snapshot update to an Amazon FSx for OpenZFS volume initiated from the Amazon FSx console, API (UpdateSnapshot), or CLI (update-snapshot).
RELEASE_NFS_V3_LOCKS - Tracks the release of Network File System (NFS) V3 locks on an Amazon FSx for OpenZFS file system.
Specifies the type of data repository task to create.
"
+ "documentation":"
Specifies the type of data repository task to create.
EXPORT_TO_REPOSITORY tasks export from your Amazon FSx for Lustre file system to a linked data repository.
IMPORT_METADATA_FROM_REPOSITORY tasks import metadata changes from a linked S3 bucket to your Amazon FSx for Lustre file system.
RELEASE_DATA_FROM_FILESYSTEM tasks release files in your Amazon FSx for Lustre file system that are archived and that meet your specified release criteria.
AUTO_RELEASE_DATA tasks automatically release files from an Amazon File Cache resource.
A list of paths for the data repository task to use when the task is processed. If a path that you provide isn't valid, the task fails.
For export tasks, the list contains paths on the Amazon FSx file system from which the files are exported to the Amazon S3 bucket. The default path is the file system root directory. The paths you provide need to be relative to the mount point of the file system. If the mount point is /mnt/fsx and /mnt/fsx/path1 is a directory or file on the file system you want to export, then the path to provide is path1.
For import tasks, the list contains paths in the Amazon S3 bucket from which POSIX metadata changes are imported to the Amazon FSx file system. The path can be an S3 bucket or prefix in the format s3://myBucket/myPrefix (where myPrefix is optional).
"
+ "documentation":"
A list of paths for the data repository task to use when the task is processed. If a path that you provide isn't valid, the task fails. If you don't provide paths, the default behavior is to export all files to S3 (for export tasks), import all files from S3 (for import tasks), or release all archived files that meet the last accessed time criteria (for release tasks).
For export tasks, the list contains paths on the FSx for Lustre file system from which the files are exported to the Amazon S3 bucket. The default path is the file system root directory. The paths you provide need to be relative to the mount point of the file system. If the mount point is /mnt/fsx and /mnt/fsx/path1 is a directory or file on the file system you want to export, then the path to provide is path1.
For import tasks, the list contains paths in the Amazon S3 bucket from which POSIX metadata changes are imported to the FSx for Lustre file system. The path can be an S3 bucket or prefix in the format s3://myBucket/myPrefix (where myPrefix is optional).
For release tasks, the list contains directory or file paths on the FSx for Lustre file system from which to release archived files. If a directory is specified, files within the directory are released. If a file path is specified, only that file is released. To release all archived files in the file system, specify a forward slash (/) as the path.
A file must also meet the last accessed time criteria specified in for the file to be released.
Specifies the file system deployment type. Single AZ deployment types are configured for redundancy within a single Availability Zone in an Amazon Web Services Region . Valid values are the following:
SINGLE_AZ_1- (Default) Creates file systems with throughput capacities of 64 - 4,096 MBps. Single_AZ_1 is available in all Amazon Web Services Regions where Amazon FSx for OpenZFS is available.
SINGLE_AZ_2- Creates file systems with throughput capacities of 160 - 10,240 MB/s using an NVMe L2ARC cache. Single_AZ_2 is available only in the US East (N. Virginia), US East (Ohio), US West (Oregon), and Europe (Ireland) Amazon Web Services Regions.
Specifies the file system deployment type. Single AZ deployment types are configured for redundancy within a single Availability Zone in an Amazon Web Services Region . Valid values are the following:
MULTI_AZ_1- Creates file systems with high availability that are configured for Multi-AZ redundancy to tolerate temporary unavailability in Availability Zones (AZs). Multi_AZ_1 is available in the following Amazon Web Services Regions:
SINGLE_AZ_1- (Default) Creates file systems with throughput capacities of 64 - 4,096 MB/s. Single_AZ_1 is available in all Amazon Web Services Regions where Amazon FSx for OpenZFS is available.
SINGLE_AZ_2- Creates file systems with throughput capacities of 160 - 10,240 MB/s using an NVMe L2ARC cache. Single_AZ_2 is available only in the US East (N. Virginia), US East (Ohio), US West (Oregon), and Europe (Ireland) Amazon Web Services Regions.
(Multi-AZ only) Specifies the IP address range in which the endpoints to access your file system will be created. By default in the Amazon FSx API and Amazon FSx console, Amazon FSx selects an available /28 IP address range for you from one of the VPC's CIDR ranges. You can have overlapping endpoint IP addresses for file systems deployed in the same VPC/route tables.
(Multi-AZ only) Specifies the virtual private cloud (VPC) route tables in which your file system's endpoints will be created. You should specify all VPC route tables associated with the subnets in which your clients are located. By default, Amazon FSx selects your VPC's default route table.
"
}
},
"documentation":"
The Amazon FSx for OpenZFS configuration properties for the file system that you are creating.
The configuration that Amazon FSx for Windows File Server uses to audit and log user accesses of files, folders, and file shares on the Amazon FSx for Windows File Server file system.
The SSD IOPS (input/output operations per second) configuration for an Amazon FSx for Windows file system. By default, Amazon FSx automatically provisions 3 IOPS per GiB of storage capacity. You can provision additional IOPS per GiB of storage, up to the maximum limit associated with your chosen throughput capacity.
"
}
},
"documentation":"
The configuration object for the Microsoft Windows file system used in CreateFileSystem and CreateFileSystemFromBackup operations.
Enables, disables, or permanently disables privileged delete on an FSx for ONTAP SnapLock Enterprise volume. Enabling privileged delete allows SnapLock administrators to delete WORM files even if they have active retention periods. PERMANENTLY_DISABLED is a terminal state. If privileged delete is permanently disabled on a SnapLock volume, you can't re-enable it. The default value is DISABLED.
Enables, disables, or permanently disables privileged delete on an FSx for ONTAP SnapLock Enterprise volume. Enabling privileged delete allows SnapLock administrators to delete WORM files even if they have active retention periods. PERMANENTLY_DISABLED is a terminal state. If privileged delete is permanently disabled on a SnapLock volume, you can't re-enable it. The default value is DISABLED.
Specifies the retention mode of an FSx for ONTAP SnapLock volume. After it is set, it can't be changed. You can choose one of the following retention modes:
COMPLIANCE: Files transitioned to write once, read many (WORM) on a Compliance volume can't be deleted until their retention periods expire. This retention mode is used to address government or industry-specific mandates or to protect against ransomware attacks. For more information, see SnapLock Compliance.
ENTERPRISE: Files transitioned to WORM on an Enterprise volume can be deleted by authorized users before their retention periods expire using privileged delete. This retention mode is used to advance an organization's data integrity and internal compliance or to test retention settings before using SnapLock Compliance. For more information, see SnapLock Enterprise.
"
+ "documentation":"
Specifies the retention mode of an FSx for ONTAP SnapLock volume. After it is set, it can't be changed. You can choose one of the following retention modes:
COMPLIANCE: Files transitioned to write once, read many (WORM) on a Compliance volume can't be deleted until their retention periods expire. This retention mode is used to address government or industry-specific mandates or to protect against ransomware attacks. For more information, see SnapLock Compliance.
ENTERPRISE: Files transitioned to WORM on an Enterprise volume can be deleted by authorized users before their retention periods expire using privileged delete. This retention mode is used to advance an organization's data integrity and internal compliance or to test retention settings before using SnapLock Compliance. For more information, see SnapLock Enterprise.
EXPORT_TO_REPOSITORY tasks export from your Amazon FSx for Lustre file system to a linked data repository.
IMPORT_METADATA_FROM_REPOSITORY tasks import metadata changes from a linked S3 bucket to your Amazon FSx for Lustre file system.
AUTO_RELEASE_DATA tasks automatically release files from an Amazon File Cache resource.
RELEASE_DATA_FROM_FILESYSTEM tasks are not supported.
"
+ "documentation":"
The type of data repository task.
EXPORT_TO_REPOSITORY tasks export from your Amazon FSx for Lustre file system to a linked data repository.
IMPORT_METADATA_FROM_REPOSITORY tasks import metadata changes from a linked S3 bucket to your Amazon FSx for Lustre file system.
RELEASE_DATA_FROM_FILESYSTEM tasks release files in your Amazon FSx for Lustre file system that are archived and that meet your specified release criteria.
AUTO_RELEASE_DATA tasks automatically release files from an Amazon File Cache resource.
The configuration that specifies the last accessed time criteria for files that will be released from an Amazon FSx for Lustre file system.
"
}
},
- "documentation":"
A description of the data repository task. You use data repository tasks to perform bulk transfer operations between an Amazon FSx for Lustre file system and a linked data repository. An Amazon File Cache resource uses a task to automatically release files from the cache.
"
+ "documentation":"
A description of the data repository task.
You use import and export data repository tasks to perform bulk transfer operations between an Amazon FSx for Lustre file system and a linked data repository.
You use release data repository tasks to release archived files from your Amazon FSx for Lustre file system.
An Amazon File Cache resource uses a task to automatically release files from the cache.
Setting this to true allows a SnapLock administrator to delete an FSx for ONTAP SnapLock Enterprise volume with unexpired write once, read many (WORM) files. The IAM permission fsx:BypassSnaplockEnterpriseRetention is also required to delete SnapLock Enterprise volumes with unexpired WORM files. The default value is false.
Setting this to true allows a SnapLock administrator to delete an FSx for ONTAP SnapLock Enterprise volume with unexpired write once, read many (WORM) files. The IAM permission fsx:BypassSnaplockEnterpriseRetention is also required to delete SnapLock Enterprise volumes with unexpired WORM files. The default value is false.
Use to specify skipping a final backup, adding tags to a final backup, or bypassing the retention period of an FSx for ONTAP SnapLock Enterprise volume when deleting an FSx for ONTAP volume.
"
@@ -3088,7 +3115,7 @@
"documentation":"
The total number of SSD IOPS provisioned for the file system.
"
}
},
- "documentation":"
The SSD IOPS (input/output operations per second) configuration for an Amazon FSx for NetApp ONTAP or FSx for OpenZFS file system. By default, Amazon FSx automatically provisions 3 IOPS per GB of storage capacity. You can provision additional IOPS per GB of storage. The configuration consists of the total number of provisioned SSD IOPS and how it is was provisioned, or the mode (by the customer or by Amazon FSx).
"
+ "documentation":"
The SSD IOPS (input/output operations per second) configuration for an Amazon FSx for NetApp ONTAP, Amazon FSx for Windows File Server, or FSx for OpenZFS file system. By default, Amazon FSx automatically provisions 3 IOPS per GB of storage capacity. You can provision additional IOPS per GB of storage. The configuration consists of the total number of provisioned SSD IOPS and how it is was provisioned, or the mode (by the customer or by Amazon FSx).
The unit of time used by the Value parameter to determine if a file can be released, based on when it was last accessed. DAYS is the only supported value. This is a required parameter.
An integer that represents the minimum amount of time (in days) since a file was last accessed in the file system. Only archived files with a MAX(atime, ctime, mtime) timestamp that is more than this amount of time in the past (relative to the task create time) will be released. The default of Value is 0. This is a required parameter.
If an archived file meets the last accessed time criteria, its file or directory path must also be specified in the Paths parameter of the operation in order for the file to be released.
"
+ }
+ },
+ "documentation":"
Defines the minimum amount of time since last access for a file to be eligible for release. Only archived files that were last accessed or modified before this point-in-time are eligible to be released from the Amazon FSx for Lustre file system.
(Multi-AZ only) Specifies the IP address range in which the endpoints to access your file system will be created. By default in the Amazon FSx API and Amazon FSx console, Amazon FSx selects an available /28 IP address range for you from one of the VPC's CIDR ranges. You can have overlapping endpoint IP addresses for file systems deployed in the same VPC/route tables.
Defines the point-in-time since an archived file was last accessed, in order for that file to be eligible for release. Only files that were last accessed before this point-in-time are eligible to be released from the file system.
"
+ }
+ },
+ "documentation":"
The configuration that specifies a minimum amount of time since last access for an archived file to be eligible for release from an Amazon FSx for Lustre file system. Only files that were last accessed before this point-in-time can be released. For example, if you specify a last accessed time criteria of 9 days, only files that were last accessed 9.00001 or more days ago can be released.
Only file data that has been archived can be released. Files that have not yet been archived, such as new or changed files that have not been exported, are not eligible for release. When files are released, their metadata stays on the file system, so they can still be accessed later. Users and applications can access a released file by reading the file again, which restores data from Amazon S3 to the FSx for Lustre file system.
If a file meets the last accessed time criteria, its file or directory path must also be specified with the Paths parameter of the operation in order for the file to be released.
Enables, disables, or permanently disables privileged delete on an FSx for ONTAP SnapLock Enterprise volume. Enabling privileged delete allows SnapLock administrators to delete write once, read many (WORM) files even if they have active retention periods. PERMANENTLY_DISABLED is a terminal state. If privileged delete is permanently disabled on a SnapLock volume, you can't re-enable it. The default value is DISABLED.
Enables, disables, or permanently disables privileged delete on an FSx for ONTAP SnapLock Enterprise volume. Enabling privileged delete allows SnapLock administrators to delete write once, read many (WORM) files even if they have active retention periods. PERMANENTLY_DISABLED is a terminal state. If privileged delete is permanently disabled on a SnapLock volume, you can't re-enable it. The default value is DISABLED.
Specifies the retention mode of an FSx for ONTAP SnapLock volume. After it is set, it can't be changed. You can choose one of the following retention modes:
COMPLIANCE: Files transitioned to write once, read many (WORM) on a Compliance volume can't be deleted until their retention periods expire. This retention mode is used to address government or industry-specific mandates or to protect against ransomware attacks. For more information, see SnapLock Compliance.
ENTERPRISE: Files transitioned to WORM on an Enterprise volume can be deleted by authorized users before their retention periods expire using privileged delete. This retention mode is used to advance an organization's data integrity and internal compliance or to test retention settings before using SnapLock Compliance. For more information, see SnapLock Enterprise.
"
+ "documentation":"
Specifies the retention mode of an FSx for ONTAP SnapLock volume. After it is set, it can't be changed. You can choose one of the following retention modes:
COMPLIANCE: Files transitioned to write once, read many (WORM) on a Compliance volume can't be deleted until their retention periods expire. This retention mode is used to address government or industry-specific mandates or to protect against ransomware attacks. For more information, see SnapLock Compliance.
ENTERPRISE: Files transitioned to WORM on an Enterprise volume can be deleted by authorized users before their retention periods expire using privileged delete. This retention mode is used to advance an organization's data integrity and internal compliance or to test retention settings before using SnapLock Compliance. For more information, see SnapLock Enterprise.
The throughput of an Amazon FSx for OpenZFS file system, measured in megabytes per second (MB/s). Valid values depend on the DeploymentType you choose, as follows:
For SINGLE_AZ_1, valid values are 64, 128, 256, 512, 1024, 2048, 3072, or 4096 MB/s.
For SINGLE_AZ_2, valid values are 160, 320, 640, 1280, 2560, 3840, 5120, 7680, or 10240 MB/s.
(Multi-AZ only) A list of IDs of existing virtual private cloud (VPC) route tables to disassociate (remove) from your Amazon FSx for OpenZFS file system. You can use the API operation to retrieve the list of VPC route table IDs for a file system.
"
+ }
},
"documentation":"
The configuration updates for an Amazon FSx for OpenZFS file system.
The configuration that Amazon FSx for Windows File Server uses to audit and log user accesses of files, folders, and file shares on the Amazon FSx for Windows File Server file system..
The SSD IOPS (input/output operations per second) configuration for an Amazon FSx for Windows file system. By default, Amazon FSx automatically provisions 3 IOPS per GiB of storage capacity. You can provision additional IOPS per GiB of storage, up to the maximum limit associated with your chosen throughput capacity.
"
}
},
"documentation":"
Updates the configuration for an existing Amazon FSx for Windows File Server file system. Amazon FSx only overwrites existing properties with non-null values provided in the request.
Enables, disables, or permanently disables privileged delete on an FSx for ONTAP SnapLock Enterprise volume. Enabling privileged delete allows SnapLock administrators to delete write once, read many (WORM) files even if they have active retention periods. PERMANENTLY_DISABLED is a terminal state. If privileged delete is permanently disabled on a SnapLock volume, you can't re-enable it. The default value is DISABLED.
Enables, disables, or permanently disables privileged delete on an FSx for ONTAP SnapLock Enterprise volume. Enabling privileged delete allows SnapLock administrators to delete write once, read many (WORM) files even if they have active retention periods. PERMANENTLY_DISABLED is a terminal state. If privileged delete is permanently disabled on a SnapLock volume, you can't re-enable it. The default value is DISABLED.
The Amazon Resource Name (ARN) for the destination of the audit logs. The destination can be any Amazon CloudWatch Logs log group ARN or Amazon Kinesis Data Firehose delivery stream ARN.
The name of the Amazon CloudWatch Logs log group must begin with the /aws/fsx prefix. The name of the Amazon Kinesis Data Firehouse delivery stream must begin with the aws-fsx prefix.
The destination ARN (either CloudWatch Logs log group or Kinesis Data Firehose delivery stream) must be in the same Amazon Web Services partition, Amazon Web Services Region, and Amazon Web Services account as your Amazon FSx file system.
"
+ "documentation":"
The Amazon Resource Name (ARN) for the destination of the audit logs. The destination can be any Amazon CloudWatch Logs log group ARN or Amazon Kinesis Data Firehose delivery stream ARN.
The name of the Amazon CloudWatch Logs log group must begin with the /aws/fsx prefix. The name of the Amazon Kinesis Data Firehose delivery stream must begin with the aws-fsx prefix.
The destination ARN (either CloudWatch Logs log group or Kinesis Data Firehose delivery stream) must be in the same Amazon Web Services partition, Amazon Web Services Region, and Amazon Web Services account as your Amazon FSx file system.
"
}
},
"documentation":"
The configuration that Amazon FSx for Windows File Server uses to audit and log user accesses of files, folders, and file shares on the Amazon FSx for Windows File Server file system. For more information, see File access auditing.
The Amazon Resource Name (ARN) that specifies the destination of the audit logs.
The destination can be any Amazon CloudWatch Logs log group ARN or Amazon Kinesis Data Firehose delivery stream ARN, with the following requirements:
The destination ARN that you provide (either CloudWatch Logs log group or Kinesis Data Firehose delivery stream) must be in the same Amazon Web Services partition, Amazon Web Services Region, and Amazon Web Services account as your Amazon FSx file system.
The name of the Amazon CloudWatch Logs log group must begin with the /aws/fsx prefix. The name of the Amazon Kinesis Data Firehouse delivery stream must begin with the aws-fsx prefix.
If you do not provide a destination in AuditLogDestination, Amazon FSx will create and use a log stream in the CloudWatch Logs /aws/fsx/windows log group.
If AuditLogDestination is provided and the resource does not exist, the request will fail with a BadRequest error.
If FileAccessAuditLogLevel and FileShareAccessAuditLogLevel are both set to DISABLED, you cannot specify a destination in AuditLogDestination.
"
+ "documentation":"
The Amazon Resource Name (ARN) that specifies the destination of the audit logs.
The destination can be any Amazon CloudWatch Logs log group ARN or Amazon Kinesis Data Firehose delivery stream ARN, with the following requirements:
The destination ARN that you provide (either CloudWatch Logs log group or Kinesis Data Firehose delivery stream) must be in the same Amazon Web Services partition, Amazon Web Services Region, and Amazon Web Services account as your Amazon FSx file system.
The name of the Amazon CloudWatch Logs log group must begin with the /aws/fsx prefix. The name of the Amazon Kinesis Data Firehose delivery stream must begin with the aws-fsx prefix.
If you do not provide a destination in AuditLogDestination, Amazon FSx will create and use a log stream in the CloudWatch Logs /aws/fsx/windows log group.
If AuditLogDestination is provided and the resource does not exist, the request will fail with a BadRequest error.
If FileAccessAuditLogLevel and FileShareAccessAuditLogLevel are both set to DISABLED, you cannot specify a destination in AuditLogDestination.
"
}
},
"documentation":"
The Windows file access auditing configuration used when creating or updating an Amazon FSx for Windows File Server file system.
The configuration that Amazon FSx for Windows File Server uses to audit and log user accesses of files, folders, and file shares on the Amazon FSx for Windows File Server file system.
The SSD IOPS (input/output operations per second) configuration for an Amazon FSx for Windows file system. By default, Amazon FSx automatically provisions 3 IOPS per GiB of storage capacity. You can provision additional IOPS per GiB of storage, up to the maximum limit associated with your chosen throughput capacity.
"
}
},
"documentation":"
The configuration for this Microsoft Windows file system.
Add endpoints to an endpoint group. The AddEndpoints API operation is the recommended option for adding endpoints. The alternative options are to add endpoints when you create an endpoint group (with the CreateEndpointGroup API) or when you update an endpoint group (with the UpdateEndpointGroup API).
There are two advantages to using AddEndpoints to add endpoints:
It's faster, because Global Accelerator only has to resolve the new endpoints that you're adding.
It's more convenient, because you don't need to specify all of the current endpoints that are already in the endpoint group in addition to the new endpoints that you want to add.
"
+ "documentation":"
Add endpoints to an endpoint group. The AddEndpoints API operation is the recommended option for adding endpoints. The alternative options are to add endpoints when you create an endpoint group (with the CreateEndpointGroup API) or when you update an endpoint group (with the UpdateEndpointGroup API).
There are two advantages to using AddEndpoints to add endpoints in Global Accelerator:
It's faster, because Global Accelerator only has to resolve the new endpoints that you're adding, rather than resolving new and existing endpoints.
It's more convenient, because you don't need to specify the current endpoints that are already in the endpoint group, in addition to the new endpoints that you want to add.
For information about endpoint types and requirements for endpoints that you can add to Global Accelerator, see Endpoints for standard accelerators in the Global Accelerator Developer Guide.
Create an accelerator. An accelerator includes one or more listeners that process inbound connections and direct traffic to one or more endpoint groups, each of which includes endpoints, such as Network Load Balancers.
Global Accelerator is a global service that supports endpoints in multiple Amazon Web Services Regions but you must specify the US West (Oregon) Region to create, update, or otherwise work with accelerators. That is, for example, specify --region us-west-2 on AWS CLI commands.
"
+ "documentation":"
Create an accelerator. An accelerator includes one or more listeners that process inbound connections and direct traffic to one or more endpoint groups, each of which includes endpoints, such as Network Load Balancers.
Global Accelerator is a global service that supports endpoints in multiple Amazon Web Services Regions but you must specify the US West (Oregon) Region to create, update, or otherwise work with accelerators. That is, for example, specify --region us-west-2 on Amazon Web Services CLI commands.
Create a custom routing accelerator. A custom routing accelerator directs traffic to one of possibly thousands of Amazon EC2 instance destinations running in a single or multiple virtual private clouds (VPC) subnet endpoints.
Be aware that, by default, all destination EC2 instances in a VPC subnet endpoint cannot receive traffic. To enable all destinations to receive traffic, or to specify individual port mappings that can receive traffic, see the AllowCustomRoutingTraffic operation.
Global Accelerator is a global service that supports endpoints in multiple Amazon Web Services Regions but you must specify the US West (Oregon) Region to create, update, or otherwise work with accelerators. That is, for example, specify --region us-west-2 on AWS CLI commands.
"
+ "documentation":"
Create a custom routing accelerator. A custom routing accelerator directs traffic to one of possibly thousands of Amazon EC2 instance destinations running in a single or multiple virtual private clouds (VPC) subnet endpoints.
Be aware that, by default, all destination EC2 instances in a VPC subnet endpoint cannot receive traffic. To enable all destinations to receive traffic, or to specify individual port mappings that can receive traffic, see the AllowCustomRoutingTraffic operation.
Global Accelerator is a global service that supports endpoints in multiple Amazon Web Services Regions but you must specify the US West (Oregon) Region to create, update, or otherwise work with accelerators. That is, for example, specify --region us-west-2 on Amazon Web Services CLI commands.
Create an endpoint group for the specified listener. An endpoint group is a collection of endpoints in one Amazon Web Services Region. A resource must be valid and active when you add it as an endpoint.
"
+ "documentation":"
Create an endpoint group for the specified listener. An endpoint group is a collection of endpoints in one Amazon Web Services Region. A resource must be valid and active when you add it as an endpoint.
For more information about endpoint types and requirements for endpoints that you can add to Global Accelerator, see Endpoints for standard accelerators in the Global Accelerator Developer Guide.
Global Accelerator is a global service that supports endpoints in multiple Amazon Web Services Regions but you must specify the US West (Oregon) Region to create, update, or otherwise work with accelerators. That is, for example, specify --region us-west-2 on AWS CLI commands.
"
+ "documentation":"
Update an accelerator to make changes, such as the following:
Change the name of the accelerator.
Disable the accelerator so that it no longer accepts or routes traffic, or so that you can delete it.
Enable the accelerator, if it is disabled.
Change the IP address type to dual-stack if it is IPv4, or change the IP address type to IPv4 if it's dual-stack.
Be aware that static IP addresses remain assigned to your accelerator for as long as it exists, even if you disable the accelerator and it no longer accepts or routes traffic. However, when you delete the accelerator, you lose the static IP addresses that are assigned to it, so you can no longer route traffic by using them.
Global Accelerator is a global service that supports endpoints in multiple Amazon Web Services Regions but you must specify the US West (Oregon) Region to create, update, or otherwise work with accelerators. That is, for example, specify --region us-west-2 on Amazon Web Services CLI commands.
This is the Global Accelerator API Reference. This guide is for developers who need detailed information about Global Accelerator API actions, data types, and errors. For more information about Global Accelerator features, see the Global Accelerator Developer Guide.
Global Accelerator is a service in which you create accelerators to improve the performance of your applications for local and global users. Depending on the type of accelerator you choose, you can gain additional benefits.
By using a standard accelerator, you can improve availability of your internet applications that are used by a global audience. With a standard accelerator, Global Accelerator directs traffic to optimal endpoints over the Amazon Web Services global network.
For other scenarios, you might choose a custom routing accelerator. With a custom routing accelerator, you can use application logic to directly map one or more users to a specific endpoint among many endpoints.
Global Accelerator is a global service that supports endpoints in multiple Amazon Web Services Regions but you must specify the US West (Oregon) Region to create, update, or otherwise work with accelerators. That is, for example, specify --region us-west-2 on AWS CLI commands.
By default, Global Accelerator provides you with static IP addresses that you associate with your accelerator. The static IP addresses are anycast from the Amazon Web Services edge network. For IPv4, Global Accelerator provides two static IPv4 addresses. For dual-stack, Global Accelerator provides a total of four addresses: two static IPv4 addresses and two static IPv6 addresses. With a standard accelerator for IPv4, instead of using the addresses that Global Accelerator provides, you can configure these entry points to be IPv4 addresses from your own IP address ranges that you bring toGlobal Accelerator (BYOIP).
For a standard accelerator, they distribute incoming application traffic across multiple endpoint resources in multiple Amazon Web Services Regions , which increases the availability of your applications. Endpoints for standard accelerators can be Network Load Balancers, Application Load Balancers, Amazon EC2 instances, or Elastic IP addresses that are located in one Amazon Web Services Region or multiple Amazon Web Services Regions. For custom routing accelerators, you map traffic that arrives to the static IP addresses to specific Amazon EC2 servers in endpoints that are virtual private cloud (VPC) subnets.
The static IP addresses remain assigned to your accelerator for as long as it exists, even if you disable the accelerator and it no longer accepts or routes traffic. However, when you delete an accelerator, you lose the static IP addresses that are assigned to it, so you can no longer route traffic by using them. You can use IAM policies like tag-based permissions with Global Accelerator to limit the users who have permissions to delete an accelerator. For more information, see Tag-based policies.
For standard accelerators, Global Accelerator uses the Amazon Web Services global network to route traffic to the optimal regional endpoint based on health, client location, and policies that you configure. The service reacts instantly to changes in health or configuration to ensure that internet traffic from clients is always directed to healthy endpoints.
This is the Global Accelerator API Reference. This guide is for developers who need detailed information about Global Accelerator API actions, data types, and errors. For more information about Global Accelerator features, see the Global Accelerator Developer Guide.
Global Accelerator is a service in which you create accelerators to improve the performance of your applications for local and global users. Depending on the type of accelerator you choose, you can gain additional benefits.
By using a standard accelerator, you can improve availability of your internet applications that are used by a global audience. With a standard accelerator, Global Accelerator directs traffic to optimal endpoints over the Amazon Web Services global network.
For other scenarios, you might choose a custom routing accelerator. With a custom routing accelerator, you can use application logic to directly map one or more users to a specific endpoint among many endpoints.
Global Accelerator is a global service that supports endpoints in multiple Amazon Web Services Regions but you must specify the US West (Oregon) Region to create, update, or otherwise work with accelerators. That is, for example, specify --region us-west-2 on Amazon Web Services CLI commands.
By default, Global Accelerator provides you with static IP addresses that you associate with your accelerator. The static IP addresses are anycast from the Amazon Web Services edge network. For IPv4, Global Accelerator provides two static IPv4 addresses. For dual-stack, Global Accelerator provides a total of four addresses: two static IPv4 addresses and two static IPv6 addresses. With a standard accelerator for IPv4, instead of using the addresses that Global Accelerator provides, you can configure these entry points to be IPv4 addresses from your own IP address ranges that you bring toGlobal Accelerator (BYOIP).
For a standard accelerator, they distribute incoming application traffic across multiple endpoint resources in multiple Amazon Web Services Regions , which increases the availability of your applications. Endpoints for standard accelerators can be Network Load Balancers, Application Load Balancers, Amazon EC2 instances, or Elastic IP addresses that are located in one Amazon Web Services Region or multiple Amazon Web Services Regions. For custom routing accelerators, you map traffic that arrives to the static IP addresses to specific Amazon EC2 servers in endpoints that are virtual private cloud (VPC) subnets.
The static IP addresses remain assigned to your accelerator for as long as it exists, even if you disable the accelerator and it no longer accepts or routes traffic. However, when you delete an accelerator, you lose the static IP addresses that are assigned to it, so you can no longer route traffic by using them. You can use IAM policies like tag-based permissions with Global Accelerator to limit the users who have permissions to delete an accelerator. For more information, see Tag-based policies.
For standard accelerators, Global Accelerator uses the Amazon Web Services global network to route traffic to the optimal regional endpoint based on health, client location, and policies that you configure. The service reacts instantly to changes in health or configuration to ensure that internet traffic from clients is always directed to healthy endpoints.
These key-value pairs define parameters for the connection:
HOST - The host URI: either the fully qualified domain name (FQDN) or the IPv4 address of the database host.
PORT - The port number, between 1024 and 65535, of the port on which the database host is listening for database connections.
USER_NAME - The name under which to log in to the database. The value string for USER_NAME is \"USERNAME\".
PASSWORD - A password, if one is used, for the user name.
ENCRYPTED_PASSWORD - When you enable connection password protection by setting ConnectionPasswordEncryption in the Data Catalog encryption settings, this field stores the encrypted password.
JDBC_DRIVER_JAR_URI - The Amazon Simple Storage Service (Amazon S3) path of the JAR file that contains the JDBC driver to use.
JDBC_DRIVER_CLASS_NAME - The class name of the JDBC driver to use.
JDBC_ENGINE - The name of the JDBC engine to use.
JDBC_ENGINE_VERSION - The version of the JDBC engine to use.
CONFIG_FILES - (Reserved for future use.)
INSTANCE_ID - The instance ID to use.
JDBC_CONNECTION_URL - The URL for connecting to a JDBC data source.
JDBC_ENFORCE_SSL - A Boolean string (true, false) specifying whether Secure Sockets Layer (SSL) with hostname matching is enforced for the JDBC connection on the client. The default is false.
CUSTOM_JDBC_CERT - An Amazon S3 location specifying the customer's root certificate. Glue uses this root certificate to validate the customer’s certificate when connecting to the customer database. Glue only handles X.509 certificates. The certificate provided must be DER-encoded and supplied in Base64 encoding PEM format.
SKIP_CUSTOM_JDBC_CERT_VALIDATION - By default, this is false. Glue validates the Signature algorithm and Subject Public Key Algorithm for the customer certificate. The only permitted algorithms for the Signature algorithm are SHA256withRSA, SHA384withRSA or SHA512withRSA. For the Subject Public Key Algorithm, the key length must be at least 2048. You can set the value of this property to true to skip Glue’s validation of the customer certificate.
CUSTOM_JDBC_CERT_STRING - A custom JDBC certificate string which is used for domain match or distinguished name match to prevent a man-in-the-middle attack. In Oracle database, this is used as the SSL_SERVER_CERT_DN; in Microsoft SQL Server, this is used as the hostNameInCertificate.
CONNECTION_URL - The URL for connecting to a general (non-JDBC) data source.
SECRET_ID - The secret ID used for the secret manager of credentials.
CONNECTOR_URL - The connector URL for a MARKETPLACE or CUSTOM connection.
CONNECTOR_TYPE - The connector type for a MARKETPLACE or CUSTOM connection.
CONNECTOR_CLASS_NAME - The connector class name for a MARKETPLACE or CUSTOM connection.
KAFKA_BOOTSTRAP_SERVERS - A comma-separated list of host and port pairs that are the addresses of the Apache Kafka brokers in a Kafka cluster to which a Kafka client will connect to and bootstrap itself.
KAFKA_SSL_ENABLED - Whether to enable or disable SSL on an Apache Kafka connection. Default value is \"true\".
KAFKA_CUSTOM_CERT - The Amazon S3 URL for the private CA cert file (.pem format). The default is an empty string.
KAFKA_SKIP_CUSTOM_CERT_VALIDATION - Whether to skip the validation of the CA cert file or not. Glue validates for three algorithms: SHA256withRSA, SHA384withRSA and SHA512withRSA. Default value is \"false\".
KAFKA_CLIENT_KEYSTORE - The Amazon S3 location of the client keystore file for Kafka client side authentication (Optional).
KAFKA_CLIENT_KEYSTORE_PASSWORD - The password to access the provided keystore (Optional).
KAFKA_CLIENT_KEY_PASSWORD - A keystore can consist of multiple keys, so this is the password to access the client key to be used with the Kafka server side key (Optional).
ENCRYPTED_KAFKA_CLIENT_KEYSTORE_PASSWORD - The encrypted version of the Kafka client keystore password (if the user has the Glue encrypt passwords setting selected).
ENCRYPTED_KAFKA_CLIENT_KEY_PASSWORD - The encrypted version of the Kafka client key password (if the user has the Glue encrypt passwords setting selected).
KAFKA_SASL_MECHANISM - \"SCRAM-SHA-512\", \"GSSAPI\", or \"AWS_MSK_IAM\". These are the supported SASL Mechanisms.
KAFKA_SASL_SCRAM_USERNAME - A plaintext username used to authenticate with the \"SCRAM-SHA-512\" mechanism.
KAFKA_SASL_SCRAM_PASSWORD - A plaintext password used to authenticate with the \"SCRAM-SHA-512\" mechanism.
ENCRYPTED_KAFKA_SASL_SCRAM_PASSWORD - The encrypted version of the Kafka SASL SCRAM password (if the user has the Glue encrypt passwords setting selected).
KAFKA_SASL_GSSAPI_KEYTAB - The S3 location of a Kerberos keytab file. A keytab stores long-term keys for one or more principals. For more information, see MIT Kerberos Documentation: Keytab.
KAFKA_SASL_GSSAPI_KRB5_CONF - The S3 location of a Kerberos krb5.conf file. A krb5.conf stores Kerberos configuration information, such as the location of the KDC server. For more information, see MIT Kerberos Documentation: krb5.conf.
KAFKA_SASL_GSSAPI_SERVICE - The Kerberos service name, as set with sasl.kerberos.service.name in your Kafka Configuration.
These key-value pairs define parameters for the connection:
HOST - The host URI: either the fully qualified domain name (FQDN) or the IPv4 address of the database host.
PORT - The port number, between 1024 and 65535, of the port on which the database host is listening for database connections.
USER_NAME - The name under which to log in to the database. The value string for USER_NAME is \"USERNAME\".
PASSWORD - A password, if one is used, for the user name.
ENCRYPTED_PASSWORD - When you enable connection password protection by setting ConnectionPasswordEncryption in the Data Catalog encryption settings, this field stores the encrypted password.
JDBC_DRIVER_JAR_URI - The Amazon Simple Storage Service (Amazon S3) path of the JAR file that contains the JDBC driver to use.
JDBC_DRIVER_CLASS_NAME - The class name of the JDBC driver to use.
JDBC_ENGINE - The name of the JDBC engine to use.
JDBC_ENGINE_VERSION - The version of the JDBC engine to use.
CONFIG_FILES - (Reserved for future use.)
INSTANCE_ID - The instance ID to use.
JDBC_CONNECTION_URL - The URL for connecting to a JDBC data source.
JDBC_ENFORCE_SSL - A Boolean string (true, false) specifying whether Secure Sockets Layer (SSL) with hostname matching is enforced for the JDBC connection on the client. The default is false.
CUSTOM_JDBC_CERT - An Amazon S3 location specifying the customer's root certificate. Glue uses this root certificate to validate the customer’s certificate when connecting to the customer database. Glue only handles X.509 certificates. The certificate provided must be DER-encoded and supplied in Base64 encoding PEM format.
SKIP_CUSTOM_JDBC_CERT_VALIDATION - By default, this is false. Glue validates the Signature algorithm and Subject Public Key Algorithm for the customer certificate. The only permitted algorithms for the Signature algorithm are SHA256withRSA, SHA384withRSA or SHA512withRSA. For the Subject Public Key Algorithm, the key length must be at least 2048. You can set the value of this property to true to skip Glue’s validation of the customer certificate.
CUSTOM_JDBC_CERT_STRING - A custom JDBC certificate string which is used for domain match or distinguished name match to prevent a man-in-the-middle attack. In Oracle database, this is used as the SSL_SERVER_CERT_DN; in Microsoft SQL Server, this is used as the hostNameInCertificate.
CONNECTION_URL - The URL for connecting to a general (non-JDBC) data source.
SECRET_ID - The secret ID used for the secret manager of credentials.
CONNECTOR_URL - The connector URL for a MARKETPLACE or CUSTOM connection.
CONNECTOR_TYPE - The connector type for a MARKETPLACE or CUSTOM connection.
CONNECTOR_CLASS_NAME - The connector class name for a MARKETPLACE or CUSTOM connection.
KAFKA_BOOTSTRAP_SERVERS - A comma-separated list of host and port pairs that are the addresses of the Apache Kafka brokers in a Kafka cluster to which a Kafka client will connect to and bootstrap itself.
KAFKA_SSL_ENABLED - Whether to enable or disable SSL on an Apache Kafka connection. Default value is \"true\".
KAFKA_CUSTOM_CERT - The Amazon S3 URL for the private CA cert file (.pem format). The default is an empty string.
KAFKA_SKIP_CUSTOM_CERT_VALIDATION - Whether to skip the validation of the CA cert file or not. Glue validates for three algorithms: SHA256withRSA, SHA384withRSA and SHA512withRSA. Default value is \"false\".
KAFKA_CLIENT_KEYSTORE - The Amazon S3 location of the client keystore file for Kafka client side authentication (Optional).
KAFKA_CLIENT_KEYSTORE_PASSWORD - The password to access the provided keystore (Optional).
KAFKA_CLIENT_KEY_PASSWORD - A keystore can consist of multiple keys, so this is the password to access the client key to be used with the Kafka server side key (Optional).
ENCRYPTED_KAFKA_CLIENT_KEYSTORE_PASSWORD - The encrypted version of the Kafka client keystore password (if the user has the Glue encrypt passwords setting selected).
ENCRYPTED_KAFKA_CLIENT_KEY_PASSWORD - The encrypted version of the Kafka client key password (if the user has the Glue encrypt passwords setting selected).
KAFKA_SASL_MECHANISM - \"SCRAM-SHA-512\", \"GSSAPI\", or \"AWS_MSK_IAM\". These are the supported SASL Mechanisms.
KAFKA_SASL_SCRAM_USERNAME - A plaintext username used to authenticate with the \"SCRAM-SHA-512\" mechanism.
KAFKA_SASL_SCRAM_PASSWORD - A plaintext password used to authenticate with the \"SCRAM-SHA-512\" mechanism.
ENCRYPTED_KAFKA_SASL_SCRAM_PASSWORD - The encrypted version of the Kafka SASL SCRAM password (if the user has the Glue encrypt passwords setting selected).
KAFKA_SASL_SCRAM_SECRETS_ARN - The Amazon Resource Name of a secret in Amazon Web Services Secrets Manager.
KAFKA_SASL_GSSAPI_KEYTAB - The S3 location of a Kerberos keytab file. A keytab stores long-term keys for one or more principals. For more information, see MIT Kerberos Documentation: Keytab.
KAFKA_SASL_GSSAPI_KRB5_CONF - The S3 location of a Kerberos krb5.conf file. A krb5.conf stores Kerberos configuration information, such as the location of the KDC server. For more information, see MIT Kerberos Documentation: krb5.conf.
KAFKA_SASL_GSSAPI_SERVICE - The Kerberos service name, as set with sasl.kerberos.service.name in your Kafka Configuration.
The type of predefined worker that is allocated when a job runs. Accepts a value of Standard, G.1X, G.2X, or G.025X for Spark jobs. Accepts the value Z.2X for Ray jobs.
For the Standard worker type, each worker provides 4 vCPU, 16 GB of memory and a 50GB disk, and 2 executors per worker.
For the G.1X worker type, each worker maps to 1 DPU (4 vCPU, 16 GB of memory, 64 GB disk), and provides 1 executor per worker. We recommend this worker type for memory-intensive jobs.
For the G.2X worker type, each worker maps to 2 DPU (8 vCPU, 32 GB of memory, 128 GB disk), and provides 1 executor per worker. We recommend this worker type for memory-intensive jobs.
For the G.025X worker type, each worker maps to 0.25 DPU (2 vCPU, 4 GB of memory, 64 GB disk), and provides 1 executor per worker. We recommend this worker type for low volume streaming jobs. This worker type is only available for Glue version 3.0 streaming jobs.
For the Z.2X worker type, each worker maps to 2 M-DPU (8vCPU, 64 GB of m emory, 128 GB disk), and provides up to 8 Ray workers based on the autoscaler.
"
+ "documentation":"
The type of predefined worker that is allocated when a job runs. Accepts a value of G.1X, G.2X, G.4X, G.8X or G.025X for Spark jobs. Accepts the value Z.2X for Ray jobs.
For the G.1X worker type, each worker maps to 1 DPU (4 vCPUs, 16 GB of memory) with 84GB disk (approximately 34GB free), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.2X worker type, each worker maps to 2 DPU (8 vCPUs, 32 GB of memory) with 128GB disk (approximately 77GB free), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.4X worker type, each worker maps to 4 DPU (16 vCPUs, 64 GB of memory) with 256GB disk (approximately 235GB free), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs in the following Amazon Web Services Regions: US East (Ohio), US East (N. Virginia), US West (Oregon), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Canada (Central), Europe (Frankfurt), Europe (Ireland), and Europe (Stockholm).
For the G.8X worker type, each worker maps to 8 DPU (32 vCPUs, 128 GB of memory) with 512GB disk (approximately 487GB free), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs, in the same Amazon Web Services Regions as supported for the G.4X worker type.
For the G.025X worker type, each worker maps to 0.25 DPU (2 vCPUs, 4 GB of memory) with 84GB disk (approximately 34GB free), and provides 1 executor per worker. We recommend this worker type for low volume streaming jobs. This worker type is only available for Glue version 3.0 streaming jobs.
For the Z.2X worker type, each worker maps to 2 M-DPU (8vCPUs, 64 GB of memory) with 128 GB disk (approximately 120GB free), and provides up to 8 Ray workers based on the autoscaler.
The type of predefined worker that is allocated to use for the session. Accepts a value of Standard, G.1X, G.2X, or G.025X.
For the Standard worker type, each worker provides 4 vCPU, 16 GB of memory and a 50GB disk, and 2 executors per worker.
For the G.1X worker type, each worker maps to 1 DPU (4 vCPU, 16 GB of memory, 64 GB disk), and provides 1 executor per worker. We recommend this worker type for memory-intensive jobs.
For the G.2X worker type, each worker maps to 2 DPU (8 vCPU, 32 GB of memory, 128 GB disk), and provides 1 executor per worker. We recommend this worker type for memory-intensive jobs.
For the G.025X worker type, each worker maps to 0.25 DPU (2 vCPU, 4 GB of memory, 64 GB disk), and provides 1 executor per worker. We recommend this worker type for low volume streaming jobs. This worker type is only available for Glue version 3.0 streaming jobs.
"
+ "documentation":"
The type of predefined worker that is allocated when a job runs. Accepts a value of G.1X, G.2X, G.4X, or G.8X for Spark jobs. Accepts the value Z.2X for Ray notebooks.
For the G.1X worker type, each worker maps to 1 DPU (4 vCPUs, 16 GB of memory) with 84GB disk (approximately 34GB free), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.2X worker type, each worker maps to 2 DPU (8 vCPUs, 32 GB of memory) with 128GB disk (approximately 77GB free), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.4X worker type, each worker maps to 4 DPU (16 vCPUs, 64 GB of memory) with 256GB disk (approximately 235GB free), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs in the following Amazon Web Services Regions: US East (Ohio), US East (N. Virginia), US West (Oregon), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Canada (Central), Europe (Frankfurt), Europe (Ireland), and Europe (Stockholm).
For the G.8X worker type, each worker maps to 8 DPU (32 vCPUs, 128 GB of memory) with 512GB disk (approximately 487GB free), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs, in the same Amazon Web Services Regions as supported for the G.4X worker type.
For the Z.2X worker type, each worker maps to 2 M-DPU (8vCPUs, 64 GB of memory) with 128 GB disk (approximately 120GB free), and provides up to 8 Ray workers based on the autoscaler.
An array of Amazon S3 location strings for Hudi, each indicating the root folder with which the metadata files for a Hudi table resides. The Hudi folder may be located in a child folder of the root folder.
The crawler will scan all folders underneath a path for a Hudi folder.
The name of the connection to use to connect to the Hudi target. If your Hudi files are stored in buckets that require VPC authorization, you can set their connection properties here.
The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Hudi metadata folder in your Amazon S3 path. Used to limit the crawler run time.
The type of predefined worker that is allocated when a job runs. Accepts a value of Standard, G.1X, G.2X, G.4X, G.8X, or G.025X for Spark jobs. Accepts the value Z.2X for Ray jobs.
For the Standard worker type, each worker provides 4 vCPU, 16 GB of memory and a 50GB disk, and 2 executors per worker.
For the G.1X worker type, each worker maps to 1 DPU (4 vCPU, 16 GB of memory, 64 GB disk), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.2X worker type, each worker maps to 2 DPU (8 vCPU, 32 GB of memory, 128 GB disk), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.4X worker type, each worker maps to 4 DPU (16 vCPU, 64 GB of memory, 256 GB disk), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs in the following Amazon Web Services Regions: US East (Ohio), US East (N. Virginia), US West (Oregon), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Canada (Central), Europe (Frankfurt), Europe (Ireland), and Europe (Stockholm).
For the G.8X worker type, each worker maps to 8 DPU (32 vCPU, 128 GB of memory, 512 GB disk), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs, in the same Amazon Web Services Regions as supported for the G.4X worker type.
For the G.025X worker type, each worker maps to 0.25 DPU (2 vCPU, 4 GB of memory, 64 GB disk), and provides 1 executor per worker. We recommend this worker type for low volume streaming jobs. This worker type is only available for Glue version 3.0 streaming jobs.
For the Z.2X worker type, each worker maps to 2 M-DPU (8vCPU, 64 GB of m emory, 128 GB disk), and provides a default of 8 Ray workers (1 per vCPU).
"
+ "documentation":"
The type of predefined worker that is allocated when a job runs. Accepts a value of G.1X, G.2X, G.4X, G.8X or G.025X for Spark jobs. Accepts the value Z.2X for Ray jobs.
For the G.1X worker type, each worker maps to 1 DPU (4 vCPUs, 16 GB of memory) with 84GB disk (approximately 34GB free), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.2X worker type, each worker maps to 2 DPU (8 vCPUs, 32 GB of memory) with 128GB disk (approximately 77GB free), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.4X worker type, each worker maps to 4 DPU (16 vCPUs, 64 GB of memory) with 256GB disk (approximately 235GB free), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs in the following Amazon Web Services Regions: US East (Ohio), US East (N. Virginia), US West (Oregon), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Canada (Central), Europe (Frankfurt), Europe (Ireland), and Europe (Stockholm).
For the G.8X worker type, each worker maps to 8 DPU (32 vCPUs, 128 GB of memory) with 512GB disk (approximately 487GB free), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs, in the same Amazon Web Services Regions as supported for the G.4X worker type.
For the G.025X worker type, each worker maps to 0.25 DPU (2 vCPUs, 4 GB of memory) with 84GB disk (approximately 34GB free), and provides 1 executor per worker. We recommend this worker type for low volume streaming jobs. This worker type is only available for Glue version 3.0 streaming jobs.
For the Z.2X worker type, each worker maps to 2 M-DPU (8vCPUs, 64 GB of memory) with 128 GB disk (approximately 120GB free), and provides up to 8 Ray workers based on the autoscaler.
The type of predefined worker that is allocated when a job runs. Accepts a value of Standard, G.1X, G.2X, or G.025X for Spark jobs. Accepts the value Z.2X for Ray jobs.
For the Standard worker type, each worker provides 4 vCPU, 16 GB of memory and a 50GB disk, and 2 executors per worker.
For the G.1X worker type, each worker maps to 1 DPU (4 vCPU, 16 GB of memory, 64 GB disk), and provides 1 executor per worker. We recommend this worker type for memory-intensive jobs.
For the G.2X worker type, each worker maps to 2 DPU (8 vCPU, 32 GB of memory, 128 GB disk), and provides 1 executor per worker. We recommend this worker type for memory-intensive jobs.
For the G.025X worker type, each worker maps to 0.25 DPU (2 vCPU, 4 GB of memory, 64 GB disk), and provides 1 executor per worker. We recommend this worker type for low volume streaming jobs. This worker type is only available for Glue version 3.0 streaming jobs.
For the Z.2X worker type, each worker maps to 2 M-DPU (8vCPU, 64 GB of m emory, 128 GB disk), and provides up to 8 Ray workers (one per vCPU) based on the autoscaler.
"
+ "documentation":"
The type of predefined worker that is allocated when a job runs. Accepts a value of G.1X, G.2X, G.4X, G.8X or G.025X for Spark jobs. Accepts the value Z.2X for Ray jobs.
For the G.1X worker type, each worker maps to 1 DPU (4 vCPUs, 16 GB of memory) with 84GB disk (approximately 34GB free), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.2X worker type, each worker maps to 2 DPU (8 vCPUs, 32 GB of memory) with 128GB disk (approximately 77GB free), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.4X worker type, each worker maps to 4 DPU (16 vCPUs, 64 GB of memory) with 256GB disk (approximately 235GB free), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs in the following Amazon Web Services Regions: US East (Ohio), US East (N. Virginia), US West (Oregon), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Canada (Central), Europe (Frankfurt), Europe (Ireland), and Europe (Stockholm).
For the G.8X worker type, each worker maps to 8 DPU (32 vCPUs, 128 GB of memory) with 512GB disk (approximately 487GB free), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs, in the same Amazon Web Services Regions as supported for the G.4X worker type.
For the G.025X worker type, each worker maps to 0.25 DPU (2 vCPUs, 4 GB of memory) with 84GB disk (approximately 34GB free), and provides 1 executor per worker. We recommend this worker type for low volume streaming jobs. This worker type is only available for Glue version 3.0 streaming jobs.
For the Z.2X worker type, each worker maps to 2 M-DPU (8vCPUs, 64 GB of memory) with 128 GB disk (approximately 120GB free), and provides up to 8 Ray workers based on the autoscaler.
The type of predefined worker that is allocated when a job runs. Accepts a value of Standard, G.1X, G.2X, or G.025X for Spark jobs. Accepts the value Z.2X for Ray jobs.
For the Standard worker type, each worker provides 4 vCPU, 16 GB of memory and a 50GB disk, and 2 executors per worker.
For the G.1X worker type, each worker maps to 1 DPU (4 vCPU, 16 GB of memory, 64 GB disk), and provides 1 executor per worker. We recommend this worker type for memory-intensive jobs.
For the G.2X worker type, each worker maps to 2 DPU (8 vCPU, 32 GB of memory, 128 GB disk), and provides 1 executor per worker. We recommend this worker type for memory-intensive jobs.
For the G.025X worker type, each worker maps to 0.25 DPU (2 vCPU, 4 GB of memory, 64 GB disk), and provides 1 executor per worker. We recommend this worker type for low volume streaming jobs. This worker type is only available for Glue version 3.0 streaming jobs.
For the Z.2X worker type, each worker maps to 2 M-DPU (8vCPU, 64 GB of m emory, 128 GB disk), and provides up to 8 Ray workers based on the autoscaler.
"
+ "documentation":"
The type of predefined worker that is allocated when a job runs. Accepts a value of G.1X, G.2X, G.4X, G.8X or G.025X for Spark jobs. Accepts the value Z.2X for Ray jobs.
For the G.1X worker type, each worker maps to 1 DPU (4 vCPUs, 16 GB of memory) with 84GB disk (approximately 34GB free), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.2X worker type, each worker maps to 2 DPU (8 vCPUs, 32 GB of memory) with 128GB disk (approximately 77GB free), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.4X worker type, each worker maps to 4 DPU (16 vCPUs, 64 GB of memory) with 256GB disk (approximately 235GB free), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs in the following Amazon Web Services Regions: US East (Ohio), US East (N. Virginia), US West (Oregon), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Canada (Central), Europe (Frankfurt), Europe (Ireland), and Europe (Stockholm).
For the G.8X worker type, each worker maps to 8 DPU (32 vCPUs, 128 GB of memory) with 512GB disk (approximately 487GB free), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs, in the same Amazon Web Services Regions as supported for the G.4X worker type.
For the G.025X worker type, each worker maps to 0.25 DPU (2 vCPUs, 4 GB of memory) with 84GB disk (approximately 34GB free), and provides 1 executor per worker. We recommend this worker type for low volume streaming jobs. This worker type is only available for Glue version 3.0 streaming jobs.
For the Z.2X worker type, each worker maps to 2 M-DPU (8vCPUs, 64 GB of memory) with 128 GB disk (approximately 120GB free), and provides up to 8 Ray workers based on the autoscaler.
Used when Action is append. Specifies the resolution behavior when a row already exists. If true, preexisting rows will be updated. If false, those rows will be inserted.
Specifies a merge action. Valid values: simple, custom. If simple, merge behavior is defined by MergeWhenMatched and MergeWhenNotMatched. If custom, defined by MergeClause.
The name of a staging table used when performing merge or upsert append actions. Data is written to this table, then moved to table by a generated postaction.
Specifies the columns combined to identify a record when detecting matches for merges and upserts. A list of structures with value, label and description keys. Each structure describes a column.
Specifies whether automatic query pushdown is enabled. If pushdown is enabled, then when a query is run on Spark, if part of the query can be \"pushed down\" to the Snowflake server, it is pushed down. This improves performance of some queries.
The type of predefined worker that is allocated when a job runs. Accepts a value of Standard, G.1X, G.2X, or G.025X for Spark jobs. Accepts the value Z.2X for Ray jobs.
For the Standard worker type, each worker provides 4 vCPU, 16 GB of memory and a 50GB disk, and 2 executors per worker.
For the G.1X worker type, each worker maps to 1 DPU (4 vCPU, 16 GB of memory, 64 GB disk), and provides 1 executor per worker. We recommend this worker type for memory-intensive jobs.
For the G.2X worker type, each worker maps to 2 DPU (8 vCPU, 32 GB of memory, 128 GB disk), and provides 1 executor per worker. We recommend this worker type for memory-intensive jobs.
For the G.025X worker type, each worker maps to 0.25 DPU (2 vCPU, 4 GB of memory, 64 GB disk), and provides 1 executor per worker. We recommend this worker type for low volume streaming jobs. This worker type is only available for Glue version 3.0 streaming jobs.
For the Z.2X worker type, each worker maps to 2 DPU (8vCPU, 64 GB of m emory, 128 GB disk), and provides up to 8 Ray workers (one per vCPU) based on the autoscaler.
"
+ "documentation":"
The type of predefined worker that is allocated when a job runs. Accepts a value of G.1X, G.2X, G.4X, G.8X or G.025X for Spark jobs. Accepts the value Z.2X for Ray jobs.
For the G.1X worker type, each worker maps to 1 DPU (4 vCPUs, 16 GB of memory) with 84GB disk (approximately 34GB free), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.2X worker type, each worker maps to 2 DPU (8 vCPUs, 32 GB of memory) with 128GB disk (approximately 77GB free), and provides 1 executor per worker. We recommend this worker type for workloads such as data transforms, joins, and queries, to offers a scalable and cost effective way to run most jobs.
For the G.4X worker type, each worker maps to 4 DPU (16 vCPUs, 64 GB of memory) with 256GB disk (approximately 235GB free), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs in the following Amazon Web Services Regions: US East (Ohio), US East (N. Virginia), US West (Oregon), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Canada (Central), Europe (Frankfurt), Europe (Ireland), and Europe (Stockholm).
For the G.8X worker type, each worker maps to 8 DPU (32 vCPUs, 128 GB of memory) with 512GB disk (approximately 487GB free), and provides 1 executor per worker. We recommend this worker type for jobs whose workloads contain your most demanding transforms, aggregations, joins, and queries. This worker type is available only for Glue version 3.0 or later Spark ETL jobs, in the same Amazon Web Services Regions as supported for the G.4X worker type.
For the G.025X worker type, each worker maps to 0.25 DPU (2 vCPUs, 4 GB of memory) with 84GB disk (approximately 34GB free), and provides 1 executor per worker. We recommend this worker type for low volume streaming jobs. This worker type is only available for Glue version 3.0 streaming jobs.
For the Z.2X worker type, each worker maps to 2 M-DPU (8vCPUs, 64 GB of memory) with 128 GB disk (approximately 120GB free), and provides up to 8 Ray workers based on the autoscaler.
Lists tags for a resource. Tagging is currently supported for detectors, finding filters, IP sets, threat intel sets, publishing destination, with a limit of 50 tags per resource. When invoked, this operation returns all assigned tags for a given resource.
"
+ "documentation":"
Lists tags for a resource. Tagging is currently supported for detectors, finding filters, IP sets, threat intel sets, and publishing destination, with a limit of 50 tags per each resource. When invoked, this operation returns all assigned tags for a given resource.
Gets the properties associated with the FHIR Data Store, including the Data Store ID, Data Store ARN, Data Store name, Data Store status, created at, Data Store type version, and Data Store endpoint.
"
+ "documentation":"
Gets the properties associated with the FHIR data store, including the data store ID, data store ARN, data store name, data store status, when the data store was created, data store type version, and the data store's endpoint.
Optional parameter to preload data upon creation of the Data Store. Currently, the only supported preloaded data is synthetic data generated from Synthea.
"
+ "documentation":"
Optional parameter to preload data upon creation of the data store. Currently, the only supported preloaded data is synthetic data generated from Synthea.
All properties associated with a Data Store, including the Data Store ID, Data Store ARN, Data Store name, Data Store status, created at, Data Store type version, and Data Store endpoint.
"
+ "documentation":"
All properties associated with a data store, including the data store ID, data store ARN, data store name, data store status, when the data store was created, data store type version, and the data store's endpoint.
Amazon HealthLake is a HIPAA eligibile service that allows customers to store, transform, query, and analyze their FHIR-formatted data in a consistent fashion in the cloud.
"
+ "documentation":"
AWS HealthLake is a HIPAA eligibile service that allows customers to store, transform, query, and analyze their FHIR-formatted data in a consistent fashion in the cloud.
Creates a monitor in Amazon CloudWatch Internet Monitor. A monitor is built based on information from the application resources that you add: Amazon Virtual Private Clouds (VPCs), Amazon CloudFront distributions, and WorkSpaces directories. Internet Monitor then publishes internet measurements from Amazon Web Services that are specific to the city-networks, that is, the locations and ASNs (typically internet service providers or ISPs), where clients access your application. For more information, see Using Amazon CloudWatch Internet Monitor in the Amazon CloudWatch User Guide.
When you create a monitor, you set a maximum limit for the number of city-networks where client traffic is monitored. The city-network maximum that you choose is the limit, but you only pay for the number of city-networks that are actually monitored. You can change the maximum at any time by updating your monitor. For more information, see Choosing a city-network maximum value in the Amazon CloudWatch User Guide.
",
+ "documentation":"
Creates a monitor in Amazon CloudWatch Internet Monitor. A monitor is built based on information from the application resources that you add: VPCs, Network Load Balancers (NLBs), Amazon CloudFront distributions, and Amazon WorkSpaces directories. Internet Monitor then publishes internet measurements from Amazon Web Services that are specific to the city-networks. That is, the locations and ASNs (typically internet service providers or ISPs), where clients access your application. For more information, see Using Amazon CloudWatch Internet Monitor in the Amazon CloudWatch User Guide.
When you create a monitor, you choose the percentage of traffic that you want to monitor. You can also set a maximum limit for the number of city-networks where client traffic is monitored, that caps the total traffic that Internet Monitor monitors. A city-network maximum is the limit of city-networks, but you only pay for the number of city-networks that are actually monitored. You can update your monitor at any time to change the percentage of traffic to monitor or the city-networks maximum. For more information, see Choosing a city-network maximum value in the Amazon CloudWatch User Guide.
Gets information the Amazon CloudWatch Internet Monitor has created and stored about a health event for a specified monitor. This information includes the impacted locations, and all of the information related to the event by location.
The information returned includes the performance, availability, and round-trip time impact, information about the network providers, the event type, and so on.
Information rolled up at the global traffic level is also returned, including the impact type and total traffic impact.
"
+ "documentation":"
Gets information the Amazon CloudWatch Internet Monitor has created and stored about a health event for a specified monitor. This information includes the impacted locations, and all the information related to the event, by location.
The information returned includes the impact on performance, availability, and round-trip time, information about the network providers (ASNs), the event type, and so on.
Information rolled up at the global traffic level is also returned, including the impact type and total traffic impact.
Lists all health events for a monitor in Amazon CloudWatch Internet Monitor. Returns all information for health events including the client location information the network cause and status, event start and end time, percentage of total traffic impacted, and status.
Health events that have start times during the time frame that is requested are not included in the list of health events.
"
+ "documentation":"
Lists all health events for a monitor in Amazon CloudWatch Internet Monitor. Returns information for health events including the event start and end time and the status.
Health events that have start times during the time frame that is requested are not included in the list of health events.
Updates a monitor. You can update a monitor to change the maximum number of city-networks (locations and ASNs or internet service providers), to add or remove resources, or to change the status of the monitor. Note that you can't change the name of a monitor.
The city-network maximum that you choose is the limit, but you only pay for the number of city-networks that are actually monitored. For more information, see Choosing a city-network maximum value in the Amazon CloudWatch User Guide.
",
+ "documentation":"
Updates a monitor. You can update a monitor to change the percentage of traffic to monitor or the maximum number of city-networks (locations and ASNs), to add or remove resources, or to change the status of the monitor. Note that you can't change the name of a monitor.
The city-network maximum that you choose is the limit, but you only pay for the number of city-networks that are actually monitored. For more information, see Choosing a city-network maximum value in the Amazon CloudWatch User Guide.
The percentage of impact caused by a health event for total traffic globally.
For information about how Internet Monitor calculates impact, see Inside Internet Monitor in the Amazon CloudWatch Internet Monitor section of the Amazon CloudWatch User Guide.
"
+ "documentation":"
The impact on total traffic that a health event has, in increased latency or reduced availability. This is the percentage of how much latency has increased or availability has decreased during the event, compared to what is typical for traffic from this client location to the Amazon Web Services location using this client network.
For information about how Internet Monitor calculates impact, see How Internet Monitor works in the Amazon CloudWatch Internet Monitor section of the Amazon CloudWatch User Guide.
The percentage of impact caused by a health event for client location traffic globally.
For information about how Internet Monitor calculates impact, see Inside Internet Monitor in the Amazon CloudWatch Internet Monitor section of the Amazon CloudWatch User Guide.
"
}
},
- "documentation":"
Measurements about the availability for your application on the internet, calculated by Amazon CloudWatch Internet Monitor. Amazon Web Services has substantial historical data about internet performance and availability between Amazon Web Services services and different network providers and geographies. By applying statistical analysis to the data, Internet Monitor can detect when the performance and availability for your application has dropped, compared to an estimated baseline that's already calculated. To make it easier to see those drops, we report that information to you in the form of health scores: a performance score and an availability score.
Availability in Internet Monitor represents the estimated percentage of traffic that is not seeing an availability drop. For example, an availability score of 99% for an end user and service location pair is equivalent to 1% of the traffic experiencing an availability drop for that pair.
Amazon CloudWatch Internet Monitor calculates measurements about the availability for your application's internet traffic between client locations and Amazon Web Services. Amazon Web Services has substantial historical data about internet performance and availability between Amazon Web Services services and different network providers and geographies. By applying statistical analysis to the data, Internet Monitor can detect when the performance and availability for your application has dropped, compared to an estimated baseline that's already calculated. To make it easier to see those drops, we report that information to you in the form of health scores: a performance score and an availability score.
Availability in Internet Monitor represents the estimated percentage of traffic that is not seeing an availability drop. For example, an availability score of 99% for an end user and service location pair is equivalent to 1% of the traffic experiencing an availability drop for that pair.
The resources to include in a monitor, which you provide as a set of Amazon Resource Names (ARNs).
You can add a combination of Amazon Virtual Private Clouds (VPCs) and Amazon CloudFront distributions, or you can add Amazon WorkSpaces directories. You can't add all three types of resources.
If you add only VPC resources, at least one VPC must have an Internet Gateway attached to it, to make sure that it has internet connectivity.
"
+ "documentation":"
The resources to include in a monitor, which you provide as a set of Amazon Resource Names (ARNs). Resources can be VPCs, NLBs, Amazon CloudFront distributions, or Amazon WorkSpaces directories.
You can add a combination of VPCs and CloudFront distributions, or you can add WorkSpaces directories, or you can add NLBs. You can't add NLBs or WorkSpaces directories together with any other resources.
If you add only Amazon VPC resources, at least one VPC must have an Internet Gateway attached to it, to make sure that it has internet connectivity.
The maximum number of city-networks to monitor for your resources. A city-network is the location (city) where clients access your application resources from and the network or ASN, such as an internet service provider (ISP), that clients access the resources through. This limit helps control billing costs.
The maximum number of city-networks to monitor for your resources. A city-network is the location (city) where clients access your application resources from and the ASN or network provider, such as an internet service provider (ISP), that clients access the resources through. Setting this limit can help control billing costs.
The percentage of the internet-facing traffic for your application that you want to monitor with this monitor.
"
+ "documentation":"
The percentage of the internet-facing traffic for your application that you want to monitor with this monitor. If you set a city-networks maximum, that limit overrides the traffic percentage that you set.
Defines the health event threshold percentages, for performance score and availability score. Internet Monitor creates a health event when there's an internet issue that affects your application end users where a health score percentage is at or below a set threshold. If you don't set a health event threshold, the default calue is 95%.
"
+ "documentation":"
Defines the threshold percentages and other configuration information for when Amazon CloudWatch Internet Monitor creates a health event. Internet Monitor creates a health event when an internet issue that affects your application end users has a health score percentage that is at or below a specific threshold, and, sometimes, when other criteria are met.
If you don't set a health event threshold, the default value is 95%.
The impact on total traffic that a health event has.
"
+ "documentation":"
The impact on total traffic that a health event has, in increased latency or reduced availability. This is the percentage of how much latency has increased or availability has decreased during the event, compared to what is typical for traffic from this client location to the Amazon Web Services location using this client network.
The threshold percentage for health events when Amazon CloudWatch Internet Monitor creates a health event.
"
+ "documentation":"
The threshold percentage for a health score that determines, along with other configuration information, when Internet Monitor creates a health event when there's an internet issue that affects your application end users.
The maximum number of city-networks to monitor for your resources. A city-network is the location (city) where clients access your application resources from and the network or ASN, such as an internet service provider (ISP), that clients access the resources through. This limit helps control billing costs.
The maximum number of city-networks to monitor for your resources. A city-network is the location (city) where clients access your application resources from and the ASN or network provider, such as an internet service provider (ISP), that clients access the resources through. This limit can help control billing costs.
The percentage of the internet-facing traffic for your application that you want to monitor with this monitor.
"
+ "documentation":"
The percentage of the internet-facing traffic for your application to monitor with this monitor. If you set a city-networks maximum, that limit overrides the traffic percentage that you set.
The list of health event thresholds. A health event threshold percentage, for performance and availability, determines the level of impact at which Amazon CloudWatch Internet Monitor creates a health event when there's an internet issue that affects your application end users.
"
+ "documentation":"
The list of health event threshold configurations. The threshold percentage for a health score determines, along with other configuration information, when Internet Monitor creates a health event when there's an internet issue that affects your application end users.
The impact on global traffic monitored by this monitor for this health event.
"
+ "documentation":"
The impact on total traffic that a health event has, in increased latency or reduced availability. This is the percentage of how much latency has increased or availability has decreased during the event, compared to what is typical for traffic from this client location to the Amazon Web Services location using this client network.
The configuration that determines the threshold and other conditions for when Internet Monitor creates a health event for a local performance issue.
"
}
},
- "documentation":"
A complex type for the configuration. Defines the health event threshold percentages, for performance score and availability score. Amazon CloudWatch Internet Monitor creates a health event when there's an internet issue that affects your application end users where a health score percentage is at or below a set threshold. If you don't set a health event threshold, the default value is 95%.
"
+ "documentation":"
A complex type with the configuration information that determines the threshold and other conditions for when Internet Monitor creates a health event for an overall performance or availability issue, across an application's geographies.
Defines the percentages, for overall performance scores and availability scores for an application, that are the thresholds for when Amazon CloudWatch Internet Monitor creates a health event. You can override the defaults to set a custom threshold for overall performance or availability scores, or both.
You can also set thresholds for local health scores,, where Internet Monitor creates a health event when scores cross a threshold for one or more city-networks, in addition to creating an event when an overall score crosses a threshold.
If you don't set a health event threshold, the default value is 95%.
For local thresholds, you also set a minimum percentage of overall traffic that is impacted by an issue before Internet Monitor creates an event. In addition, you can disable local thresholds, for performance scores, availability scores, or both.
Performance in Internet Monitor represents the estimated percentage of traffic that is not seeing a performance drop. For example, a performance score of 99% for an end user and service location pair is equivalent to 1% of the traffic experiencing a performance drop for that pair.
Internet health includes measurements calculated by Amazon CloudWatch Internet Monitor about the performance and availability for your application on the internet. Amazon Web Services has substantial historical data about internet performance and availability between Amazon Web Services services and different network providers and geographies. By applying statistical analysis to the data, Internet Monitor can detect when the performance and availability for your application has dropped, compared to an estimated baseline that's already calculated. To make it easier to see those drops, we report that information to you in the form of health scores: a performance score and an availability score.
"
+ "documentation":"
Internet health includes measurements calculated by Amazon CloudWatch Internet Monitor about the performance and availability for your application on the internet. Amazon Web Services has substantial historical data about internet performance and availability between Amazon Web Services services and different network providers and geographies. By applying statistical analysis to the data, Internet Monitor can detect when the performance and availability for your application has dropped, compared to an estimated baseline that's already calculated. To make it easier to see those drops, Internet Monitor reports the information to you in the form of health scores: a performance score and an availability score.
The status of whether Internet Monitor creates a health event based on a threshold percentage set for a local health score. The status can be ENABLED or DISABLED.
The minimum percentage of overall traffic for an application that must be impacted by an issue before Internet Monitor creates an event when a threshold is crossed for a local health score.
"
+ }
+ },
+ "documentation":"
A complex type with the configuration information that determines the threshold and other conditions for when Internet Monitor creates a health event for a local performance or availability issue, when scores cross a threshold for one or more city-networks.
Defines the percentages, for performance scores or availability scores, that are the local thresholds for when Amazon CloudWatch Internet Monitor creates a health event. Also defines whether a local threshold is enabled or disabled, and the minimum percentage of overall traffic that must be impacted by an issue before Internet Monitor creates an event when a threshold is crossed for a local health score.
How much performance impact was caused by a health event for total traffic globally. For performance, this is the percentage of how much latency increased during the event compared to typical performance for your application traffic globally.
The impact on total traffic that a health event has, in increased latency or reduced availability. This is the percentage of how much latency has increased or availability has decreased during the event, compared to what is typical for traffic from this client location to the Amazon Web Services location using this client network.
Measurements about the performance for your application on the internet calculated by Amazon CloudWatch Internet Monitor. Amazon Web Services has substantial historical data about internet performance and availability between Amazon Web Services services and different network providers and geographies. By applying statistical analysis to the data, Internet Monitor can detect when the performance and availability for your application has dropped, compared to an estimated baseline that's already calculated. To make it easier to see those drops, we report that information to you in the form of health scores: a performance score and an availability score.
Performance in Internet Monitor represents the estimated percentage of traffic that is not seeing a performance drop. For example, a performance score of 99% for an end user and service location pair is equivalent to 1% of the traffic experiencing a performance drop for that pair.
Amazon CloudWatch Internet Monitor calculates measurements about the performance for your application's internet traffic between client locations and Amazon Web Services. Amazon Web Services has substantial historical data about internet performance and availability between Amazon Web Services services and different network providers and geographies. By applying statistical analysis to the data, Internet Monitor can detect when the performance and availability for your application has dropped, compared to an estimated baseline that's already calculated. To make it easier to see those drops, we report that information to you in the form of health scores: a performance score and an availability score.
Performance in Internet Monitor represents the estimated percentage of traffic that is not seeing a performance drop. For example, a performance score of 99% for an end user and service location pair is equivalent to 1% of the traffic experiencing a performance drop for that pair.
The resources to include in a monitor, which you provide as a set of Amazon Resource Names (ARNs).
You can add a combination of Amazon Virtual Private Clouds (VPCs) and Amazon CloudFront distributions, or you can add Amazon WorkSpaces directories. You can't add all three types of resources.
If you add only VPC resources, at least one VPC must have an Internet Gateway attached to it, to make sure that it has internet connectivity.
"
+ "documentation":"
The resources to include in a monitor, which you provide as a set of Amazon Resource Names (ARNs). Resources can be VPCs, NLBs, Amazon CloudFront distributions, or Amazon WorkSpaces directories.
You can add a combination of VPCs and CloudFront distributions, or you can add WorkSpaces directories, or you can add NLBs. You can't add NLBs or WorkSpaces directories together with any other resources.
If you add only Amazon Virtual Private Clouds resources, at least one VPC must have an Internet Gateway attached to it, to make sure that it has internet connectivity.
The maximum number of city-networks to monitor for your resources. A city-network is the location (city) where clients access your application resources from and the network or ASN, such as an internet service provider, that clients access the resources through.
"
+ "documentation":"
The maximum number of city-networks to monitor for your application. A city-network is the location (city) where clients access your application resources from and the ASN or network provider, such as an internet service provider (ISP), that clients access the resources through. Setting this limit can help control billing costs.
The percentage of the internet-facing traffic for your application that you want to monitor with this monitor.
"
+ "documentation":"
The percentage of the internet-facing traffic for your application that you want to monitor with this monitor. If you set a city-networks maximum, that limit overrides the traffic percentage that you set.
The list of health event thresholds. A health event threshold percentage, for performance and availability, determines when Internet Monitor creates a health event when there's an internet issue that affects your application end users.
"
+ "documentation":"
The list of health score thresholds. A threshold percentage for health scores, along with other configuration information, determines when Internet Monitor creates a health event when there's an internet issue that affects your application end users.
Amazon CloudWatch Internet Monitor provides visibility into how internet issues impact the performance and availability between your applications hosted on Amazon Web Services and your end users. It reduces the time it takes for you to diagnose internet issues from days to minutes. Internet Monitor uses the connectivity data that Amazon Web Services captures from its global networking footprint to calculate a baseline of performance and availability for internet traffic. This is the same data that Amazon Web Services uses to monitor internet uptime and availability. With those measurements as a baseline, Internet Monitor raises awareness for you when there are significant problems for your end users in the different geographic locations where your application runs.
Internet Monitor publishes internet measurements to CloudWatch Logs and CloudWatch Metrics, to easily support using CloudWatch tools with health information for geographies and networks specific to your application. Internet Monitor sends health events to Amazon EventBridge so that you can set up notifications. If an issue is caused by the Amazon Web Services network, you also automatically receive an Amazon Web Services Health Dashboard notification with the steps that Amazon Web Services is taking to mitigate the problem.
To use Internet Monitor, you create a monitor and associate your application's resources with it, VPCs, CloudFront distributions, or WorkSpaces directories, to enable Internet Monitor to know where your application's internet traffic is. Internet Monitor then provides internet measurements from Amazon Web Services that are specific to the locations and networks that communicate with your application.
Amazon CloudWatch Internet Monitor provides visibility into how internet issues impact the performance and availability between your applications hosted on Amazon Web Services and your end users. It can reduce the time it takes for you to diagnose internet issues from days to minutes. Internet Monitor uses the connectivity data that Amazon Web Services captures from its global networking footprint to calculate a baseline of performance and availability for internet traffic. This is the same data that Amazon Web Services uses to monitor internet uptime and availability. With those measurements as a baseline, Internet Monitor raises awareness for you when there are significant problems for your end users in the different geographic locations where your application runs.
Internet Monitor publishes internet measurements to CloudWatch Logs and CloudWatch Metrics, to easily support using CloudWatch tools with health information for geographies and networks specific to your application. Internet Monitor sends health events to Amazon EventBridge so that you can set up notifications. If an issue is caused by the Amazon Web Services network, you also automatically receive an Amazon Web Services Health Dashboard notification with the steps that Amazon Web Services is taking to mitigate the problem.
To use Internet Monitor, you create a monitor and associate your application's resources with it - VPCs, NLBs, CloudFront distributions, or WorkSpaces directories - so Internet Monitor can determine where your application's internet traffic is. Internet Monitor then provides internet measurements from Amazon Web Services that are specific to the locations and ASNs (typically, internet service providers or ISPs) that communicate with your application.
"
}
diff --git a/services/iot/pom.xml b/services/iot/pom.xml
index 65175d3b45b..49a840587a1 100644
--- a/services/iot/pom.xml
+++ b/services/iot/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiotAWS Java SDK :: Services :: AWS IoT
diff --git a/services/iot1clickdevices/pom.xml b/services/iot1clickdevices/pom.xml
index 345db741056..d7fef23a201 100644
--- a/services/iot1clickdevices/pom.xml
+++ b/services/iot1clickdevices/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiot1clickdevicesAWS Java SDK :: Services :: IoT 1Click Devices Service
diff --git a/services/iot1clickprojects/pom.xml b/services/iot1clickprojects/pom.xml
index 9b804f31be7..be8cd289fa6 100644
--- a/services/iot1clickprojects/pom.xml
+++ b/services/iot1clickprojects/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiot1clickprojectsAWS Java SDK :: Services :: IoT 1Click Projects
diff --git a/services/iotanalytics/pom.xml b/services/iotanalytics/pom.xml
index 3257f9c6fac..d4917204760 100644
--- a/services/iotanalytics/pom.xml
+++ b/services/iotanalytics/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiotanalyticsAWS Java SDK :: Services :: IoTAnalytics
diff --git a/services/iotdataplane/pom.xml b/services/iotdataplane/pom.xml
index 9c71b107772..3907df5f297 100644
--- a/services/iotdataplane/pom.xml
+++ b/services/iotdataplane/pom.xml
@@ -22,7 +22,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiotdataplaneAWS Java SDK :: Services :: AWS IoT Data Plane
diff --git a/services/iotdeviceadvisor/pom.xml b/services/iotdeviceadvisor/pom.xml
index 9198d93a98d..d9329f3b343 100644
--- a/services/iotdeviceadvisor/pom.xml
+++ b/services/iotdeviceadvisor/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiotdeviceadvisorAWS Java SDK :: Services :: Iot Device Advisor
diff --git a/services/iotevents/pom.xml b/services/iotevents/pom.xml
index 095cc9ea56d..fd0fb0d0722 100644
--- a/services/iotevents/pom.xml
+++ b/services/iotevents/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTioteventsAWS Java SDK :: Services :: IoT Events
diff --git a/services/ioteventsdata/pom.xml b/services/ioteventsdata/pom.xml
index d86f9302dd3..adba4e284f7 100644
--- a/services/ioteventsdata/pom.xml
+++ b/services/ioteventsdata/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTioteventsdataAWS Java SDK :: Services :: IoT Events Data
diff --git a/services/iotfleethub/pom.xml b/services/iotfleethub/pom.xml
index f5942d95292..809a9f85c9a 100644
--- a/services/iotfleethub/pom.xml
+++ b/services/iotfleethub/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiotfleethubAWS Java SDK :: Services :: Io T Fleet Hub
diff --git a/services/iotfleetwise/pom.xml b/services/iotfleetwise/pom.xml
index 26e85a0278c..7b12421d9a4 100644
--- a/services/iotfleetwise/pom.xml
+++ b/services/iotfleetwise/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiotfleetwiseAWS Java SDK :: Services :: Io T Fleet Wise
diff --git a/services/iotjobsdataplane/pom.xml b/services/iotjobsdataplane/pom.xml
index 0a478c6c67e..f5e4f9dbb0a 100644
--- a/services/iotjobsdataplane/pom.xml
+++ b/services/iotjobsdataplane/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiotjobsdataplaneAWS Java SDK :: Services :: IoT Jobs Data Plane
diff --git a/services/iotroborunner/pom.xml b/services/iotroborunner/pom.xml
index b18056ba29c..1dab7a56f86 100644
--- a/services/iotroborunner/pom.xml
+++ b/services/iotroborunner/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiotroborunnerAWS Java SDK :: Services :: IoT Robo Runner
diff --git a/services/iotsecuretunneling/pom.xml b/services/iotsecuretunneling/pom.xml
index 0a2ca5ed5ea..8dbff726202 100644
--- a/services/iotsecuretunneling/pom.xml
+++ b/services/iotsecuretunneling/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiotsecuretunnelingAWS Java SDK :: Services :: IoTSecureTunneling
diff --git a/services/iotsitewise/pom.xml b/services/iotsitewise/pom.xml
index 98085044812..98c61b98f5e 100644
--- a/services/iotsitewise/pom.xml
+++ b/services/iotsitewise/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiotsitewiseAWS Java SDK :: Services :: Io T Site Wise
diff --git a/services/iotthingsgraph/pom.xml b/services/iotthingsgraph/pom.xml
index 4c0132b9891..68abfb48dfa 100644
--- a/services/iotthingsgraph/pom.xml
+++ b/services/iotthingsgraph/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiotthingsgraphAWS Java SDK :: Services :: IoTThingsGraph
diff --git a/services/iottwinmaker/pom.xml b/services/iottwinmaker/pom.xml
index d38e5b47714..832a73a6296 100644
--- a/services/iottwinmaker/pom.xml
+++ b/services/iottwinmaker/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiottwinmakerAWS Java SDK :: Services :: Io T Twin Maker
diff --git a/services/iotwireless/pom.xml b/services/iotwireless/pom.xml
index a09e368b455..6d9977ddbda 100644
--- a/services/iotwireless/pom.xml
+++ b/services/iotwireless/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTiotwirelessAWS Java SDK :: Services :: IoT Wireless
diff --git a/services/ivs/pom.xml b/services/ivs/pom.xml
index 20a61d57c49..c7c8008a62c 100644
--- a/services/ivs/pom.xml
+++ b/services/ivs/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTivsAWS Java SDK :: Services :: Ivs
diff --git a/services/ivschat/pom.xml b/services/ivschat/pom.xml
index 1ec01195a32..654e44d2c1a 100644
--- a/services/ivschat/pom.xml
+++ b/services/ivschat/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTivschatAWS Java SDK :: Services :: Ivschat
diff --git a/services/ivsrealtime/pom.xml b/services/ivsrealtime/pom.xml
index 6e9dc94bb2c..655c74f7324 100644
--- a/services/ivsrealtime/pom.xml
+++ b/services/ivsrealtime/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTivsrealtimeAWS Java SDK :: Services :: IVS Real Time
diff --git a/services/ivsrealtime/src/main/resources/codegen-resources/endpoint-tests.json b/services/ivsrealtime/src/main/resources/codegen-resources/endpoint-tests.json
index aa2d6d179c1..d51b2c7cff9 100644
--- a/services/ivsrealtime/src/main/resources/codegen-resources/endpoint-tests.json
+++ b/services/ivsrealtime/src/main/resources/codegen-resources/endpoint-tests.json
@@ -1,54 +1,54 @@
{
"testCases": [
{
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://ivsrealtime-fips.us-gov-east-1.api.aws"
+ "url": "https://ivsrealtime-fips.us-east-1.api.aws"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": true,
- "Region": "us-gov-east-1",
"UseDualStack": true
}
},
{
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://ivsrealtime-fips.us-gov-east-1.amazonaws.com"
+ "url": "https://ivsrealtime-fips.us-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": true,
- "Region": "us-gov-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://ivsrealtime.us-gov-east-1.api.aws"
+ "url": "https://ivsrealtime.us-east-1.api.aws"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": false,
- "Region": "us-gov-east-1",
"UseDualStack": true
}
},
{
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://ivsrealtime.us-gov-east-1.amazonaws.com"
+ "url": "https://ivsrealtime.us-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": false,
- "Region": "us-gov-east-1",
"UseDualStack": false
}
},
@@ -60,8 +60,8 @@
}
},
"params": {
- "UseFIPS": true,
"Region": "cn-north-1",
+ "UseFIPS": true,
"UseDualStack": true
}
},
@@ -73,8 +73,8 @@
}
},
"params": {
- "UseFIPS": true,
"Region": "cn-north-1",
+ "UseFIPS": true,
"UseDualStack": false
}
},
@@ -86,8 +86,8 @@
}
},
"params": {
- "UseFIPS": false,
"Region": "cn-north-1",
+ "UseFIPS": false,
"UseDualStack": true
}
},
@@ -99,108 +99,108 @@
}
},
"params": {
- "UseFIPS": false,
"Region": "cn-north-1",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
+ "endpoint": {
+ "url": "https://ivsrealtime-fips.us-gov-east-1.api.aws"
+ }
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": true,
- "Region": "us-iso-east-1",
"UseDualStack": true
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://ivsrealtime-fips.us-iso-east-1.c2s.ic.gov"
+ "url": "https://ivsrealtime-fips.us-gov-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": true,
- "Region": "us-iso-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "error": "DualStack is enabled but this partition does not support DualStack"
+ "endpoint": {
+ "url": "https://ivsrealtime.us-gov-east-1.api.aws"
+ }
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": false,
- "Region": "us-iso-east-1",
"UseDualStack": true
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://ivsrealtime.us-iso-east-1.c2s.ic.gov"
+ "url": "https://ivsrealtime.us-gov-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": false,
- "Region": "us-iso-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://ivsrealtime-fips.us-east-1.api.aws"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": true,
- "Region": "us-east-1",
"UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://ivsrealtime-fips.us-east-1.amazonaws.com"
+ "url": "https://ivsrealtime-fips.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": true,
- "Region": "us-east-1",
"UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://ivsrealtime.us-east-1.api.aws"
- }
+ "error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": false,
- "Region": "us-east-1",
"UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://ivsrealtime.us-east-1.amazonaws.com"
+ "url": "https://ivsrealtime.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": false,
- "Region": "us-east-1",
"UseDualStack": false
}
},
@@ -210,8 +210,8 @@
"error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
- "UseFIPS": true,
"Region": "us-isob-east-1",
+ "UseFIPS": true,
"UseDualStack": true
}
},
@@ -223,8 +223,8 @@
}
},
"params": {
- "UseFIPS": true,
"Region": "us-isob-east-1",
+ "UseFIPS": true,
"UseDualStack": false
}
},
@@ -234,8 +234,8 @@
"error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
- "UseFIPS": false,
"Region": "us-isob-east-1",
+ "UseFIPS": false,
"UseDualStack": true
}
},
@@ -247,21 +247,34 @@
}
},
"params": {
- "UseFIPS": false,
"Region": "us-isob-east-1",
+ "UseFIPS": false,
"UseDualStack": false
}
},
{
- "documentation": "For custom endpoint with fips disabled and dualstack disabled",
+ "documentation": "For custom endpoint with region set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
"url": "https://example.com"
}
},
"params": {
- "UseFIPS": false,
"Region": "us-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false,
+ "Endpoint": "https://example.com"
+ }
+ },
+ {
+ "documentation": "For custom endpoint with region not set and fips disabled and dualstack disabled",
+ "expect": {
+ "endpoint": {
+ "url": "https://example.com"
+ }
+ },
+ "params": {
+ "UseFIPS": false,
"UseDualStack": false,
"Endpoint": "https://example.com"
}
@@ -272,8 +285,8 @@
"error": "Invalid Configuration: FIPS and custom endpoint are not supported"
},
"params": {
- "UseFIPS": true,
"Region": "us-east-1",
+ "UseFIPS": true,
"UseDualStack": false,
"Endpoint": "https://example.com"
}
@@ -284,11 +297,17 @@
"error": "Invalid Configuration: Dualstack and custom endpoint are not supported"
},
"params": {
- "UseFIPS": false,
"Region": "us-east-1",
+ "UseFIPS": false,
"UseDualStack": true,
"Endpoint": "https://example.com"
}
+ },
+ {
+ "documentation": "Missing region",
+ "expect": {
+ "error": "Invalid Configuration: Missing Region"
+ }
}
],
"version": "1.0"
diff --git a/services/ivsrealtime/src/main/resources/codegen-resources/service-2.json b/services/ivsrealtime/src/main/resources/codegen-resources/service-2.json
index edc0b300714..2146dc2e7a9 100644
--- a/services/ivsrealtime/src/main/resources/codegen-resources/service-2.json
+++ b/services/ivsrealtime/src/main/resources/codegen-resources/service-2.json
@@ -407,7 +407,7 @@
"members":{
"errorCode":{
"shape":"EventErrorCode",
- "documentation":"
If the event is an error event, the error code is provided to give insight into the specific error that occurred. If the event is not an error event, this field is null. INSUFFICIENT_CAPABILITIES indicates that the participant tried to take an action that the participant’s token is not allowed to do. For more information about participant capabilities, see the capabilities field in CreateParticipantToken.
"
+ "documentation":"
If the event is an error event, the error code is provided to give insight into the specific error that occurred. If the event is not an error event, this field is null. INSUFFICIENT_CAPABILITIES indicates that the participant tried to take an action that the participant’s token is not allowed to do. For more information about participant capabilities, see the capabilities field in CreateParticipantToken. QUOTA_EXCEEDED indicates that the number of participants who want to publish/subscribe to a stage exceeds the quota; for more information, see Service Quotas. PUBLISHER_NOT_FOUND indicates that the participant tried to subscribe to a publisher that doesn’t exist.
The Amazon Interactive Video Service (IVS) stage API is REST compatible, using a standard HTTP API and an AWS EventBridge event stream for responses. JSON is used for both requests and responses, including errors.
Terminology:
The IVS stage API sometimes is referred to as the IVS RealTime API.
A participant token is an authorization token used to publish/subscribe to a stage.
A participant object represents participants (people) in the stage and contains information about them. When a token is created, it includes a participant ID; when a participant uses that token to join a stage, the participant is associated with that participant ID There is a 1:1 mapping between participant tokens and participants.
Stage — A stage is a virtual space where multiple participants can exchange audio and video in real time.
Tagging
A tag is a metadata label that you assign to an AWS resource. A tag comprises a key and a value, both set by you. For example, you might set a tag as topic:nature to label a particular video category. See Tagging AWS Resources for more information, including restrictions that apply to tags and \"Tag naming limits and requirements\"; Amazon IVS stages has no service-specific constraints beyond what is documented there.
Tags can help you identify and organize your AWS resources. For example, you can use the same tag for different resources to indicate that they are related. You can also use tags to manage access (see Access Tags).
ListTagsForResource — Gets information about AWS tags for the specified ARN.
TagResource — Adds or updates tags for the AWS resource with the specified ARN.
UntagResource — Removes tags from the resource with the specified ARN.
"
+ "documentation":"
Introduction
The Amazon Interactive Video Service (IVS) real-time API is REST compatible, using a standard HTTP API and an AWS EventBridge event stream for responses. JSON is used for both requests and responses, including errors.
Terminology:
A stage is a virtual space where participants can exchange video in real time.
A participant token is a token that authenticates a participant when they join a stage.
A participant object represents participants (people) in the stage and contains information about them. When a token is created, it includes a participant ID; when a participant uses that token to join a stage, the participant is associated with that participant ID There is a 1:1 mapping between participant tokens and participants.
Stage — A stage is a virtual space where participants can exchange video in real time.
Tagging
A tag is a metadata label that you assign to an AWS resource. A tag comprises a key and a value, both set by you. For example, you might set a tag as topic:nature to label a particular video category. See Tagging AWS Resources for more information, including restrictions that apply to tags and \"Tag naming limits and requirements\"; Amazon IVS stages has no service-specific constraints beyond what is documented there.
Tags can help you identify and organize your AWS resources. For example, you can use the same tag for different resources to indicate that they are related. You can also use tags to manage access (see Access Tags).
If the response of ListClusterOperationsV2 is truncated, it returns a NextToken in the response. This NextToken should be sent in the subsequent request to ListClusterOperationsV2.
Associates a SignalingChannel to a stream to store the media. There are two signaling modes that can specified :
If the StorageStatus is disabled, no data will be stored, and the StreamARN parameter will not be needed.
If the StorageStatus is enabled, the data will be stored in the StreamARN provided.
"
+ "documentation":"
Associates a SignalingChannel to a stream to store the media. There are two signaling modes that can specified :
If the StorageStatus is disabled, no data will be stored, and the StreamARN parameter will not be needed.
If the StorageStatus is enabled, the data will be stored in the StreamARN provided.
If StorageStatus is enabled, direct peer-to-peer (master-viewer) connections no longer occur. Peers connect directly to the storage session. You must call the JoinStorageSession API to trigger an SDP offer send and establish a connection between a peer and the storage session.
"
},
"UpdateNotificationConfiguration":{
"name":"UpdateNotificationConfiguration",
@@ -1754,11 +1754,7 @@
"max":50,
"min":1
},
- "SamplingInterval":{
- "type":"integer",
- "max":20000,
- "min":3000
- },
+ "SamplingInterval":{"type":"integer"},
"ScheduleConfig":{
"type":"structure",
"required":[
diff --git a/services/kinesisvideoarchivedmedia/pom.xml b/services/kinesisvideoarchivedmedia/pom.xml
index a906003899f..cb479896aee 100644
--- a/services/kinesisvideoarchivedmedia/pom.xml
+++ b/services/kinesisvideoarchivedmedia/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTkinesisvideoarchivedmediaAWS Java SDK :: Services :: Kinesis Video Archived Media
diff --git a/services/kinesisvideoarchivedmedia/src/main/resources/codegen-resources/endpoint-rule-set.json b/services/kinesisvideoarchivedmedia/src/main/resources/codegen-resources/endpoint-rule-set.json
index 33f59063d56..f61682a8f1f 100644
--- a/services/kinesisvideoarchivedmedia/src/main/resources/codegen-resources/endpoint-rule-set.json
+++ b/services/kinesisvideoarchivedmedia/src/main/resources/codegen-resources/endpoint-rule-set.json
@@ -32,13 +32,12 @@
{
"conditions": [
{
- "fn": "aws.partition",
+ "fn": "isSet",
"argv": [
{
- "ref": "Region"
+ "ref": "Endpoint"
}
- ],
- "assign": "PartitionResult"
+ ]
}
],
"type": "tree",
@@ -46,23 +45,20 @@
{
"conditions": [
{
- "fn": "isSet",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "Endpoint"
- }
+ "ref": "UseFIPS"
+ },
+ true
]
- },
- {
- "fn": "parseURL",
- "argv": [
- {
- "ref": "Endpoint"
- }
- ],
- "assign": "url"
}
],
+ "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "type": "error"
+ },
+ {
+ "conditions": [],
"type": "tree",
"rules": [
{
@@ -71,67 +67,42 @@
"fn": "booleanEquals",
"argv": [
{
- "ref": "UseFIPS"
+ "ref": "UseDualStack"
},
true
]
}
],
- "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
"type": "error"
},
{
"conditions": [],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
- },
- true
- ]
- }
- ],
- "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
- "type": "error"
+ "endpoint": {
+ "url": {
+ "ref": "Endpoint"
},
- {
- "conditions": [],
- "endpoint": {
- "url": {
- "ref": "Endpoint"
- },
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
}
]
- },
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
{
"conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
- },
- true
- ]
- },
- {
- "fn": "booleanEquals",
+ "fn": "isSet",
"argv": [
{
- "ref": "UseDualStack"
- },
- true
+ "ref": "Region"
+ }
]
}
],
@@ -140,90 +111,215 @@
{
"conditions": [
{
- "fn": "booleanEquals",
+ "fn": "aws.partition",
"argv": [
- true,
{
- "fn": "getAttr",
+ "ref": "Region"
+ }
+ ],
+ "assign": "PartitionResult"
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseFIPS"
},
- "supportsFIPS"
+ true
+ ]
+ },
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ {
+ "ref": "UseDualStack"
+ },
+ true
]
}
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsFIPS"
+ ]
+ }
+ ]
+ },
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://kinesisvideo-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
+ "type": "error"
+ }
]
},
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseFIPS"
},
- "supportsDualStack"
+ true
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsFIPS"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://kinesisvideo-fips.{Region}.{PartitionResult#dnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
]
+ },
+ {
+ "conditions": [],
+ "error": "FIPS is enabled but this partition does not support FIPS",
+ "type": "error"
}
]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
- "endpoint": {
- "url": "https://kinesisvideo-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
- },
- {
- "conditions": [],
- "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
},
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseDualStack"
},
- "supportsFIPS"
+ true
]
}
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://kinesisvideo.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "error": "DualStack is enabled but this partition does not support DualStack",
+ "type": "error"
+ }
]
- }
- ],
- "type": "tree",
- "rules": [
+ },
{
"conditions": [],
"type": "tree",
@@ -231,7 +327,7 @@
{
"conditions": [],
"endpoint": {
- "url": "https://kinesisvideo-fips.{Region}.{PartitionResult#dnsSuffix}",
+ "url": "https://kinesisvideo.{Region}.{PartitionResult#dnsSuffix}",
"properties": {},
"headers": {}
},
@@ -240,74 +336,13 @@
]
}
]
- },
- {
- "conditions": [],
- "error": "FIPS is enabled but this partition does not support FIPS",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
- },
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- true,
- {
- "fn": "getAttr",
- "argv": [
- {
- "ref": "PartitionResult"
- },
- "supportsDualStack"
- ]
- }
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
- "endpoint": {
- "url": "https://kinesisvideo.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
- },
- {
- "conditions": [],
- "error": "DualStack is enabled but this partition does not support DualStack",
- "type": "error"
}
]
},
{
"conditions": [],
- "endpoint": {
- "url": "https://kinesisvideo.{Region}.{PartitionResult#dnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "error": "Invalid Configuration: Missing Region",
+ "type": "error"
}
]
}
diff --git a/services/kinesisvideoarchivedmedia/src/main/resources/codegen-resources/endpoint-tests.json b/services/kinesisvideoarchivedmedia/src/main/resources/codegen-resources/endpoint-tests.json
index 286590f5a81..3e71d1d389b 100644
--- a/services/kinesisvideoarchivedmedia/src/main/resources/codegen-resources/endpoint-tests.json
+++ b/services/kinesisvideoarchivedmedia/src/main/resources/codegen-resources/endpoint-tests.json
@@ -1,1191 +1,402 @@
{
"testCases": [
{
- "documentation": "For region ap-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.ap-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.ap-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.ap-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.ap-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.us-gov-east-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-gov-east-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.us-gov-east-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-gov-east-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.us-gov-east-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-gov-east-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.us-gov-east-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-gov-east-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.ca-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ca-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.ca-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ca-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.ca-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ca-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.ca-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ca-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.eu-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.eu-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.eu-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.eu-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-iso-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-iso-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-iso-west-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.us-iso-west-1.c2s.ic.gov"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-iso-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-iso-west-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "error": "DualStack is enabled but this partition does not support DualStack"
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-iso-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-iso-west-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.us-iso-west-1.c2s.ic.gov"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-iso-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.us-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.us-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.us-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.us-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.af-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "af-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.af-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "af-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.af-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "af-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.af-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "af-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.eu-north-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-north-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.eu-north-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-north-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.eu-north-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-north-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.eu-north-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-north-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.eu-west-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.eu-west-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.eu-west-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.eu-west-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.eu-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.eu-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.eu-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.eu-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.eu-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.eu-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.eu-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.eu-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.ap-northeast-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.ap-northeast-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.ap-northeast-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.ap-northeast-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.ap-northeast-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.ap-northeast-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.ap-northeast-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.ap-northeast-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region me-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.me-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "me-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region me-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.me-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "me-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region me-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.me-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "me-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.me-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "me-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region sa-east-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.sa-east-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "sa-east-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region sa-east-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.sa-east-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "sa-east-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region sa-east-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.sa-east-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "sa-east-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://kinesisvideo.sa-east-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "sa-east-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.ap-east-1.api.aws"
+ "url": "https://kinesisvideo.af-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-east-1",
- "UseFIPS": true
+ "Region": "af-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.ap-east-1.amazonaws.com"
+ "url": "https://kinesisvideo.ap-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
"Region": "ap-east-1",
- "UseFIPS": true
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.ap-east-1.api.aws"
+ "url": "https://kinesisvideo.ap-northeast-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-east-1",
- "UseFIPS": false
+ "Region": "ap-northeast-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.ap-east-1.amazonaws.com"
+ "url": "https://kinesisvideo.ap-northeast-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-east-1",
- "UseFIPS": false
+ "Region": "ap-northeast-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.cn-north-1.api.amazonwebservices.com.cn"
+ "url": "https://kinesisvideo.ap-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "cn-north-1",
- "UseFIPS": true
+ "Region": "ap-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.cn-north-1.amazonaws.com.cn"
+ "url": "https://kinesisvideo.ap-southeast-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "cn-north-1",
- "UseFIPS": true
+ "Region": "ap-southeast-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.cn-north-1.api.amazonwebservices.com.cn"
+ "url": "https://kinesisvideo.ap-southeast-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "cn-north-1",
- "UseFIPS": false
+ "Region": "ap-southeast-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.cn-north-1.amazonaws.com.cn"
+ "url": "https://kinesisvideo.ca-central-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "cn-north-1",
- "UseFIPS": false
+ "Region": "ca-central-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.us-gov-west-1.api.aws"
+ "url": "https://kinesisvideo.eu-central-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-gov-west-1",
- "UseFIPS": true
+ "Region": "eu-central-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.us-gov-west-1.amazonaws.com"
+ "url": "https://kinesisvideo.eu-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-gov-west-1",
- "UseFIPS": true
+ "Region": "eu-west-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.us-gov-west-1.api.aws"
+ "url": "https://kinesisvideo.eu-west-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-gov-west-1",
- "UseFIPS": false
+ "Region": "eu-west-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.us-gov-west-1.amazonaws.com"
+ "url": "https://kinesisvideo.eu-west-3.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-gov-west-1",
- "UseFIPS": false
+ "Region": "eu-west-3",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.ap-southeast-1.api.aws"
+ "url": "https://kinesisvideo.sa-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-1",
- "UseFIPS": true
+ "Region": "sa-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.ap-southeast-1.amazonaws.com"
+ "url": "https://kinesisvideo.us-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-1",
- "UseFIPS": true
+ "Region": "us-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.ap-southeast-1.api.aws"
+ "url": "https://kinesisvideo.us-east-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-1",
- "UseFIPS": false
+ "Region": "us-east-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.ap-southeast-1.amazonaws.com"
+ "url": "https://kinesisvideo.us-west-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-1",
- "UseFIPS": false
+ "Region": "us-west-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.ap-southeast-2.api.aws"
+ "url": "https://kinesisvideo-fips.us-east-1.api.aws"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-2",
- "UseFIPS": true
+ "Region": "us-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.ap-southeast-2.amazonaws.com"
+ "url": "https://kinesisvideo-fips.us-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-2",
- "UseFIPS": true
+ "Region": "us-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.ap-southeast-2.api.aws"
+ "url": "https://kinesisvideo.us-east-1.api.aws"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-2",
- "UseFIPS": false
+ "Region": "us-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.ap-southeast-2.amazonaws.com"
+ "url": "https://kinesisvideo-fips.cn-north-1.api.amazonwebservices.com.cn"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-iso-east-1",
- "UseFIPS": true
+ "Region": "cn-north-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.us-iso-east-1.c2s.ic.gov"
+ "url": "https://kinesisvideo-fips.cn-north-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-iso-east-1",
- "UseFIPS": true
+ "Region": "cn-north-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled",
"expect": {
- "error": "DualStack is enabled but this partition does not support DualStack"
+ "endpoint": {
+ "url": "https://kinesisvideo.cn-north-1.api.amazonwebservices.com.cn"
+ }
},
"params": {
- "UseDualStack": true,
- "Region": "us-iso-east-1",
- "UseFIPS": false
+ "Region": "cn-north-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.us-iso-east-1.c2s.ic.gov"
+ "url": "https://kinesisvideo.cn-north-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-iso-east-1",
- "UseFIPS": false
+ "Region": "cn-north-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.us-east-1.api.aws"
+ "url": "https://kinesisvideo-fips.us-gov-east-1.api.aws"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-1",
- "UseFIPS": true
+ "Region": "us-gov-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.us-east-1.amazonaws.com"
+ "url": "https://kinesisvideo-fips.us-gov-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-1",
- "UseFIPS": true
+ "Region": "us-gov-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.us-east-1.api.aws"
+ "url": "https://kinesisvideo.us-gov-east-1.api.aws"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-1",
- "UseFIPS": false
+ "Region": "us-gov-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.us-east-1.amazonaws.com"
+ "url": "https://kinesisvideo.us-gov-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-1",
- "UseFIPS": false
+ "Region": "us-gov-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://kinesisvideo-fips.us-east-2.api.aws"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-2",
- "UseFIPS": true
+ "Region": "us-iso-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo-fips.us-east-2.amazonaws.com"
+ "url": "https://kinesisvideo-fips.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-2",
- "UseFIPS": true
+ "Region": "us-iso-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://kinesisvideo.us-east-2.api.aws"
- }
+ "error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-2",
- "UseFIPS": false
+ "Region": "us-iso-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://kinesisvideo.us-east-2.amazonaws.com"
+ "url": "https://kinesisvideo.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-2",
- "UseFIPS": false
+ "Region": "us-iso-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
@@ -1194,9 +405,9 @@
"error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
- "UseDualStack": true,
"Region": "us-isob-east-1",
- "UseFIPS": true
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
@@ -1207,9 +418,9 @@
}
},
"params": {
- "UseDualStack": false,
"Region": "us-isob-east-1",
- "UseFIPS": true
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
@@ -1218,9 +429,9 @@
"error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
- "UseDualStack": true,
"Region": "us-isob-east-1",
- "UseFIPS": false
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
@@ -1231,22 +442,35 @@
}
},
"params": {
- "UseDualStack": false,
"Region": "us-isob-east-1",
- "UseFIPS": false
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For custom endpoint with fips disabled and dualstack disabled",
+ "documentation": "For custom endpoint with region set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
"url": "https://example.com"
}
},
"params": {
- "UseDualStack": false,
"Region": "us-east-1",
"UseFIPS": false,
+ "UseDualStack": false,
+ "Endpoint": "https://example.com"
+ }
+ },
+ {
+ "documentation": "For custom endpoint with region not set and fips disabled and dualstack disabled",
+ "expect": {
+ "endpoint": {
+ "url": "https://example.com"
+ }
+ },
+ "params": {
+ "UseFIPS": false,
+ "UseDualStack": false,
"Endpoint": "https://example.com"
}
},
@@ -1256,9 +480,9 @@
"error": "Invalid Configuration: FIPS and custom endpoint are not supported"
},
"params": {
- "UseDualStack": false,
"Region": "us-east-1",
"UseFIPS": true,
+ "UseDualStack": false,
"Endpoint": "https://example.com"
}
},
@@ -1268,11 +492,17 @@
"error": "Invalid Configuration: Dualstack and custom endpoint are not supported"
},
"params": {
- "UseDualStack": true,
"Region": "us-east-1",
"UseFIPS": false,
+ "UseDualStack": true,
"Endpoint": "https://example.com"
}
+ },
+ {
+ "documentation": "Missing region",
+ "expect": {
+ "error": "Invalid Configuration: Missing Region"
+ }
}
],
"version": "1.0"
diff --git a/services/kinesisvideoarchivedmedia/src/main/resources/codegen-resources/service-2.json b/services/kinesisvideoarchivedmedia/src/main/resources/codegen-resources/service-2.json
index a074f379800..88ff5a3f413 100644
--- a/services/kinesisvideoarchivedmedia/src/main/resources/codegen-resources/service-2.json
+++ b/services/kinesisvideoarchivedmedia/src/main/resources/codegen-resources/service-2.json
@@ -30,7 +30,7 @@
{"shape":"InvalidMediaFrameException"},
{"shape":"NoDataRetentionException"}
],
- "documentation":"
Downloads an MP4 file (clip) containing the archived, on-demand media from the specified video stream over the specified time range.
Both the StreamName and the StreamARN parameters are optional, but you must specify either the StreamName or the StreamARN when invoking this API operation.
As a prerequisite to using GetCLip API, you must obtain an endpoint using GetDataEndpoint, specifying GET_CLIP for the APIName parameter.
An Amazon Kinesis video stream has the following requirements for providing data through MP4:
The media must contain h.264 or h.265 encoded video and, optionally, AAC or G.711 encoded audio. Specifically, the codec ID of track 1 should be V_MPEG/ISO/AVC (for h.264) or V_MPEGH/ISO/HEVC (for H.265). Optionally, the codec ID of track 2 should be A_AAC (for AAC) or A_MS/ACM (for G.711).
Data retention must be greater than 0.
The video track of each fragment must contain codec private data in the Advanced Video Coding (AVC) for H.264 format and HEVC for H.265 format. For more information, see MPEG-4 specification ISO/IEC 14496-15. For information about adapting stream data to a given format, see NAL Adaptation Flags.
You can monitor the amount of outgoing data by monitoring the GetClip.OutgoingBytes Amazon CloudWatch metric. For information about using CloudWatch to monitor Kinesis Video Streams, see Monitoring Kinesis Video Streams. For pricing information, see Amazon Kinesis Video Streams Pricing and AWS Pricing. Charges for outgoing AWS data apply.
"
+ "documentation":"
Downloads an MP4 file (clip) containing the archived, on-demand media from the specified video stream over the specified time range.
Both the StreamName and the StreamARN parameters are optional, but you must specify either the StreamName or the StreamARN when invoking this API operation.
As a prerequisite to using GetCLip API, you must obtain an endpoint using GetDataEndpoint, specifying GET_CLIP for the APIName parameter.
An Amazon Kinesis video stream has the following requirements for providing data through MP4:
The media must contain h.264 or h.265 encoded video and, optionally, AAC or G.711 encoded audio. Specifically, the codec ID of track 1 should be V_MPEG/ISO/AVC (for h.264) or V_MPEGH/ISO/HEVC (for H.265). Optionally, the codec ID of track 2 should be A_AAC (for AAC) or A_MS/ACM (for G.711).
Data retention must be greater than 0.
The video track of each fragment must contain codec private data in the Advanced Video Coding (AVC) for H.264 format and HEVC for H.265 format. For more information, see MPEG-4 specification ISO/IEC 14496-15. For information about adapting stream data to a given format, see NAL Adaptation Flags.
Retrieves an MPEG Dynamic Adaptive Streaming over HTTP (DASH) URL for the stream. You can then open the URL in a media player to view the stream contents.
Both the StreamName and the StreamARN parameters are optional, but you must specify either the StreamName or the StreamARN when invoking this API operation.
An Amazon Kinesis video stream has the following requirements for providing data through MPEG-DASH:
The media must contain h.264 or h.265 encoded video and, optionally, AAC or G.711 encoded audio. Specifically, the codec ID of track 1 should be V_MPEG/ISO/AVC (for h.264) or V_MPEGH/ISO/HEVC (for H.265). Optionally, the codec ID of track 2 should be A_AAC (for AAC) or A_MS/ACM (for G.711).
Data retention must be greater than 0.
The video track of each fragment must contain codec private data in the Advanced Video Coding (AVC) for H.264 format and HEVC for H.265 format. For more information, see MPEG-4 specification ISO/IEC 14496-15. For information about adapting stream data to a given format, see NAL Adaptation Flags.
The following procedure shows how to use MPEG-DASH with Kinesis Video Streams:
Get an endpoint using GetDataEndpoint, specifying GET_DASH_STREAMING_SESSION_URL for the APIName parameter.
Retrieve the MPEG-DASH URL using GetDASHStreamingSessionURL. Kinesis Video Streams creates an MPEG-DASH streaming session to be used for accessing content in a stream using the MPEG-DASH protocol. GetDASHStreamingSessionURL returns an authenticated URL (that includes an encrypted session token) for the session's MPEG-DASH manifest (the root resource needed for streaming with MPEG-DASH).
Don't share or store this token where an unauthorized entity can access it. The token provides access to the content of the stream. Safeguard the token with the same measures that you use with your AWS credentials.
The media that is made available through the manifest consists only of the requested stream, time range, and format. No other media data (such as frames outside the requested window or alternate bitrates) is made available.
Provide the URL (containing the encrypted session token) for the MPEG-DASH manifest to a media player that supports the MPEG-DASH protocol. Kinesis Video Streams makes the initialization fragment and media fragments available through the manifest URL. The initialization fragment contains the codec private data for the stream, and other data needed to set up the video or audio decoder and renderer. The media fragments contain encoded video frames or encoded audio samples.
The media player receives the authenticated URL and requests stream metadata and media data normally. When the media player requests data, it calls the following actions:
GetDASHManifest: Retrieves an MPEG DASH manifest, which contains the metadata for the media that you want to playback.
GetMP4InitFragment: Retrieves the MP4 initialization fragment. The media player typically loads the initialization fragment before loading any media fragments. This fragment contains the \"fytp\" and \"moov\" MP4 atoms, and the child atoms that are needed to initialize the media player decoder.
The initialization fragment does not correspond to a fragment in a Kinesis video stream. It contains only the codec private data for the stream and respective track, which the media player needs to decode the media frames.
GetMP4MediaFragment: Retrieves MP4 media fragments. These fragments contain the \"moof\" and \"mdat\" MP4 atoms and their child atoms, containing the encoded fragment's media frames and their timestamps.
After the first media fragment is made available in a streaming session, any fragments that don't contain the same codec private data cause an error to be returned when those different media fragments are loaded. Therefore, the codec private data should not change between fragments in a session. This also means that the session fails if the fragments in a stream change from having only video to having both audio and video.
Data retrieved with this action is billable. See Pricing for details.
You can monitor the amount of data that the media player consumes by monitoring the GetMP4MediaFragment.OutgoingBytes Amazon CloudWatch metric. For information about using CloudWatch to monitor Kinesis Video Streams, see Monitoring Kinesis Video Streams. For pricing information, see Amazon Kinesis Video Streams Pricing and AWS Pricing. Charges for both HLS sessions and outgoing AWS data apply.
If an error is thrown after invoking a Kinesis Video Streams archived media API, in addition to the HTTP status code and the response body, it includes the following pieces of information:
x-amz-ErrorType HTTP header – contains a more specific error type in addition to what the HTTP status code provides.
x-amz-RequestId HTTP header – if you want to report an issue to AWS, the support team can better diagnose the problem if given the Request Id.
Both the HTTP status code and the ErrorType header can be utilized to make programmatic decisions about whether errors are retry-able and under what conditions, as well as provide information on what actions the client programmer might need to take in order to successfully try again.
For more information, see the Errors section at the bottom of this topic, as well as Common Errors.
"
+ "documentation":"
Retrieves an MPEG Dynamic Adaptive Streaming over HTTP (DASH) URL for the stream. You can then open the URL in a media player to view the stream contents.
Both the StreamName and the StreamARN parameters are optional, but you must specify either the StreamName or the StreamARN when invoking this API operation.
An Amazon Kinesis video stream has the following requirements for providing data through MPEG-DASH:
The media must contain h.264 or h.265 encoded video and, optionally, AAC or G.711 encoded audio. Specifically, the codec ID of track 1 should be V_MPEG/ISO/AVC (for h.264) or V_MPEGH/ISO/HEVC (for H.265). Optionally, the codec ID of track 2 should be A_AAC (for AAC) or A_MS/ACM (for G.711).
Data retention must be greater than 0.
The video track of each fragment must contain codec private data in the Advanced Video Coding (AVC) for H.264 format and HEVC for H.265 format. For more information, see MPEG-4 specification ISO/IEC 14496-15. For information about adapting stream data to a given format, see NAL Adaptation Flags.
The following procedure shows how to use MPEG-DASH with Kinesis Video Streams:
Get an endpoint using GetDataEndpoint, specifying GET_DASH_STREAMING_SESSION_URL for the APIName parameter.
Retrieve the MPEG-DASH URL using GetDASHStreamingSessionURL. Kinesis Video Streams creates an MPEG-DASH streaming session to be used for accessing content in a stream using the MPEG-DASH protocol. GetDASHStreamingSessionURL returns an authenticated URL (that includes an encrypted session token) for the session's MPEG-DASH manifest (the root resource needed for streaming with MPEG-DASH).
Don't share or store this token where an unauthorized entity can access it. The token provides access to the content of the stream. Safeguard the token with the same measures that you use with your Amazon Web Services credentials.
The media that is made available through the manifest consists only of the requested stream, time range, and format. No other media data (such as frames outside the requested window or alternate bitrates) is made available.
Provide the URL (containing the encrypted session token) for the MPEG-DASH manifest to a media player that supports the MPEG-DASH protocol. Kinesis Video Streams makes the initialization fragment and media fragments available through the manifest URL. The initialization fragment contains the codec private data for the stream, and other data needed to set up the video or audio decoder and renderer. The media fragments contain encoded video frames or encoded audio samples.
The media player receives the authenticated URL and requests stream metadata and media data normally. When the media player requests data, it calls the following actions:
GetDASHManifest: Retrieves an MPEG DASH manifest, which contains the metadata for the media that you want to playback.
GetMP4InitFragment: Retrieves the MP4 initialization fragment. The media player typically loads the initialization fragment before loading any media fragments. This fragment contains the \"fytp\" and \"moov\" MP4 atoms, and the child atoms that are needed to initialize the media player decoder.
The initialization fragment does not correspond to a fragment in a Kinesis video stream. It contains only the codec private data for the stream and respective track, which the media player needs to decode the media frames.
GetMP4MediaFragment: Retrieves MP4 media fragments. These fragments contain the \"moof\" and \"mdat\" MP4 atoms and their child atoms, containing the encoded fragment's media frames and their timestamps.
After the first media fragment is made available in a streaming session, any fragments that don't contain the same codec private data cause an error to be returned when those different media fragments are loaded. Therefore, the codec private data should not change between fragments in a session. This also means that the session fails if the fragments in a stream change from having only video to having both audio and video.
Data retrieved with this action is billable. See Pricing for details.
You can monitor the amount of data that the media player consumes by monitoring the GetMP4MediaFragment.OutgoingBytes Amazon CloudWatch metric. For information about using CloudWatch to monitor Kinesis Video Streams, see Monitoring Kinesis Video Streams. For pricing information, see Amazon Kinesis Video Streams Pricing and Amazon Web Services Pricing. Charges for both HLS sessions and outgoing Amazon Web Services data apply.
If an error is thrown after invoking a Kinesis Video Streams archived media API, in addition to the HTTP status code and the response body, it includes the following pieces of information:
x-amz-ErrorType HTTP header – contains a more specific error type in addition to what the HTTP status code provides.
x-amz-RequestId HTTP header – if you want to report an issue to Amazon Web Services the support team can better diagnose the problem if given the Request Id.
Both the HTTP status code and the ErrorType header can be utilized to make programmatic decisions about whether errors are retry-able and under what conditions, as well as provide information on what actions the client programmer might need to take in order to successfully try again.
For more information, see the Errors section at the bottom of this topic, as well as Common Errors.
Retrieves an HTTP Live Streaming (HLS) URL for the stream. You can then open the URL in a browser or media player to view the stream contents.
Both the StreamName and the StreamARN parameters are optional, but you must specify either the StreamName or the StreamARN when invoking this API operation.
An Amazon Kinesis video stream has the following requirements for providing data through HLS:
The media must contain h.264 or h.265 encoded video and, optionally, AAC encoded audio. Specifically, the codec ID of track 1 should be V_MPEG/ISO/AVC (for h.264) or V_MPEG/ISO/HEVC (for h.265). Optionally, the codec ID of track 2 should be A_AAC.
Data retention must be greater than 0.
The video track of each fragment must contain codec private data in the Advanced Video Coding (AVC) for H.264 format or HEVC for H.265 format (MPEG-4 specification ISO/IEC 14496-15). For information about adapting stream data to a given format, see NAL Adaptation Flags.
The audio track (if present) of each fragment must contain codec private data in the AAC format (AAC specification ISO/IEC 13818-7).
Kinesis Video Streams HLS sessions contain fragments in the fragmented MPEG-4 form (also called fMP4 or CMAF) or the MPEG-2 form (also called TS chunks, which the HLS specification also supports). For more information about HLS fragment types, see the HLS specification.
The following procedure shows how to use HLS with Kinesis Video Streams:
Get an endpoint using GetDataEndpoint, specifying GET_HLS_STREAMING_SESSION_URL for the APIName parameter.
Retrieve the HLS URL using GetHLSStreamingSessionURL. Kinesis Video Streams creates an HLS streaming session to be used for accessing content in a stream using the HLS protocol. GetHLSStreamingSessionURL returns an authenticated URL (that includes an encrypted session token) for the session's HLS master playlist (the root resource needed for streaming with HLS).
Don't share or store this token where an unauthorized entity could access it. The token provides access to the content of the stream. Safeguard the token with the same measures that you would use with your AWS credentials.
The media that is made available through the playlist consists only of the requested stream, time range, and format. No other media data (such as frames outside the requested window or alternate bitrates) is made available.
Provide the URL (containing the encrypted session token) for the HLS master playlist to a media player that supports the HLS protocol. Kinesis Video Streams makes the HLS media playlist, initialization fragment, and media fragments available through the master playlist URL. The initialization fragment contains the codec private data for the stream, and other data needed to set up the video or audio decoder and renderer. The media fragments contain H.264-encoded video frames or AAC-encoded audio samples.
The media player receives the authenticated URL and requests stream metadata and media data normally. When the media player requests data, it calls the following actions:
GetHLSMasterPlaylist: Retrieves an HLS master playlist, which contains a URL for the GetHLSMediaPlaylist action for each track, and additional metadata for the media player, including estimated bitrate and resolution.
GetHLSMediaPlaylist: Retrieves an HLS media playlist, which contains a URL to access the MP4 initialization fragment with the GetMP4InitFragment action, and URLs to access the MP4 media fragments with the GetMP4MediaFragment actions. The HLS media playlist also contains metadata about the stream that the player needs to play it, such as whether the PlaybackMode is LIVE or ON_DEMAND. The HLS media playlist is typically static for sessions with a PlaybackType of ON_DEMAND. The HLS media playlist is continually updated with new fragments for sessions with a PlaybackType of LIVE. There is a distinct HLS media playlist for the video track and the audio track (if applicable) that contains MP4 media URLs for the specific track.
GetMP4InitFragment: Retrieves the MP4 initialization fragment. The media player typically loads the initialization fragment before loading any media fragments. This fragment contains the \"fytp\" and \"moov\" MP4 atoms, and the child atoms that are needed to initialize the media player decoder.
The initialization fragment does not correspond to a fragment in a Kinesis video stream. It contains only the codec private data for the stream and respective track, which the media player needs to decode the media frames.
GetMP4MediaFragment: Retrieves MP4 media fragments. These fragments contain the \"moof\" and \"mdat\" MP4 atoms and their child atoms, containing the encoded fragment's media frames and their timestamps.
After the first media fragment is made available in a streaming session, any fragments that don't contain the same codec private data cause an error to be returned when those different media fragments are loaded. Therefore, the codec private data should not change between fragments in a session. This also means that the session fails if the fragments in a stream change from having only video to having both audio and video.
Data retrieved with this action is billable. See Pricing for details.
GetTSFragment: Retrieves MPEG TS fragments containing both initialization and media data for all tracks in the stream.
If the ContainerFormat is MPEG_TS, this API is used instead of GetMP4InitFragment and GetMP4MediaFragment to retrieve stream media.
A streaming session URL must not be shared between players. The service might throttle a session if multiple media players are sharing it. For connection limits, see Kinesis Video Streams Limits.
You can monitor the amount of data that the media player consumes by monitoring the GetMP4MediaFragment.OutgoingBytes Amazon CloudWatch metric. For information about using CloudWatch to monitor Kinesis Video Streams, see Monitoring Kinesis Video Streams. For pricing information, see Amazon Kinesis Video Streams Pricing and AWS Pricing. Charges for both HLS sessions and outgoing AWS data apply.
If an error is thrown after invoking a Kinesis Video Streams archived media API, in addition to the HTTP status code and the response body, it includes the following pieces of information:
x-amz-ErrorType HTTP header – contains a more specific error type in addition to what the HTTP status code provides.
x-amz-RequestId HTTP header – if you want to report an issue to AWS, the support team can better diagnose the problem if given the Request Id.
Both the HTTP status code and the ErrorType header can be utilized to make programmatic decisions about whether errors are retry-able and under what conditions, as well as provide information on what actions the client programmer might need to take in order to successfully try again.
For more information, see the Errors section at the bottom of this topic, as well as Common Errors.
"
+ "documentation":"
Retrieves an HTTP Live Streaming (HLS) URL for the stream. You can then open the URL in a browser or media player to view the stream contents.
Both the StreamName and the StreamARN parameters are optional, but you must specify either the StreamName or the StreamARN when invoking this API operation.
An Amazon Kinesis video stream has the following requirements for providing data through HLS:
For streaming video, the media must contain H.264 or H.265 encoded video and, optionally, AAC encoded audio. Specifically, the codec ID of track 1 should be V_MPEG/ISO/AVC (for H.264) or V_MPEG/ISO/HEVC (for H.265). Optionally, the codec ID of track 2 should be A_AAC. For audio only streaming, the codec ID of track 1 should be A_AAC.
Data retention must be greater than 0.
The video track of each fragment must contain codec private data in the Advanced Video Coding (AVC) for H.264 format or HEVC for H.265 format (MPEG-4 specification ISO/IEC 14496-15). For information about adapting stream data to a given format, see NAL Adaptation Flags.
The audio track (if present) of each fragment must contain codec private data in the AAC format (AAC specification ISO/IEC 13818-7).
Kinesis Video Streams HLS sessions contain fragments in the fragmented MPEG-4 form (also called fMP4 or CMAF) or the MPEG-2 form (also called TS chunks, which the HLS specification also supports). For more information about HLS fragment types, see the HLS specification.
The following procedure shows how to use HLS with Kinesis Video Streams:
Get an endpoint using GetDataEndpoint, specifying GET_HLS_STREAMING_SESSION_URL for the APIName parameter.
Retrieve the HLS URL using GetHLSStreamingSessionURL. Kinesis Video Streams creates an HLS streaming session to be used for accessing content in a stream using the HLS protocol. GetHLSStreamingSessionURL returns an authenticated URL (that includes an encrypted session token) for the session's HLS master playlist (the root resource needed for streaming with HLS).
Don't share or store this token where an unauthorized entity could access it. The token provides access to the content of the stream. Safeguard the token with the same measures that you would use with your Amazon Web Services credentials.
The media that is made available through the playlist consists only of the requested stream, time range, and format. No other media data (such as frames outside the requested window or alternate bitrates) is made available.
Provide the URL (containing the encrypted session token) for the HLS master playlist to a media player that supports the HLS protocol. Kinesis Video Streams makes the HLS media playlist, initialization fragment, and media fragments available through the master playlist URL. The initialization fragment contains the codec private data for the stream, and other data needed to set up the video or audio decoder and renderer. The media fragments contain H.264-encoded video frames or AAC-encoded audio samples.
The media player receives the authenticated URL and requests stream metadata and media data normally. When the media player requests data, it calls the following actions:
GetHLSMasterPlaylist: Retrieves an HLS master playlist, which contains a URL for the GetHLSMediaPlaylist action for each track, and additional metadata for the media player, including estimated bitrate and resolution.
GetHLSMediaPlaylist: Retrieves an HLS media playlist, which contains a URL to access the MP4 initialization fragment with the GetMP4InitFragment action, and URLs to access the MP4 media fragments with the GetMP4MediaFragment actions. The HLS media playlist also contains metadata about the stream that the player needs to play it, such as whether the PlaybackMode is LIVE or ON_DEMAND. The HLS media playlist is typically static for sessions with a PlaybackType of ON_DEMAND. The HLS media playlist is continually updated with new fragments for sessions with a PlaybackType of LIVE. There is a distinct HLS media playlist for the video track and the audio track (if applicable) that contains MP4 media URLs for the specific track.
GetMP4InitFragment: Retrieves the MP4 initialization fragment. The media player typically loads the initialization fragment before loading any media fragments. This fragment contains the \"fytp\" and \"moov\" MP4 atoms, and the child atoms that are needed to initialize the media player decoder.
The initialization fragment does not correspond to a fragment in a Kinesis video stream. It contains only the codec private data for the stream and respective track, which the media player needs to decode the media frames.
GetMP4MediaFragment: Retrieves MP4 media fragments. These fragments contain the \"moof\" and \"mdat\" MP4 atoms and their child atoms, containing the encoded fragment's media frames and their timestamps.
For the HLS streaming session, in-track codec private data (CPD) changes are supported. After the first media fragment is made available in a streaming session, fragments can contain CPD changes for each track. Therefore, the fragments in a session can have a different resolution, bit rate, or other information in the CPD without interrupting playback. However, any change made in the track number or track codec format can return an error when those different media fragments are loaded. For example, streaming will fail if the fragments in the stream change from having only video to having both audio and video, or if an AAC audio track is changed to an ALAW audio track. For each streaming session, only 500 CPD changes are allowed.
Data retrieved with this action is billable. For information, see Pricing.
GetTSFragment: Retrieves MPEG TS fragments containing both initialization and media data for all tracks in the stream.
If the ContainerFormat is MPEG_TS, this API is used instead of GetMP4InitFragment and GetMP4MediaFragment to retrieve stream media.
A streaming session URL must not be shared between players. The service might throttle a session if multiple media players are sharing it. For connection limits, see Kinesis Video Streams Limits.
You can monitor the amount of data that the media player consumes by monitoring the GetMP4MediaFragment.OutgoingBytes Amazon CloudWatch metric. For information about using CloudWatch to monitor Kinesis Video Streams, see Monitoring Kinesis Video Streams. For pricing information, see Amazon Kinesis Video Streams Pricing and Amazon Web Services Pricing. Charges for both HLS sessions and outgoing Amazon Web Services data apply.
If an error is thrown after invoking a Kinesis Video Streams archived media API, in addition to the HTTP status code and the response body, it includes the following pieces of information:
x-amz-ErrorType HTTP header – contains a more specific error type in addition to what the HTTP status code provides.
x-amz-RequestId HTTP header – if you want to report an issue to Amazon Web Services, the support team can better diagnose the problem if given the Request Id.
Both the HTTP status code and the ErrorType header can be utilized to make programmatic decisions about whether errors are retry-able and under what conditions, as well as provide information on what actions the client programmer might need to take in order to successfully try again.
For more information, see the Errors section at the bottom of this topic, as well as Common Errors.
Gets media for a list of fragments (specified by fragment number) from the archived data in an Amazon Kinesis video stream.
You must first call the GetDataEndpoint API to get an endpoint. Then send the GetMediaForFragmentList requests to this endpoint using the --endpoint-url parameter.
If an error is thrown after invoking a Kinesis Video Streams archived media API, in addition to the HTTP status code and the response body, it includes the following pieces of information:
x-amz-ErrorType HTTP header – contains a more specific error type in addition to what the HTTP status code provides.
x-amz-RequestId HTTP header – if you want to report an issue to AWS, the support team can better diagnose the problem if given the Request Id.
Both the HTTP status code and the ErrorType header can be utilized to make programmatic decisions about whether errors are retry-able and under what conditions, as well as provide information on what actions the client programmer might need to take in order to successfully try again.
For more information, see the Errors section at the bottom of this topic, as well as Common Errors.
"
+ "documentation":"
Gets media for a list of fragments (specified by fragment number) from the archived data in an Amazon Kinesis video stream.
You must first call the GetDataEndpoint API to get an endpoint. Then send the GetMediaForFragmentList requests to this endpoint using the --endpoint-url parameter.
If an error is thrown after invoking a Kinesis Video Streams archived media API, in addition to the HTTP status code and the response body, it includes the following pieces of information:
x-amz-ErrorType HTTP header – contains a more specific error type in addition to what the HTTP status code provides.
x-amz-RequestId HTTP header – if you want to report an issue to Amazon Web Services, the support team can better diagnose the problem if given the Request Id.
Both the HTTP status code and the ErrorType header can be utilized to make programmatic decisions about whether errors are retry-able and under what conditions, as well as provide information on what actions the client programmer might need to take in order to successfully try again.
For more information, see the Errors section at the bottom of this topic, as well as Common Errors.
Returns a list of Fragment objects from the specified stream and timestamp range within the archived data.
Listing fragments is eventually consistent. This means that even if the producer receives an acknowledgment that a fragment is persisted, the result might not be returned immediately from a request to ListFragments. However, results are typically available in less than one second.
You must first call the GetDataEndpoint API to get an endpoint. Then send the ListFragments requests to this endpoint using the --endpoint-url parameter.
If an error is thrown after invoking a Kinesis Video Streams archived media API, in addition to the HTTP status code and the response body, it includes the following pieces of information:
x-amz-ErrorType HTTP header – contains a more specific error type in addition to what the HTTP status code provides.
x-amz-RequestId HTTP header – if you want to report an issue to AWS, the support team can better diagnose the problem if given the Request Id.
Both the HTTP status code and the ErrorType header can be utilized to make programmatic decisions about whether errors are retry-able and under what conditions, as well as provide information on what actions the client programmer might need to take in order to successfully try again.
For more information, see the Errors section at the bottom of this topic, as well as Common Errors.
"
+ "documentation":"
Returns a list of Fragment objects from the specified stream and timestamp range within the archived data.
Listing fragments is eventually consistent. This means that even if the producer receives an acknowledgment that a fragment is persisted, the result might not be returned immediately from a request to ListFragments. However, results are typically available in less than one second.
You must first call the GetDataEndpoint API to get an endpoint. Then send the ListFragments requests to this endpoint using the --endpoint-url parameter.
If an error is thrown after invoking a Kinesis Video Streams archived media API, in addition to the HTTP status code and the response body, it includes the following pieces of information:
x-amz-ErrorType HTTP header – contains a more specific error type in addition to what the HTTP status code provides.
x-amz-RequestId HTTP header – if you want to report an issue to Amazon Web Services, the support team can better diagnose the problem if given the Request Id.
Both the HTTP status code and the ErrorType header can be utilized to make programmatic decisions about whether errors are retry-able and under what conditions, as well as provide information on what actions the client programmer might need to take in order to successfully try again.
For more information, see the Errors section at the bottom of this topic, as well as Common Errors.
The end timestamp for the range of images to be generated.
"
+ "documentation":"
The end timestamp for the range of images to be generated. If the time range between StartTimestamp and EndTimestamp is more than 300 seconds above StartTimestamp, you will receive an IllegalArgumentException.
The time interval in milliseconds (ms) at which the images need to be generated from the stream. The minimum value that can be provided is 3000 ms. If the timestamp range is less than the sampling interval, the Image from the startTimestamp will be returned if available.
The minimum value of 3000 ms is a soft limit. If needed, a lower sampling frequency can be requested.
"
+ "documentation":"
The time interval in milliseconds (ms) at which the images need to be generated from the stream, with a default of 3000 ms. The minimum value that can be provided is 200 ms. If the timestamp range is less than the sampling interval, the Image from the startTimestamp will be returned if available.
The maximum number of images to be returned by the API.
The default limit is 100 images per API response. The additional results will be paginated.
"
+ "documentation":"
The maximum number of images to be returned by the API.
The default limit is 25 images per API response. Providing a MaxResults greater than this value will result in a page size of 25. Any additional results will be paginated.
The payload that Kinesis Video Streams returns is a sequence of chunks from the specified stream. For information about the chunks, see PutMedia. The chunks that Kinesis Video Streams returns in the GetMediaForFragmentList call also include the following additional Matroska (MKV) tags:
AWS_KINESISVIDEO_FRAGMENT_NUMBER - Fragment number returned in the chunk.
AWS_KINESISVIDEO_SERVER_SIDE_TIMESTAMP - Server-side timestamp of the fragment.
AWS_KINESISVIDEO_PRODUCER_SIDE_TIMESTAMP - Producer-side timestamp of the fragment.
The following tags will be included if an exception occurs:
AWS_KINESISVIDEO_FRAGMENT_NUMBER - The number of the fragment that threw the exception
AWS_KINESISVIDEO_EXCEPTION_ERROR_CODE - The integer code of the exception
AWS_KINESISVIDEO_EXCEPTION_MESSAGE - A text description of the exception
"
+ "documentation":"
The payload that Kinesis Video Streams returns is a sequence of chunks from the specified stream. For information about the chunks, see PutMedia. The chunks that Kinesis Video Streams returns in the GetMediaForFragmentList call also include the following additional Matroska (MKV) tags:
AWS_KINESISVIDEO_FRAGMENT_NUMBER - Fragment number returned in the chunk.
AWS_KINESISVIDEO_SERVER_SIDE_TIMESTAMP - Server-side timestamp of the fragment.
AWS_KINESISVIDEO_PRODUCER_SIDE_TIMESTAMP - Producer-side timestamp of the fragment.
The following tags will be included if an exception occurs:
AWS_KINESISVIDEO_FRAGMENT_NUMBER - The number of the fragment that threw the exception
AWS_KINESISVIDEO_EXCEPTION_ERROR_CODE - The integer code of the
AWS_KINESISVIDEO_EXCEPTION_MESSAGE - A text description of the exception
GetMedia throws this error when Kinesis Video Streams can't find the stream that you specified.
GetHLSStreamingSessionURL and GetDASHStreamingSessionURL throw this error if a session with a PlaybackMode of ON_DEMAND or LIVE_REPLAYis requested for a stream that has no fragments within the requested time range, or if a session with a PlaybackMode of LIVE is requested for a stream that has no fragments within the last 30 seconds.
",
+ "documentation":"
GetImages will throw this error when Kinesis Video Streams can't find the stream that you specified.
GetHLSStreamingSessionURL and GetDASHStreamingSessionURL throw this error if a session with a PlaybackMode of ON_DEMAND or LIVE_REPLAYis requested for a stream that has no fragments within the requested time range, or if a session with a PlaybackMode of LIVE is requested for a stream that has no fragments within the last 30 seconds.
Creates a container for a collection of data being ingested for analysis. The dataset contains the metadata describing where the data is and what the data actually looks like. In other words, it contains the location of the data source, the data schema, and other information. A dataset also contains any tags associated with the ingested data.
"
+ "documentation":"
Creates a container for a collection of data being ingested for analysis. The dataset contains the metadata describing where the data is and what the data actually looks like. For example, it contains the location of the data source, the data schema, and other information. A dataset also contains any tags associated with the ingested data.
Deletes an ML model currently available for Amazon Lookout for Equipment. This will prevent it from being used with an inference scheduler, even one that is already set up.
Provides a JSON containing the overall information about a specific ML model, including model name and ARN, dataset, training and evaluation information, status, and so on.
Generates a list of all model versions for a given model, including the model version, model version ARN, and status. To list a subset of versions, use the MaxModelVersion and MinModelVersion fields.
Indicates that the asset associated with this sensor has been shut off. As long as this condition is met, Lookout for Equipment will not use data from this asset for training, evaluation, or inference.
The date on which the training data began being gathered. If you imported the version, this is the date that the training data in the source version began being gathered.
The date on which the training data finished being gathered. If you imported the version, this is the date that the training data in the source version finished being gathered.
The date on which the data in the evaluation set began being gathered. If you imported the version, this is the date that the evaluation set data in the source version began being gathered.
The date on which the data in the evaluation set began being gathered. If you imported the version, this is the date that the evaluation set data in the source version finished being gathered.
Shows an aggregated summary, in JSON format, of the model's performance within the evaluation time range. These metrics are created when evaluating the model.
Indicates that the asset associated with this sensor has been shut off. As long as this condition is met, Lookout for Equipment will not use data from this asset for training, evaluation, or inference.
The name of the machine learning dataset to be created. If the dataset already exists, Amazon Lookout for Equipment overwrites the existing dataset. If you don't specify this field, it is filled with the name of the source dataset.
The name for the machine learning model to be created. If the model already exists, Amazon Lookout for Equipment creates a new version. If you do not specify this field, it is filled with the name of the source model.
Pattern for matching the Amazon S3 files which will be used for ingestion. If no KeyPattern is provided, we will use the default hierarchy file structure, which is same as KeyPattern {prefix}/{component_name}/*
"
+ "documentation":"
The pattern for matching the Amazon S3 files that will be used for ingestion. If the schema was created previously without any KeyPattern, then the default KeyPattern {prefix}/{component_name}/* is used to download files from Amazon S3 according to the schema. This field is required when ingestion is being done for the first time.
If the total number of results exceeds the limit that the response can display, the response returns an opaque pagination token indicating where to continue the listing of machine learning model versions. Use this token in the NextToken field in the request to list the next page of results.
If the total number of results exceeds the limit that the response can display, the response returns an opaque pagination token indicating where to continue the listing of machine learning model versions. Use this token in the NextToken field in the request to list the next page of results.
The Amazon Resource Name (ARN) of the model version that is set as active. The active model version is the model version that the inference scheduler uses to run an inference execution.
"
}
},
"documentation":"
Provides information about the specified ML model, including dataset and model names and ARNs, as well as status.
The AWS Key Management Service (AWS KMS) key being used to encrypt the S3 object. Without this key, data in the bucket is not accessible.
"
+ "documentation":"
The Amazon Web Services Key Management Service (KMS key) key being used to encrypt the S3 object. Without this key, data in the bucket is not accessible.
The contract or wallet address on the blockchain network by which to filter the request. You must specify the address property of the ownerFilter when listing balances of tokens owned by the address.
The contract address or a token identifier on the blockchain network by which to filter the request. You must specify the contractAddress property of this container when listing tokens minted by a contract.
You must always specify the network property of this container when using this operation.
Even if additional results can be retrieved, the request can return less results than maxResults or an empty array of results.
To retrieve the next set of results, make another request with the returned nextToken value. The value of nextToken is null when there are no more results to return
Even if additional results can be retrieved, the request can return less results than maxResults or an empty array of results.
To retrieve the next set of results, make another request with the returned nextToken value. The value of nextToken is null when there are no more results to return
The container for the sort order for ListTransactions. The SortOrder field only accepts the values ASCENDING and DESCENDING. Not providing SortOrder will default to ASCENDING.
"
+ }
+ },
+ "documentation":"
The container for determining how the list transaction result will be sorted.
The request or operation couldn't be performed because a service is throttling requests. The most common source of throttling errors is when you create resources that exceed your service limit for this resource type. Request a limit increase or delete unused resources, if possible.
The block hash is a unique identifier for a block. It is a fixed-size string that is calculated by using the information in the block. The block hash is used to verify the integrity of the data in the block.
The unique identifier of the transaction. It is generated whenever a transaction is verified and added to the blockchain.
"
+ }
+ },
+ "documentation":"
There are two possible types of transactions used for this data type:
A Bitcoin transaction is a movement of BTC from one address to another.
An Ethereum transaction refers to an action initiated by an externally owned account, which is an account managed by a human, not a contract. For example, if Bob sends Alice 1 ETH, Bob's account must be debited and Alice's must be credited. This state-changing action occurs within a transaction.
Amazon Managed Blockchain (AMB) Query provides you with convenient access to multi-blockchain network data, which makes it easier for you to extract contextual data related to blockchain activity. You can use AMB Query to read data from public blockchain networks, such as Bitcoin Mainnet and Ethereum Mainnet. You can also get information such as the current and historical balances of addresses, or you can get a list of blockchain transactions for a given time period. Additionally, you can get details of a given transaction, such as transaction events, which you can further analyze or use in business logic for your applications.
"
+}
diff --git a/services/managedblockchainquery/src/main/resources/codegen-resources/waiters-2.json b/services/managedblockchainquery/src/main/resources/codegen-resources/waiters-2.json
new file mode 100644
index 00000000000..13f60ee66be
--- /dev/null
+++ b/services/managedblockchainquery/src/main/resources/codegen-resources/waiters-2.json
@@ -0,0 +1,5 @@
+{
+ "version": 2,
+ "waiters": {
+ }
+}
diff --git a/services/marketplacecatalog/pom.xml b/services/marketplacecatalog/pom.xml
index c4be7dac003..56bf88cc674 100644
--- a/services/marketplacecatalog/pom.xml
+++ b/services/marketplacecatalog/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmarketplacecatalogAWS Java SDK :: Services :: Marketplace Catalog
diff --git a/services/marketplacecommerceanalytics/pom.xml b/services/marketplacecommerceanalytics/pom.xml
index a137dc77b48..0a5d01b6bb2 100644
--- a/services/marketplacecommerceanalytics/pom.xml
+++ b/services/marketplacecommerceanalytics/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmarketplacecommerceanalyticsAWS Java SDK :: Services :: AWS Marketplace Commerce Analytics
diff --git a/services/marketplaceentitlement/pom.xml b/services/marketplaceentitlement/pom.xml
index 4d82ce7955c..20b908726c8 100644
--- a/services/marketplaceentitlement/pom.xml
+++ b/services/marketplaceentitlement/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmarketplaceentitlementAWS Java SDK :: Services :: AWS Marketplace Entitlement
diff --git a/services/marketplacemetering/pom.xml b/services/marketplacemetering/pom.xml
index 53e40936bf0..8711e7594cc 100644
--- a/services/marketplacemetering/pom.xml
+++ b/services/marketplacemetering/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmarketplacemeteringAWS Java SDK :: Services :: AWS Marketplace Metering Service
diff --git a/services/mediaconnect/pom.xml b/services/mediaconnect/pom.xml
index 31e0c926e23..b4ebc5183db 100644
--- a/services/mediaconnect/pom.xml
+++ b/services/mediaconnect/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmediaconnectAWS Java SDK :: Services :: MediaConnect
diff --git a/services/mediaconvert/pom.xml b/services/mediaconvert/pom.xml
index 8db63780f75..e3cfe40b7a4 100644
--- a/services/mediaconvert/pom.xml
+++ b/services/mediaconvert/pom.xml
@@ -20,7 +20,7 @@
servicessoftware.amazon.awssdk
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOT4.0.0mediaconvert
diff --git a/services/mediaconvert/src/main/resources/codegen-resources/service-2.json b/services/mediaconvert/src/main/resources/codegen-resources/service-2.json
index 0b7a288c566..c6ee5fd641d 100644
--- a/services/mediaconvert/src/main/resources/codegen-resources/service-2.json
+++ b/services/mediaconvert/src/main/resources/codegen-resources/service-2.json
@@ -1217,7 +1217,7 @@
"Bitrate": {
"shape": "__integerMin6000Max1024000",
"locationName": "bitrate",
- "documentation": "Specify the average bitrate in bits per second. The set of valid values for this setting is: 6000, 8000, 10000, 12000, 14000, 16000, 20000, 24000, 28000, 32000, 40000, 48000, 56000, 64000, 80000, 96000, 112000, 128000, 160000, 192000, 224000, 256000, 288000, 320000, 384000, 448000, 512000, 576000, 640000, 768000, 896000, 1024000. The value you set is also constrained by the values that you choose for Profile (codecProfile), Bitrate control mode (codingMode), and Sample rate (sampleRate). Default values depend on Bitrate control mode and Profile."
+ "documentation": "Specify the average bitrate in bits per second. The set of valid values for this setting is: 6000, 8000, 10000, 12000, 14000, 16000, 20000, 24000, 28000, 32000, 40000, 48000, 56000, 64000, 80000, 96000, 112000, 128000, 160000, 192000, 224000, 256000, 288000, 320000, 384000, 448000, 512000, 576000, 640000, 768000, 896000, 1024000. The value you set is also constrained by the values that you choose for Profile, Bitrate control mode, and Sample rate. Default values depend on Bitrate control mode and Profile."
},
"CodecProfile": {
"shape": "AacCodecProfile",
@@ -1255,7 +1255,7 @@
"documentation": "VBR Quality Level - Only used if rate_control_mode is VBR."
}
},
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value AAC. The service accepts one of two mutually exclusive groups of AAC settings--VBR and CBR. To select one of these modes, set the value of Bitrate control mode (rateControlMode) to \"VBR\" or \"CBR\". In VBR mode, you control the audio quality with the setting VBR quality (vbrQuality). In CBR mode, you use the setting Bitrate (bitrate). Defaults and valid values depend on the rate control mode."
+ "documentation": "Required when you set Codec to the value AAC. The service accepts one of two mutually exclusive groups of AAC settings--VBR and CBR. To select one of these modes, set the value of Bitrate control mode to \"VBR\" or \"CBR\". In VBR mode, you control the audio quality with the setting VBR quality. In CBR mode, you use the setting Bitrate. Defaults and valid values depend on the rate control mode."
},
"AacSpecification": {
"type": "string",
@@ -1301,7 +1301,7 @@
},
"Ac3DynamicRangeCompressionLine": {
"type": "string",
- "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the line operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile (DynamicRangeCompressionProfile). For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.",
+ "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the line operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile. For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.",
"enum": [
"FILM_STANDARD",
"FILM_LIGHT",
@@ -1313,7 +1313,7 @@
},
"Ac3DynamicRangeCompressionProfile": {
"type": "string",
- "documentation": "When you want to add Dolby dynamic range compression (DRC) signaling to your output stream, we recommend that you use the mode-specific settings instead of Dynamic range compression profile (DynamicRangeCompressionProfile). The mode-specific settings are Dynamic range compression profile, line mode (dynamicRangeCompressionLine) and Dynamic range compression profile, RF mode (dynamicRangeCompressionRf). Note that when you specify values for all three settings, MediaConvert ignores the value of this setting in favor of the mode-specific settings. If you do use this setting instead of the mode-specific settings, choose None (NONE) to leave out DRC signaling. Keep the default Film standard (FILM_STANDARD) to set the profile to Dolby's film standard profile for all operating modes.",
+ "documentation": "When you want to add Dolby dynamic range compression (DRC) signaling to your output stream, we recommend that you use the mode-specific settings instead of Dynamic range compression profile. The mode-specific settings are Dynamic range compression profile, line mode and Dynamic range compression profile, RF mode. Note that when you specify values for all three settings, MediaConvert ignores the value of this setting in favor of the mode-specific settings. If you do use this setting instead of the mode-specific settings, choose None to leave out DRC signaling. Keep the default Film standard to set the profile to Dolby's film standard profile for all operating modes.",
"enum": [
"FILM_STANDARD",
"NONE"
@@ -1321,7 +1321,7 @@
},
"Ac3DynamicRangeCompressionRf": {
"type": "string",
- "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the RF operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile (DynamicRangeCompressionProfile). For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.",
+ "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the RF operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile. For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.",
"enum": [
"FILM_STANDARD",
"FILM_LIGHT",
@@ -1373,17 +1373,17 @@
"DynamicRangeCompressionLine": {
"shape": "Ac3DynamicRangeCompressionLine",
"locationName": "dynamicRangeCompressionLine",
- "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the line operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile (DynamicRangeCompressionProfile). For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf."
+ "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the line operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile. For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf."
},
"DynamicRangeCompressionProfile": {
"shape": "Ac3DynamicRangeCompressionProfile",
"locationName": "dynamicRangeCompressionProfile",
- "documentation": "When you want to add Dolby dynamic range compression (DRC) signaling to your output stream, we recommend that you use the mode-specific settings instead of Dynamic range compression profile (DynamicRangeCompressionProfile). The mode-specific settings are Dynamic range compression profile, line mode (dynamicRangeCompressionLine) and Dynamic range compression profile, RF mode (dynamicRangeCompressionRf). Note that when you specify values for all three settings, MediaConvert ignores the value of this setting in favor of the mode-specific settings. If you do use this setting instead of the mode-specific settings, choose None (NONE) to leave out DRC signaling. Keep the default Film standard (FILM_STANDARD) to set the profile to Dolby's film standard profile for all operating modes."
+ "documentation": "When you want to add Dolby dynamic range compression (DRC) signaling to your output stream, we recommend that you use the mode-specific settings instead of Dynamic range compression profile. The mode-specific settings are Dynamic range compression profile, line mode and Dynamic range compression profile, RF mode. Note that when you specify values for all three settings, MediaConvert ignores the value of this setting in favor of the mode-specific settings. If you do use this setting instead of the mode-specific settings, choose None to leave out DRC signaling. Keep the default Film standard to set the profile to Dolby's film standard profile for all operating modes."
},
"DynamicRangeCompressionRf": {
"shape": "Ac3DynamicRangeCompressionRf",
"locationName": "dynamicRangeCompressionRf",
- "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the RF operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile (DynamicRangeCompressionProfile). For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf."
+ "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the RF operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile. For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf."
},
"LfeFilter": {
"shape": "Ac3LfeFilter",
@@ -1401,7 +1401,7 @@
"documentation": "This value is always 48000. It represents the sample rate in Hz."
}
},
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value AC3."
+ "documentation": "Required when you set Codec to the value AC3."
},
"AccelerationMode": {
"type": "string",
@@ -1479,7 +1479,7 @@
},
"AfdSignaling": {
"type": "string",
- "documentation": "This setting only applies to H.264, H.265, and MPEG2 outputs. Use Insert AFD signaling (AfdSignaling) to specify whether the service includes AFD values in the output video data and what those values are. * Choose None to remove all AFD values from this output. * Choose Fixed to ignore input AFD values and instead encode the value specified in the job. * Choose Auto to calculate output AFD values based on the input AFD scaler data.",
+ "documentation": "This setting only applies to H.264, H.265, and MPEG2 outputs. Use Insert AFD signaling to specify whether the service includes AFD values in the output video data and what those values are. * Choose None to remove all AFD values from this output. * Choose Fixed to ignore input AFD values and instead encode the value specified in the job. * Choose Auto to calculate output AFD values based on the input AFD scaler data.",
"enum": [
"NONE",
"AUTO",
@@ -1492,7 +1492,7 @@
"BitDepth": {
"shape": "__integerMin16Max24",
"locationName": "bitDepth",
- "documentation": "Specify Bit depth (BitDepth), in bits per sample, to choose the encoding quality for this audio track."
+ "documentation": "Specify Bit depth, in bits per sample, to choose the encoding quality for this audio track."
},
"Channels": {
"shape": "__integerMin1Max64",
@@ -1505,7 +1505,7 @@
"documentation": "Sample rate in hz."
}
},
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value AIFF."
+ "documentation": "Required when you set Codec to the value AIFF."
},
"AllowedRenditionSize": {
"type": "structure",
@@ -1538,7 +1538,7 @@
},
"AncillaryConvert608To708": {
"type": "string",
- "documentation": "Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert (UPCONVERT), MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708.",
+ "documentation": "Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert, MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708.",
"enum": [
"UPCONVERT",
"DISABLED"
@@ -1550,7 +1550,7 @@
"Convert608To708": {
"shape": "AncillaryConvert608To708",
"locationName": "convert608To708",
- "documentation": "Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert (UPCONVERT), MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708."
+ "documentation": "Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert, MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708."
},
"SourceAncillaryChannelNumber": {
"shape": "__integerMin1Max4",
@@ -1633,7 +1633,7 @@
},
"AudioCodec": {
"type": "string",
- "documentation": "Choose the audio codec for this output. Note that the option Dolby Digital passthrough (PASSTHROUGH) applies only to Dolby Digital and Dolby Digital Plus audio inputs. Make sure that you choose a codec that's supported with your output container: https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers.html#reference-codecs-containers-output-audio For audio-only outputs, make sure that both your input audio codec and your output audio codec are supported for audio-only workflows. For more information, see: https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers-input.html#reference-codecs-containers-input-audio-only and https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers.html#audio-only-output",
+ "documentation": "Choose the audio codec for this output. Note that the option Dolby Digital passthrough applies only to Dolby Digital and Dolby Digital Plus audio inputs. Make sure that you choose a codec that's supported with your output container: https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers.html#reference-codecs-containers-output-audio For audio-only outputs, make sure that both your input audio codec and your output audio codec are supported for audio-only workflows. For more information, see: https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers-input.html#reference-codecs-containers-input-audio-only and https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers.html#audio-only-output",
"enum": [
"AAC",
"MP2",
@@ -1654,37 +1654,37 @@
"AacSettings": {
"shape": "AacSettings",
"locationName": "aacSettings",
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value AAC. The service accepts one of two mutually exclusive groups of AAC settings--VBR and CBR. To select one of these modes, set the value of Bitrate control mode (rateControlMode) to \"VBR\" or \"CBR\". In VBR mode, you control the audio quality with the setting VBR quality (vbrQuality). In CBR mode, you use the setting Bitrate (bitrate). Defaults and valid values depend on the rate control mode."
+ "documentation": "Required when you set Codec to the value AAC. The service accepts one of two mutually exclusive groups of AAC settings--VBR and CBR. To select one of these modes, set the value of Bitrate control mode to \"VBR\" or \"CBR\". In VBR mode, you control the audio quality with the setting VBR quality. In CBR mode, you use the setting Bitrate. Defaults and valid values depend on the rate control mode."
},
"Ac3Settings": {
"shape": "Ac3Settings",
"locationName": "ac3Settings",
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value AC3."
+ "documentation": "Required when you set Codec to the value AC3."
},
"AiffSettings": {
"shape": "AiffSettings",
"locationName": "aiffSettings",
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value AIFF."
+ "documentation": "Required when you set Codec to the value AIFF."
},
"Codec": {
"shape": "AudioCodec",
"locationName": "codec",
- "documentation": "Choose the audio codec for this output. Note that the option Dolby Digital passthrough (PASSTHROUGH) applies only to Dolby Digital and Dolby Digital Plus audio inputs. Make sure that you choose a codec that's supported with your output container: https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers.html#reference-codecs-containers-output-audio For audio-only outputs, make sure that both your input audio codec and your output audio codec are supported for audio-only workflows. For more information, see: https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers-input.html#reference-codecs-containers-input-audio-only and https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers.html#audio-only-output"
+ "documentation": "Choose the audio codec for this output. Note that the option Dolby Digital passthrough applies only to Dolby Digital and Dolby Digital Plus audio inputs. Make sure that you choose a codec that's supported with your output container: https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers.html#reference-codecs-containers-output-audio For audio-only outputs, make sure that both your input audio codec and your output audio codec are supported for audio-only workflows. For more information, see: https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers-input.html#reference-codecs-containers-input-audio-only and https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers.html#audio-only-output"
},
"Eac3AtmosSettings": {
"shape": "Eac3AtmosSettings",
"locationName": "eac3AtmosSettings",
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value EAC3_ATMOS."
+ "documentation": "Required when you set Codec to the value EAC3_ATMOS."
},
"Eac3Settings": {
"shape": "Eac3Settings",
"locationName": "eac3Settings",
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value EAC3."
+ "documentation": "Required when you set Codec to the value EAC3."
},
"Mp2Settings": {
"shape": "Mp2Settings",
"locationName": "mp2Settings",
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value MP2."
+ "documentation": "Required when you set Codec to the value MP2."
},
"Mp3Settings": {
"shape": "Mp3Settings",
@@ -1704,7 +1704,7 @@
"WavSettings": {
"shape": "WavSettings",
"locationName": "wavSettings",
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value WAV."
+ "documentation": "Required when you set Codec to the value WAV."
}
},
"documentation": "Settings related to audio encoding. The settings in this group vary depending on the value that you choose for your audio codec."
@@ -1753,7 +1753,7 @@
"CustomLanguageCode": {
"shape": "__stringPatternAZaZ23AZaZ",
"locationName": "customLanguageCode",
- "documentation": "Specify the language for this audio output track. The service puts this language code into your output audio track when you set Language code control (AudioLanguageCodeControl) to Use configured (USE_CONFIGURED). The service also uses your specified custom language code when you set Language code control (AudioLanguageCodeControl) to Follow input (FOLLOW_INPUT), but your input file doesn't specify a language code. For all outputs, you can use an ISO 639-2 or ISO 639-3 code. For streaming outputs, you can also use any other code in the full RFC-5646 specification. Streaming outputs are those that are in one of the following output groups: CMAF, DASH ISO, Apple HLS, or Microsoft Smooth Streaming."
+ "documentation": "Specify the language for this audio output track. The service puts this language code into your output audio track when you set Language code control to Use configured. The service also uses your specified custom language code when you set Language code control to Follow input, but your input file doesn't specify a language code. For all outputs, you can use an ISO 639-2 or ISO 639-3 code. For streaming outputs, you can also use any other code in the full RFC-5646 specification. Streaming outputs are those that are in one of the following output groups: CMAF, DASH ISO, Apple HLS, or Microsoft Smooth Streaming."
},
"LanguageCode": {
"shape": "LanguageCode",
@@ -1763,7 +1763,7 @@
"LanguageCodeControl": {
"shape": "AudioLanguageCodeControl",
"locationName": "languageCodeControl",
- "documentation": "Specify which source for language code takes precedence for this audio track. When you choose Follow input (FOLLOW_INPUT), the service uses the language code from the input track if it's present. If there's no languge code on the input track, the service uses the code that you specify in the setting Language code (languageCode or customLanguageCode). When you choose Use configured (USE_CONFIGURED), the service uses the language code that you specify."
+ "documentation": "Specify which source for language code takes precedence for this audio track. When you choose Follow input, the service uses the language code from the input track if it's present. If there's no languge code on the input track, the service uses the code that you specify in the setting Language code. When you choose Use configured, the service uses the language code that you specify."
},
"RemixSettings": {
"shape": "RemixSettings",
@@ -1790,7 +1790,7 @@
},
"AudioLanguageCodeControl": {
"type": "string",
- "documentation": "Specify which source for language code takes precedence for this audio track. When you choose Follow input (FOLLOW_INPUT), the service uses the language code from the input track if it's present. If there's no languge code on the input track, the service uses the code that you specify in the setting Language code (languageCode or customLanguageCode). When you choose Use configured (USE_CONFIGURED), the service uses the language code that you specify.",
+ "documentation": "Specify which source for language code takes precedence for this audio track. When you choose Follow input, the service uses the language code from the input track if it's present. If there's no languge code on the input track, the service uses the code that you specify in the setting Language code. When you choose Use configured, the service uses the language code that you specify.",
"enum": [
"FOLLOW_INPUT",
"USE_CONFIGURED"
@@ -1861,7 +1861,7 @@
"TargetLkfs": {
"shape": "__doubleMinNegative59Max0",
"locationName": "targetLkfs",
- "documentation": "When you use Audio normalization (AudioNormalizationSettings), optionally use this setting to specify a target loudness. If you don't specify a value here, the encoder chooses a value for you, based on the algorithm that you choose for Algorithm (algorithm). If you choose algorithm 1770-1, the encoder will choose -24 LKFS; otherwise, the encoder will choose -23 LKFS."
+ "documentation": "When you use Audio normalization, optionally use this setting to specify a target loudness. If you don't specify a value here, the encoder chooses a value for you, based on the algorithm that you choose for Algorithm. If you choose algorithm 1770-1, the encoder will choose -24 LKFS; otherwise, the encoder will choose -23 LKFS."
},
"TruePeakLimiterThreshold": {
"shape": "__doubleMinNegative8Max0",
@@ -1917,7 +1917,7 @@
"ProgramSelection": {
"shape": "__integerMin0Max8",
"locationName": "programSelection",
- "documentation": "Use this setting for input streams that contain Dolby E, to have the service extract specific program data from the track. To select multiple programs, create multiple selectors with the same Track and different Program numbers. In the console, this setting is visible when you set Selector type to Track. Choose the program number from the dropdown list. If you are sending a JSON file, provide the program ID, which is part of the audio metadata. If your input file has incorrect metadata, you can choose All channels instead of a program number to have the service ignore the program IDs and include all the programs in the track."
+ "documentation": "Use this setting for input streams that contain Dolby E, to have the service extract specific program data from the track. To select multiple programs, create multiple selectors with the same Track and different Program numbers. In the console, this setting is visible when you set Selector type to Track. Choose the program number from the dropdown list. If your input file has incorrect metadata, you can choose All channels instead of a program number to have the service ignore the program IDs and include all the programs in the track."
},
"RemixSettings": {
"shape": "RemixSettings",
@@ -1932,10 +1932,10 @@
"Tracks": {
"shape": "__listOf__integerMin1Max2147483647",
"locationName": "tracks",
- "documentation": "Identify a track from the input audio to include in this selector by entering the track index number. To include several tracks in a single audio selector, specify multiple tracks as follows. Using the console, enter a comma-separated list. For examle, type \"1,2,3\" to include tracks 1 through 3. Specifying directly in your JSON job file, provide the track numbers in an array. For example, \"tracks\": [1,2,3]."
+ "documentation": "Identify a track from the input audio to include in this selector by entering the track index number. To include several tracks in a single audio selector, specify multiple tracks as follows. Using the console, enter a comma-separated list. For example, type \"1,2,3\" to include tracks 1 through 3."
}
},
- "documentation": "Use Audio selectors (AudioSelectors) to specify a track or set of tracks from the input that you will use in your outputs. You can use multiple Audio selectors per input."
+ "documentation": "Use Audio selectors to specify a track or set of tracks from the input that you will use in your outputs. You can use multiple Audio selectors per input."
},
"AudioSelectorGroup": {
"type": "structure",
@@ -1946,7 +1946,7 @@
"documentation": "Name of an Audio Selector within the same input to include in the group. Audio selector names are standardized, based on their order within the input (e.g., \"Audio Selector 1\"). The audio selector name parameter can be repeated to add any number of audio selectors to the group."
}
},
- "documentation": "Use audio selector groups to combine multiple sidecar audio inputs so that you can assign them to a single output audio tab (AudioDescription). Note that, if you're working with embedded audio, it's simpler to assign multiple input tracks into a single audio selector rather than use an audio selector group."
+ "documentation": "Use audio selector groups to combine multiple sidecar audio inputs so that you can assign them to a single output audio tab. Note that, if you're working with embedded audio, it's simpler to assign multiple input tracks into a single audio selector rather than use an audio selector group."
},
"AudioSelectorType": {
"type": "string",
@@ -2036,7 +2036,7 @@
},
"Av1AdaptiveQuantization": {
"type": "string",
- "documentation": "Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to Spatial adaptive quantization (spatialAdaptiveQuantization).",
+ "documentation": "Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to Spatial adaptive quantization.",
"enum": [
"OFF",
"LOW",
@@ -2048,7 +2048,7 @@
},
"Av1BitDepth": {
"type": "string",
- "documentation": "Specify the Bit depth (Av1BitDepth). You can choose 8-bit (BIT_8) or 10-bit (BIT_10).",
+ "documentation": "Specify the Bit depth. You can choose 8-bit or 10-bit.",
"enum": [
"BIT_8",
"BIT_10"
@@ -2056,7 +2056,7 @@
},
"Av1FramerateControl": {
"type": "string",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator.",
+ "documentation": "Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -2077,7 +2077,7 @@
"QvbrQualityLevel": {
"shape": "__integerMin1Max10",
"locationName": "qvbrQualityLevel",
- "documentation": "Use this setting only when you set Rate control mode (RateControlMode) to QVBR. Specify the target quality level for this output. MediaConvert determines the right number of bits to use for each part of the video to maintain the video quality that you specify. When you keep the default value, AUTO, MediaConvert picks a quality level for you, based on characteristics of your input video. If you prefer to specify a quality level, specify a number from 1 through 10. Use higher numbers for greater quality. Level 10 results in nearly lossless compression. The quality level for most broadcast-quality transcodes is between 6 and 9. Optionally, to specify a value between whole numbers, also provide a value for the setting qvbrQualityLevelFineTune. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33."
+ "documentation": "Use this setting only when you set Rate control mode to QVBR. Specify the target quality level for this output. MediaConvert determines the right number of bits to use for each part of the video to maintain the video quality that you specify. When you keep the default value, AUTO, MediaConvert picks a quality level for you, based on characteristics of your input video. If you prefer to specify a quality level, specify a number from 1 through 10. Use higher numbers for greater quality. Level 10 results in nearly lossless compression. The quality level for most broadcast-quality transcodes is between 6 and 9. Optionally, to specify a value between whole numbers, also provide a value for the setting qvbrQualityLevelFineTune. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33."
},
"QvbrQualityLevelFineTune": {
"shape": "__doubleMin0Max1",
@@ -2085,7 +2085,7 @@
"documentation": "Optional. Specify a value here to set the QVBR quality to a level that is between whole numbers. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33. MediaConvert rounds your QVBR quality level to the nearest third of a whole number. For example, if you set qvbrQualityLevel to 7 and you set qvbrQualityLevelFineTune to .25, your actual QVBR quality level is 7.33."
}
},
- "documentation": "Settings for quality-defined variable bitrate encoding with the AV1 codec. Use these settings only when you set QVBR for Rate control mode (RateControlMode)."
+ "documentation": "Settings for quality-defined variable bitrate encoding with the AV1 codec. Use these settings only when you set QVBR for Rate control mode."
},
"Av1RateControlMode": {
"type": "string",
@@ -2100,17 +2100,17 @@
"AdaptiveQuantization": {
"shape": "Av1AdaptiveQuantization",
"locationName": "adaptiveQuantization",
- "documentation": "Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to Spatial adaptive quantization (spatialAdaptiveQuantization)."
+ "documentation": "Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to Spatial adaptive quantization."
},
"BitDepth": {
"shape": "Av1BitDepth",
"locationName": "bitDepth",
- "documentation": "Specify the Bit depth (Av1BitDepth). You can choose 8-bit (BIT_8) or 10-bit (BIT_10)."
+ "documentation": "Specify the Bit depth. You can choose 8-bit or 10-bit."
},
"FramerateControl": {
"shape": "Av1FramerateControl",
"locationName": "framerateControl",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator."
+ "documentation": "Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction."
},
"FramerateConversionAlgorithm": {
"shape": "Av1FramerateConversionAlgorithm",
@@ -2145,7 +2145,7 @@
"QvbrSettings": {
"shape": "Av1QvbrSettings",
"locationName": "qvbrSettings",
- "documentation": "Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode (RateControlMode)."
+ "documentation": "Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode."
},
"RateControlMode": {
"shape": "Av1RateControlMode",
@@ -2160,14 +2160,14 @@
"SpatialAdaptiveQuantization": {
"shape": "Av1SpatialAdaptiveQuantization",
"locationName": "spatialAdaptiveQuantization",
- "documentation": "Keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization (adaptiveQuantization) depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher."
+ "documentation": "Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher."
}
},
"documentation": "Required when you set Codec, under VideoDescription>CodecSettings to the value AV1."
},
"Av1SpatialAdaptiveQuantization": {
"type": "string",
- "documentation": "Keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization (adaptiveQuantization) depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.",
+ "documentation": "Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.",
"enum": [
"DISABLED",
"ENABLED"
@@ -2196,7 +2196,7 @@
},
"AvcIntraFramerateControl": {
"type": "string",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator.",
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -2213,7 +2213,7 @@
},
"AvcIntraInterlaceMode": {
"type": "string",
- "documentation": "Choose the scan line type for the output. Keep the default value, Progressive (PROGRESSIVE) to create a progressive output, regardless of the scan type of your input. Use Top field first (TOP_FIELD) or Bottom field first (BOTTOM_FIELD) to create an output that's interlaced with the same field polarity throughout. Use Follow, default top (FOLLOW_TOP_FIELD) or Follow, default bottom (FOLLOW_BOTTOM_FIELD) to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.",
+ "documentation": "Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.",
"enum": [
"PROGRESSIVE",
"TOP_FIELD",
@@ -2224,7 +2224,7 @@
},
"AvcIntraScanTypeConversionMode": {
"type": "string",
- "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing (INTERLACED_OPTIMIZE) to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing (INTERLACED), for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine (telecine) to None (NONE) or Soft (SOFT). You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode (interlaceMode) to a value other than Progressive (PROGRESSIVE).",
+ "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.",
"enum": [
"INTERLACED",
"INTERLACED_OPTIMIZE"
@@ -2241,12 +2241,12 @@
"AvcIntraUhdSettings": {
"shape": "AvcIntraUhdSettings",
"locationName": "avcIntraUhdSettings",
- "documentation": "Optional when you set AVC-Intra class (avcIntraClass) to Class 4K/2K (CLASS_4K_2K). When you set AVC-Intra class to a different value, this object isn't allowed."
+ "documentation": "Optional when you set AVC-Intra class to Class 4K/2K. When you set AVC-Intra class to a different value, this object isn't allowed."
},
"FramerateControl": {
"shape": "AvcIntraFramerateControl",
"locationName": "framerateControl",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator."
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction."
},
"FramerateConversionAlgorithm": {
"shape": "AvcIntraFramerateConversionAlgorithm",
@@ -2266,29 +2266,29 @@
"InterlaceMode": {
"shape": "AvcIntraInterlaceMode",
"locationName": "interlaceMode",
- "documentation": "Choose the scan line type for the output. Keep the default value, Progressive (PROGRESSIVE) to create a progressive output, regardless of the scan type of your input. Use Top field first (TOP_FIELD) or Bottom field first (BOTTOM_FIELD) to create an output that's interlaced with the same field polarity throughout. Use Follow, default top (FOLLOW_TOP_FIELD) or Follow, default bottom (FOLLOW_BOTTOM_FIELD) to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose."
+ "documentation": "Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose."
},
"ScanTypeConversionMode": {
"shape": "AvcIntraScanTypeConversionMode",
"locationName": "scanTypeConversionMode",
- "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing (INTERLACED_OPTIMIZE) to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing (INTERLACED), for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine (telecine) to None (NONE) or Soft (SOFT). You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode (interlaceMode) to a value other than Progressive (PROGRESSIVE)."
+ "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive."
},
"SlowPal": {
"shape": "AvcIntraSlowPal",
"locationName": "slowPal",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1."
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25."
},
"Telecine": {
"shape": "AvcIntraTelecine",
"locationName": "telecine",
- "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine (HARD) to create a smoother picture. When you keep the default value, None (NONE), MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture."
+ "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture."
}
},
"documentation": "Required when you choose AVC-Intra for your output video codec. For more information about the AVC-Intra settings, see the relevant specification. For detailed information about SD and HD in AVC-Intra, see https://ieeexplore.ieee.org/document/7290936. For information about 4K/2K in AVC-Intra, see https://pro-av.panasonic.net/en/avc-ultra/AVC-ULTRAoverview.pdf."
},
"AvcIntraSlowPal": {
"type": "string",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1.",
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.",
"enum": [
"DISABLED",
"ENABLED"
@@ -2296,7 +2296,7 @@
},
"AvcIntraTelecine": {
"type": "string",
- "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine (HARD) to create a smoother picture. When you keep the default value, None (NONE), MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.",
+ "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.",
"enum": [
"NONE",
"HARD"
@@ -2304,7 +2304,7 @@
},
"AvcIntraUhdQualityTuningLevel": {
"type": "string",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how many transcoding passes MediaConvert does with your video. When you choose Multi-pass (MULTI_PASS), your video quality is better and your output bitrate is more accurate. That is, the actual bitrate of your output is closer to the target bitrate defined in the specification. When you choose Single-pass (SINGLE_PASS), your encoding time is faster. The default behavior is Single-pass (SINGLE_PASS).",
+ "documentation": "Optional. Use Quality tuning level to choose how many transcoding passes MediaConvert does with your video. When you choose Multi-pass, your video quality is better and your output bitrate is more accurate. That is, the actual bitrate of your output is closer to the target bitrate defined in the specification. When you choose Single-pass, your encoding time is faster. The default behavior is Single-pass.",
"enum": [
"SINGLE_PASS",
"MULTI_PASS"
@@ -2316,10 +2316,10 @@
"QualityTuningLevel": {
"shape": "AvcIntraUhdQualityTuningLevel",
"locationName": "qualityTuningLevel",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how many transcoding passes MediaConvert does with your video. When you choose Multi-pass (MULTI_PASS), your video quality is better and your output bitrate is more accurate. That is, the actual bitrate of your output is closer to the target bitrate defined in the specification. When you choose Single-pass (SINGLE_PASS), your encoding time is faster. The default behavior is Single-pass (SINGLE_PASS)."
+ "documentation": "Optional. Use Quality tuning level to choose how many transcoding passes MediaConvert does with your video. When you choose Multi-pass, your video quality is better and your output bitrate is more accurate. That is, the actual bitrate of your output is closer to the target bitrate defined in the specification. When you choose Single-pass, your encoding time is faster. The default behavior is Single-pass."
}
},
- "documentation": "Optional when you set AVC-Intra class (avcIntraClass) to Class 4K/2K (CLASS_4K_2K). When you set AVC-Intra class to a different value, this object isn't allowed."
+ "documentation": "Optional when you set AVC-Intra class to Class 4K/2K. When you set AVC-Intra class to a different value, this object isn't allowed."
},
"BadRequestException": {
"type": "structure",
@@ -2384,7 +2384,7 @@
},
"BurnInSubtitleStylePassthrough": {
"type": "string",
- "documentation": "Set Style passthrough (StylePassthrough) to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use default settings: white text with black outlining, bottom-center positioning, and automatic sizing. Whether you set Style passthrough to enabled or not, you can also choose to manually override any of the individual style and position settings.",
+ "documentation": "Set Style passthrough to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use default settings: white text with black outlining, bottom-center positioning, and automatic sizing. Whether you set Style passthrough to enabled or not, you can also choose to manually override any of the individual style and position settings.",
"enum": [
"ENABLED",
"DISABLED"
@@ -2401,27 +2401,27 @@
"ApplyFontColor": {
"shape": "BurninSubtitleApplyFontColor",
"locationName": "applyFontColor",
- "documentation": "Ignore this setting unless Style passthrough (StylePassthrough) is set to Enabled and Font color (FontColor) set to Black, Yellow, Red, Green, Blue, or Hex. Use Apply font color (ApplyFontColor) for additional font color controls. When you choose White text only (WHITE_TEXT_ONLY), or leave blank, your font color setting only applies to white text in your input captions. For example, if your font color setting is Yellow, and your input captions have red and white text, your output captions will have red and yellow text. When you choose ALL_TEXT, your font color setting applies to all of your output captions text."
+ "documentation": "Ignore this setting unless Style passthrough is set to Enabled and Font color set to Black, Yellow, Red, Green, Blue, or Hex. Use Apply font color for additional font color controls. When you choose White text only, or leave blank, your font color setting only applies to white text in your input captions. For example, if your font color setting is Yellow, and your input captions have red and white text, your output captions will have red and yellow text. When you choose ALL_TEXT, your font color setting applies to all of your output captions text."
},
"BackgroundColor": {
"shape": "BurninSubtitleBackgroundColor",
"locationName": "backgroundColor",
- "documentation": "Specify the color of the rectangle behind the captions. Leave background color (BackgroundColor) blank and set Style passthrough (StylePassthrough) to enabled to use the background color data from your input captions, if present."
+ "documentation": "Specify the color of the rectangle behind the captions. Leave background color blank and set Style passthrough to enabled to use the background color data from your input captions, if present."
},
"BackgroundOpacity": {
"shape": "__integerMin0Max255",
"locationName": "backgroundOpacity",
- "documentation": "Specify the opacity of the background rectangle. Enter a value from 0 to 255, where 0 is transparent and 255 is opaque. If Style passthrough (StylePassthrough) is set to enabled, leave blank to pass through the background style information in your input captions to your output captions. If Style passthrough is set to disabled, leave blank to use a value of 0 and remove all backgrounds from your output captions."
+ "documentation": "Specify the opacity of the background rectangle. Enter a value from 0 to 255, where 0 is transparent and 255 is opaque. If Style passthrough is set to enabled, leave blank to pass through the background style information in your input captions to your output captions. If Style passthrough is set to disabled, leave blank to use a value of 0 and remove all backgrounds from your output captions."
},
"FallbackFont": {
"shape": "BurninSubtitleFallbackFont",
"locationName": "fallbackFont",
- "documentation": "Specify the font that you want the service to use for your burn in captions when your input captions specify a font that MediaConvert doesn't support. When you set Fallback font (FallbackFont) to best match (BEST_MATCH), or leave blank, MediaConvert uses a supported font that most closely matches the font that your input captions specify. When there are multiple unsupported fonts in your input captions, MediaConvert matches each font with the supported font that matches best. When you explicitly choose a replacement font, MediaConvert uses that font to replace all unsupported fonts from your input."
+ "documentation": "Specify the font that you want the service to use for your burn in captions when your input captions specify a font that MediaConvert doesn't support. When you set Fallback font to best match, or leave blank, MediaConvert uses a supported font that most closely matches the font that your input captions specify. When there are multiple unsupported fonts in your input captions, MediaConvert matches each font with the supported font that matches best. When you explicitly choose a replacement font, MediaConvert uses that font to replace all unsupported fonts from your input."
},
"FontColor": {
"shape": "BurninSubtitleFontColor",
"locationName": "fontColor",
- "documentation": "Specify the color of the burned-in captions text. Leave Font color (FontColor) blank and set Style passthrough (StylePassthrough) to enabled to use the font color data from your input captions, if present."
+ "documentation": "Specify the color of the burned-in captions text. Leave Font color blank and set Style passthrough to enabled to use the font color data from your input captions, if present."
},
"FontOpacity": {
"shape": "__integerMin0Max255",
@@ -2431,17 +2431,17 @@
"FontResolution": {
"shape": "__integerMin96Max600",
"locationName": "fontResolution",
- "documentation": "Specify the Font resolution (FontResolution) in DPI (dots per inch)."
+ "documentation": "Specify the Font resolution in DPI (dots per inch)."
},
"FontScript": {
"shape": "FontScript",
"locationName": "fontScript",
- "documentation": "Set Font script (FontScript) to Automatically determined (AUTOMATIC), or leave blank, to automatically determine the font script in your input captions. Otherwise, set to Simplified Chinese (HANS) or Traditional Chinese (HANT) if your input font script uses Simplified or Traditional Chinese."
+ "documentation": "Set Font script to Automatically determined, or leave blank, to automatically determine the font script in your input captions. Otherwise, set to Simplified Chinese (HANS) or Traditional Chinese (HANT) if your input font script uses Simplified or Traditional Chinese."
},
"FontSize": {
"shape": "__integerMin0Max96",
"locationName": "fontSize",
- "documentation": "Specify the Font size (FontSize) in pixels. Must be a positive integer. Set to 0, or leave blank, for automatic font size."
+ "documentation": "Specify the Font size in pixels. Must be a positive integer. Set to 0, or leave blank, for automatic font size."
},
"HexFontColor": {
"shape": "__stringMin6Max8Pattern09aFAF609aFAF2",
@@ -2451,22 +2451,22 @@
"OutlineColor": {
"shape": "BurninSubtitleOutlineColor",
"locationName": "outlineColor",
- "documentation": "Specify font outline color. Leave Outline color (OutlineColor) blank and set Style passthrough (StylePassthrough) to enabled to use the font outline color data from your input captions, if present."
+ "documentation": "Specify font outline color. Leave Outline color blank and set Style passthrough to enabled to use the font outline color data from your input captions, if present."
},
"OutlineSize": {
"shape": "__integerMin0Max10",
"locationName": "outlineSize",
- "documentation": "Specify the Outline size (OutlineSize) of the caption text, in pixels. Leave Outline size blank and set Style passthrough (StylePassthrough) to enabled to use the outline size data from your input captions, if present."
+ "documentation": "Specify the Outline size of the caption text, in pixels. Leave Outline size blank and set Style passthrough to enabled to use the outline size data from your input captions, if present."
},
"ShadowColor": {
"shape": "BurninSubtitleShadowColor",
"locationName": "shadowColor",
- "documentation": "Specify the color of the shadow cast by the captions. Leave Shadow color (ShadowColor) blank and set Style passthrough (StylePassthrough) to enabled to use the shadow color data from your input captions, if present."
+ "documentation": "Specify the color of the shadow cast by the captions. Leave Shadow color blank and set Style passthrough to enabled to use the shadow color data from your input captions, if present."
},
"ShadowOpacity": {
"shape": "__integerMin0Max255",
"locationName": "shadowOpacity",
- "documentation": "Specify the opacity of the shadow. Enter a value from 0 to 255, where 0 is transparent and 255 is opaque. If Style passthrough (StylePassthrough) is set to Enabled, leave Shadow opacity (ShadowOpacity) blank to pass through the shadow style information in your input captions to your output captions. If Style passthrough is set to disabled, leave blank to use a value of 0 and remove all shadows from your output captions."
+ "documentation": "Specify the opacity of the shadow. Enter a value from 0 to 255, where 0 is transparent and 255 is opaque. If Style passthrough is set to Enabled, leave Shadow opacity blank to pass through the shadow style information in your input captions to your output captions. If Style passthrough is set to disabled, leave blank to use a value of 0 and remove all shadows from your output captions."
},
"ShadowXOffset": {
"shape": "__integerMinNegative2147483648Max2147483647",
@@ -2476,30 +2476,30 @@
"ShadowYOffset": {
"shape": "__integerMinNegative2147483648Max2147483647",
"locationName": "shadowYOffset",
- "documentation": "Specify the vertical offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels above the text. Leave Shadow y-offset (ShadowYOffset) blank and set Style passthrough (StylePassthrough) to enabled to use the shadow y-offset data from your input captions, if present."
+ "documentation": "Specify the vertical offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels above the text. Leave Shadow y-offset blank and set Style passthrough to enabled to use the shadow y-offset data from your input captions, if present."
},
"StylePassthrough": {
"shape": "BurnInSubtitleStylePassthrough",
"locationName": "stylePassthrough",
- "documentation": "Set Style passthrough (StylePassthrough) to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use default settings: white text with black outlining, bottom-center positioning, and automatic sizing. Whether you set Style passthrough to enabled or not, you can also choose to manually override any of the individual style and position settings."
+ "documentation": "Set Style passthrough to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use default settings: white text with black outlining, bottom-center positioning, and automatic sizing. Whether you set Style passthrough to enabled or not, you can also choose to manually override any of the individual style and position settings."
},
"TeletextSpacing": {
"shape": "BurninSubtitleTeletextSpacing",
"locationName": "teletextSpacing",
- "documentation": "Specify whether the text spacing (TeletextSpacing) in your captions is set by the captions grid, or varies depending on letter width. Choose fixed grid (FIXED_GRID) to conform to the spacing specified in the captions file more accurately. Choose proportional (PROPORTIONAL) to make the text easier to read for closed captions."
+ "documentation": "Specify whether the text spacing in your captions is set by the captions grid, or varies depending on letter width. Choose fixed grid to conform to the spacing specified in the captions file more accurately. Choose proportional to make the text easier to read for closed captions."
},
"XPosition": {
"shape": "__integerMin0Max2147483647",
"locationName": "xPosition",
- "documentation": "Specify the horizontal position (XPosition) of the captions, relative to the left side of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the left of the output. If no explicit x_position is provided, the horizontal caption position will be determined by the alignment parameter."
+ "documentation": "Specify the horizontal position of the captions, relative to the left side of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the left of the output. If no explicit x_position is provided, the horizontal caption position will be determined by the alignment parameter."
},
"YPosition": {
"shape": "__integerMin0Max2147483647",
"locationName": "yPosition",
- "documentation": "Specify the vertical position (YPosition) of the captions, relative to the top of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the top of the output. If no explicit y_position is provided, the caption will be positioned towards the bottom of the output."
+ "documentation": "Specify the vertical position of the captions, relative to the top of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the top of the output. If no explicit y_position is provided, the caption will be positioned towards the bottom of the output."
}
},
- "documentation": "Burn-in is a captions delivery method, rather than a captions format. Burn-in writes the captions directly on your video frames, replacing pixels of video content with the captions. Set up burn-in captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/burn-in-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to BURN_IN."
+ "documentation": "Burn-in is a captions delivery method, rather than a captions format. Burn-in writes the captions directly on your video frames, replacing pixels of video content with the captions. Set up burn-in captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/burn-in-output-captions.html."
},
"BurninSubtitleAlignment": {
"type": "string",
@@ -2512,7 +2512,7 @@
},
"BurninSubtitleApplyFontColor": {
"type": "string",
- "documentation": "Ignore this setting unless Style passthrough (StylePassthrough) is set to Enabled and Font color (FontColor) set to Black, Yellow, Red, Green, Blue, or Hex. Use Apply font color (ApplyFontColor) for additional font color controls. When you choose White text only (WHITE_TEXT_ONLY), or leave blank, your font color setting only applies to white text in your input captions. For example, if your font color setting is Yellow, and your input captions have red and white text, your output captions will have red and yellow text. When you choose ALL_TEXT, your font color setting applies to all of your output captions text.",
+ "documentation": "Ignore this setting unless Style passthrough is set to Enabled and Font color set to Black, Yellow, Red, Green, Blue, or Hex. Use Apply font color for additional font color controls. When you choose White text only, or leave blank, your font color setting only applies to white text in your input captions. For example, if your font color setting is Yellow, and your input captions have red and white text, your output captions will have red and yellow text. When you choose ALL_TEXT, your font color setting applies to all of your output captions text.",
"enum": [
"WHITE_TEXT_ONLY",
"ALL_TEXT"
@@ -2520,7 +2520,7 @@
},
"BurninSubtitleBackgroundColor": {
"type": "string",
- "documentation": "Specify the color of the rectangle behind the captions. Leave background color (BackgroundColor) blank and set Style passthrough (StylePassthrough) to enabled to use the background color data from your input captions, if present.",
+ "documentation": "Specify the color of the rectangle behind the captions. Leave background color blank and set Style passthrough to enabled to use the background color data from your input captions, if present.",
"enum": [
"NONE",
"BLACK",
@@ -2530,7 +2530,7 @@
},
"BurninSubtitleFallbackFont": {
"type": "string",
- "documentation": "Specify the font that you want the service to use for your burn in captions when your input captions specify a font that MediaConvert doesn't support. When you set Fallback font (FallbackFont) to best match (BEST_MATCH), or leave blank, MediaConvert uses a supported font that most closely matches the font that your input captions specify. When there are multiple unsupported fonts in your input captions, MediaConvert matches each font with the supported font that matches best. When you explicitly choose a replacement font, MediaConvert uses that font to replace all unsupported fonts from your input.",
+ "documentation": "Specify the font that you want the service to use for your burn in captions when your input captions specify a font that MediaConvert doesn't support. When you set Fallback font to best match, or leave blank, MediaConvert uses a supported font that most closely matches the font that your input captions specify. When there are multiple unsupported fonts in your input captions, MediaConvert matches each font with the supported font that matches best. When you explicitly choose a replacement font, MediaConvert uses that font to replace all unsupported fonts from your input.",
"enum": [
"BEST_MATCH",
"MONOSPACED_SANSSERIF",
@@ -2541,7 +2541,7 @@
},
"BurninSubtitleFontColor": {
"type": "string",
- "documentation": "Specify the color of the burned-in captions text. Leave Font color (FontColor) blank and set Style passthrough (StylePassthrough) to enabled to use the font color data from your input captions, if present.",
+ "documentation": "Specify the color of the burned-in captions text. Leave Font color blank and set Style passthrough to enabled to use the font color data from your input captions, if present.",
"enum": [
"WHITE",
"BLACK",
@@ -2555,7 +2555,7 @@
},
"BurninSubtitleOutlineColor": {
"type": "string",
- "documentation": "Specify font outline color. Leave Outline color (OutlineColor) blank and set Style passthrough (StylePassthrough) to enabled to use the font outline color data from your input captions, if present.",
+ "documentation": "Specify font outline color. Leave Outline color blank and set Style passthrough to enabled to use the font outline color data from your input captions, if present.",
"enum": [
"BLACK",
"WHITE",
@@ -2568,7 +2568,7 @@
},
"BurninSubtitleShadowColor": {
"type": "string",
- "documentation": "Specify the color of the shadow cast by the captions. Leave Shadow color (ShadowColor) blank and set Style passthrough (StylePassthrough) to enabled to use the shadow color data from your input captions, if present.",
+ "documentation": "Specify the color of the shadow cast by the captions. Leave Shadow color blank and set Style passthrough to enabled to use the shadow color data from your input captions, if present.",
"enum": [
"NONE",
"BLACK",
@@ -2578,7 +2578,7 @@
},
"BurninSubtitleTeletextSpacing": {
"type": "string",
- "documentation": "Specify whether the text spacing (TeletextSpacing) in your captions is set by the captions grid, or varies depending on letter width. Choose fixed grid (FIXED_GRID) to conform to the spacing specified in the captions file more accurately. Choose proportional (PROPORTIONAL) to make the text easier to read for closed captions.",
+ "documentation": "Specify whether the text spacing in your captions is set by the captions grid, or varies depending on letter width. Choose fixed grid to conform to the spacing specified in the captions file more accurately. Choose proportional to make the text easier to read for closed captions.",
"enum": [
"FIXED_GRID",
"PROPORTIONAL",
@@ -2620,7 +2620,7 @@
"DestinationSettings": {
"shape": "CaptionDestinationSettings",
"locationName": "destinationSettings",
- "documentation": "Settings related to one captions tab on the MediaConvert console. In your job JSON, an instance of captions DestinationSettings is equivalent to one captions tab in the console. Usually, one captions tab corresponds to one output captions track. Depending on your output captions format, one tab might correspond to a set of output captions tracks. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/including-captions.html."
+ "documentation": "Settings related to one captions tab on the MediaConvert console. Usually, one captions tab corresponds to one output captions track. Depending on your output captions format, one tab might correspond to a set of output captions tracks. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/including-captions.html."
},
"LanguageCode": {
"shape": "LanguageCode",
@@ -2646,7 +2646,7 @@
"DestinationSettings": {
"shape": "CaptionDestinationSettings",
"locationName": "destinationSettings",
- "documentation": "Settings related to one captions tab on the MediaConvert console. In your job JSON, an instance of captions DestinationSettings is equivalent to one captions tab in the console. Usually, one captions tab corresponds to one output captions track. Depending on your output captions format, one tab might correspond to a set of output captions tracks. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/including-captions.html."
+ "documentation": "Settings related to one captions tab on the MediaConvert console. Usually, one captions tab corresponds to one output captions track. Depending on your output captions format, one tab might correspond to a set of output captions tracks. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/including-captions.html."
},
"LanguageCode": {
"shape": "LanguageCode",
@@ -2667,59 +2667,59 @@
"BurninDestinationSettings": {
"shape": "BurninDestinationSettings",
"locationName": "burninDestinationSettings",
- "documentation": "Burn-in is a captions delivery method, rather than a captions format. Burn-in writes the captions directly on your video frames, replacing pixels of video content with the captions. Set up burn-in captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/burn-in-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to BURN_IN."
+ "documentation": "Burn-in is a captions delivery method, rather than a captions format. Burn-in writes the captions directly on your video frames, replacing pixels of video content with the captions. Set up burn-in captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/burn-in-output-captions.html."
},
"DestinationType": {
"shape": "CaptionDestinationType",
"locationName": "destinationType",
- "documentation": "Specify the format for this set of captions on this output. The default format is embedded without SCTE-20. Note that your choice of video output container constrains your choice of output captions format. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/captions-support-tables.html. If you are using SCTE-20 and you want to create an output that complies with the SCTE-43 spec, choose SCTE-20 plus embedded (SCTE20_PLUS_EMBEDDED). To create a non-compliant output where the embedded captions come first, choose Embedded plus SCTE-20 (EMBEDDED_PLUS_SCTE20)."
+ "documentation": "Specify the format for this set of captions on this output. The default format is embedded without SCTE-20. Note that your choice of video output container constrains your choice of output captions format. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/captions-support-tables.html. If you are using SCTE-20 and you want to create an output that complies with the SCTE-43 spec, choose SCTE-20 plus embedded. To create a non-compliant output where the embedded captions come first, choose Embedded plus SCTE-20."
},
"DvbSubDestinationSettings": {
"shape": "DvbSubDestinationSettings",
"locationName": "dvbSubDestinationSettings",
- "documentation": "Settings related to DVB-Sub captions. Set up DVB-Sub captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/dvb-sub-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to DVB_SUB."
+ "documentation": "Settings related to DVB-Sub captions. Set up DVB-Sub captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/dvb-sub-output-captions.html."
},
"EmbeddedDestinationSettings": {
"shape": "EmbeddedDestinationSettings",
"locationName": "embeddedDestinationSettings",
- "documentation": "Settings related to CEA/EIA-608 and CEA/EIA-708 (also called embedded or ancillary) captions. Set up embedded captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/embedded-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to EMBEDDED, EMBEDDED_PLUS_SCTE20, or SCTE20_PLUS_EMBEDDED."
+ "documentation": "Settings related to CEA/EIA-608 and CEA/EIA-708 (also called embedded or ancillary) captions. Set up embedded captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/embedded-output-captions.html."
},
"ImscDestinationSettings": {
"shape": "ImscDestinationSettings",
"locationName": "imscDestinationSettings",
- "documentation": "Settings related to IMSC captions. IMSC is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to IMSC."
+ "documentation": "Settings related to IMSC captions. IMSC is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html."
},
"SccDestinationSettings": {
"shape": "SccDestinationSettings",
"locationName": "sccDestinationSettings",
- "documentation": "Settings related to SCC captions. SCC is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/scc-srt-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to SCC."
+ "documentation": "Settings related to SCC captions. SCC is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/scc-srt-output-captions.html."
},
"SrtDestinationSettings": {
"shape": "SrtDestinationSettings",
"locationName": "srtDestinationSettings",
- "documentation": "Settings related to SRT captions. SRT is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to SRT."
+ "documentation": "Settings related to SRT captions. SRT is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video."
},
"TeletextDestinationSettings": {
"shape": "TeletextDestinationSettings",
"locationName": "teletextDestinationSettings",
- "documentation": "Settings related to teletext captions. Set up teletext captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/teletext-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to TELETEXT."
+ "documentation": "Settings related to teletext captions. Set up teletext captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/teletext-output-captions.html."
},
"TtmlDestinationSettings": {
"shape": "TtmlDestinationSettings",
"locationName": "ttmlDestinationSettings",
- "documentation": "Settings related to TTML captions. TTML is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to TTML."
+ "documentation": "Settings related to TTML captions. TTML is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html."
},
"WebvttDestinationSettings": {
"shape": "WebvttDestinationSettings",
"locationName": "webvttDestinationSettings",
- "documentation": "Settings related to WebVTT captions. WebVTT is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to WebVTT."
+ "documentation": "Settings related to WebVTT captions. WebVTT is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html."
}
},
- "documentation": "Settings related to one captions tab on the MediaConvert console. In your job JSON, an instance of captions DestinationSettings is equivalent to one captions tab in the console. Usually, one captions tab corresponds to one output captions track. Depending on your output captions format, one tab might correspond to a set of output captions tracks. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/including-captions.html."
+ "documentation": "Settings related to one captions tab on the MediaConvert console. Usually, one captions tab corresponds to one output captions track. Depending on your output captions format, one tab might correspond to a set of output captions tracks. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/including-captions.html."
},
"CaptionDestinationType": {
"type": "string",
- "documentation": "Specify the format for this set of captions on this output. The default format is embedded without SCTE-20. Note that your choice of video output container constrains your choice of output captions format. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/captions-support-tables.html. If you are using SCTE-20 and you want to create an output that complies with the SCTE-43 spec, choose SCTE-20 plus embedded (SCTE20_PLUS_EMBEDDED). To create a non-compliant output where the embedded captions come first, choose Embedded plus SCTE-20 (EMBEDDED_PLUS_SCTE20).",
+ "documentation": "Specify the format for this set of captions on this output. The default format is embedded without SCTE-20. Note that your choice of video output container constrains your choice of output captions format. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/captions-support-tables.html. If you are using SCTE-20 and you want to create an output that complies with the SCTE-43 spec, choose SCTE-20 plus embedded. To create a non-compliant output where the embedded captions come first, choose Embedded plus SCTE-20.",
"enum": [
"BURN_IN",
"DVB_SUB",
@@ -2770,15 +2770,15 @@
"FramerateDenominator": {
"shape": "__integerMin1Max1001",
"locationName": "framerateDenominator",
- "documentation": "Specify the denominator of the fraction that represents the frame rate for the setting Caption source frame rate (CaptionSourceFramerate). Use this setting along with the setting Framerate numerator (framerateNumerator)."
+ "documentation": "Specify the denominator of the fraction that represents the frame rate for the setting Caption source frame rate. Use this setting along with the setting Framerate numerator."
},
"FramerateNumerator": {
"shape": "__integerMin1Max60000",
"locationName": "framerateNumerator",
- "documentation": "Specify the numerator of the fraction that represents the frame rate for the setting Caption source frame rate (CaptionSourceFramerate). Use this setting along with the setting Framerate denominator (framerateDenominator)."
+ "documentation": "Specify the numerator of the fraction that represents the frame rate for the setting Caption source frame rate. Use this setting along with the setting Framerate denominator."
}
},
- "documentation": "Ignore this setting unless your input captions format is SCC. To have the service compensate for differing frame rates between your input captions and input video, specify the frame rate of the captions file. Specify this value as a fraction. When you work directly in your JSON job specification, use the settings framerateNumerator and framerateDenominator. For example, you might specify 24 / 1 for 24 fps, 25 / 1 for 25 fps, 24000 / 1001 for 23.976 fps, or 30000 / 1001 for 29.97 fps."
+ "documentation": "Ignore this setting unless your input captions format is SCC. To have the service compensate for differing frame rates between your input captions and input video, specify the frame rate of the captions file. Specify this value as a fraction. For example, you might specify 24 / 1 for 24 fps, 25 / 1 for 25 fps, 24000 / 1001 for 23.976 fps, or 30000 / 1001 for 29.97 fps."
},
"CaptionSourceSettings": {
"type": "structure",
@@ -2806,7 +2806,7 @@
"SourceType": {
"shape": "CaptionSourceType",
"locationName": "sourceType",
- "documentation": "Use Source (SourceType) to identify the format of your input captions. The service cannot auto-detect caption format."
+ "documentation": "Use Source to identify the format of your input captions. The service cannot auto-detect caption format."
},
"TeletextSourceSettings": {
"shape": "TeletextSourceSettings",
@@ -2828,7 +2828,7 @@
},
"CaptionSourceType": {
"type": "string",
- "documentation": "Use Source (SourceType) to identify the format of your input captions. The service cannot auto-detect caption format.",
+ "documentation": "Use Source to identify the format of your input captions. The service cannot auto-detect caption format.",
"enum": [
"ANCILLARY",
"DVB_SUB",
@@ -2855,7 +2855,7 @@
"documentation": "In your JSON job specification, include one child of OutputChannels for each audio channel that you want in your output. Each child should contain one instance of InputChannels or InputChannelsFineTune."
}
},
- "documentation": "Channel mapping (ChannelMapping) contains the group of fields that hold the remixing value for each channel, in dB. Specify remix values to indicate how much of the content from your input audio channel you want in your output audio channels. Each instance of the InputChannels or InputChannelsFineTune array specifies these values for one output channel. Use one instance of this array for each output channel. In the console, each array corresponds to a column in the graphical depiction of the mapping matrix. The rows of the graphical matrix correspond to input channels. Valid values are within the range from -60 (mute) through 6. A setting of 0 passes the input channel unchanged to the output channel (no attenuation or amplification). Use InputChannels or InputChannelsFineTune to specify your remix values. Don't use both."
+ "documentation": "Channel mapping contains the group of fields that hold the remixing value for each channel, in dB. Specify remix values to indicate how much of the content from your input audio channel you want in your output audio channels. Each instance of the InputChannels or InputChannelsFineTune array specifies these values for one output channel. Use one instance of this array for each output channel. In the console, each array corresponds to a column in the graphical depiction of the mapping matrix. The rows of the graphical matrix correspond to input channels. Valid values are within the range from -60 (mute) through 6. A setting of 0 passes the input channel unchanged to the output channel (no attenuation or amplification). Use InputChannels or InputChannelsFineTune to specify your remix values. Don't use both."
},
"ClipLimits": {
"type": "structure",
@@ -2901,7 +2901,7 @@
},
"CmafClientCache": {
"type": "string",
- "documentation": "Disable this setting only when your workflow requires the #EXT-X-ALLOW-CACHE:no tag. Otherwise, keep the default value Enabled (ENABLED) and control caching in your video distribution set up. For example, use the Cache-Control http header.",
+ "documentation": "Disable this setting only when your workflow requires the #EXT-X-ALLOW-CACHE:no tag. Otherwise, keep the default value Enabled and control caching in your video distribution set up. For example, use the Cache-Control http header.",
"enum": [
"DISABLED",
"ENABLED"
@@ -2926,7 +2926,7 @@
"EncryptionMethod": {
"shape": "CmafEncryptionType",
"locationName": "encryptionMethod",
- "documentation": "Specify the encryption scheme that you want the service to use when encrypting your CMAF segments. Choose AES-CBC subsample (SAMPLE-AES) or AES_CTR (AES-CTR)."
+ "documentation": "Specify the encryption scheme that you want the service to use when encrypting your CMAF segments. Choose AES-CBC subsample or AES_CTR."
},
"InitializationVectorInManifest": {
"shape": "CmafInitializationVectorInManifest",
@@ -2953,7 +2953,7 @@
},
"CmafEncryptionType": {
"type": "string",
- "documentation": "Specify the encryption scheme that you want the service to use when encrypting your CMAF segments. Choose AES-CBC subsample (SAMPLE-AES) or AES_CTR (AES-CTR).",
+ "documentation": "Specify the encryption scheme that you want the service to use when encrypting your CMAF segments. Choose AES-CBC subsample or AES_CTR.",
"enum": [
"SAMPLE_AES",
"AES_CTR"
@@ -2975,7 +2975,7 @@
"ClientCache": {
"shape": "CmafClientCache",
"locationName": "clientCache",
- "documentation": "Disable this setting only when your workflow requires the #EXT-X-ALLOW-CACHE:no tag. Otherwise, keep the default value Enabled (ENABLED) and control caching in your video distribution set up. For example, use the Cache-Control http header."
+ "documentation": "Disable this setting only when your workflow requires the #EXT-X-ALLOW-CACHE:no tag. Otherwise, keep the default value Enabled and control caching in your video distribution set up. For example, use the Cache-Control http header."
},
"CodecSpecification": {
"shape": "CmafCodecSpecification",
@@ -2990,7 +2990,7 @@
"Destination": {
"shape": "__stringPatternS3",
"locationName": "destination",
- "documentation": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file."
+ "documentation": "Use Destination to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file."
},
"DestinationSettings": {
"shape": "DestinationSettings",
@@ -3005,12 +3005,12 @@
"FragmentLength": {
"shape": "__integerMin1Max2147483647",
"locationName": "fragmentLength",
- "documentation": "Specify the length, in whole seconds, of the mp4 fragments. When you don't specify a value, MediaConvert defaults to 2. Related setting: Use Fragment length control (FragmentLengthControl) to specify whether the encoder enforces this value strictly."
+ "documentation": "Specify the length, in whole seconds, of the mp4 fragments. When you don't specify a value, MediaConvert defaults to 2. Related setting: Use Fragment length control to specify whether the encoder enforces this value strictly."
},
"ImageBasedTrickPlay": {
"shape": "CmafImageBasedTrickPlay",
"locationName": "imageBasedTrickPlay",
- "documentation": "Specify whether MediaConvert generates images for trick play. Keep the default value, None (NONE), to not generate any images. Choose Thumbnail (THUMBNAIL) to generate tiled thumbnails. Choose Thumbnail and full frame (THUMBNAIL_AND_FULLFRAME) to generate tiled thumbnails and full-resolution images of single frames. When you enable Write HLS manifest (WriteHlsManifest), MediaConvert creates a child manifest for each set of images that you generate and adds corresponding entries to the parent manifest. When you enable Write DASH manifest (WriteDashManifest), MediaConvert adds an entry in the .mpd manifest for each set of images that you generate. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md"
+ "documentation": "Specify whether MediaConvert generates images for trick play. Keep the default value, None, to not generate any images. Choose Thumbnail to generate tiled thumbnails. Choose Thumbnail and full frame to generate tiled thumbnails and full-resolution images of single frames. When you enable Write HLS manifest, MediaConvert creates a child manifest for each set of images that you generate and adds corresponding entries to the parent manifest. When you enable Write DASH manifest, MediaConvert adds an entry in the .mpd manifest for each set of images that you generate. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md"
},
"ImageBasedTrickPlaySettings": {
"shape": "CmafImageBasedTrickPlaySettings",
@@ -3045,12 +3045,12 @@
"MpdProfile": {
"shape": "CmafMpdProfile",
"locationName": "mpdProfile",
- "documentation": "Specify whether your DASH profile is on-demand or main. When you choose Main profile (MAIN_PROFILE), the service signals urn:mpeg:dash:profile:isoff-main:2011 in your .mpd DASH manifest. When you choose On-demand (ON_DEMAND_PROFILE), the service signals urn:mpeg:dash:profile:isoff-on-demand:2011 in your .mpd. When you choose On-demand, you must also set the output group setting Segment control (SegmentControl) to Single file (SINGLE_FILE)."
+ "documentation": "Specify whether your DASH profile is on-demand or main. When you choose Main profile, the service signals urn:mpeg:dash:profile:isoff-main:2011 in your .mpd DASH manifest. When you choose On-demand, the service signals urn:mpeg:dash:profile:isoff-on-demand:2011 in your .mpd. When you choose On-demand, you must also set the output group setting Segment control to Single file."
},
"PtsOffsetHandlingForBFrames": {
"shape": "CmafPtsOffsetHandlingForBFrames",
"locationName": "ptsOffsetHandlingForBFrames",
- "documentation": "Use this setting only when your output video stream has B-frames, which causes the initial presentation time stamp (PTS) to be offset from the initial decode time stamp (DTS). Specify how MediaConvert handles PTS when writing time stamps in output DASH manifests. Choose Match initial PTS (MATCH_INITIAL_PTS) when you want MediaConvert to use the initial PTS as the first time stamp in the manifest. Choose Zero-based (ZERO_BASED) to have MediaConvert ignore the initial PTS in the video stream and instead write the initial time stamp as zero in the manifest. For outputs that don't have B-frames, the time stamps in your DASH manifests start at zero regardless of your choice here."
+ "documentation": "Use this setting only when your output video stream has B-frames, which causes the initial presentation time stamp (PTS) to be offset from the initial decode time stamp (DTS). Specify how MediaConvert handles PTS when writing time stamps in output DASH manifests. Choose Match initial PTS when you want MediaConvert to use the initial PTS as the first time stamp in the manifest. Choose Zero-based to have MediaConvert ignore the initial PTS in the video stream and instead write the initial time stamp as zero in the manifest. For outputs that don't have B-frames, the time stamps in your DASH manifests start at zero regardless of your choice here."
},
"SegmentControl": {
"shape": "CmafSegmentControl",
@@ -3060,12 +3060,12 @@
"SegmentLength": {
"shape": "__integerMin1Max2147483647",
"locationName": "segmentLength",
- "documentation": "Specify the length, in whole seconds, of each segment. When you don't specify a value, MediaConvert defaults to 10. Related settings: Use Segment length control (SegmentLengthControl) to specify whether the encoder enforces this value strictly. Use Segment control (CmafSegmentControl) to specify whether MediaConvert creates separate segment files or one content file that has metadata to mark the segment boundaries."
+ "documentation": "Specify the length, in whole seconds, of each segment. When you don't specify a value, MediaConvert defaults to 10. Related settings: Use Segment length control to specify whether the encoder enforces this value strictly. Use Segment control to specify whether MediaConvert creates separate segment files or one content file that has metadata to mark the segment boundaries."
},
"SegmentLengthControl": {
"shape": "CmafSegmentLengthControl",
"locationName": "segmentLengthControl",
- "documentation": "Specify how you want MediaConvert to determine the segment length. Choose Exact (EXACT) to have the encoder use the exact length that you specify with the setting Segment length (SegmentLength). This might result in extra I-frames. Choose Multiple of GOP (GOP_MULTIPLE) to have the encoder round up the segment lengths to match the next GOP boundary."
+ "documentation": "Specify how you want MediaConvert to determine the segment length. Choose Exact to have the encoder use the exact length that you specify with the setting Segment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary."
},
"StreamInfResolution": {
"shape": "CmafStreamInfResolution",
@@ -3095,14 +3095,14 @@
"WriteSegmentTimelineInRepresentation": {
"shape": "CmafWriteSegmentTimelineInRepresentation",
"locationName": "writeSegmentTimelineInRepresentation",
- "documentation": "When you enable Precise segment duration in DASH manifests (writeSegmentTimelineInRepresentation), your DASH manifest shows precise segment durations. The segment duration information appears inside the SegmentTimeline element, inside SegmentTemplate at the Representation level. When this feature isn't enabled, the segment durations in your DASH manifest are approximate. The segment duration information appears in the duration attribute of the SegmentTemplate element."
+ "documentation": "When you enable Precise segment duration in DASH manifests, your DASH manifest shows precise segment durations. The segment duration information appears inside the SegmentTimeline element, inside SegmentTemplate at the Representation level. When this feature isn't enabled, the segment durations in your DASH manifest are approximate. The segment duration information appears in the duration attribute of the SegmentTemplate element."
}
},
- "documentation": "Settings related to your CMAF output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html. When you work directly in your JSON job specification, include this object and any required children when you set Type, under OutputGroupSettings, to CMAF_GROUP_SETTINGS."
+ "documentation": "Settings related to your CMAF output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html."
},
"CmafImageBasedTrickPlay": {
"type": "string",
- "documentation": "Specify whether MediaConvert generates images for trick play. Keep the default value, None (NONE), to not generate any images. Choose Thumbnail (THUMBNAIL) to generate tiled thumbnails. Choose Thumbnail and full frame (THUMBNAIL_AND_FULLFRAME) to generate tiled thumbnails and full-resolution images of single frames. When you enable Write HLS manifest (WriteHlsManifest), MediaConvert creates a child manifest for each set of images that you generate and adds corresponding entries to the parent manifest. When you enable Write DASH manifest (WriteDashManifest), MediaConvert adds an entry in the .mpd manifest for each set of images that you generate. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md",
+ "documentation": "Specify whether MediaConvert generates images for trick play. Keep the default value, None, to not generate any images. Choose Thumbnail to generate tiled thumbnails. Choose Thumbnail and full frame to generate tiled thumbnails and full-resolution images of single frames. When you enable Write HLS manifest, MediaConvert creates a child manifest for each set of images that you generate and adds corresponding entries to the parent manifest. When you enable Write DASH manifest, MediaConvert adds an entry in the .mpd manifest for each set of images that you generate. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md",
"enum": [
"NONE",
"THUMBNAIL",
@@ -3196,7 +3196,7 @@
},
"CmafMpdProfile": {
"type": "string",
- "documentation": "Specify whether your DASH profile is on-demand or main. When you choose Main profile (MAIN_PROFILE), the service signals urn:mpeg:dash:profile:isoff-main:2011 in your .mpd DASH manifest. When you choose On-demand (ON_DEMAND_PROFILE), the service signals urn:mpeg:dash:profile:isoff-on-demand:2011 in your .mpd. When you choose On-demand, you must also set the output group setting Segment control (SegmentControl) to Single file (SINGLE_FILE).",
+ "documentation": "Specify whether your DASH profile is on-demand or main. When you choose Main profile, the service signals urn:mpeg:dash:profile:isoff-main:2011 in your .mpd DASH manifest. When you choose On-demand, the service signals urn:mpeg:dash:profile:isoff-on-demand:2011 in your .mpd. When you choose On-demand, you must also set the output group setting Segment control to Single file.",
"enum": [
"MAIN_PROFILE",
"ON_DEMAND_PROFILE"
@@ -3204,7 +3204,7 @@
},
"CmafPtsOffsetHandlingForBFrames": {
"type": "string",
- "documentation": "Use this setting only when your output video stream has B-frames, which causes the initial presentation time stamp (PTS) to be offset from the initial decode time stamp (DTS). Specify how MediaConvert handles PTS when writing time stamps in output DASH manifests. Choose Match initial PTS (MATCH_INITIAL_PTS) when you want MediaConvert to use the initial PTS as the first time stamp in the manifest. Choose Zero-based (ZERO_BASED) to have MediaConvert ignore the initial PTS in the video stream and instead write the initial time stamp as zero in the manifest. For outputs that don't have B-frames, the time stamps in your DASH manifests start at zero regardless of your choice here.",
+ "documentation": "Use this setting only when your output video stream has B-frames, which causes the initial presentation time stamp (PTS) to be offset from the initial decode time stamp (DTS). Specify how MediaConvert handles PTS when writing time stamps in output DASH manifests. Choose Match initial PTS when you want MediaConvert to use the initial PTS as the first time stamp in the manifest. Choose Zero-based to have MediaConvert ignore the initial PTS in the video stream and instead write the initial time stamp as zero in the manifest. For outputs that don't have B-frames, the time stamps in your DASH manifests start at zero regardless of your choice here.",
"enum": [
"ZERO_BASED",
"MATCH_INITIAL_PTS"
@@ -3220,7 +3220,7 @@
},
"CmafSegmentLengthControl": {
"type": "string",
- "documentation": "Specify how you want MediaConvert to determine the segment length. Choose Exact (EXACT) to have the encoder use the exact length that you specify with the setting Segment length (SegmentLength). This might result in extra I-frames. Choose Multiple of GOP (GOP_MULTIPLE) to have the encoder round up the segment lengths to match the next GOP boundary.",
+ "documentation": "Specify how you want MediaConvert to determine the segment length. Choose Exact to have the encoder use the exact length that you specify with the setting Segment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary.",
"enum": [
"EXACT",
"GOP_MULTIPLE"
@@ -3268,7 +3268,7 @@
},
"CmafWriteSegmentTimelineInRepresentation": {
"type": "string",
- "documentation": "When you enable Precise segment duration in DASH manifests (writeSegmentTimelineInRepresentation), your DASH manifest shows precise segment durations. The segment duration information appears inside the SegmentTimeline element, inside SegmentTemplate at the Representation level. When this feature isn't enabled, the segment durations in your DASH manifest are approximate. The segment duration information appears in the duration attribute of the SegmentTemplate element.",
+ "documentation": "When you enable Precise segment duration in DASH manifests, your DASH manifest shows precise segment durations. The segment duration information appears inside the SegmentTimeline element, inside SegmentTemplate at the Representation level. When this feature isn't enabled, the segment durations in your DASH manifest are approximate. The segment duration information appears in the duration attribute of the SegmentTemplate element.",
"enum": [
"ENABLED",
"DISABLED"
@@ -3276,7 +3276,7 @@
},
"CmfcAudioDuration": {
"type": "string",
- "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration (MATCH_VIDEO_DURATION). In all other cases, keep the default value, Default codec duration (DEFAULT_CODEC_DURATION). When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.",
+ "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.",
"enum": [
"DEFAULT_CODEC_DURATION",
"MATCH_VIDEO_DURATION"
@@ -3284,7 +3284,7 @@
},
"CmfcAudioTrackType": {
"type": "string",
- "documentation": "Use this setting to control the values that MediaConvert puts in your HLS parent playlist to control how the client player selects which audio track to play. The other options for this setting determine the values that MediaConvert writes for the DEFAULT and AUTOSELECT attributes of the EXT-X-MEDIA entry for the audio variant. For more information about these attributes, see the Apple documentation article https://developer.apple.com/documentation/http_live_streaming/example_playlists_for_http_live_streaming/adding_alternate_media_to_a_playlist. Choose Alternate audio, auto select, default (ALTERNATE_AUDIO_AUTO_SELECT_DEFAULT) to set DEFAULT=YES and AUTOSELECT=YES. Choose this value for only one variant in your output group. Choose Alternate audio, auto select, not default (ALTERNATE_AUDIO_AUTO_SELECT) to set DEFAULT=NO and AUTOSELECT=YES. Choose Alternate Audio, Not Auto Select to set DEFAULT=NO and AUTOSELECT=NO. When you don't specify a value for this setting, MediaConvert defaults to Alternate audio, auto select, default. When there is more than one variant in your output group, you must explicitly choose a value for this setting.",
+ "documentation": "Use this setting to control the values that MediaConvert puts in your HLS parent playlist to control how the client player selects which audio track to play. The other options for this setting determine the values that MediaConvert writes for the DEFAULT and AUTOSELECT attributes of the EXT-X-MEDIA entry for the audio variant. For more information about these attributes, see the Apple documentation article https://developer.apple.com/documentation/http_live_streaming/example_playlists_for_http_live_streaming/adding_alternate_media_to_a_playlist. Choose Alternate audio, auto select, default to set DEFAULT=YES and AUTOSELECT=YES. Choose this value for only one variant in your output group. Choose Alternate audio, auto select, not default to set DEFAULT=NO and AUTOSELECT=YES. Choose Alternate Audio, Not Auto Select to set DEFAULT=NO and AUTOSELECT=NO. When you don't specify a value for this setting, MediaConvert defaults to Alternate audio, auto select, default. When there is more than one variant in your output group, you must explicitly choose a value for this setting.",
"enum": [
"ALTERNATE_AUDIO_AUTO_SELECT_DEFAULT",
"ALTERNATE_AUDIO_AUTO_SELECT",
@@ -3293,7 +3293,7 @@
},
"CmfcDescriptiveVideoServiceFlag": {
"type": "string",
- "documentation": "Specify whether to flag this audio track as descriptive video service (DVS) in your HLS parent manifest. When you choose Flag (FLAG), MediaConvert includes the parameter CHARACTERISTICS=\"public.accessibility.describes-video\" in the EXT-X-MEDIA entry for this track. When you keep the default choice, Don't flag (DONT_FLAG), MediaConvert leaves this parameter out. The DVS flag can help with accessibility on Apple devices. For more information, see the Apple documentation.",
+ "documentation": "Specify whether to flag this audio track as descriptive video service (DVS) in your HLS parent manifest. When you choose Flag, MediaConvert includes the parameter CHARACTERISTICS=\"public.accessibility.describes-video\" in the EXT-X-MEDIA entry for this track. When you keep the default choice, Don't flag, MediaConvert leaves this parameter out. The DVS flag can help with accessibility on Apple devices. For more information, see the Apple documentation.",
"enum": [
"DONT_FLAG",
"FLAG"
@@ -3301,7 +3301,7 @@
},
"CmfcIFrameOnlyManifest": {
"type": "string",
- "documentation": "Choose Include (INCLUDE) to have MediaConvert generate an HLS child manifest that lists only the I-frames for this rendition, in addition to your regular manifest for this rendition. You might use this manifest as part of a workflow that creates preview functions for your video. MediaConvert adds both the I-frame only child manifest and the regular child manifest to the parent manifest. When you don't need the I-frame only child manifest, keep the default value Exclude (EXCLUDE).",
+ "documentation": "Choose Include to have MediaConvert generate an HLS child manifest that lists only the I-frames for this rendition, in addition to your regular manifest for this rendition. You might use this manifest as part of a workflow that creates preview functions for your video. MediaConvert adds both the I-frame only child manifest and the regular child manifest to the parent manifest. When you don't need the I-frame only child manifest, keep the default value Exclude.",
"enum": [
"INCLUDE",
"EXCLUDE"
@@ -3317,7 +3317,7 @@
},
"CmfcManifestMetadataSignaling": {
"type": "string",
- "documentation": "To add an InbandEventStream element in your output MPD manifest for each type of event message, set Manifest metadata signaling to Enabled. For ID3 event messages, the InbandEventStream element schemeIdUri will be same value that you specify for ID3 metadata scheme ID URI. For SCTE35 event messages, the InbandEventStream element schemeIdUri will be \"urn:scte:scte35:2013:bin\". To leave these elements out of your output MPD manifest, set Manifest metadata signaling to Disabled. To enable Manifest metadata signaling, you must also set SCTE-35 source to Passthrough, ESAM SCTE-35 to insert, or ID3 metadata (TimedMetadata) to Passthrough.",
+ "documentation": "To add an InbandEventStream element in your output MPD manifest for each type of event message, set Manifest metadata signaling to Enabled. For ID3 event messages, the InbandEventStream element schemeIdUri will be same value that you specify for ID3 metadata scheme ID URI. For SCTE35 event messages, the InbandEventStream element schemeIdUri will be \"urn:scte:scte35:2013:bin\". To leave these elements out of your output MPD manifest, set Manifest metadata signaling to Disabled. To enable Manifest metadata signaling, you must also set SCTE-35 source to Passthrough, ESAM SCTE-35 to insert, or ID3 metadata to Passthrough.",
"enum": [
"ENABLED",
"DISABLED"
@@ -3325,7 +3325,7 @@
},
"CmfcScte35Esam": {
"type": "string",
- "documentation": "Use this setting only when you specify SCTE-35 markers from ESAM. Choose INSERT to put SCTE-35 markers in this output at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML (sccXml).",
+ "documentation": "Use this setting only when you specify SCTE-35 markers from ESAM. Choose INSERT to put SCTE-35 markers in this output at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML.",
"enum": [
"INSERT",
"NONE"
@@ -3333,7 +3333,7 @@
},
"CmfcScte35Source": {
"type": "string",
- "documentation": "Ignore this setting unless you have SCTE-35 markers in your input video file. Choose Passthrough (PASSTHROUGH) if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None (NONE) if you don't want those SCTE-35 markers in this output.",
+ "documentation": "Ignore this setting unless you have SCTE-35 markers in your input video file. Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want those SCTE-35 markers in this output.",
"enum": [
"PASSTHROUGH",
"NONE"
@@ -3345,32 +3345,32 @@
"AudioDuration": {
"shape": "CmfcAudioDuration",
"locationName": "audioDuration",
- "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration (MATCH_VIDEO_DURATION). In all other cases, keep the default value, Default codec duration (DEFAULT_CODEC_DURATION). When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec."
+ "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec."
},
"AudioGroupId": {
"shape": "__string",
"locationName": "audioGroupId",
- "documentation": "Specify the audio rendition group for this audio rendition. Specify up to one value for each audio output in your output group. This value appears in your HLS parent manifest in the EXT-X-MEDIA tag of TYPE=AUDIO, as the value for the GROUP-ID attribute. For example, if you specify \"audio_aac_1\" for Audio group ID, it appears in your manifest like this: #EXT-X-MEDIA:TYPE=AUDIO,GROUP-ID=\"audio_aac_1\". Related setting: To associate the rendition group that this audio track belongs to with a video rendition, include the same value that you provide here for that video output's setting Audio rendition sets (audioRenditionSets)."
+ "documentation": "Specify the audio rendition group for this audio rendition. Specify up to one value for each audio output in your output group. This value appears in your HLS parent manifest in the EXT-X-MEDIA tag of TYPE=AUDIO, as the value for the GROUP-ID attribute. For example, if you specify \"audio_aac_1\" for Audio group ID, it appears in your manifest like this: #EXT-X-MEDIA:TYPE=AUDIO,GROUP-ID=\"audio_aac_1\". Related setting: To associate the rendition group that this audio track belongs to with a video rendition, include the same value that you provide here for that video output's setting Audio rendition sets."
},
"AudioRenditionSets": {
"shape": "__string",
"locationName": "audioRenditionSets",
- "documentation": "List the audio rendition groups that you want included with this video rendition. Use a comma-separated list. For example, say you want to include the audio rendition groups that have the audio group IDs \"audio_aac_1\" and \"audio_dolby\". Then you would specify this value: \"audio_aac_1,audio_dolby\". Related setting: The rendition groups that you include in your comma-separated list should all match values that you specify in the setting Audio group ID (AudioGroupId) for audio renditions in the same output group as this video rendition. Default behavior: If you don't specify anything here and for Audio group ID, MediaConvert puts each audio variant in its own audio rendition group and associates it with every video variant. Each value in your list appears in your HLS parent manifest in the EXT-X-STREAM-INF tag as the value for the AUDIO attribute. To continue the previous example, say that the file name for the child manifest for your video rendition is \"amazing_video_1.m3u8\". Then, in your parent manifest, each value will appear on separate lines, like this: #EXT-X-STREAM-INF:AUDIO=\"audio_aac_1\"... amazing_video_1.m3u8 #EXT-X-STREAM-INF:AUDIO=\"audio_dolby\"... amazing_video_1.m3u8"
+ "documentation": "List the audio rendition groups that you want included with this video rendition. Use a comma-separated list. For example, say you want to include the audio rendition groups that have the audio group IDs \"audio_aac_1\" and \"audio_dolby\". Then you would specify this value: \"audio_aac_1,audio_dolby\". Related setting: The rendition groups that you include in your comma-separated list should all match values that you specify in the setting Audio group ID for audio renditions in the same output group as this video rendition. Default behavior: If you don't specify anything here and for Audio group ID, MediaConvert puts each audio variant in its own audio rendition group and associates it with every video variant. Each value in your list appears in your HLS parent manifest in the EXT-X-STREAM-INF tag as the value for the AUDIO attribute. To continue the previous example, say that the file name for the child manifest for your video rendition is \"amazing_video_1.m3u8\". Then, in your parent manifest, each value will appear on separate lines, like this: #EXT-X-STREAM-INF:AUDIO=\"audio_aac_1\"... amazing_video_1.m3u8 #EXT-X-STREAM-INF:AUDIO=\"audio_dolby\"... amazing_video_1.m3u8"
},
"AudioTrackType": {
"shape": "CmfcAudioTrackType",
"locationName": "audioTrackType",
- "documentation": "Use this setting to control the values that MediaConvert puts in your HLS parent playlist to control how the client player selects which audio track to play. The other options for this setting determine the values that MediaConvert writes for the DEFAULT and AUTOSELECT attributes of the EXT-X-MEDIA entry for the audio variant. For more information about these attributes, see the Apple documentation article https://developer.apple.com/documentation/http_live_streaming/example_playlists_for_http_live_streaming/adding_alternate_media_to_a_playlist. Choose Alternate audio, auto select, default (ALTERNATE_AUDIO_AUTO_SELECT_DEFAULT) to set DEFAULT=YES and AUTOSELECT=YES. Choose this value for only one variant in your output group. Choose Alternate audio, auto select, not default (ALTERNATE_AUDIO_AUTO_SELECT) to set DEFAULT=NO and AUTOSELECT=YES. Choose Alternate Audio, Not Auto Select to set DEFAULT=NO and AUTOSELECT=NO. When you don't specify a value for this setting, MediaConvert defaults to Alternate audio, auto select, default. When there is more than one variant in your output group, you must explicitly choose a value for this setting."
+ "documentation": "Use this setting to control the values that MediaConvert puts in your HLS parent playlist to control how the client player selects which audio track to play. The other options for this setting determine the values that MediaConvert writes for the DEFAULT and AUTOSELECT attributes of the EXT-X-MEDIA entry for the audio variant. For more information about these attributes, see the Apple documentation article https://developer.apple.com/documentation/http_live_streaming/example_playlists_for_http_live_streaming/adding_alternate_media_to_a_playlist. Choose Alternate audio, auto select, default to set DEFAULT=YES and AUTOSELECT=YES. Choose this value for only one variant in your output group. Choose Alternate audio, auto select, not default to set DEFAULT=NO and AUTOSELECT=YES. Choose Alternate Audio, Not Auto Select to set DEFAULT=NO and AUTOSELECT=NO. When you don't specify a value for this setting, MediaConvert defaults to Alternate audio, auto select, default. When there is more than one variant in your output group, you must explicitly choose a value for this setting."
},
"DescriptiveVideoServiceFlag": {
"shape": "CmfcDescriptiveVideoServiceFlag",
"locationName": "descriptiveVideoServiceFlag",
- "documentation": "Specify whether to flag this audio track as descriptive video service (DVS) in your HLS parent manifest. When you choose Flag (FLAG), MediaConvert includes the parameter CHARACTERISTICS=\"public.accessibility.describes-video\" in the EXT-X-MEDIA entry for this track. When you keep the default choice, Don't flag (DONT_FLAG), MediaConvert leaves this parameter out. The DVS flag can help with accessibility on Apple devices. For more information, see the Apple documentation."
+ "documentation": "Specify whether to flag this audio track as descriptive video service (DVS) in your HLS parent manifest. When you choose Flag, MediaConvert includes the parameter CHARACTERISTICS=\"public.accessibility.describes-video\" in the EXT-X-MEDIA entry for this track. When you keep the default choice, Don't flag, MediaConvert leaves this parameter out. The DVS flag can help with accessibility on Apple devices. For more information, see the Apple documentation."
},
"IFrameOnlyManifest": {
"shape": "CmfcIFrameOnlyManifest",
"locationName": "iFrameOnlyManifest",
- "documentation": "Choose Include (INCLUDE) to have MediaConvert generate an HLS child manifest that lists only the I-frames for this rendition, in addition to your regular manifest for this rendition. You might use this manifest as part of a workflow that creates preview functions for your video. MediaConvert adds both the I-frame only child manifest and the regular child manifest to the parent manifest. When you don't need the I-frame only child manifest, keep the default value Exclude (EXCLUDE)."
+ "documentation": "Choose Include to have MediaConvert generate an HLS child manifest that lists only the I-frames for this rendition, in addition to your regular manifest for this rendition. You might use this manifest as part of a workflow that creates preview functions for your video. MediaConvert adds both the I-frame only child manifest and the regular child manifest to the parent manifest. When you don't need the I-frame only child manifest, keep the default value Exclude."
},
"KlvMetadata": {
"shape": "CmfcKlvMetadata",
@@ -3380,44 +3380,44 @@
"ManifestMetadataSignaling": {
"shape": "CmfcManifestMetadataSignaling",
"locationName": "manifestMetadataSignaling",
- "documentation": "To add an InbandEventStream element in your output MPD manifest for each type of event message, set Manifest metadata signaling to Enabled. For ID3 event messages, the InbandEventStream element schemeIdUri will be same value that you specify for ID3 metadata scheme ID URI. For SCTE35 event messages, the InbandEventStream element schemeIdUri will be \"urn:scte:scte35:2013:bin\". To leave these elements out of your output MPD manifest, set Manifest metadata signaling to Disabled. To enable Manifest metadata signaling, you must also set SCTE-35 source to Passthrough, ESAM SCTE-35 to insert, or ID3 metadata (TimedMetadata) to Passthrough."
+ "documentation": "To add an InbandEventStream element in your output MPD manifest for each type of event message, set Manifest metadata signaling to Enabled. For ID3 event messages, the InbandEventStream element schemeIdUri will be same value that you specify for ID3 metadata scheme ID URI. For SCTE35 event messages, the InbandEventStream element schemeIdUri will be \"urn:scte:scte35:2013:bin\". To leave these elements out of your output MPD manifest, set Manifest metadata signaling to Disabled. To enable Manifest metadata signaling, you must also set SCTE-35 source to Passthrough, ESAM SCTE-35 to insert, or ID3 metadata to Passthrough."
},
"Scte35Esam": {
"shape": "CmfcScte35Esam",
"locationName": "scte35Esam",
- "documentation": "Use this setting only when you specify SCTE-35 markers from ESAM. Choose INSERT to put SCTE-35 markers in this output at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML (sccXml)."
+ "documentation": "Use this setting only when you specify SCTE-35 markers from ESAM. Choose INSERT to put SCTE-35 markers in this output at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML."
},
"Scte35Source": {
"shape": "CmfcScte35Source",
"locationName": "scte35Source",
- "documentation": "Ignore this setting unless you have SCTE-35 markers in your input video file. Choose Passthrough (PASSTHROUGH) if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None (NONE) if you don't want those SCTE-35 markers in this output."
+ "documentation": "Ignore this setting unless you have SCTE-35 markers in your input video file. Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want those SCTE-35 markers in this output."
},
"TimedMetadata": {
"shape": "CmfcTimedMetadata",
"locationName": "timedMetadata",
- "documentation": "To include ID3 metadata in this output: Set ID3 metadata (timedMetadata) to Passthrough (PASSTHROUGH). Specify this ID3 metadata in Custom ID3 metadata inserter (timedMetadataInsertion). MediaConvert writes each instance of ID3 metadata in a separate Event Message (eMSG) box. To exclude this ID3 metadata: Set ID3 metadata to None (NONE) or leave blank."
+ "documentation": "To include ID3 metadata in this output: Set ID3 metadata to Passthrough. Specify this ID3 metadata in Custom ID3 metadata inserter. MediaConvert writes each instance of ID3 metadata in a separate Event Message (eMSG) box. To exclude this ID3 metadata: Set ID3 metadata to None or leave blank."
},
"TimedMetadataBoxVersion": {
"shape": "CmfcTimedMetadataBoxVersion",
"locationName": "timedMetadataBoxVersion",
- "documentation": "Specify the event message box (eMSG) version for ID3 timed metadata in your output.\nFor more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.3 Syntax.\nLeave blank to use the default value Version 0.\nWhen you specify Version 1, you must also set ID3 metadata (timedMetadata) to Passthrough."
+ "documentation": "Specify the event message box (eMSG) version for ID3 timed metadata in your output.\nFor more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.3 Syntax.\nLeave blank to use the default value Version 0.\nWhen you specify Version 1, you must also set ID3 metadata to Passthrough."
},
"TimedMetadataSchemeIdUri": {
"shape": "__stringMax1000",
"locationName": "timedMetadataSchemeIdUri",
- "documentation": "Specify the event message box (eMSG) scheme ID URI (scheme_id_uri) for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. Leave blank to use the default value: https://aomedia.org/emsg/ID3 When you specify a value for ID3 metadata scheme ID URI, you must also set ID3 metadata (timedMetadata) to Passthrough."
+ "documentation": "Specify the event message box (eMSG) scheme ID URI for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. Leave blank to use the default value: https://aomedia.org/emsg/ID3 When you specify a value for ID3 metadata scheme ID URI, you must also set ID3 metadata to Passthrough."
},
"TimedMetadataValue": {
"shape": "__stringMax1000",
"locationName": "timedMetadataValue",
- "documentation": "Specify the event message box (eMSG) value for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. When you specify a value for ID3 Metadata Value, you must also set ID3 metadata (timedMetadata) to Passthrough."
+ "documentation": "Specify the event message box (eMSG) value for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. When you specify a value for ID3 Metadata Value, you must also set ID3 metadata to Passthrough."
}
},
"documentation": "These settings relate to the fragmented MP4 container for the segments in your CMAF outputs."
},
"CmfcTimedMetadata": {
"type": "string",
- "documentation": "To include ID3 metadata in this output: Set ID3 metadata (timedMetadata) to Passthrough (PASSTHROUGH). Specify this ID3 metadata in Custom ID3 metadata inserter (timedMetadataInsertion). MediaConvert writes each instance of ID3 metadata in a separate Event Message (eMSG) box. To exclude this ID3 metadata: Set ID3 metadata to None (NONE) or leave blank.",
+ "documentation": "To include ID3 metadata in this output: Set ID3 metadata to Passthrough. Specify this ID3 metadata in Custom ID3 metadata inserter. MediaConvert writes each instance of ID3 metadata in a separate Event Message (eMSG) box. To exclude this ID3 metadata: Set ID3 metadata to None or leave blank.",
"enum": [
"PASSTHROUGH",
"NONE"
@@ -3425,7 +3425,7 @@
},
"CmfcTimedMetadataBoxVersion": {
"type": "string",
- "documentation": "Specify the event message box (eMSG) version for ID3 timed metadata in your output.\nFor more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.3 Syntax.\nLeave blank to use the default value Version 0.\nWhen you specify Version 1, you must also set ID3 metadata (timedMetadata) to Passthrough.",
+ "documentation": "Specify the event message box (eMSG) version for ID3 timed metadata in your output.\nFor more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.3 Syntax.\nLeave blank to use the default value Version 0.\nWhen you specify Version 1, you must also set ID3 metadata to Passthrough.",
"enum": [
"VERSION_0",
"VERSION_1"
@@ -3457,7 +3457,7 @@
"Hdr10Metadata": {
"shape": "Hdr10Metadata",
"locationName": "hdr10Metadata",
- "documentation": "Use these settings when you convert to the HDR 10 color space. Specify the SMPTE ST 2086 Mastering Display Color Volume static metadata that you want signaled in the output. These values don't affect the pixel values that are encoded in the video stream. They are intended to help the downstream video player display content in a way that reflects the intentions of the the content creator. When you set Color space conversion (ColorSpaceConversion) to HDR 10 (FORCE_HDR10), these settings are required. You must set values for Max frame average light level (maxFrameAverageLightLevel) and Max content light level (maxContentLightLevel); these settings don't have a default value. The default values for the other HDR 10 metadata settings are defined by the P3D65 color space. For more information about MediaConvert HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr."
+ "documentation": "Use these settings when you convert to the HDR 10 color space. Specify the SMPTE ST 2086 Mastering Display Color Volume static metadata that you want signaled in the output. These values don't affect the pixel values that are encoded in the video stream. They are intended to help the downstream video player display content in a way that reflects the intentions of the the content creator. When you set Color space conversion to HDR 10, these settings are required. You must set values for Max frame average light level and Max content light level; these settings don't have a default value. The default values for the other HDR 10 metadata settings are defined by the P3D65 color space. For more information about MediaConvert HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr."
},
"HdrToSdrToneMapper": {
"shape": "HDRToSDRToneMapper",
@@ -3489,7 +3489,7 @@
},
"ColorMetadata": {
"type": "string",
- "documentation": "Choose Insert (INSERT) for this setting to include color metadata in this output. Choose Ignore (IGNORE) to exclude color metadata from this output. If you don't specify a value, the service sets this to Insert by default.",
+ "documentation": "Choose Insert for this setting to include color metadata in this output. Choose Ignore to exclude color metadata from this output. If you don't specify a value, the service sets this to Insert by default.",
"enum": [
"IGNORE",
"INSERT"
@@ -3525,7 +3525,7 @@
},
"ColorSpaceUsage": {
"type": "string",
- "documentation": "There are two sources for color metadata, the input file and the job input settings Color space (ColorSpace) and HDR master display information settings(Hdr10Metadata). The Color space usage setting determines which takes precedence. Choose Force (FORCE) to use color metadata from the input job settings. If you don't specify values for those settings, the service defaults to using metadata from your input. FALLBACK - Choose Fallback (FALLBACK) to use color metadata from the source when it is present. If there's no color metadata in your input file, the service defaults to using values you specify in the input settings.",
+ "documentation": "There are two sources for color metadata, the input file and the job input settings Color space and HDR master display information settings. The Color space usage setting determines which takes precedence. Choose Force to use color metadata from the input job settings. If you don't specify values for those settings, the service defaults to using metadata from your input. FALLBACK - Choose Fallback to use color metadata from the source when it is present. If there's no color metadata in your input file, the service defaults to using values you specify in the input settings.",
"enum": [
"FORCE",
"FALLBACK"
@@ -3573,7 +3573,7 @@
"M2tsSettings": {
"shape": "M2tsSettings",
"locationName": "m2tsSettings",
- "documentation": "MPEG-2 TS container settings. These apply to outputs in a File output group when the output's container (ContainerType) is MPEG-2 Transport Stream (M2TS). In these assets, data is organized by the program map table (PMT). Each transport stream program contains subsets of data, including audio, video, and metadata. Each of these subsets of data has a numerical label called a packet identifier (PID). Each transport stream program corresponds to one MediaConvert output. The PMT lists the types of data in a program along with their PID. Downstream systems and players use the program map table to look up the PID for each type of data it accesses and then uses the PIDs to locate specific data within the asset."
+ "documentation": "MPEG-2 TS container settings. These apply to outputs in a File output group when the output's container is MPEG-2 Transport Stream (M2TS). In these assets, data is organized by the program map table (PMT). Each transport stream program contains subsets of data, including audio, video, and metadata. Each of these subsets of data has a numerical label called a packet identifier (PID). Each transport stream program corresponds to one MediaConvert output. The PMT lists the types of data in a program along with their PID. Downstream systems and players use the program map table to look up the PID for each type of data it accesses and then uses the PIDs to locate specific data within the asset."
},
"M3u8Settings": {
"shape": "M3u8Settings",
@@ -3896,7 +3896,7 @@
"PlaybackDeviceCompatibility": {
"shape": "DashIsoPlaybackDeviceCompatibility",
"locationName": "playbackDeviceCompatibility",
- "documentation": "This setting can improve the compatibility of your output with video players on obsolete devices. It applies only to DASH H.264 outputs with DRM encryption. Choose Unencrypted SEI (UNENCRYPTED_SEI) only to correct problems with playback on older devices. Otherwise, keep the default setting CENC v1 (CENC_V1). If you choose Unencrypted SEI, for that output, the service will exclude the access unit delimiter and will leave the SEI NAL units unencrypted."
+ "documentation": "This setting can improve the compatibility of your output with video players on obsolete devices. It applies only to DASH H.264 outputs with DRM encryption. Choose Unencrypted SEI only to correct problems with playback on older devices. Otherwise, keep the default setting CENC v1. If you choose Unencrypted SEI, for that output, the service will exclude the access unit delimiter and will leave the SEI NAL units unencrypted."
},
"SpekeKeyProvider": {
"shape": "SpekeKeyProvider",
@@ -3908,7 +3908,7 @@
},
"DashIsoGroupAudioChannelConfigSchemeIdUri": {
"type": "string",
- "documentation": "Use this setting only when your audio codec is a Dolby one (AC3, EAC3, or Atmos) and your downstream workflow requires that your DASH manifest use the Dolby channel configuration tag, rather than the MPEG one. For example, you might need to use this to make dynamic ad insertion work. Specify which audio channel configuration scheme ID URI MediaConvert writes in your DASH manifest. Keep the default value, MPEG channel configuration (MPEG_CHANNEL_CONFIGURATION), to have MediaConvert write this: urn:mpeg:mpegB:cicp:ChannelConfiguration. Choose Dolby channel configuration (DOLBY_CHANNEL_CONFIGURATION) to have MediaConvert write this instead: tag:dolby.com,2014:dash:audio_channel_configuration:2011.",
+ "documentation": "Use this setting only when your audio codec is a Dolby one (AC3, EAC3, or Atmos) and your downstream workflow requires that your DASH manifest use the Dolby channel configuration tag, rather than the MPEG one. For example, you might need to use this to make dynamic ad insertion work. Specify which audio channel configuration scheme ID URI MediaConvert writes in your DASH manifest. Keep the default value, MPEG channel configuration, to have MediaConvert write this: urn:mpeg:mpegB:cicp:ChannelConfiguration. Choose Dolby channel configuration to have MediaConvert write this instead: tag:dolby.com,2014:dash:audio_channel_configuration:2011.",
"enum": [
"MPEG_CHANNEL_CONFIGURATION",
"DOLBY_CHANNEL_CONFIGURATION"
@@ -3925,7 +3925,7 @@
"AudioChannelConfigSchemeIdUri": {
"shape": "DashIsoGroupAudioChannelConfigSchemeIdUri",
"locationName": "audioChannelConfigSchemeIdUri",
- "documentation": "Use this setting only when your audio codec is a Dolby one (AC3, EAC3, or Atmos) and your downstream workflow requires that your DASH manifest use the Dolby channel configuration tag, rather than the MPEG one. For example, you might need to use this to make dynamic ad insertion work. Specify which audio channel configuration scheme ID URI MediaConvert writes in your DASH manifest. Keep the default value, MPEG channel configuration (MPEG_CHANNEL_CONFIGURATION), to have MediaConvert write this: urn:mpeg:mpegB:cicp:ChannelConfiguration. Choose Dolby channel configuration (DOLBY_CHANNEL_CONFIGURATION) to have MediaConvert write this instead: tag:dolby.com,2014:dash:audio_channel_configuration:2011."
+ "documentation": "Use this setting only when your audio codec is a Dolby one (AC3, EAC3, or Atmos) and your downstream workflow requires that your DASH manifest use the Dolby channel configuration tag, rather than the MPEG one. For example, you might need to use this to make dynamic ad insertion work. Specify which audio channel configuration scheme ID URI MediaConvert writes in your DASH manifest. Keep the default value, MPEG channel configuration, to have MediaConvert write this: urn:mpeg:mpegB:cicp:ChannelConfiguration. Choose Dolby channel configuration to have MediaConvert write this instead: tag:dolby.com,2014:dash:audio_channel_configuration:2011."
},
"BaseUrl": {
"shape": "__string",
@@ -3940,7 +3940,7 @@
"Destination": {
"shape": "__stringPatternS3",
"locationName": "destination",
- "documentation": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file."
+ "documentation": "Use Destination to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file."
},
"DestinationSettings": {
"shape": "DestinationSettings",
@@ -3965,7 +3965,7 @@
"ImageBasedTrickPlay": {
"shape": "DashIsoImageBasedTrickPlay",
"locationName": "imageBasedTrickPlay",
- "documentation": "Specify whether MediaConvert generates images for trick play. Keep the default value, None (NONE), to not generate any images. Choose Thumbnail (THUMBNAIL) to generate tiled thumbnails. Choose Thumbnail and full frame (THUMBNAIL_AND_FULLFRAME) to generate tiled thumbnails and full-resolution images of single frames. MediaConvert adds an entry in the .mpd manifest for each set of images that you generate. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md"
+ "documentation": "Specify whether MediaConvert generates images for trick play. Keep the default value, None, to not generate any images. Choose Thumbnail to generate tiled thumbnails. Choose Thumbnail and full frame to generate tiled thumbnails and full-resolution images of single frames. MediaConvert adds an entry in the .mpd manifest for each set of images that you generate. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md"
},
"ImageBasedTrickPlaySettings": {
"shape": "DashIsoImageBasedTrickPlaySettings",
@@ -3990,12 +3990,12 @@
"MpdProfile": {
"shape": "DashIsoMpdProfile",
"locationName": "mpdProfile",
- "documentation": "Specify whether your DASH profile is on-demand or main. When you choose Main profile (MAIN_PROFILE), the service signals urn:mpeg:dash:profile:isoff-main:2011 in your .mpd DASH manifest. When you choose On-demand (ON_DEMAND_PROFILE), the service signals urn:mpeg:dash:profile:isoff-on-demand:2011 in your .mpd. When you choose On-demand, you must also set the output group setting Segment control (SegmentControl) to Single file (SINGLE_FILE)."
+ "documentation": "Specify whether your DASH profile is on-demand or main. When you choose Main profile, the service signals urn:mpeg:dash:profile:isoff-main:2011 in your .mpd DASH manifest. When you choose On-demand, the service signals urn:mpeg:dash:profile:isoff-on-demand:2011 in your .mpd. When you choose On-demand, you must also set the output group setting Segment control to Single file."
},
"PtsOffsetHandlingForBFrames": {
"shape": "DashIsoPtsOffsetHandlingForBFrames",
"locationName": "ptsOffsetHandlingForBFrames",
- "documentation": "Use this setting only when your output video stream has B-frames, which causes the initial presentation time stamp (PTS) to be offset from the initial decode time stamp (DTS). Specify how MediaConvert handles PTS when writing time stamps in output DASH manifests. Choose Match initial PTS (MATCH_INITIAL_PTS) when you want MediaConvert to use the initial PTS as the first time stamp in the manifest. Choose Zero-based (ZERO_BASED) to have MediaConvert ignore the initial PTS in the video stream and instead write the initial time stamp as zero in the manifest. For outputs that don't have B-frames, the time stamps in your DASH manifests start at zero regardless of your choice here."
+ "documentation": "Use this setting only when your output video stream has B-frames, which causes the initial presentation time stamp (PTS) to be offset from the initial decode time stamp (DTS). Specify how MediaConvert handles PTS when writing time stamps in output DASH manifests. Choose Match initial PTS when you want MediaConvert to use the initial PTS as the first time stamp in the manifest. Choose Zero-based to have MediaConvert ignore the initial PTS in the video stream and instead write the initial time stamp as zero in the manifest. For outputs that don't have B-frames, the time stamps in your DASH manifests start at zero regardless of your choice here."
},
"SegmentControl": {
"shape": "DashIsoSegmentControl",
@@ -4005,12 +4005,12 @@
"SegmentLength": {
"shape": "__integerMin1Max2147483647",
"locationName": "segmentLength",
- "documentation": "Specify the length, in whole seconds, of each segment. When you don't specify a value, MediaConvert defaults to 30. Related settings: Use Segment length control (SegmentLengthControl) to specify whether the encoder enforces this value strictly. Use Segment control (DashIsoSegmentControl) to specify whether MediaConvert creates separate segment files or one content file that has metadata to mark the segment boundaries."
+ "documentation": "Specify the length, in whole seconds, of each segment. When you don't specify a value, MediaConvert defaults to 30. Related settings: Use Segment length control to specify whether the encoder enforces this value strictly. Use Segment control to specify whether MediaConvert creates separate segment files or one content file that has metadata to mark the segment boundaries."
},
"SegmentLengthControl": {
"shape": "DashIsoSegmentLengthControl",
"locationName": "segmentLengthControl",
- "documentation": "Specify how you want MediaConvert to determine the segment length. Choose Exact (EXACT) to have the encoder use the exact length that you specify with the setting Segment length (SegmentLength). This might result in extra I-frames. Choose Multiple of GOP (GOP_MULTIPLE) to have the encoder round up the segment lengths to match the next GOP boundary."
+ "documentation": "Specify how you want MediaConvert to determine the segment length. Choose Exact to have the encoder use the exact length that you specify with the setting Segment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary."
},
"VideoCompositionOffsets": {
"shape": "DashIsoVideoCompositionOffsets",
@@ -4023,7 +4023,7 @@
"documentation": "If you get an HTTP error in the 400 range when you play back your DASH output, enable this setting and run your transcoding job again. When you enable this setting, the service writes precise segment durations in the DASH manifest. The segment duration information appears inside the SegmentTimeline element, inside SegmentTemplate at the Representation level. When you don't enable this setting, the service writes approximate segment durations in your DASH manifest."
}
},
- "documentation": "Settings related to your DASH output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html. When you work directly in your JSON job specification, include this object and any required children when you set Type, under OutputGroupSettings, to DASH_ISO_GROUP_SETTINGS."
+ "documentation": "Settings related to your DASH output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html."
},
"DashIsoHbbtvCompliance": {
"type": "string",
@@ -4035,7 +4035,7 @@
},
"DashIsoImageBasedTrickPlay": {
"type": "string",
- "documentation": "Specify whether MediaConvert generates images for trick play. Keep the default value, None (NONE), to not generate any images. Choose Thumbnail (THUMBNAIL) to generate tiled thumbnails. Choose Thumbnail and full frame (THUMBNAIL_AND_FULLFRAME) to generate tiled thumbnails and full-resolution images of single frames. MediaConvert adds an entry in the .mpd manifest for each set of images that you generate. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md",
+ "documentation": "Specify whether MediaConvert generates images for trick play. Keep the default value, None, to not generate any images. Choose Thumbnail to generate tiled thumbnails. Choose Thumbnail and full frame to generate tiled thumbnails and full-resolution images of single frames. MediaConvert adds an entry in the .mpd manifest for each set of images that you generate. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md",
"enum": [
"NONE",
"THUMBNAIL",
@@ -4097,7 +4097,7 @@
},
"DashIsoMpdProfile": {
"type": "string",
- "documentation": "Specify whether your DASH profile is on-demand or main. When you choose Main profile (MAIN_PROFILE), the service signals urn:mpeg:dash:profile:isoff-main:2011 in your .mpd DASH manifest. When you choose On-demand (ON_DEMAND_PROFILE), the service signals urn:mpeg:dash:profile:isoff-on-demand:2011 in your .mpd. When you choose On-demand, you must also set the output group setting Segment control (SegmentControl) to Single file (SINGLE_FILE).",
+ "documentation": "Specify whether your DASH profile is on-demand or main. When you choose Main profile, the service signals urn:mpeg:dash:profile:isoff-main:2011 in your .mpd DASH manifest. When you choose On-demand, the service signals urn:mpeg:dash:profile:isoff-on-demand:2011 in your .mpd. When you choose On-demand, you must also set the output group setting Segment control to Single file.",
"enum": [
"MAIN_PROFILE",
"ON_DEMAND_PROFILE"
@@ -4105,7 +4105,7 @@
},
"DashIsoPlaybackDeviceCompatibility": {
"type": "string",
- "documentation": "This setting can improve the compatibility of your output with video players on obsolete devices. It applies only to DASH H.264 outputs with DRM encryption. Choose Unencrypted SEI (UNENCRYPTED_SEI) only to correct problems with playback on older devices. Otherwise, keep the default setting CENC v1 (CENC_V1). If you choose Unencrypted SEI, for that output, the service will exclude the access unit delimiter and will leave the SEI NAL units unencrypted.",
+ "documentation": "This setting can improve the compatibility of your output with video players on obsolete devices. It applies only to DASH H.264 outputs with DRM encryption. Choose Unencrypted SEI only to correct problems with playback on older devices. Otherwise, keep the default setting CENC v1. If you choose Unencrypted SEI, for that output, the service will exclude the access unit delimiter and will leave the SEI NAL units unencrypted.",
"enum": [
"CENC_V1",
"UNENCRYPTED_SEI"
@@ -4113,7 +4113,7 @@
},
"DashIsoPtsOffsetHandlingForBFrames": {
"type": "string",
- "documentation": "Use this setting only when your output video stream has B-frames, which causes the initial presentation time stamp (PTS) to be offset from the initial decode time stamp (DTS). Specify how MediaConvert handles PTS when writing time stamps in output DASH manifests. Choose Match initial PTS (MATCH_INITIAL_PTS) when you want MediaConvert to use the initial PTS as the first time stamp in the manifest. Choose Zero-based (ZERO_BASED) to have MediaConvert ignore the initial PTS in the video stream and instead write the initial time stamp as zero in the manifest. For outputs that don't have B-frames, the time stamps in your DASH manifests start at zero regardless of your choice here.",
+ "documentation": "Use this setting only when your output video stream has B-frames, which causes the initial presentation time stamp (PTS) to be offset from the initial decode time stamp (DTS). Specify how MediaConvert handles PTS when writing time stamps in output DASH manifests. Choose Match initial PTS when you want MediaConvert to use the initial PTS as the first time stamp in the manifest. Choose Zero-based to have MediaConvert ignore the initial PTS in the video stream and instead write the initial time stamp as zero in the manifest. For outputs that don't have B-frames, the time stamps in your DASH manifests start at zero regardless of your choice here.",
"enum": [
"ZERO_BASED",
"MATCH_INITIAL_PTS"
@@ -4129,7 +4129,7 @@
},
"DashIsoSegmentLengthControl": {
"type": "string",
- "documentation": "Specify how you want MediaConvert to determine the segment length. Choose Exact (EXACT) to have the encoder use the exact length that you specify with the setting Segment length (SegmentLength). This might result in extra I-frames. Choose Multiple of GOP (GOP_MULTIPLE) to have the encoder round up the segment lengths to match the next GOP boundary.",
+ "documentation": "Specify how you want MediaConvert to determine the segment length. Choose Exact to have the encoder use the exact length that you specify with the setting Segment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary.",
"enum": [
"EXACT",
"GOP_MULTIPLE"
@@ -4145,7 +4145,7 @@
},
"DashIsoWriteSegmentTimelineInRepresentation": {
"type": "string",
- "documentation": "When you enable Precise segment duration in manifests (writeSegmentTimelineInRepresentation), your DASH manifest shows precise segment durations. The segment duration information appears inside the SegmentTimeline element, inside SegmentTemplate at the Representation level. When this feature isn't enabled, the segment durations in your DASH manifest are approximate. The segment duration information appears in the duration attribute of the SegmentTemplate element.",
+ "documentation": "When you enable Precise segment duration in manifests, your DASH manifest shows precise segment durations. The segment duration information appears inside the SegmentTimeline element, inside SegmentTemplate at the Representation level. When this feature isn't enabled, the segment durations in your DASH manifest are approximate. The segment duration information appears in the duration attribute of the SegmentTemplate element.",
"enum": [
"ENABLED",
"DISABLED"
@@ -4196,7 +4196,7 @@
"Mode": {
"shape": "DeinterlacerMode",
"locationName": "mode",
- "documentation": "Use Deinterlacer (DeinterlaceMode) to choose how the service will do deinterlacing. Default is Deinterlace. - Deinterlace converts interlaced to progressive. - Inverse telecine converts Hard Telecine 29.97i to progressive 23.976p. - Adaptive auto-detects and converts to progressive."
+ "documentation": "Use Deinterlacer to choose how the service will do deinterlacing. Default is Deinterlace.\n- Deinterlace converts interlaced to progressive.\n- Inverse telecine converts Hard Telecine 29.97i to progressive 23.976p.\n- Adaptive auto-detects and converts to progressive."
}
},
"documentation": "Settings for deinterlacer"
@@ -4211,7 +4211,7 @@
},
"DeinterlacerMode": {
"type": "string",
- "documentation": "Use Deinterlacer (DeinterlaceMode) to choose how the service will do deinterlacing. Default is Deinterlace. - Deinterlace converts interlaced to progressive. - Inverse telecine converts Hard Telecine 29.97i to progressive 23.976p. - Adaptive auto-detects and converts to progressive.",
+ "documentation": "Use Deinterlacer to choose how the service will do deinterlacing. Default is Deinterlace.\n- Deinterlace converts interlaced to progressive.\n- Inverse telecine converts Hard Telecine 29.97i to progressive 23.976p.\n- Adaptive auto-detects and converts to progressive.",
"enum": [
"DEINTERLACE",
"INVERSE_TELECINE",
@@ -4428,7 +4428,7 @@
},
"DropFrameTimecode": {
"type": "string",
- "documentation": "Applies only to 29.97 fps outputs. When this feature is enabled, the service will use drop-frame timecode on outputs. If it is not possible to use drop-frame timecode, the system will fall back to non-drop-frame. This setting is enabled by default when Timecode insertion (TimecodeInsertion) is enabled.",
+ "documentation": "Applies only to 29.97 fps outputs. When this feature is enabled, the service will use drop-frame timecode on outputs. If it is not possible to use drop-frame timecode, the system will fall back to non-drop-frame. This setting is enabled by default when Timecode insertion is enabled.",
"enum": [
"DISABLED",
"ENABLED"
@@ -4453,7 +4453,7 @@
"documentation": "The number of milliseconds between instances of this table in the output transport stream."
}
},
- "documentation": "Use these settings to insert a DVB Network Information Table (NIT) in the transport stream of this output. When you work directly in your JSON job specification, include this object only when your job has a transport stream output and the container settings contain the object M2tsSettings."
+ "documentation": "Use these settings to insert a DVB Network Information Table (NIT) in the transport stream of this output."
},
"DvbSdtSettings": {
"type": "structure",
@@ -4479,7 +4479,7 @@
"documentation": "The service provider name placed in the service_descriptor in the Service Description Table. Maximum length is 256 characters."
}
},
- "documentation": "Use these settings to insert a DVB Service Description Table (SDT) in the transport stream of this output. When you work directly in your JSON job specification, include this object only when your job has a transport stream output and the container settings contain the object M2tsSettings."
+ "documentation": "Use these settings to insert a DVB Service Description Table (SDT) in the transport stream of this output."
},
"DvbSubDestinationSettings": {
"type": "structure",
@@ -4492,17 +4492,17 @@
"ApplyFontColor": {
"shape": "DvbSubtitleApplyFontColor",
"locationName": "applyFontColor",
- "documentation": "Ignore this setting unless Style Passthrough (StylePassthrough) is set to Enabled and Font color (FontColor) set to Black, Yellow, Red, Green, Blue, or Hex. Use Apply font color (ApplyFontColor) for additional font color controls. When you choose White text only (WHITE_TEXT_ONLY), or leave blank, your font color setting only applies to white text in your input captions. For example, if your font color setting is Yellow, and your input captions have red and white text, your output captions will have red and yellow text. When you choose ALL_TEXT, your font color setting applies to all of your output captions text."
+ "documentation": "Ignore this setting unless Style Passthrough is set to Enabled and Font color set to Black, Yellow, Red, Green, Blue, or Hex. Use Apply font color for additional font color controls. When you choose White text only, or leave blank, your font color setting only applies to white text in your input captions. For example, if your font color setting is Yellow, and your input captions have red and white text, your output captions will have red and yellow text. When you choose ALL_TEXT, your font color setting applies to all of your output captions text."
},
"BackgroundColor": {
"shape": "DvbSubtitleBackgroundColor",
"locationName": "backgroundColor",
- "documentation": "Specify the color of the rectangle behind the captions. Leave background color (BackgroundColor) blank and set Style passthrough (StylePassthrough) to enabled to use the background color data from your input captions, if present."
+ "documentation": "Specify the color of the rectangle behind the captions. Leave background color blank and set Style passthrough to enabled to use the background color data from your input captions, if present."
},
"BackgroundOpacity": {
"shape": "__integerMin0Max255",
"locationName": "backgroundOpacity",
- "documentation": "Specify the opacity of the background rectangle. Enter a value from 0 to 255, where 0 is transparent and 255 is opaque. If Style passthrough (StylePassthrough) is set to enabled, leave blank to pass through the background style information in your input captions to your output captions. If Style passthrough is set to disabled, leave blank to use a value of 0 and remove all backgrounds from your output captions. Within your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Specify the opacity of the background rectangle. Enter a value from 0 to 255, where 0 is transparent and 255 is opaque. If Style passthrough is set to enabled, leave blank to pass through the background style information in your input captions to your output captions. If Style passthrough is set to disabled, leave blank to use a value of 0 and remove all backgrounds from your output captions. Within your job settings, all of your DVB-Sub settings must be identical."
},
"DdsHandling": {
"shape": "DvbddsHandling",
@@ -4512,22 +4512,22 @@
"DdsXCoordinate": {
"shape": "__integerMin0Max2147483647",
"locationName": "ddsXCoordinate",
- "documentation": "Use this setting, along with DDS y-coordinate (ddsYCoordinate), to specify the upper left corner of the display definition segment (DDS) display window. With this setting, specify the distance, in pixels, between the left side of the frame and the left side of the DDS display window. Keep the default value, 0, to have MediaConvert automatically choose this offset. Related setting: When you use this setting, you must set DDS handling (ddsHandling) to a value other than None (NONE). MediaConvert uses these values to determine whether to write page position data to the DDS or to the page composition segment (PCS). All burn-in and DVB-Sub font settings must match."
+ "documentation": "Use this setting, along with DDS y-coordinate, to specify the upper left corner of the display definition segment (DDS) display window. With this setting, specify the distance, in pixels, between the left side of the frame and the left side of the DDS display window. Keep the default value, 0, to have MediaConvert automatically choose this offset. Related setting: When you use this setting, you must set DDS handling to a value other than None. MediaConvert uses these values to determine whether to write page position data to the DDS or to the page composition segment. All burn-in and DVB-Sub font settings must match."
},
"DdsYCoordinate": {
"shape": "__integerMin0Max2147483647",
"locationName": "ddsYCoordinate",
- "documentation": "Use this setting, along with DDS x-coordinate (ddsXCoordinate), to specify the upper left corner of the display definition segment (DDS) display window. With this setting, specify the distance, in pixels, between the top of the frame and the top of the DDS display window. Keep the default value, 0, to have MediaConvert automatically choose this offset. Related setting: When you use this setting, you must set DDS handling (ddsHandling) to a value other than None (NONE). MediaConvert uses these values to determine whether to write page position data to the DDS or to the page composition segment (PCS). All burn-in and DVB-Sub font settings must match."
+ "documentation": "Use this setting, along with DDS x-coordinate, to specify the upper left corner of the display definition segment (DDS) display window. With this setting, specify the distance, in pixels, between the top of the frame and the top of the DDS display window. Keep the default value, 0, to have MediaConvert automatically choose this offset. Related setting: When you use this setting, you must set DDS handling to a value other than None. MediaConvert uses these values to determine whether to write page position data to the DDS or to the page composition segment (PCS). All burn-in and DVB-Sub font settings must match."
},
"FallbackFont": {
"shape": "DvbSubSubtitleFallbackFont",
"locationName": "fallbackFont",
- "documentation": "Specify the font that you want the service to use for your burn in captions when your input captions specify a font that MediaConvert doesn't support. When you set Fallback font (FallbackFont) to best match (BEST_MATCH), or leave blank, MediaConvert uses a supported font that most closely matches the font that your input captions specify. When there are multiple unsupported fonts in your input captions, MediaConvert matches each font with the supported font that matches best. When you explicitly choose a replacement font, MediaConvert uses that font to replace all unsupported fonts from your input."
+ "documentation": "Specify the font that you want the service to use for your burn in captions when your input captions specify a font that MediaConvert doesn't support. When you set Fallback font to best match, or leave blank, MediaConvert uses a supported font that most closely matches the font that your input captions specify. When there are multiple unsupported fonts in your input captions, MediaConvert matches each font with the supported font that matches best. When you explicitly choose a replacement font, MediaConvert uses that font to replace all unsupported fonts from your input."
},
"FontColor": {
"shape": "DvbSubtitleFontColor",
"locationName": "fontColor",
- "documentation": "Specify the color of the captions text. Leave Font color (FontColor) blank and set Style passthrough (StylePassthrough) to enabled to use the font color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Specify the color of the captions text. Leave Font color blank and set Style passthrough to enabled to use the font color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical."
},
"FontOpacity": {
"shape": "__integerMin0Max255",
@@ -4537,22 +4537,22 @@
"FontResolution": {
"shape": "__integerMin96Max600",
"locationName": "fontResolution",
- "documentation": "Specify the Font resolution (FontResolution) in DPI (dots per inch).\nWithin your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Specify the Font resolution in DPI (dots per inch).\nWithin your job settings, all of your DVB-Sub settings must be identical."
},
"FontScript": {
"shape": "FontScript",
"locationName": "fontScript",
- "documentation": "Set Font script (FontScript) to Automatically determined (AUTOMATIC), or leave blank, to automatically determine the font script in your input captions. Otherwise, set to Simplified Chinese (HANS) or Traditional Chinese (HANT) if your input font script uses Simplified or Traditional Chinese. Within your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Set Font script to Automatically determined, or leave blank, to automatically determine the font script in your input captions. Otherwise, set to Simplified Chinese (HANS) or Traditional Chinese (HANT) if your input font script uses Simplified or Traditional Chinese. Within your job settings, all of your DVB-Sub settings must be identical."
},
"FontSize": {
"shape": "__integerMin0Max96",
"locationName": "fontSize",
- "documentation": "Specify the Font size (FontSize) in pixels. Must be a positive integer. Set to 0, or leave blank, for automatic font size. Within your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Specify the Font size in pixels. Must be a positive integer. Set to 0, or leave blank, for automatic font size. Within your job settings, all of your DVB-Sub settings must be identical."
},
"Height": {
"shape": "__integerMin1Max2147483647",
"locationName": "height",
- "documentation": "Specify the height, in pixels, of this set of DVB-Sub captions. The default value is 576 pixels. Related setting: When you use this setting, you must set DDS handling (ddsHandling) to a value other than None (NONE). All burn-in and DVB-Sub font settings must match."
+ "documentation": "Specify the height, in pixels, of this set of DVB-Sub captions. The default value is 576 pixels. Related setting: When you use this setting, you must set DDS handling to a value other than None. All burn-in and DVB-Sub font settings must match."
},
"HexFontColor": {
"shape": "__stringMin6Max8Pattern09aFAF609aFAF2",
@@ -4562,22 +4562,22 @@
"OutlineColor": {
"shape": "DvbSubtitleOutlineColor",
"locationName": "outlineColor",
- "documentation": "Specify font outline color. Leave Outline color (OutlineColor) blank and set Style passthrough (StylePassthrough) to enabled to use the font outline color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Specify font outline color. Leave Outline color blank and set Style passthrough to enabled to use the font outline color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical."
},
"OutlineSize": {
"shape": "__integerMin0Max10",
"locationName": "outlineSize",
- "documentation": "Specify the Outline size (OutlineSize) of the caption text, in pixels. Leave Outline size blank and set Style passthrough (StylePassthrough) to enabled to use the outline size data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Specify the Outline size of the caption text, in pixels. Leave Outline size blank and set Style passthrough to enabled to use the outline size data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical."
},
"ShadowColor": {
"shape": "DvbSubtitleShadowColor",
"locationName": "shadowColor",
- "documentation": "Specify the color of the shadow cast by the captions. Leave Shadow color (ShadowColor) blank and set Style passthrough (StylePassthrough) to enabled to use the shadow color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Specify the color of the shadow cast by the captions. Leave Shadow color blank and set Style passthrough to enabled to use the shadow color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical."
},
"ShadowOpacity": {
"shape": "__integerMin0Max255",
"locationName": "shadowOpacity",
- "documentation": "Specify the opacity of the shadow. Enter a value from 0 to 255, where 0 is transparent and 255 is opaque. If Style passthrough (StylePassthrough) is set to Enabled, leave Shadow opacity (ShadowOpacity) blank to pass through the shadow style information in your input captions to your output captions. If Style passthrough is set to disabled, leave blank to use a value of 0 and remove all shadows from your output captions. Within your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Specify the opacity of the shadow. Enter a value from 0 to 255, where 0 is transparent and 255 is opaque. If Style passthrough is set to Enabled, leave Shadow opacity blank to pass through the shadow style information in your input captions to your output captions. If Style passthrough is set to disabled, leave blank to use a value of 0 and remove all shadows from your output captions. Within your job settings, all of your DVB-Sub settings must be identical."
},
"ShadowXOffset": {
"shape": "__integerMinNegative2147483648Max2147483647",
@@ -4587,12 +4587,12 @@
"ShadowYOffset": {
"shape": "__integerMinNegative2147483648Max2147483647",
"locationName": "shadowYOffset",
- "documentation": "Specify the vertical offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels above the text. Leave Shadow y-offset (ShadowYOffset) blank and set Style passthrough (StylePassthrough) to enabled to use the shadow y-offset data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Specify the vertical offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels above the text. Leave Shadow y-offset blank and set Style passthrough to enabled to use the shadow y-offset data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical."
},
"StylePassthrough": {
"shape": "DvbSubtitleStylePassthrough",
"locationName": "stylePassthrough",
- "documentation": "Set Style passthrough (StylePassthrough) to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use default settings: white text with black outlining, bottom-center positioning, and automatic sizing. Whether you set Style passthrough to enabled or not, you can also choose to manually override any of the individual style and position settings."
+ "documentation": "Set Style passthrough to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use default settings: white text with black outlining, bottom-center positioning, and automatic sizing. Whether you set Style passthrough to enabled or not, you can also choose to manually override any of the individual style and position settings."
},
"SubtitlingType": {
"shape": "DvbSubtitlingType",
@@ -4602,25 +4602,25 @@
"TeletextSpacing": {
"shape": "DvbSubtitleTeletextSpacing",
"locationName": "teletextSpacing",
- "documentation": "Specify whether the Text spacing (TeletextSpacing) in your captions is set by the captions grid, or varies depending on letter width. Choose fixed grid (FIXED_GRID) to conform to the spacing specified in the captions file more accurately. Choose proportional (PROPORTIONAL) to make the text easier to read for closed captions. Within your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Specify whether the Text spacing in your captions is set by the captions grid, or varies depending on letter width. Choose fixed grid to conform to the spacing specified in the captions file more accurately. Choose proportional to make the text easier to read for closed captions. Within your job settings, all of your DVB-Sub settings must be identical."
},
"Width": {
"shape": "__integerMin1Max2147483647",
"locationName": "width",
- "documentation": "Specify the width, in pixels, of this set of DVB-Sub captions. The default value is 720 pixels. Related setting: When you use this setting, you must set DDS handling (ddsHandling) to a value other than None (NONE). All burn-in and DVB-Sub font settings must match."
+ "documentation": "Specify the width, in pixels, of this set of DVB-Sub captions. The default value is 720 pixels. Related setting: When you use this setting, you must set DDS handling to a value other than None. All burn-in and DVB-Sub font settings must match."
},
"XPosition": {
"shape": "__integerMin0Max2147483647",
"locationName": "xPosition",
- "documentation": "Specify the horizontal position (XPosition) of the captions, relative to the left side of the outputin pixels. A value of 10 would result in the captions starting 10 pixels from the left ofthe output. If no explicit x_position is provided, the horizontal caption position will bedetermined by the alignment parameter. Within your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Specify the horizontal position of the captions, relative to the left side of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the left of the output. If no explicit x_position is provided, the horizontal caption position will be determined by the alignment parameter. Within your job settings, all of your DVB-Sub settings must be identical."
},
"YPosition": {
"shape": "__integerMin0Max2147483647",
"locationName": "yPosition",
- "documentation": "Specify the vertical position (YPosition) of the captions, relative to the top of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the top of the output. If no explicit y_position is provided, the caption will be positioned towards the bottom of the output. Within your job settings, all of your DVB-Sub settings must be identical."
+ "documentation": "Specify the vertical position of the captions, relative to the top of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the top of the output. If no explicit y_position is provided, the caption will be positioned towards the bottom of the output. Within your job settings, all of your DVB-Sub settings must be identical."
}
},
- "documentation": "Settings related to DVB-Sub captions. Set up DVB-Sub captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/dvb-sub-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to DVB_SUB."
+ "documentation": "Settings related to DVB-Sub captions. Set up DVB-Sub captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/dvb-sub-output-captions.html."
},
"DvbSubSourceSettings": {
"type": "structure",
@@ -4635,7 +4635,7 @@
},
"DvbSubSubtitleFallbackFont": {
"type": "string",
- "documentation": "Specify the font that you want the service to use for your burn in captions when your input captions specify a font that MediaConvert doesn't support. When you set Fallback font (FallbackFont) to best match (BEST_MATCH), or leave blank, MediaConvert uses a supported font that most closely matches the font that your input captions specify. When there are multiple unsupported fonts in your input captions, MediaConvert matches each font with the supported font that matches best. When you explicitly choose a replacement font, MediaConvert uses that font to replace all unsupported fonts from your input.",
+ "documentation": "Specify the font that you want the service to use for your burn in captions when your input captions specify a font that MediaConvert doesn't support. When you set Fallback font to best match, or leave blank, MediaConvert uses a supported font that most closely matches the font that your input captions specify. When there are multiple unsupported fonts in your input captions, MediaConvert matches each font with the supported font that matches best. When you explicitly choose a replacement font, MediaConvert uses that font to replace all unsupported fonts from your input.",
"enum": [
"BEST_MATCH",
"MONOSPACED_SANSSERIF",
@@ -4655,7 +4655,7 @@
},
"DvbSubtitleApplyFontColor": {
"type": "string",
- "documentation": "Ignore this setting unless Style Passthrough (StylePassthrough) is set to Enabled and Font color (FontColor) set to Black, Yellow, Red, Green, Blue, or Hex. Use Apply font color (ApplyFontColor) for additional font color controls. When you choose White text only (WHITE_TEXT_ONLY), or leave blank, your font color setting only applies to white text in your input captions. For example, if your font color setting is Yellow, and your input captions have red and white text, your output captions will have red and yellow text. When you choose ALL_TEXT, your font color setting applies to all of your output captions text.",
+ "documentation": "Ignore this setting unless Style Passthrough is set to Enabled and Font color set to Black, Yellow, Red, Green, Blue, or Hex. Use Apply font color for additional font color controls. When you choose White text only, or leave blank, your font color setting only applies to white text in your input captions. For example, if your font color setting is Yellow, and your input captions have red and white text, your output captions will have red and yellow text. When you choose ALL_TEXT, your font color setting applies to all of your output captions text.",
"enum": [
"WHITE_TEXT_ONLY",
"ALL_TEXT"
@@ -4663,7 +4663,7 @@
},
"DvbSubtitleBackgroundColor": {
"type": "string",
- "documentation": "Specify the color of the rectangle behind the captions. Leave background color (BackgroundColor) blank and set Style passthrough (StylePassthrough) to enabled to use the background color data from your input captions, if present.",
+ "documentation": "Specify the color of the rectangle behind the captions. Leave background color blank and set Style passthrough to enabled to use the background color data from your input captions, if present.",
"enum": [
"NONE",
"BLACK",
@@ -4673,7 +4673,7 @@
},
"DvbSubtitleFontColor": {
"type": "string",
- "documentation": "Specify the color of the captions text. Leave Font color (FontColor) blank and set Style passthrough (StylePassthrough) to enabled to use the font color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical.",
+ "documentation": "Specify the color of the captions text. Leave Font color blank and set Style passthrough to enabled to use the font color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical.",
"enum": [
"WHITE",
"BLACK",
@@ -4687,7 +4687,7 @@
},
"DvbSubtitleOutlineColor": {
"type": "string",
- "documentation": "Specify font outline color. Leave Outline color (OutlineColor) blank and set Style passthrough (StylePassthrough) to enabled to use the font outline color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical.",
+ "documentation": "Specify font outline color. Leave Outline color blank and set Style passthrough to enabled to use the font outline color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical.",
"enum": [
"BLACK",
"WHITE",
@@ -4700,7 +4700,7 @@
},
"DvbSubtitleShadowColor": {
"type": "string",
- "documentation": "Specify the color of the shadow cast by the captions. Leave Shadow color (ShadowColor) blank and set Style passthrough (StylePassthrough) to enabled to use the shadow color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical.",
+ "documentation": "Specify the color of the shadow cast by the captions. Leave Shadow color blank and set Style passthrough to enabled to use the shadow color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical.",
"enum": [
"NONE",
"BLACK",
@@ -4710,7 +4710,7 @@
},
"DvbSubtitleStylePassthrough": {
"type": "string",
- "documentation": "Set Style passthrough (StylePassthrough) to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use default settings: white text with black outlining, bottom-center positioning, and automatic sizing. Whether you set Style passthrough to enabled or not, you can also choose to manually override any of the individual style and position settings.",
+ "documentation": "Set Style passthrough to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use default settings: white text with black outlining, bottom-center positioning, and automatic sizing. Whether you set Style passthrough to enabled or not, you can also choose to manually override any of the individual style and position settings.",
"enum": [
"ENABLED",
"DISABLED"
@@ -4718,7 +4718,7 @@
},
"DvbSubtitleTeletextSpacing": {
"type": "string",
- "documentation": "Specify whether the Text spacing (TeletextSpacing) in your captions is set by the captions grid, or varies depending on letter width. Choose fixed grid (FIXED_GRID) to conform to the spacing specified in the captions file more accurately. Choose proportional (PROPORTIONAL) to make the text easier to read for closed captions. Within your job settings, all of your DVB-Sub settings must be identical.",
+ "documentation": "Specify whether the Text spacing in your captions is set by the captions grid, or varies depending on letter width. Choose fixed grid to conform to the spacing specified in the captions file more accurately. Choose proportional to make the text easier to read for closed captions. Within your job settings, all of your DVB-Sub settings must be identical.",
"enum": [
"FIXED_GRID",
"PROPORTIONAL",
@@ -4742,7 +4742,7 @@
"documentation": "The number of milliseconds between instances of this table in the output transport stream."
}
},
- "documentation": "Use these settings to insert a DVB Time and Date Table (TDT) in the transport stream of this output. When you work directly in your JSON job specification, include this object only when your job has a transport stream output and the container settings contain the object M2tsSettings."
+ "documentation": "Use these settings to insert a DVB Time and Date Table (TDT) in the transport stream of this output."
},
"DvbddsHandling": {
"type": "string",
@@ -4780,7 +4780,7 @@
},
"Eac3AtmosDownmixControl": {
"type": "string",
- "documentation": "Specify whether MediaConvert should use any downmix metadata from your input file. Keep the default value, Custom (SPECIFIED) to provide downmix values in your job settings. Choose Follow source (INITIALIZE_FROM_SOURCE) to use the metadata from your input. Related settings--Use these settings to specify your downmix values: Left only/Right only surround (LoRoSurroundMixLevel), Left total/Right total surround (LtRtSurroundMixLevel), Left total/Right total center (LtRtCenterMixLevel), Left only/Right only center (LoRoCenterMixLevel), and Stereo downmix (StereoDownmix). When you keep Custom (SPECIFIED) for Downmix control (DownmixControl) and you don't specify values for the related settings, MediaConvert uses default values for those settings.",
+ "documentation": "Specify whether MediaConvert should use any downmix metadata from your input file. Keep the default value, Custom to provide downmix values in your job settings. Choose Follow source to use the metadata from your input. Related settings--Use these settings to specify your downmix values: Left only/Right only surround, Left total/Right total surround, Left total/Right total center, Left only/Right only center, and Stereo downmix. When you keep Custom for Downmix control and you don't specify values for the related settings, MediaConvert uses default values for those settings.",
"enum": [
"SPECIFIED",
"INITIALIZE_FROM_SOURCE"
@@ -4788,7 +4788,7 @@
},
"Eac3AtmosDynamicRangeCompressionLine": {
"type": "string",
- "documentation": "Choose the Dolby dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby stream for the line operating mode. Default value: Film light (ATMOS_STORAGE_DDP_COMPR_FILM_LIGHT) Related setting: To have MediaConvert use the value you specify here, keep the default value, Custom (SPECIFIED) for the setting Dynamic range control (DynamicRangeControl). Otherwise, MediaConvert ignores Dynamic range compression line (DynamicRangeCompressionLine). For information about the Dolby DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.",
+ "documentation": "Choose the Dolby dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby stream for the line operating mode. Default value: Film light Related setting: To have MediaConvert use the value you specify here, keep the default value, Custom for the setting Dynamic range control. Otherwise, MediaConvert ignores Dynamic range compression line. For information about the Dolby DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.",
"enum": [
"NONE",
"FILM_STANDARD",
@@ -4800,7 +4800,7 @@
},
"Eac3AtmosDynamicRangeCompressionRf": {
"type": "string",
- "documentation": "Choose the Dolby dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby stream for the RF operating mode. Default value: Film light (ATMOS_STORAGE_DDP_COMPR_FILM_LIGHT) Related setting: To have MediaConvert use the value you specify here, keep the default value, Custom (SPECIFIED) for the setting Dynamic range control (DynamicRangeControl). Otherwise, MediaConvert ignores Dynamic range compression RF (DynamicRangeCompressionRf). For information about the Dolby DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.",
+ "documentation": "Choose the Dolby dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby stream for the RF operating mode. Default value: Film light Related setting: To have MediaConvert use the value you specify here, keep the default value, Custom for the setting Dynamic range control. Otherwise, MediaConvert ignores Dynamic range compression RF. For information about the Dolby DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.",
"enum": [
"NONE",
"FILM_STANDARD",
@@ -4812,7 +4812,7 @@
},
"Eac3AtmosDynamicRangeControl": {
"type": "string",
- "documentation": "Specify whether MediaConvert should use any dynamic range control metadata from your input file. Keep the default value, Custom (SPECIFIED), to provide dynamic range control values in your job settings. Choose Follow source (INITIALIZE_FROM_SOURCE) to use the metadata from your input. Related settings--Use these settings to specify your dynamic range control values: Dynamic range compression line (DynamicRangeCompressionLine) and Dynamic range compression RF (DynamicRangeCompressionRf). When you keep the value Custom (SPECIFIED) for Dynamic range control (DynamicRangeControl) and you don't specify values for the related settings, MediaConvert uses default values for those settings.",
+ "documentation": "Specify whether MediaConvert should use any dynamic range control metadata from your input file. Keep the default value, Custom, to provide dynamic range control values in your job settings. Choose Follow source to use the metadata from your input. Related settings--Use these settings to specify your dynamic range control values: Dynamic range compression line and Dynamic range compression RF. When you keep the value Custom for Dynamic range control and you don't specify values for the related settings, MediaConvert uses default values for those settings.",
"enum": [
"SPECIFIED",
"INITIALIZE_FROM_SOURCE"
@@ -4855,42 +4855,42 @@
"DownmixControl": {
"shape": "Eac3AtmosDownmixControl",
"locationName": "downmixControl",
- "documentation": "Specify whether MediaConvert should use any downmix metadata from your input file. Keep the default value, Custom (SPECIFIED) to provide downmix values in your job settings. Choose Follow source (INITIALIZE_FROM_SOURCE) to use the metadata from your input. Related settings--Use these settings to specify your downmix values: Left only/Right only surround (LoRoSurroundMixLevel), Left total/Right total surround (LtRtSurroundMixLevel), Left total/Right total center (LtRtCenterMixLevel), Left only/Right only center (LoRoCenterMixLevel), and Stereo downmix (StereoDownmix). When you keep Custom (SPECIFIED) for Downmix control (DownmixControl) and you don't specify values for the related settings, MediaConvert uses default values for those settings."
+ "documentation": "Specify whether MediaConvert should use any downmix metadata from your input file. Keep the default value, Custom to provide downmix values in your job settings. Choose Follow source to use the metadata from your input. Related settings--Use these settings to specify your downmix values: Left only/Right only surround, Left total/Right total surround, Left total/Right total center, Left only/Right only center, and Stereo downmix. When you keep Custom for Downmix control and you don't specify values for the related settings, MediaConvert uses default values for those settings."
},
"DynamicRangeCompressionLine": {
"shape": "Eac3AtmosDynamicRangeCompressionLine",
"locationName": "dynamicRangeCompressionLine",
- "documentation": "Choose the Dolby dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby stream for the line operating mode. Default value: Film light (ATMOS_STORAGE_DDP_COMPR_FILM_LIGHT) Related setting: To have MediaConvert use the value you specify here, keep the default value, Custom (SPECIFIED) for the setting Dynamic range control (DynamicRangeControl). Otherwise, MediaConvert ignores Dynamic range compression line (DynamicRangeCompressionLine). For information about the Dolby DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf."
+ "documentation": "Choose the Dolby dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby stream for the line operating mode. Default value: Film light Related setting: To have MediaConvert use the value you specify here, keep the default value, Custom for the setting Dynamic range control. Otherwise, MediaConvert ignores Dynamic range compression line. For information about the Dolby DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf."
},
"DynamicRangeCompressionRf": {
"shape": "Eac3AtmosDynamicRangeCompressionRf",
"locationName": "dynamicRangeCompressionRf",
- "documentation": "Choose the Dolby dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby stream for the RF operating mode. Default value: Film light (ATMOS_STORAGE_DDP_COMPR_FILM_LIGHT) Related setting: To have MediaConvert use the value you specify here, keep the default value, Custom (SPECIFIED) for the setting Dynamic range control (DynamicRangeControl). Otherwise, MediaConvert ignores Dynamic range compression RF (DynamicRangeCompressionRf). For information about the Dolby DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf."
+ "documentation": "Choose the Dolby dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby stream for the RF operating mode. Default value: Film light Related setting: To have MediaConvert use the value you specify here, keep the default value, Custom for the setting Dynamic range control. Otherwise, MediaConvert ignores Dynamic range compression RF. For information about the Dolby DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf."
},
"DynamicRangeControl": {
"shape": "Eac3AtmosDynamicRangeControl",
"locationName": "dynamicRangeControl",
- "documentation": "Specify whether MediaConvert should use any dynamic range control metadata from your input file. Keep the default value, Custom (SPECIFIED), to provide dynamic range control values in your job settings. Choose Follow source (INITIALIZE_FROM_SOURCE) to use the metadata from your input. Related settings--Use these settings to specify your dynamic range control values: Dynamic range compression line (DynamicRangeCompressionLine) and Dynamic range compression RF (DynamicRangeCompressionRf). When you keep the value Custom (SPECIFIED) for Dynamic range control (DynamicRangeControl) and you don't specify values for the related settings, MediaConvert uses default values for those settings."
+ "documentation": "Specify whether MediaConvert should use any dynamic range control metadata from your input file. Keep the default value, Custom, to provide dynamic range control values in your job settings. Choose Follow source to use the metadata from your input. Related settings--Use these settings to specify your dynamic range control values: Dynamic range compression line and Dynamic range compression RF. When you keep the value Custom for Dynamic range control and you don't specify values for the related settings, MediaConvert uses default values for those settings."
},
"LoRoCenterMixLevel": {
"shape": "__doubleMinNegative6Max3",
"locationName": "loRoCenterMixLevel",
- "documentation": "Specify a value for the following Dolby Atmos setting: Left only/Right only center mix (Lo/Ro center). MediaConvert uses this value for downmixing. Default value: -3 dB (ATMOS_STORAGE_DDP_MIXLEV_MINUS_3_DB). Valid values: 3.0, 1.5, 0.0, -1.5, -3.0, -4.5, and -6.0. Related setting: How the service uses this value depends on the value that you choose for Stereo downmix (Eac3AtmosStereoDownmix). Related setting: To have MediaConvert use this value, keep the default value, Custom (SPECIFIED) for the setting Downmix control (DownmixControl). Otherwise, MediaConvert ignores Left only/Right only center (LoRoCenterMixLevel)."
+ "documentation": "Specify a value for the following Dolby Atmos setting: Left only/Right only center mix (Lo/Ro center). MediaConvert uses this value for downmixing. Default value: -3 dB. Valid values: 3.0, 1.5, 0.0, -1.5, -3.0, -4.5, and -6.0. Related setting: How the service uses this value depends on the value that you choose for Stereo downmix. Related setting: To have MediaConvert use this value, keep the default value, Custom for the setting Downmix control. Otherwise, MediaConvert ignores Left only/Right only center."
},
"LoRoSurroundMixLevel": {
"shape": "__doubleMinNegative60MaxNegative1",
"locationName": "loRoSurroundMixLevel",
- "documentation": "Specify a value for the following Dolby Atmos setting: Left only/Right only (Lo/Ro surround). MediaConvert uses this value for downmixing. Default value: -3 dB (ATMOS_STORAGE_DDP_MIXLEV_MINUS_3_DB). Valid values: -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. Related setting: How the service uses this value depends on the value that you choose for Stereo downmix (Eac3AtmosStereoDownmix). Related setting: To have MediaConvert use this value, keep the default value, Custom (SPECIFIED) for the setting Downmix control (DownmixControl). Otherwise, MediaConvert ignores Left only/Right only surround (LoRoSurroundMixLevel)."
+ "documentation": "Specify a value for the following Dolby Atmos setting: Left only/Right only. MediaConvert uses this value for downmixing. Default value: -3 dB. Valid values: -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. Related setting: How the service uses this value depends on the value that you choose for Stereo downmix. Related setting: To have MediaConvert use this value, keep the default value, Custom for the setting Downmix control. Otherwise, MediaConvert ignores Left only/Right only surround."
},
"LtRtCenterMixLevel": {
"shape": "__doubleMinNegative6Max3",
"locationName": "ltRtCenterMixLevel",
- "documentation": "Specify a value for the following Dolby Atmos setting: Left total/Right total center mix (Lt/Rt center). MediaConvert uses this value for downmixing. Default value: -3 dB (ATMOS_STORAGE_DDP_MIXLEV_MINUS_3_DB) Valid values: 3.0, 1.5, 0.0, -1.5, -3.0, -4.5, and -6.0. Related setting: How the service uses this value depends on the value that you choose for Stereo downmix (Eac3AtmosStereoDownmix). Related setting: To have MediaConvert use this value, keep the default value, Custom (SPECIFIED) for the setting Downmix control (DownmixControl). Otherwise, MediaConvert ignores Left total/Right total center (LtRtCenterMixLevel)."
+ "documentation": "Specify a value for the following Dolby Atmos setting: Left total/Right total center mix (Lt/Rt center). MediaConvert uses this value for downmixing. Default value: -3 dB Valid values: 3.0, 1.5, 0.0, -1.5, -3.0, -4.5, and -6.0. Related setting: How the service uses this value depends on the value that you choose for Stereo downmix. Related setting: To have MediaConvert use this value, keep the default value, Custom for the setting Downmix control. Otherwise, MediaConvert ignores Left total/Right total center."
},
"LtRtSurroundMixLevel": {
"shape": "__doubleMinNegative60MaxNegative1",
"locationName": "ltRtSurroundMixLevel",
- "documentation": "Specify a value for the following Dolby Atmos setting: Left total/Right total surround mix (Lt/Rt surround). MediaConvert uses this value for downmixing. Default value: -3 dB (ATMOS_STORAGE_DDP_MIXLEV_MINUS_3_DB) Valid values: -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. Related setting: How the service uses this value depends on the value that you choose for Stereo downmix (Eac3AtmosStereoDownmix). Related setting: To have MediaConvert use this value, keep the default value, Custom (SPECIFIED) for the setting Downmix control (DownmixControl). Otherwise, the service ignores Left total/Right total surround (LtRtSurroundMixLevel)."
+ "documentation": "Specify a value for the following Dolby Atmos setting: Left total/Right total surround mix (Lt/Rt surround). MediaConvert uses this value for downmixing. Default value: -3 dB Valid values: -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. Related setting: How the service uses this value depends on the value that you choose for Stereo downmix. Related setting: To have MediaConvert use this value, keep the default value, Custom for the setting Downmix control. Otherwise, the service ignores Left total/Right total surround."
},
"MeteringMode": {
"shape": "Eac3AtmosMeteringMode",
@@ -4910,7 +4910,7 @@
"StereoDownmix": {
"shape": "Eac3AtmosStereoDownmix",
"locationName": "stereoDownmix",
- "documentation": "Choose how the service does stereo downmixing. Default value: Not indicated (ATMOS_STORAGE_DDP_DMIXMOD_NOT_INDICATED) Related setting: To have MediaConvert use this value, keep the default value, Custom (SPECIFIED) for the setting Downmix control (DownmixControl). Otherwise, MediaConvert ignores Stereo downmix (StereoDownmix)."
+ "documentation": "Choose how the service does stereo downmixing. Default value: Not indicated Related setting: To have MediaConvert use this value, keep the default value, Custom for the setting Downmix control. Otherwise, MediaConvert ignores Stereo downmix."
},
"SurroundExMode": {
"shape": "Eac3AtmosSurroundExMode",
@@ -4918,11 +4918,11 @@
"documentation": "Specify whether your input audio has an additional center rear surround channel matrix encoded into your left and right surround channels."
}
},
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value EAC3_ATMOS."
+ "documentation": "Required when you set Codec to the value EAC3_ATMOS."
},
"Eac3AtmosStereoDownmix": {
"type": "string",
- "documentation": "Choose how the service does stereo downmixing. Default value: Not indicated (ATMOS_STORAGE_DDP_DMIXMOD_NOT_INDICATED) Related setting: To have MediaConvert use this value, keep the default value, Custom (SPECIFIED) for the setting Downmix control (DownmixControl). Otherwise, MediaConvert ignores Stereo downmix (StereoDownmix).",
+ "documentation": "Choose how the service does stereo downmixing. Default value: Not indicated Related setting: To have MediaConvert use this value, keep the default value, Custom for the setting Downmix control. Otherwise, MediaConvert ignores Stereo downmix.",
"enum": [
"NOT_INDICATED",
"STEREO",
@@ -4977,7 +4977,7 @@
},
"Eac3DynamicRangeCompressionLine": {
"type": "string",
- "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the line operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile (DynamicRangeCompressionProfile). For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.",
+ "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the line operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile. For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.",
"enum": [
"NONE",
"FILM_STANDARD",
@@ -4989,7 +4989,7 @@
},
"Eac3DynamicRangeCompressionRf": {
"type": "string",
- "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the RF operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile (DynamicRangeCompressionProfile). For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.",
+ "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the RF operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile. For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.",
"enum": [
"NONE",
"FILM_STANDARD",
@@ -5075,12 +5075,12 @@
"DynamicRangeCompressionLine": {
"shape": "Eac3DynamicRangeCompressionLine",
"locationName": "dynamicRangeCompressionLine",
- "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the line operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile (DynamicRangeCompressionProfile). For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf."
+ "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the line operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile. For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf."
},
"DynamicRangeCompressionRf": {
"shape": "Eac3DynamicRangeCompressionRf",
"locationName": "dynamicRangeCompressionRf",
- "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the RF operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile (DynamicRangeCompressionProfile). For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf."
+ "documentation": "Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the RF operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile. For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf."
},
"LfeControl": {
"shape": "Eac3LfeControl",
@@ -5095,22 +5095,22 @@
"LoRoCenterMixLevel": {
"shape": "__doubleMinNegative60Max3",
"locationName": "loRoCenterMixLevel",
- "documentation": "Specify a value for the following Dolby Digital Plus setting: Left only/Right only center mix (Lo/Ro center). MediaConvert uses this value for downmixing. How the service uses this value depends on the value that you choose for Stereo downmix (Eac3StereoDownmix). Valid values: 3.0, 1.5, 0.0, -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. This setting applies only if you keep the default value of 3/2 - L, R, C, Ls, Rs (CODING_MODE_3_2) for the setting Coding mode (Eac3CodingMode). If you choose a different value for Coding mode, the service ignores Left only/Right only center (loRoCenterMixLevel)."
+ "documentation": "Specify a value for the following Dolby Digital Plus setting: Left only/Right only center mix. MediaConvert uses this value for downmixing. How the service uses this value depends on the value that you choose for Stereo downmix. Valid values: 3.0, 1.5, 0.0, -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. This setting applies only if you keep the default value of 3/2 - L, R, C, Ls, Rs for the setting Coding mode. If you choose a different value for Coding mode, the service ignores Left only/Right only center."
},
"LoRoSurroundMixLevel": {
"shape": "__doubleMinNegative60MaxNegative1",
"locationName": "loRoSurroundMixLevel",
- "documentation": "Specify a value for the following Dolby Digital Plus setting: Left only/Right only (Lo/Ro surround). MediaConvert uses this value for downmixing. How the service uses this value depends on the value that you choose for Stereo downmix (Eac3StereoDownmix). Valid values: -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. This setting applies only if you keep the default value of 3/2 - L, R, C, Ls, Rs (CODING_MODE_3_2) for the setting Coding mode (Eac3CodingMode). If you choose a different value for Coding mode, the service ignores Left only/Right only surround (loRoSurroundMixLevel)."
+ "documentation": "Specify a value for the following Dolby Digital Plus setting: Left only/Right only. MediaConvert uses this value for downmixing. How the service uses this value depends on the value that you choose for Stereo downmix. Valid values: -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. This setting applies only if you keep the default value of 3/2 - L, R, C, Ls, Rs for the setting Coding mode. If you choose a different value for Coding mode, the service ignores Left only/Right only surround."
},
"LtRtCenterMixLevel": {
"shape": "__doubleMinNegative60Max3",
"locationName": "ltRtCenterMixLevel",
- "documentation": "Specify a value for the following Dolby Digital Plus setting: Left total/Right total center mix (Lt/Rt center). MediaConvert uses this value for downmixing. How the service uses this value depends on the value that you choose for Stereo downmix (Eac3StereoDownmix). Valid values: 3.0, 1.5, 0.0, -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. This setting applies only if you keep the default value of 3/2 - L, R, C, Ls, Rs (CODING_MODE_3_2) for the setting Coding mode (Eac3CodingMode). If you choose a different value for Coding mode, the service ignores Left total/Right total center (ltRtCenterMixLevel)."
+ "documentation": "Specify a value for the following Dolby Digital Plus setting: Left total/Right total center mix. MediaConvert uses this value for downmixing. How the service uses this value depends on the value that you choose for Stereo downmix. Valid values: 3.0, 1.5, 0.0, -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. This setting applies only if you keep the default value of 3/2 - L, R, C, Ls, Rs for the setting Coding mode. If you choose a different value for Coding mode, the service ignores Left total/Right total center."
},
"LtRtSurroundMixLevel": {
"shape": "__doubleMinNegative60MaxNegative1",
"locationName": "ltRtSurroundMixLevel",
- "documentation": "Specify a value for the following Dolby Digital Plus setting: Left total/Right total surround mix (Lt/Rt surround). MediaConvert uses this value for downmixing. How the service uses this value depends on the value that you choose for Stereo downmix (Eac3StereoDownmix). Valid values: -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. This setting applies only if you keep the default value of 3/2 - L, R, C, Ls, Rs (CODING_MODE_3_2) for the setting Coding mode (Eac3CodingMode). If you choose a different value for Coding mode, the service ignores Left total/Right total surround (ltRtSurroundMixLevel)."
+ "documentation": "Specify a value for the following Dolby Digital Plus setting: Left total/Right total surround mix. MediaConvert uses this value for downmixing. How the service uses this value depends on the value that you choose for Stereo downmix. Valid values: -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. This setting applies only if you keep the default value of 3/2 - L, R, C, Ls, Rs for the setting Coding mode. If you choose a different value for Coding mode, the service ignores Left total/Right total surround."
},
"MetadataControl": {
"shape": "Eac3MetadataControl",
@@ -5135,7 +5135,7 @@
"StereoDownmix": {
"shape": "Eac3StereoDownmix",
"locationName": "stereoDownmix",
- "documentation": "Choose how the service does stereo downmixing. This setting only applies if you keep the default value of 3/2 - L, R, C, Ls, Rs (CODING_MODE_3_2) for the setting Coding mode (Eac3CodingMode). If you choose a different value for Coding mode, the service ignores Stereo downmix (Eac3StereoDownmix)."
+ "documentation": "Choose how the service does stereo downmixing. This setting only applies if you keep the default value of 3/2 - L, R, C, Ls, Rs for the setting Coding mode. If you choose a different value for Coding mode, the service ignores Stereo downmix."
},
"SurroundExMode": {
"shape": "Eac3SurroundExMode",
@@ -5148,11 +5148,11 @@
"documentation": "When encoding 2/0 audio, sets whether Dolby Surround is matrix encoded into the two channels."
}
},
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value EAC3."
+ "documentation": "Required when you set Codec to the value EAC3."
},
"Eac3StereoDownmix": {
"type": "string",
- "documentation": "Choose how the service does stereo downmixing. This setting only applies if you keep the default value of 3/2 - L, R, C, Ls, Rs (CODING_MODE_3_2) for the setting Coding mode (Eac3CodingMode). If you choose a different value for Coding mode, the service ignores Stereo downmix (Eac3StereoDownmix).",
+ "documentation": "Choose how the service does stereo downmixing. This setting only applies if you keep the default value of 3/2 - L, R, C, Ls, Rs for the setting Coding mode. If you choose a different value for Coding mode, the service ignores Stereo downmix.",
"enum": [
"NOT_INDICATED",
"LO_RO",
@@ -5180,7 +5180,7 @@
},
"EmbeddedConvert608To708": {
"type": "string",
- "documentation": "Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert (UPCONVERT), MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708.",
+ "documentation": "Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert, MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708.",
"enum": [
"UPCONVERT",
"DISABLED"
@@ -5197,10 +5197,10 @@
"Destination708ServiceNumber": {
"shape": "__integerMin1Max6",
"locationName": "destination708ServiceNumber",
- "documentation": "Ignore this setting unless your input captions are SCC format and you want both 608 and 708 captions embedded in your output stream. Optionally, specify the 708 service number for each output captions channel. Choose a different number for each channel. To use this setting, also set Force 608 to 708 upconvert (Convert608To708) to Upconvert (UPCONVERT) in your input captions selector settings. If you choose to upconvert but don't specify a 708 service number, MediaConvert uses the number that you specify for CC channel number (destination608ChannelNumber) for the 708 service number. For more information, see https://docs.aws.amazon.com/console/mediaconvert/dual-scc-to-embedded."
+ "documentation": "Ignore this setting unless your input captions are SCC format and you want both 608 and 708 captions embedded in your output stream. Optionally, specify the 708 service number for each output captions channel. Choose a different number for each channel. To use this setting, also set Force 608 to 708 upconvert to Upconvert in your input captions selector settings. If you choose to upconvert but don't specify a 708 service number, MediaConvert uses the number that you specify for CC channel number for the 708 service number. For more information, see https://docs.aws.amazon.com/console/mediaconvert/dual-scc-to-embedded."
}
},
- "documentation": "Settings related to CEA/EIA-608 and CEA/EIA-708 (also called embedded or ancillary) captions. Set up embedded captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/embedded-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to EMBEDDED, EMBEDDED_PLUS_SCTE20, or SCTE20_PLUS_EMBEDDED."
+ "documentation": "Settings related to CEA/EIA-608 and CEA/EIA-708 (also called embedded or ancillary) captions. Set up embedded captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/embedded-output-captions.html."
},
"EmbeddedSourceSettings": {
"type": "structure",
@@ -5208,7 +5208,7 @@
"Convert608To708": {
"shape": "EmbeddedConvert608To708",
"locationName": "convert608To708",
- "documentation": "Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert (UPCONVERT), MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708."
+ "documentation": "Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert, MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708."
},
"Source608ChannelNumber": {
"shape": "__integerMin1Max4",
@@ -5238,7 +5238,7 @@
},
"EmbeddedTimecodeOverride": {
"type": "string",
- "documentation": "Set Embedded timecode override (embeddedTimecodeOverride) to Use MDPM (USE_MDPM) when your AVCHD input contains timecode tag data in the Modified Digital Video Pack Metadata (MDPM). When you do, we recommend you also set Timecode source (inputTimecodeSource) to Embedded (EMBEDDED). Leave Embedded timecode override blank, or set to None (NONE), when your input does not contain MDPM timecode.",
+ "documentation": "Set Embedded timecode override to Use MDPM when your AVCHD input contains timecode tag data in the Modified Digital Video Pack Metadata. When you do, we recommend you also set Timecode source to Embedded. Leave Embedded timecode override blank, or set to None, when your input does not contain MDPM timecode.",
"enum": [
"NONE",
"USE_MDPM"
@@ -5272,7 +5272,7 @@
"ManifestConfirmConditionNotification": {
"shape": "EsamManifestConfirmConditionNotification",
"locationName": "manifestConfirmConditionNotification",
- "documentation": "Specifies an ESAM ManifestConfirmConditionNotification XML as per OC-SP-ESAM-API-I03-131025. The transcoder uses the manifest conditioning instructions that you provide in the setting MCC XML (mccXml)."
+ "documentation": "Specifies an ESAM ManifestConfirmConditionNotification XML as per OC-SP-ESAM-API-I03-131025. The transcoder uses the manifest conditioning instructions that you provide in the setting MCC XML."
},
"ResponseSignalPreroll": {
"shape": "__integerMin0Max30000",
@@ -5282,7 +5282,7 @@
"SignalProcessingNotification": {
"shape": "EsamSignalProcessingNotification",
"locationName": "signalProcessingNotification",
- "documentation": "Specifies an ESAM SignalProcessingNotification XML as per OC-SP-ESAM-API-I03-131025. The transcoder uses the signal processing instructions that you provide in the setting SCC XML (sccXml)."
+ "documentation": "Specifies an ESAM SignalProcessingNotification XML as per OC-SP-ESAM-API-I03-131025. The transcoder uses the signal processing instructions that you provide in the setting SCC XML."
}
},
"documentation": "Settings for Event Signaling And Messaging (ESAM). If you don't do ad insertion, you can ignore these settings."
@@ -5293,7 +5293,7 @@
"SccXml": {
"shape": "__stringPatternSNSignalProcessingNotificationNS",
"locationName": "sccXml",
- "documentation": "Provide your ESAM SignalProcessingNotification XML document inside your JSON job settings. Form the XML document as per OC-SP-ESAM-API-I03-131025. The transcoder will use the signal processing instructions in the message that you supply. Provide your ESAM SignalProcessingNotification XML document inside your JSON job settings. For your MPEG2-TS file outputs, if you want the service to place SCTE-35 markers at the insertion points you specify in the XML document, you must also enable SCTE-35 ESAM (scte35Esam). Note that you can either specify an ESAM XML document or enable SCTE-35 passthrough. You can't do both."
+ "documentation": "Provide your ESAM SignalProcessingNotification XML document inside your JSON job settings. Form the XML document as per OC-SP-ESAM-API-I03-131025. The transcoder will use the signal processing instructions in the message that you supply. For your MPEG2-TS file outputs, if you want the service to place SCTE-35 markers at the insertion points you specify in the XML document, you must also enable SCTE-35 ESAM. Note that you can either specify an ESAM XML document or enable SCTE-35 passthrough. You can't do both."
}
},
"documentation": "ESAM SignalProcessingNotification data defined by OC-SP-ESAM-API-I03-131025."
@@ -5348,7 +5348,7 @@
"Destination": {
"shape": "__stringPatternS3",
"locationName": "destination",
- "documentation": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file."
+ "documentation": "Use Destination to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file."
},
"DestinationSettings": {
"shape": "DestinationSettings",
@@ -5356,11 +5356,11 @@
"documentation": "Settings associated with the destination. Will vary based on the type of destination"
}
},
- "documentation": "Settings related to your File output group. MediaConvert uses this group of settings to generate a single standalone file, rather than a streaming package. When you work directly in your JSON job specification, include this object and any required children when you set Type, under OutputGroupSettings, to FILE_GROUP_SETTINGS."
+ "documentation": "Settings related to your File output group. MediaConvert uses this group of settings to generate a single standalone file, rather than a streaming package."
},
"FileSourceConvert608To708": {
"type": "string",
- "documentation": "Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert (UPCONVERT), MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708.",
+ "documentation": "Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert, MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708.",
"enum": [
"UPCONVERT",
"DISABLED"
@@ -5372,7 +5372,7 @@
"Convert608To708": {
"shape": "FileSourceConvert608To708",
"locationName": "convert608To708",
- "documentation": "Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert (UPCONVERT), MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708."
+ "documentation": "Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert, MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708."
},
"ConvertPaintToPop": {
"shape": "CaptionSourceConvertPaintOnToPopOn",
@@ -5382,7 +5382,7 @@
"Framerate": {
"shape": "CaptionSourceFramerate",
"locationName": "framerate",
- "documentation": "Ignore this setting unless your input captions format is SCC. To have the service compensate for differing frame rates between your input captions and input video, specify the frame rate of the captions file. Specify this value as a fraction. When you work directly in your JSON job specification, use the settings framerateNumerator and framerateDenominator. For example, you might specify 24 / 1 for 24 fps, 25 / 1 for 25 fps, 24000 / 1001 for 23.976 fps, or 30000 / 1001 for 29.97 fps."
+ "documentation": "Ignore this setting unless your input captions format is SCC. To have the service compensate for differing frame rates between your input captions and input video, specify the frame rate of the captions file. Specify this value as a fraction. For example, you might specify 24 / 1 for 24 fps, 25 / 1 for 25 fps, 24000 / 1001 for 23.976 fps, or 30000 / 1001 for 29.97 fps."
},
"SourceFile": {
"shape": "__stringMin14PatternS3SccSCCTtmlTTMLDfxpDFXPStlSTLSrtSRTXmlXMLSmiSMIVttVTTWebvttWEBVTTHttpsSccSCCTtmlTTMLDfxpDFXPStlSTLSrtSRTXmlXMLSmiSMIVttVTTWebvttWEBVTT",
@@ -5392,19 +5392,19 @@
"TimeDelta": {
"shape": "__integerMinNegative2147483648Max2147483647",
"locationName": "timeDelta",
- "documentation": "Optional. Use this setting when you need to adjust the sync between your sidecar captions and your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/time-delta-use-cases.html. Enter a positive or negative number to modify the times in the captions file. For example, type 15 to add 15 seconds to all the times in the captions file. Type -5 to subtract 5 seconds from the times in the captions file. You can optionally specify your time delta in milliseconds instead of seconds. When you do so, set the related setting, Time delta units (TimeDeltaUnits) to Milliseconds (MILLISECONDS). Note that, when you specify a time delta for timecode-based caption sources, such as SCC and STL, and your time delta isn't a multiple of the input frame rate, MediaConvert snaps the captions to the nearest frame. For example, when your input video frame rate is 25 fps and you specify 1010ms for time delta, MediaConvert delays your captions by 1000 ms."
+ "documentation": "Optional. Use this setting when you need to adjust the sync between your sidecar captions and your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/time-delta-use-cases.html. Enter a positive or negative number to modify the times in the captions file. For example, type 15 to add 15 seconds to all the times in the captions file. Type -5 to subtract 5 seconds from the times in the captions file. You can optionally specify your time delta in milliseconds instead of seconds. When you do so, set the related setting, Time delta units to Milliseconds. Note that, when you specify a time delta for timecode-based caption sources, such as SCC and STL, and your time delta isn't a multiple of the input frame rate, MediaConvert snaps the captions to the nearest frame. For example, when your input video frame rate is 25 fps and you specify 1010ms for time delta, MediaConvert delays your captions by 1000 ms."
},
"TimeDeltaUnits": {
"shape": "FileSourceTimeDeltaUnits",
"locationName": "timeDeltaUnits",
- "documentation": "When you use the setting Time delta (TimeDelta) to adjust the sync between your sidecar captions and your video, use this setting to specify the units for the delta that you specify. When you don't specify a value for Time delta units (TimeDeltaUnits), MediaConvert uses seconds by default."
+ "documentation": "When you use the setting Time delta to adjust the sync between your sidecar captions and your video, use this setting to specify the units for the delta that you specify. When you don't specify a value for Time delta units, MediaConvert uses seconds by default."
}
},
"documentation": "If your input captions are SCC, SMI, SRT, STL, TTML, WebVTT, or IMSC 1.1 in an xml file, specify the URI of the input caption source file. If your caption source is IMSC in an IMF package, use TrackSourceSettings instead of FileSoureSettings."
},
"FileSourceTimeDeltaUnits": {
"type": "string",
- "documentation": "When you use the setting Time delta (TimeDelta) to adjust the sync between your sidecar captions and your video, use this setting to specify the units for the delta that you specify. When you don't specify a value for Time delta units (TimeDeltaUnits), MediaConvert uses seconds by default.",
+ "documentation": "When you use the setting Time delta to adjust the sync between your sidecar captions and your video, use this setting to specify the units for the delta that you specify. When you don't specify a value for Time delta units, MediaConvert uses seconds by default.",
"enum": [
"SECONDS",
"MILLISECONDS"
@@ -5473,7 +5473,7 @@
"documentation": "JPEG Quality - a higher value equals higher quality."
}
},
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value FRAME_CAPTURE."
+ "documentation": "Required when you set Codec to the value FRAME_CAPTURE."
},
"GetJobRequest": {
"type": "structure",
@@ -5588,7 +5588,7 @@
},
"H264AdaptiveQuantization": {
"type": "string",
- "documentation": "Keep the default value, Auto (AUTO), for this setting to have MediaConvert automatically apply the best types of quantization for your video content. When you want to apply your quantization settings manually, you must set H264AdaptiveQuantization to a value other than Auto (AUTO). Use this setting to specify the strength of any adaptive quantization filters that you enable. If you don't want MediaConvert to do any adaptive quantization in this transcode, set Adaptive quantization (H264AdaptiveQuantization) to Off (OFF). Related settings: The value that you choose here applies to the following settings: H264FlickerAdaptiveQuantization, H264SpatialAdaptiveQuantization, and H264TemporalAdaptiveQuantization.",
+ "documentation": "Keep the default value, Auto, for this setting to have MediaConvert automatically apply the best types of quantization for your video content. When you want to apply your quantization settings manually, you must set H264AdaptiveQuantization to a value other than Auto. Use this setting to specify the strength of any adaptive quantization filters that you enable. If you don't want MediaConvert to do any adaptive quantization in this transcode, set Adaptive quantization to Off. Related settings: The value that you choose here applies to the following settings: H264FlickerAdaptiveQuantization, H264SpatialAdaptiveQuantization, and H264TemporalAdaptiveQuantization.",
"enum": [
"OFF",
"AUTO",
@@ -5601,7 +5601,7 @@
},
"H264CodecLevel": {
"type": "string",
- "documentation": "Specify an H.264 level that is consistent with your output video settings. If you aren't sure what level to specify, choose Auto (AUTO).",
+ "documentation": "Specify an H.264 level that is consistent with your output video settings. If you aren't sure what level to specify, choose Auto.",
"enum": [
"AUTO",
"LEVEL_1",
@@ -5636,7 +5636,7 @@
},
"H264DynamicSubGop": {
"type": "string",
- "documentation": "Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames (numberBFramesBetweenReferenceFrames).",
+ "documentation": "Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames.",
"enum": [
"ADAPTIVE",
"STATIC"
@@ -5652,7 +5652,7 @@
},
"H264FieldEncoding": {
"type": "string",
- "documentation": "The video encoding method for your MPEG-4 AVC output. Keep the default value, PAFF, to have MediaConvert use PAFF encoding for interlaced outputs. Choose Force field (FORCE_FIELD) to disable PAFF encoding and create separate interlaced fields. Choose MBAFF to disable PAFF and have MediaConvert use MBAFF encoding for interlaced outputs.",
+ "documentation": "The video encoding method for your MPEG-4 AVC output. Keep the default value, PAFF, to have MediaConvert use PAFF encoding for interlaced outputs. Choose Force field to disable PAFF encoding and create separate interlaced fields. Choose MBAFF to disable PAFF and have MediaConvert use MBAFF encoding for interlaced outputs.",
"enum": [
"PAFF",
"FORCE_FIELD",
@@ -5661,7 +5661,7 @@
},
"H264FlickerAdaptiveQuantization": {
"type": "string",
- "documentation": "Only use this setting when you change the default value, AUTO, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264FlickerAdaptiveQuantization is Disabled (DISABLED). Change this value to Enabled (ENABLED) to reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. To manually enable or disable H264FlickerAdaptiveQuantization, you must set Adaptive quantization (H264AdaptiveQuantization) to a value other than AUTO.",
+ "documentation": "Only use this setting when you change the default value, AUTO, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264FlickerAdaptiveQuantization is Disabled. Change this value to Enabled to reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. To manually enable or disable H264FlickerAdaptiveQuantization, you must set Adaptive quantization to a value other than AUTO.",
"enum": [
"DISABLED",
"ENABLED"
@@ -5669,7 +5669,7 @@
},
"H264FramerateControl": {
"type": "string",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator.",
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -5694,7 +5694,7 @@
},
"H264GopSizeUnits": {
"type": "string",
- "documentation": "Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto (AUTO) and and leave GOP size (GopSize) blank. By default, if you don't specify GOP mode control (GopSizeUnits), MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames (FRAMES) or Specified, seconds (SECONDS) and then provide the GOP length in the related setting GOP size (GopSize).",
+ "documentation": "Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto and and leave GOP size blank. By default, if you don't specify GOP mode control, MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames or Specified, seconds and then provide the GOP length in the related setting GOP size.",
"enum": [
"FRAMES",
"SECONDS",
@@ -5703,7 +5703,7 @@
},
"H264InterlaceMode": {
"type": "string",
- "documentation": "Choose the scan line type for the output. Keep the default value, Progressive (PROGRESSIVE) to create a progressive output, regardless of the scan type of your input. Use Top field first (TOP_FIELD) or Bottom field first (BOTTOM_FIELD) to create an output that's interlaced with the same field polarity throughout. Use Follow, default top (FOLLOW_TOP_FIELD) or Follow, default bottom (FOLLOW_BOTTOM_FIELD) to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.",
+ "documentation": "Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.",
"enum": [
"PROGRESSIVE",
"TOP_FIELD",
@@ -5714,7 +5714,7 @@
},
"H264ParControl": {
"type": "string",
- "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source (INITIALIZE_FROM_SOURCE), uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. To specify a different PAR by editing the JSON job specification, choose SPECIFIED. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.",
+ "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -5740,7 +5740,7 @@
"QvbrQualityLevel": {
"shape": "__integerMin1Max10",
"locationName": "qvbrQualityLevel",
- "documentation": "Use this setting only when you set Rate control mode (RateControlMode) to QVBR. Specify the target quality level for this output. MediaConvert determines the right number of bits to use for each part of the video to maintain the video quality that you specify. When you keep the default value, AUTO, MediaConvert picks a quality level for you, based on characteristics of your input video. If you prefer to specify a quality level, specify a number from 1 through 10. Use higher numbers for greater quality. Level 10 results in nearly lossless compression. The quality level for most broadcast-quality transcodes is between 6 and 9. Optionally, to specify a value between whole numbers, also provide a value for the setting qvbrQualityLevelFineTune. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33."
+ "documentation": "Use this setting only when you set Rate control mode to QVBR. Specify the target quality level for this output. MediaConvert determines the right number of bits to use for each part of the video to maintain the video quality that you specify. When you keep the default value, AUTO, MediaConvert picks a quality level for you, based on characteristics of your input video. If you prefer to specify a quality level, specify a number from 1 through 10. Use higher numbers for greater quality. Level 10 results in nearly lossless compression. The quality level for most broadcast-quality transcodes is between 6 and 9. Optionally, to specify a value between whole numbers, also provide a value for the setting qvbrQualityLevelFineTune. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33."
},
"QvbrQualityLevelFineTune": {
"shape": "__doubleMin0Max1",
@@ -5748,7 +5748,7 @@
"documentation": "Optional. Specify a value here to set the QVBR quality to a level that is between whole numbers. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33. MediaConvert rounds your QVBR quality level to the nearest third of a whole number. For example, if you set qvbrQualityLevel to 7 and you set qvbrQualityLevelFineTune to .25, your actual QVBR quality level is 7.33."
}
},
- "documentation": "Settings for quality-defined variable bitrate encoding with the H.264 codec. Use these settings only when you set QVBR for Rate control mode (RateControlMode)."
+ "documentation": "Settings for quality-defined variable bitrate encoding with the H.264 codec. Use these settings only when you set QVBR for Rate control mode."
},
"H264RateControlMode": {
"type": "string",
@@ -5769,7 +5769,7 @@
},
"H264ScanTypeConversionMode": {
"type": "string",
- "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing (INTERLACED_OPTIMIZE) to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing (INTERLACED), for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine (telecine) to None (NONE) or Soft (SOFT). You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode (interlaceMode) to a value other than Progressive (PROGRESSIVE).",
+ "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.",
"enum": [
"INTERLACED",
"INTERLACED_OPTIMIZE"
@@ -5777,7 +5777,7 @@
},
"H264SceneChangeDetect": {
"type": "string",
- "documentation": "Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection (TRANSITION_DETECTION) for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr.",
+ "documentation": "Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr.",
"enum": [
"DISABLED",
"ENABLED",
@@ -5790,7 +5790,7 @@
"AdaptiveQuantization": {
"shape": "H264AdaptiveQuantization",
"locationName": "adaptiveQuantization",
- "documentation": "Keep the default value, Auto (AUTO), for this setting to have MediaConvert automatically apply the best types of quantization for your video content. When you want to apply your quantization settings manually, you must set H264AdaptiveQuantization to a value other than Auto (AUTO). Use this setting to specify the strength of any adaptive quantization filters that you enable. If you don't want MediaConvert to do any adaptive quantization in this transcode, set Adaptive quantization (H264AdaptiveQuantization) to Off (OFF). Related settings: The value that you choose here applies to the following settings: H264FlickerAdaptiveQuantization, H264SpatialAdaptiveQuantization, and H264TemporalAdaptiveQuantization."
+ "documentation": "Keep the default value, Auto, for this setting to have MediaConvert automatically apply the best types of quantization for your video content. When you want to apply your quantization settings manually, you must set H264AdaptiveQuantization to a value other than Auto. Use this setting to specify the strength of any adaptive quantization filters that you enable. If you don't want MediaConvert to do any adaptive quantization in this transcode, set Adaptive quantization to Off. Related settings: The value that you choose here applies to the following settings: H264FlickerAdaptiveQuantization, H264SpatialAdaptiveQuantization, and H264TemporalAdaptiveQuantization."
},
"BandwidthReductionFilter": {
"shape": "BandwidthReductionFilter",
@@ -5805,7 +5805,7 @@
"CodecLevel": {
"shape": "H264CodecLevel",
"locationName": "codecLevel",
- "documentation": "Specify an H.264 level that is consistent with your output video settings. If you aren't sure what level to specify, choose Auto (AUTO)."
+ "documentation": "Specify an H.264 level that is consistent with your output video settings. If you aren't sure what level to specify, choose Auto."
},
"CodecProfile": {
"shape": "H264CodecProfile",
@@ -5825,17 +5825,17 @@
"FieldEncoding": {
"shape": "H264FieldEncoding",
"locationName": "fieldEncoding",
- "documentation": "The video encoding method for your MPEG-4 AVC output. Keep the default value, PAFF, to have MediaConvert use PAFF encoding for interlaced outputs. Choose Force field (FORCE_FIELD) to disable PAFF encoding and create separate interlaced fields. Choose MBAFF to disable PAFF and have MediaConvert use MBAFF encoding for interlaced outputs."
+ "documentation": "The video encoding method for your MPEG-4 AVC output. Keep the default value, PAFF, to have MediaConvert use PAFF encoding for interlaced outputs. Choose Force field to disable PAFF encoding and create separate interlaced fields. Choose MBAFF to disable PAFF and have MediaConvert use MBAFF encoding for interlaced outputs."
},
"FlickerAdaptiveQuantization": {
"shape": "H264FlickerAdaptiveQuantization",
"locationName": "flickerAdaptiveQuantization",
- "documentation": "Only use this setting when you change the default value, AUTO, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264FlickerAdaptiveQuantization is Disabled (DISABLED). Change this value to Enabled (ENABLED) to reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. To manually enable or disable H264FlickerAdaptiveQuantization, you must set Adaptive quantization (H264AdaptiveQuantization) to a value other than AUTO."
+ "documentation": "Only use this setting when you change the default value, AUTO, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264FlickerAdaptiveQuantization is Disabled. Change this value to Enabled to reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. To manually enable or disable H264FlickerAdaptiveQuantization, you must set Adaptive quantization to a value other than AUTO."
},
"FramerateControl": {
"shape": "H264FramerateControl",
"locationName": "framerateControl",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator."
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction."
},
"FramerateConversionAlgorithm": {
"shape": "H264FramerateConversionAlgorithm",
@@ -5860,17 +5860,17 @@
"GopClosedCadence": {
"shape": "__integerMin0Max2147483647",
"locationName": "gopClosedCadence",
- "documentation": "Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, keep the default value by leaving this setting out of your JSON job specification. In the console, do this by keeping the default empty value. If you do explicitly specify a value, for segmented outputs, don't set this value to 0."
+ "documentation": "Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. In the console, do this by keeping the default empty value. If you do explicitly specify a value, for segmented outputs, don't set this value to 0."
},
"GopSize": {
"shape": "__doubleMin0",
"locationName": "gopSize",
- "documentation": "Use this setting only when you set GOP mode control (GopSizeUnits) to Specified, frames (FRAMES) or Specified, seconds (SECONDS). Specify the GOP length using a whole number of frames or a decimal value of seconds. MediaConvert will interpret this value as frames or seconds depending on the value you choose for GOP mode control (GopSizeUnits). If you want to allow MediaConvert to automatically determine GOP size, leave GOP size blank and set GOP mode control to Auto (AUTO). If your output group specifies HLS, DASH, or CMAF, leave GOP size blank and set GOP mode control to Auto in each output in your output group."
+ "documentation": "Use this setting only when you set GOP mode control to Specified, frames or Specified, seconds. Specify the GOP length using a whole number of frames or a decimal value of seconds. MediaConvert will interpret this value as frames or seconds depending on the value you choose for GOP mode control. If you want to allow MediaConvert to automatically determine GOP size, leave GOP size blank and set GOP mode control to Auto. If your output group specifies HLS, DASH, or CMAF, leave GOP size blank and set GOP mode control to Auto in each output in your output group."
},
"GopSizeUnits": {
"shape": "H264GopSizeUnits",
"locationName": "gopSizeUnits",
- "documentation": "Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto (AUTO) and and leave GOP size (GopSize) blank. By default, if you don't specify GOP mode control (GopSizeUnits), MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames (FRAMES) or Specified, seconds (SECONDS) and then provide the GOP length in the related setting GOP size (GopSize)."
+ "documentation": "Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto and and leave GOP size blank. By default, if you don't specify GOP mode control, MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames or Specified, seconds and then provide the GOP length in the related setting GOP size."
},
"HrdBufferFinalFillPercentage": {
"shape": "__integerMin0Max100",
@@ -5890,7 +5890,7 @@
"InterlaceMode": {
"shape": "H264InterlaceMode",
"locationName": "interlaceMode",
- "documentation": "Choose the scan line type for the output. Keep the default value, Progressive (PROGRESSIVE) to create a progressive output, regardless of the scan type of your input. Use Top field first (TOP_FIELD) or Bottom field first (BOTTOM_FIELD) to create an output that's interlaced with the same field polarity throughout. Use Follow, default top (FOLLOW_TOP_FIELD) or Follow, default bottom (FOLLOW_BOTTOM_FIELD) to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose."
+ "documentation": "Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose."
},
"MaxBitrate": {
"shape": "__integerMin1000Max1152000000",
@@ -5900,7 +5900,7 @@
"MinIInterval": {
"shape": "__integerMin0Max30",
"locationName": "minIInterval",
- "documentation": "Use this setting only when you also enable Scene change detection (SceneChangeDetect). This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, keep the default value by leaving this setting out of your JSON job specification. In the console, do this by keeping the default empty value. When you explicitly specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval (minIInterval) to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs."
+ "documentation": "Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. When you explicitly specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs."
},
"NumberBFramesBetweenReferenceFrames": {
"shape": "__integerMin0Max7",
@@ -5915,17 +5915,17 @@
"ParControl": {
"shape": "H264ParControl",
"locationName": "parControl",
- "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source (INITIALIZE_FROM_SOURCE), uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. To specify a different PAR by editing the JSON job specification, choose SPECIFIED. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings."
+ "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings."
},
"ParDenominator": {
"shape": "__integerMin1Max2147483647",
"locationName": "parDenominator",
- "documentation": "Required when you set Pixel aspect ratio (parControl) to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33."
+ "documentation": "Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33."
},
"ParNumerator": {
"shape": "__integerMin1Max2147483647",
"locationName": "parNumerator",
- "documentation": "Required when you set Pixel aspect ratio (parControl) to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40."
+ "documentation": "Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40."
},
"QualityTuningLevel": {
"shape": "H264QualityTuningLevel",
@@ -5935,7 +5935,7 @@
"QvbrSettings": {
"shape": "H264QvbrSettings",
"locationName": "qvbrSettings",
- "documentation": "Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode (RateControlMode)."
+ "documentation": "Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode."
},
"RateControlMode": {
"shape": "H264RateControlMode",
@@ -5950,12 +5950,12 @@
"ScanTypeConversionMode": {
"shape": "H264ScanTypeConversionMode",
"locationName": "scanTypeConversionMode",
- "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing (INTERLACED_OPTIMIZE) to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing (INTERLACED), for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine (telecine) to None (NONE) or Soft (SOFT). You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode (interlaceMode) to a value other than Progressive (PROGRESSIVE)."
+ "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive."
},
"SceneChangeDetect": {
"shape": "H264SceneChangeDetect",
"locationName": "sceneChangeDetect",
- "documentation": "Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection (TRANSITION_DETECTION) for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr."
+ "documentation": "Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr."
},
"Slices": {
"shape": "__integerMin1Max32",
@@ -5965,17 +5965,17 @@
"SlowPal": {
"shape": "H264SlowPal",
"locationName": "slowPal",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1."
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25."
},
"Softness": {
"shape": "__integerMin0Max128",
"locationName": "softness",
- "documentation": "Ignore this setting unless you need to comply with a specification that requires a specific value. If you don't have a specification requirement, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness (sharpness) or by enabling a noise reducer filter (noiseReducerFilter). The Softness (softness) setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, for flat quantization. Choose the value 1 or 16 to use the default JVT softening quantization matricies from the H.264 specification. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video."
+ "documentation": "Ignore this setting unless you need to comply with a specification that requires a specific value. If you don't have a specification requirement, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness or by enabling a noise reducer filter. The Softness setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, for flat quantization. Choose the value 1 or 16 to use the default JVT softening quantization matricies from the H.264 specification. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video."
},
"SpatialAdaptiveQuantization": {
"shape": "H264SpatialAdaptiveQuantization",
"locationName": "spatialAdaptiveQuantization",
- "documentation": "Only use this setting when you change the default value, Auto (AUTO), for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264SpatialAdaptiveQuantization is Enabled (ENABLED). Keep this default value to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to set H264SpatialAdaptiveQuantization to Disabled (DISABLED). Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization (H264AdaptiveQuantization) depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher. To manually enable or disable H264SpatialAdaptiveQuantization, you must set Adaptive quantization (H264AdaptiveQuantization) to a value other than AUTO."
+ "documentation": "Only use this setting when you change the default value, Auto, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264SpatialAdaptiveQuantization is Enabled. Keep this default value to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to set H264SpatialAdaptiveQuantization to Disabled. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher. To manually enable or disable H264SpatialAdaptiveQuantization, you must set Adaptive quantization to a value other than AUTO."
},
"Syntax": {
"shape": "H264Syntax",
@@ -5985,12 +5985,12 @@
"Telecine": {
"shape": "H264Telecine",
"locationName": "telecine",
- "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine (HARD) produces a 29.97i output. Soft telecine (SOFT) produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None (NONE), MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture."
+ "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine produces a 29.97i output. Soft telecine produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture."
},
"TemporalAdaptiveQuantization": {
"shape": "H264TemporalAdaptiveQuantization",
"locationName": "temporalAdaptiveQuantization",
- "documentation": "Only use this setting when you change the default value, AUTO, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264TemporalAdaptiveQuantization is Enabled (ENABLED). Keep this default value to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to set H264TemporalAdaptiveQuantization to Disabled (DISABLED). Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization (adaptiveQuantization). To manually enable or disable H264TemporalAdaptiveQuantization, you must set Adaptive quantization (H264AdaptiveQuantization) to a value other than AUTO."
+ "documentation": "Only use this setting when you change the default value, AUTO, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264TemporalAdaptiveQuantization is Enabled. Keep this default value to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to set H264TemporalAdaptiveQuantization to Disabled. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization. To manually enable or disable H264TemporalAdaptiveQuantization, you must set Adaptive quantization to a value other than AUTO."
},
"UnregisteredSeiTimecode": {
"shape": "H264UnregisteredSeiTimecode",
@@ -5998,11 +5998,11 @@
"documentation": "Inserts timecode for each frame as 4 bytes of an unregistered SEI message."
}
},
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value H_264."
+ "documentation": "Required when you set Codec to the value H_264."
},
"H264SlowPal": {
"type": "string",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1.",
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.",
"enum": [
"DISABLED",
"ENABLED"
@@ -6010,7 +6010,7 @@
},
"H264SpatialAdaptiveQuantization": {
"type": "string",
- "documentation": "Only use this setting when you change the default value, Auto (AUTO), for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264SpatialAdaptiveQuantization is Enabled (ENABLED). Keep this default value to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to set H264SpatialAdaptiveQuantization to Disabled (DISABLED). Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization (H264AdaptiveQuantization) depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher. To manually enable or disable H264SpatialAdaptiveQuantization, you must set Adaptive quantization (H264AdaptiveQuantization) to a value other than AUTO.",
+ "documentation": "Only use this setting when you change the default value, Auto, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264SpatialAdaptiveQuantization is Enabled. Keep this default value to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to set H264SpatialAdaptiveQuantization to Disabled. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher. To manually enable or disable H264SpatialAdaptiveQuantization, you must set Adaptive quantization to a value other than AUTO.",
"enum": [
"DISABLED",
"ENABLED"
@@ -6026,7 +6026,7 @@
},
"H264Telecine": {
"type": "string",
- "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine (HARD) produces a 29.97i output. Soft telecine (SOFT) produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None (NONE), MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.",
+ "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine produces a 29.97i output. Soft telecine produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.",
"enum": [
"NONE",
"SOFT",
@@ -6035,7 +6035,7 @@
},
"H264TemporalAdaptiveQuantization": {
"type": "string",
- "documentation": "Only use this setting when you change the default value, AUTO, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264TemporalAdaptiveQuantization is Enabled (ENABLED). Keep this default value to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to set H264TemporalAdaptiveQuantization to Disabled (DISABLED). Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization (adaptiveQuantization). To manually enable or disable H264TemporalAdaptiveQuantization, you must set Adaptive quantization (H264AdaptiveQuantization) to a value other than AUTO.",
+ "documentation": "Only use this setting when you change the default value, AUTO, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264TemporalAdaptiveQuantization is Enabled. Keep this default value to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to set H264TemporalAdaptiveQuantization to Disabled. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization. To manually enable or disable H264TemporalAdaptiveQuantization, you must set Adaptive quantization to a value other than AUTO.",
"enum": [
"DISABLED",
"ENABLED"
@@ -6051,7 +6051,7 @@
},
"H265AdaptiveQuantization": {
"type": "string",
- "documentation": "When you set Adaptive Quantization (H265AdaptiveQuantization) to Auto (AUTO), or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output. Set Adaptive Quantization to Low (LOW), Medium (MEDIUM), High (HIGH), Higher (HIGHER), or Max (MAX) to manually control the strength of the quantization filter. When you do, you can specify a value for Spatial Adaptive Quantization (H265SpatialAdaptiveQuantization), Temporal Adaptive Quantization (H265TemporalAdaptiveQuantization), and Flicker Adaptive Quantization (H265FlickerAdaptiveQuantization), to further control the quantization filter. Set Adaptive Quantization to Off (OFF) to apply no quantization to your output.",
+ "documentation": "When you set Adaptive Quantization to Auto, or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output. Set Adaptive Quantization to Low, Medium, High, Higher, or Max to manually control the strength of the quantization filter. When you do, you can specify a value for Spatial Adaptive Quantization, Temporal Adaptive Quantization, and Flicker Adaptive Quantization, to further control the quantization filter. Set Adaptive Quantization to Off to apply no quantization to your output.",
"enum": [
"OFF",
"LOW",
@@ -6106,7 +6106,7 @@
},
"H265DynamicSubGop": {
"type": "string",
- "documentation": "Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames (numberBFramesBetweenReferenceFrames).",
+ "documentation": "Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames.",
"enum": [
"ADAPTIVE",
"STATIC"
@@ -6114,7 +6114,7 @@
},
"H265FlickerAdaptiveQuantization": {
"type": "string",
- "documentation": "Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set adaptiveQuantization to a value other than Off (OFF).",
+ "documentation": "Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set adaptiveQuantization to a value other than Off.",
"enum": [
"DISABLED",
"ENABLED"
@@ -6122,7 +6122,7 @@
},
"H265FramerateControl": {
"type": "string",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator.",
+ "documentation": "Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -6147,7 +6147,7 @@
},
"H265GopSizeUnits": {
"type": "string",
- "documentation": "Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto (AUTO) and and leave GOP size (GopSize) blank. By default, if you don't specify GOP mode control (GopSizeUnits), MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames (FRAMES) or Specified, seconds (SECONDS) and then provide the GOP length in the related setting GOP size (GopSize).",
+ "documentation": "Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto and and leave GOP size blank. By default, if you don't specify GOP mode control, MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames or Specified, seconds and then provide the GOP length in the related setting GOP size.",
"enum": [
"FRAMES",
"SECONDS",
@@ -6156,7 +6156,7 @@
},
"H265InterlaceMode": {
"type": "string",
- "documentation": "Choose the scan line type for the output. Keep the default value, Progressive (PROGRESSIVE) to create a progressive output, regardless of the scan type of your input. Use Top field first (TOP_FIELD) or Bottom field first (BOTTOM_FIELD) to create an output that's interlaced with the same field polarity throughout. Use Follow, default top (FOLLOW_TOP_FIELD) or Follow, default bottom (FOLLOW_BOTTOM_FIELD) to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.",
+ "documentation": "Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.",
"enum": [
"PROGRESSIVE",
"TOP_FIELD",
@@ -6167,7 +6167,7 @@
},
"H265ParControl": {
"type": "string",
- "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source (INITIALIZE_FROM_SOURCE), uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. To specify a different PAR by editing the JSON job specification, choose SPECIFIED. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.",
+ "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -6175,7 +6175,7 @@
},
"H265QualityTuningLevel": {
"type": "string",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.",
+ "documentation": "Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.",
"enum": [
"SINGLE_PASS",
"SINGLE_PASS_HQ",
@@ -6193,7 +6193,7 @@
"QvbrQualityLevel": {
"shape": "__integerMin1Max10",
"locationName": "qvbrQualityLevel",
- "documentation": "Use this setting only when you set Rate control mode (RateControlMode) to QVBR. Specify the target quality level for this output. MediaConvert determines the right number of bits to use for each part of the video to maintain the video quality that you specify. When you keep the default value, AUTO, MediaConvert picks a quality level for you, based on characteristics of your input video. If you prefer to specify a quality level, specify a number from 1 through 10. Use higher numbers for greater quality. Level 10 results in nearly lossless compression. The quality level for most broadcast-quality transcodes is between 6 and 9. Optionally, to specify a value between whole numbers, also provide a value for the setting qvbrQualityLevelFineTune. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33."
+ "documentation": "Use this setting only when you set Rate control mode to QVBR. Specify the target quality level for this output. MediaConvert determines the right number of bits to use for each part of the video to maintain the video quality that you specify. When you keep the default value, AUTO, MediaConvert picks a quality level for you, based on characteristics of your input video. If you prefer to specify a quality level, specify a number from 1 through 10. Use higher numbers for greater quality. Level 10 results in nearly lossless compression. The quality level for most broadcast-quality transcodes is between 6 and 9. Optionally, to specify a value between whole numbers, also provide a value for the setting qvbrQualityLevelFineTune. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33."
},
"QvbrQualityLevelFineTune": {
"shape": "__doubleMin0Max1",
@@ -6201,7 +6201,7 @@
"documentation": "Optional. Specify a value here to set the QVBR quality to a level that is between whole numbers. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33. MediaConvert rounds your QVBR quality level to the nearest third of a whole number. For example, if you set qvbrQualityLevel to 7 and you set qvbrQualityLevelFineTune to .25, your actual QVBR quality level is 7.33."
}
},
- "documentation": "Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode (RateControlMode)."
+ "documentation": "Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode."
},
"H265RateControlMode": {
"type": "string",
@@ -6223,7 +6223,7 @@
},
"H265ScanTypeConversionMode": {
"type": "string",
- "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing (INTERLACED_OPTIMIZE) to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing (INTERLACED), for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine (telecine) to None (NONE) or Soft (SOFT). You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode (interlaceMode) to a value other than Progressive (PROGRESSIVE).",
+ "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.",
"enum": [
"INTERLACED",
"INTERLACED_OPTIMIZE"
@@ -6231,7 +6231,7 @@
},
"H265SceneChangeDetect": {
"type": "string",
- "documentation": "Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection (TRANSITION_DETECTION) for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr.",
+ "documentation": "Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr.",
"enum": [
"DISABLED",
"ENABLED",
@@ -6244,7 +6244,7 @@
"AdaptiveQuantization": {
"shape": "H265AdaptiveQuantization",
"locationName": "adaptiveQuantization",
- "documentation": "When you set Adaptive Quantization (H265AdaptiveQuantization) to Auto (AUTO), or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output. Set Adaptive Quantization to Low (LOW), Medium (MEDIUM), High (HIGH), Higher (HIGHER), or Max (MAX) to manually control the strength of the quantization filter. When you do, you can specify a value for Spatial Adaptive Quantization (H265SpatialAdaptiveQuantization), Temporal Adaptive Quantization (H265TemporalAdaptiveQuantization), and Flicker Adaptive Quantization (H265FlickerAdaptiveQuantization), to further control the quantization filter. Set Adaptive Quantization to Off (OFF) to apply no quantization to your output."
+ "documentation": "When you set Adaptive Quantization to Auto, or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output. Set Adaptive Quantization to Low, Medium, High, Higher, or Max to manually control the strength of the quantization filter. When you do, you can specify a value for Spatial Adaptive Quantization, Temporal Adaptive Quantization, and Flicker Adaptive Quantization, to further control the quantization filter. Set Adaptive Quantization to Off to apply no quantization to your output."
},
"AlternateTransferFunctionSei": {
"shape": "H265AlternateTransferFunctionSei",
@@ -6279,12 +6279,12 @@
"FlickerAdaptiveQuantization": {
"shape": "H265FlickerAdaptiveQuantization",
"locationName": "flickerAdaptiveQuantization",
- "documentation": "Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set adaptiveQuantization to a value other than Off (OFF)."
+ "documentation": "Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set adaptiveQuantization to a value other than Off."
},
"FramerateControl": {
"shape": "H265FramerateControl",
"locationName": "framerateControl",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator."
+ "documentation": "Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction."
},
"FramerateConversionAlgorithm": {
"shape": "H265FramerateConversionAlgorithm",
@@ -6309,17 +6309,17 @@
"GopClosedCadence": {
"shape": "__integerMin0Max2147483647",
"locationName": "gopClosedCadence",
- "documentation": "Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, keep the default value by leaving this setting out of your JSON job specification. In the console, do this by keeping the default empty value. If you do explicitly specify a value, for segmented outputs, don't set this value to 0."
+ "documentation": "Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. If you do explicitly specify a value, for segmented outputs, don't set this value to 0."
},
"GopSize": {
"shape": "__doubleMin0",
"locationName": "gopSize",
- "documentation": "Use this setting only when you set GOP mode control (GopSizeUnits) to Specified, frames (FRAMES) or Specified, seconds (SECONDS). Specify the GOP length using a whole number of frames or a decimal value of seconds. MediaConvert will interpret this value as frames or seconds depending on the value you choose for GOP mode control (GopSizeUnits). If you want to allow MediaConvert to automatically determine GOP size, leave GOP size blank and set GOP mode control to Auto (AUTO). If your output group specifies HLS, DASH, or CMAF, leave GOP size blank and set GOP mode control to Auto in each output in your output group."
+ "documentation": "Use this setting only when you set GOP mode control to Specified, frames or Specified, seconds. Specify the GOP length using a whole number of frames or a decimal value of seconds. MediaConvert will interpret this value as frames or seconds depending on the value you choose for GOP mode control. If you want to allow MediaConvert to automatically determine GOP size, leave GOP size blank and set GOP mode control to Auto. If your output group specifies HLS, DASH, or CMAF, leave GOP size blank and set GOP mode control to Auto in each output in your output group."
},
"GopSizeUnits": {
"shape": "H265GopSizeUnits",
"locationName": "gopSizeUnits",
- "documentation": "Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto (AUTO) and and leave GOP size (GopSize) blank. By default, if you don't specify GOP mode control (GopSizeUnits), MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames (FRAMES) or Specified, seconds (SECONDS) and then provide the GOP length in the related setting GOP size (GopSize)."
+ "documentation": "Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto and and leave GOP size blank. By default, if you don't specify GOP mode control, MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames or Specified, seconds and then provide the GOP length in the related setting GOP size."
},
"HrdBufferFinalFillPercentage": {
"shape": "__integerMin0Max100",
@@ -6339,7 +6339,7 @@
"InterlaceMode": {
"shape": "H265InterlaceMode",
"locationName": "interlaceMode",
- "documentation": "Choose the scan line type for the output. Keep the default value, Progressive (PROGRESSIVE) to create a progressive output, regardless of the scan type of your input. Use Top field first (TOP_FIELD) or Bottom field first (BOTTOM_FIELD) to create an output that's interlaced with the same field polarity throughout. Use Follow, default top (FOLLOW_TOP_FIELD) or Follow, default bottom (FOLLOW_BOTTOM_FIELD) to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose."
+ "documentation": "Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose."
},
"MaxBitrate": {
"shape": "__integerMin1000Max1466400000",
@@ -6349,7 +6349,7 @@
"MinIInterval": {
"shape": "__integerMin0Max30",
"locationName": "minIInterval",
- "documentation": "Use this setting only when you also enable Scene change detection (SceneChangeDetect). This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, keep the default value by leaving this setting out of your JSON job specification. In the console, do this by keeping the default empty value. When you explicitly specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval (minIInterval) to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs."
+ "documentation": "Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. When you explicitly specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs."
},
"NumberBFramesBetweenReferenceFrames": {
"shape": "__integerMin0Max7",
@@ -6364,27 +6364,27 @@
"ParControl": {
"shape": "H265ParControl",
"locationName": "parControl",
- "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source (INITIALIZE_FROM_SOURCE), uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. To specify a different PAR by editing the JSON job specification, choose SPECIFIED. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings."
+ "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings."
},
"ParDenominator": {
"shape": "__integerMin1Max2147483647",
"locationName": "parDenominator",
- "documentation": "Required when you set Pixel aspect ratio (parControl) to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33."
+ "documentation": "Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33."
},
"ParNumerator": {
"shape": "__integerMin1Max2147483647",
"locationName": "parNumerator",
- "documentation": "Required when you set Pixel aspect ratio (parControl) to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40."
+ "documentation": "Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40."
},
"QualityTuningLevel": {
"shape": "H265QualityTuningLevel",
"locationName": "qualityTuningLevel",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding."
+ "documentation": "Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding."
},
"QvbrSettings": {
"shape": "H265QvbrSettings",
"locationName": "qvbrSettings",
- "documentation": "Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode (RateControlMode)."
+ "documentation": "Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode."
},
"RateControlMode": {
"shape": "H265RateControlMode",
@@ -6399,12 +6399,12 @@
"ScanTypeConversionMode": {
"shape": "H265ScanTypeConversionMode",
"locationName": "scanTypeConversionMode",
- "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing (INTERLACED_OPTIMIZE) to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing (INTERLACED), for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine (telecine) to None (NONE) or Soft (SOFT). You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode (interlaceMode) to a value other than Progressive (PROGRESSIVE)."
+ "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive."
},
"SceneChangeDetect": {
"shape": "H265SceneChangeDetect",
"locationName": "sceneChangeDetect",
- "documentation": "Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection (TRANSITION_DETECTION) for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr."
+ "documentation": "Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr."
},
"Slices": {
"shape": "__integerMin1Max32",
@@ -6414,22 +6414,22 @@
"SlowPal": {
"shape": "H265SlowPal",
"locationName": "slowPal",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1."
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25."
},
"SpatialAdaptiveQuantization": {
"shape": "H265SpatialAdaptiveQuantization",
"locationName": "spatialAdaptiveQuantization",
- "documentation": "Keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization (adaptiveQuantization) depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher."
+ "documentation": "Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher."
},
"Telecine": {
"shape": "H265Telecine",
"locationName": "telecine",
- "documentation": "This field applies only if the Streams > Advanced > Framerate (framerate) field is set to 29.970. This field works with the Streams > Advanced > Preprocessors > Deinterlacer field (deinterlace_mode) and the Streams > Advanced > Interlaced Mode field (interlace_mode) to identify the scan type for the output: Progressive, Interlaced, Hard Telecine or Soft Telecine. - Hard: produces 29.97i output from 23.976 input. - Soft: produces 23.976; the player converts this output to 29.97i."
+ "documentation": "This field applies only if the Streams > Advanced > Framerate field is set to 29.970. This field works with the Streams > Advanced > Preprocessors > Deinterlacer field and the Streams > Advanced > Interlaced Mode field to identify the scan type for the output: Progressive, Interlaced, Hard Telecine or Soft Telecine. - Hard: produces 29.97i output from 23.976 input. - Soft: produces 23.976; the player converts this output to 29.97i."
},
"TemporalAdaptiveQuantization": {
"shape": "H265TemporalAdaptiveQuantization",
"locationName": "temporalAdaptiveQuantization",
- "documentation": "Keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization (adaptiveQuantization)."
+ "documentation": "Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization."
},
"TemporalIds": {
"shape": "H265TemporalIds",
@@ -6456,7 +6456,7 @@
},
"H265SlowPal": {
"type": "string",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1.",
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.",
"enum": [
"DISABLED",
"ENABLED"
@@ -6464,7 +6464,7 @@
},
"H265SpatialAdaptiveQuantization": {
"type": "string",
- "documentation": "Keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization (adaptiveQuantization) depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.",
+ "documentation": "Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.",
"enum": [
"DISABLED",
"ENABLED"
@@ -6472,7 +6472,7 @@
},
"H265Telecine": {
"type": "string",
- "documentation": "This field applies only if the Streams > Advanced > Framerate (framerate) field is set to 29.970. This field works with the Streams > Advanced > Preprocessors > Deinterlacer field (deinterlace_mode) and the Streams > Advanced > Interlaced Mode field (interlace_mode) to identify the scan type for the output: Progressive, Interlaced, Hard Telecine or Soft Telecine. - Hard: produces 29.97i output from 23.976 input. - Soft: produces 23.976; the player converts this output to 29.97i.",
+ "documentation": "This field applies only if the Streams > Advanced > Framerate field is set to 29.970. This field works with the Streams > Advanced > Preprocessors > Deinterlacer field and the Streams > Advanced > Interlaced Mode field to identify the scan type for the output: Progressive, Interlaced, Hard Telecine or Soft Telecine. - Hard: produces 29.97i output from 23.976 input. - Soft: produces 23.976; the player converts this output to 29.97i.",
"enum": [
"NONE",
"SOFT",
@@ -6481,7 +6481,7 @@
},
"H265TemporalAdaptiveQuantization": {
"type": "string",
- "documentation": "Keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization (adaptiveQuantization).",
+ "documentation": "Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.",
"enum": [
"DISABLED",
"ENABLED"
@@ -6635,7 +6635,7 @@
},
"HlsAudioOnlyContainer": {
"type": "string",
- "documentation": "Use this setting only in audio-only outputs. Choose MPEG-2 Transport Stream (M2TS) to create a file in an MPEG2-TS container. Keep the default value Automatic (AUTOMATIC) to create a raw audio-only file with no container. Regardless of the value that you specify here, if this output has video, the service will place outputs into an MPEG2-TS container.",
+ "documentation": "Use this setting only in audio-only outputs. Choose MPEG-2 Transport Stream (M2TS) to create a file in an MPEG2-TS container. Keep the default value Automatic to create a raw audio-only file with no container. Regardless of the value that you specify here, if this output has video, the service will place outputs into an MPEG2-TS container.",
"enum": [
"AUTOMATIC",
"M2TS"
@@ -6643,7 +6643,7 @@
},
"HlsAudioOnlyHeader": {
"type": "string",
- "documentation": "Ignore this setting unless you are using FairPlay DRM with Verimatrix and you encounter playback issues. Keep the default value, Include (INCLUDE), to output audio-only headers. Choose Exclude (EXCLUDE) to remove the audio-only headers from your audio segments.",
+ "documentation": "Ignore this setting unless you are using FairPlay DRM with Verimatrix and you encounter playback issues. Keep the default value, Include, to output audio-only headers. Choose Exclude to remove the audio-only headers from your audio segments.",
"enum": [
"INCLUDE",
"EXCLUDE"
@@ -6696,7 +6696,7 @@
},
"HlsCaptionSegmentLengthControl": {
"type": "string",
- "documentation": "Set Caption segment length control (CaptionSegmentLengthControl) to Match video (MATCH_VIDEO) to create caption segments that align with the video segments from the first video output in this output group. For example, if the video segments are 2 seconds long, your WebVTT segments will also be 2 seconds long. Keep the default setting, Large segments (LARGE_SEGMENTS) to create caption segments that are 300 seconds long.",
+ "documentation": "Set Caption segment length control to Match video to create caption segments that align with the video segments from the first video output in this output group. For example, if the video segments are 2 seconds long, your WebVTT segments will also be 2 seconds long. Keep the default setting, Large segments to create caption segments that are 300 seconds long.",
"enum": [
"LARGE_SEGMENTS",
"MATCH_VIDEO"
@@ -6704,7 +6704,7 @@
},
"HlsClientCache": {
"type": "string",
- "documentation": "Disable this setting only when your workflow requires the #EXT-X-ALLOW-CACHE:no tag. Otherwise, keep the default value Enabled (ENABLED) and control caching in your video distribution set up. For example, use the Cache-Control http header.",
+ "documentation": "Disable this setting only when your workflow requires the #EXT-X-ALLOW-CACHE:no tag. Otherwise, keep the default value Enabled and control caching in your video distribution set up. For example, use the Cache-Control http header.",
"enum": [
"DISABLED",
"ENABLED"
@@ -6720,7 +6720,7 @@
},
"HlsDescriptiveVideoServiceFlag": {
"type": "string",
- "documentation": "Specify whether to flag this audio track as descriptive video service (DVS) in your HLS parent manifest. When you choose Flag (FLAG), MediaConvert includes the parameter CHARACTERISTICS=\"public.accessibility.describes-video\" in the EXT-X-MEDIA entry for this track. When you keep the default choice, Don't flag (DONT_FLAG), MediaConvert leaves this parameter out. The DVS flag can help with accessibility on Apple devices. For more information, see the Apple documentation.",
+ "documentation": "Specify whether to flag this audio track as descriptive video service (DVS) in your HLS parent manifest. When you choose Flag, MediaConvert includes the parameter CHARACTERISTICS=\"public.accessibility.describes-video\" in the EXT-X-MEDIA entry for this track. When you keep the default choice, Don't flag, MediaConvert leaves this parameter out. The DVS flag can help with accessibility on Apple devices. For more information, see the Apple documentation.",
"enum": [
"DONT_FLAG",
"FLAG"
@@ -6799,7 +6799,7 @@
"AudioOnlyHeader": {
"shape": "HlsAudioOnlyHeader",
"locationName": "audioOnlyHeader",
- "documentation": "Ignore this setting unless you are using FairPlay DRM with Verimatrix and you encounter playback issues. Keep the default value, Include (INCLUDE), to output audio-only headers. Choose Exclude (EXCLUDE) to remove the audio-only headers from your audio segments."
+ "documentation": "Ignore this setting unless you are using FairPlay DRM with Verimatrix and you encounter playback issues. Keep the default value, Include, to output audio-only headers. Choose Exclude to remove the audio-only headers from your audio segments."
},
"BaseUrl": {
"shape": "__string",
@@ -6819,12 +6819,12 @@
"CaptionSegmentLengthControl": {
"shape": "HlsCaptionSegmentLengthControl",
"locationName": "captionSegmentLengthControl",
- "documentation": "Set Caption segment length control (CaptionSegmentLengthControl) to Match video (MATCH_VIDEO) to create caption segments that align with the video segments from the first video output in this output group. For example, if the video segments are 2 seconds long, your WebVTT segments will also be 2 seconds long. Keep the default setting, Large segments (LARGE_SEGMENTS) to create caption segments that are 300 seconds long."
+ "documentation": "Set Caption segment length control to Match video to create caption segments that align with the video segments from the first video output in this output group. For example, if the video segments are 2 seconds long, your WebVTT segments will also be 2 seconds long. Keep the default setting, Large segments to create caption segments that are 300 seconds long."
},
"ClientCache": {
"shape": "HlsClientCache",
"locationName": "clientCache",
- "documentation": "Disable this setting only when your workflow requires the #EXT-X-ALLOW-CACHE:no tag. Otherwise, keep the default value Enabled (ENABLED) and control caching in your video distribution set up. For example, use the Cache-Control http header."
+ "documentation": "Disable this setting only when your workflow requires the #EXT-X-ALLOW-CACHE:no tag. Otherwise, keep the default value Enabled and control caching in your video distribution set up. For example, use the Cache-Control http header."
},
"CodecSpecification": {
"shape": "HlsCodecSpecification",
@@ -6834,7 +6834,7 @@
"Destination": {
"shape": "__stringPatternS3",
"locationName": "destination",
- "documentation": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file."
+ "documentation": "Use Destination to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file."
},
"DestinationSettings": {
"shape": "DestinationSettings",
@@ -6854,7 +6854,7 @@
"ImageBasedTrickPlay": {
"shape": "HlsImageBasedTrickPlay",
"locationName": "imageBasedTrickPlay",
- "documentation": "Specify whether MediaConvert generates images for trick play. Keep the default value, None (NONE), to not generate any images. Choose Thumbnail (THUMBNAIL) to generate tiled thumbnails. Choose Thumbnail and full frame (THUMBNAIL_AND_FULLFRAME) to generate tiled thumbnails and full-resolution images of single frames. MediaConvert creates a child manifest for each set of images that you generate and adds corresponding entries to the parent manifest. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md"
+ "documentation": "Specify whether MediaConvert generates images for trick play. Keep the default value, None, to not generate any images. Choose Thumbnail to generate tiled thumbnails. Choose Thumbnail and full frame to generate tiled thumbnails and full-resolution images of single frames. MediaConvert creates a child manifest for each set of images that you generate and adds corresponding entries to the parent manifest. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md"
},
"ImageBasedTrickPlaySettings": {
"shape": "HlsImageBasedTrickPlaySettings",
@@ -6909,12 +6909,12 @@
"SegmentLength": {
"shape": "__integerMin1Max2147483647",
"locationName": "segmentLength",
- "documentation": "Specify the length, in whole seconds, of each segment. When you don't specify a value, MediaConvert defaults to 10. Related settings: Use Segment length control (SegmentLengthControl) to specify whether the encoder enforces this value strictly. Use Segment control (HlsSegmentControl) to specify whether MediaConvert creates separate segment files or one content file that has metadata to mark the segment boundaries."
+ "documentation": "Specify the length, in whole seconds, of each segment. When you don't specify a value, MediaConvert defaults to 10. Related settings: Use Segment length control to specify whether the encoder enforces this value strictly. Use Segment control to specify whether MediaConvert creates separate segment files or one content file that has metadata to mark the segment boundaries."
},
"SegmentLengthControl": {
"shape": "HlsSegmentLengthControl",
"locationName": "segmentLengthControl",
- "documentation": "Specify how you want MediaConvert to determine the segment length. Choose Exact (EXACT) to have the encoder use the exact length that you specify with the setting Segment length (SegmentLength). This might result in extra I-frames. Choose Multiple of GOP (GOP_MULTIPLE) to have the encoder round up the segment lengths to match the next GOP boundary."
+ "documentation": "Specify how you want MediaConvert to determine the segment length. Choose Exact to have the encoder use the exact length that you specify with the setting Segment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary."
},
"SegmentsPerSubdirectory": {
"shape": "__integerMin1Max2147483647",
@@ -6934,12 +6934,12 @@
"TimedMetadataId3Frame": {
"shape": "HlsTimedMetadataId3Frame",
"locationName": "timedMetadataId3Frame",
- "documentation": "Specify the type of the ID3 frame (timedMetadataId3Frame) to use for ID3 timestamps (timedMetadataId3Period) in your output. To include ID3 timestamps: Specify PRIV (PRIV) or TDRL (TDRL) and set ID3 metadata (timedMetadata) to Passthrough (PASSTHROUGH). To exclude ID3 timestamps: Set ID3 timestamp frame type to None (NONE)."
+ "documentation": "Specify the type of the ID3 frame to use for ID3 timestamps in your output. To include ID3 timestamps: Specify PRIV or TDRL and set ID3 metadata to Passthrough. To exclude ID3 timestamps: Set ID3 timestamp frame type to None."
},
"TimedMetadataId3Period": {
"shape": "__integerMinNegative2147483648Max2147483647",
"locationName": "timedMetadataId3Period",
- "documentation": "Specify the interval in seconds to write ID3 timestamps in your output. The first timestamp starts at the output timecode and date, and increases incrementally with each ID3 timestamp. To use the default interval of 10 seconds: Leave blank. To include this metadata in your output: Set ID3 timestamp frame type (timedMetadataId3Frame) to PRIV (PRIV) or TDRL (TDRL), and set ID3 metadata (timedMetadata) to Passthrough (PASSTHROUGH)."
+ "documentation": "Specify the interval in seconds to write ID3 timestamps in your output. The first timestamp starts at the output timecode and date, and increases incrementally with each ID3 timestamp. To use the default interval of 10 seconds: Leave blank. To include this metadata in your output: Set ID3 timestamp frame type to PRIV or TDRL, and set ID3 metadata to Passthrough."
},
"TimestampDeltaMilliseconds": {
"shape": "__integerMinNegative2147483648Max2147483647",
@@ -6947,11 +6947,11 @@
"documentation": "Provides an extra millisecond delta offset to fine tune the timestamps."
}
},
- "documentation": "Settings related to your HLS output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html. When you work directly in your JSON job specification, include this object and any required children when you set Type, under OutputGroupSettings, to HLS_GROUP_SETTINGS."
+ "documentation": "Settings related to your HLS output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html."
},
"HlsIFrameOnlyManifest": {
"type": "string",
- "documentation": "Choose Include (INCLUDE) to have MediaConvert generate a child manifest that lists only the I-frames for this rendition, in addition to your regular manifest for this rendition. You might use this manifest as part of a workflow that creates preview functions for your video. MediaConvert adds both the I-frame only child manifest and the regular child manifest to the parent manifest. When you don't need the I-frame only child manifest, keep the default value Exclude (EXCLUDE).",
+ "documentation": "Choose Include to have MediaConvert generate a child manifest that lists only the I-frames for this rendition, in addition to your regular manifest for this rendition. You might use this manifest as part of a workflow that creates preview functions for your video. MediaConvert adds both the I-frame only child manifest and the regular child manifest to the parent manifest. When you don't need the I-frame only child manifest, keep the default value Exclude.",
"enum": [
"INCLUDE",
"EXCLUDE"
@@ -6959,7 +6959,7 @@
},
"HlsImageBasedTrickPlay": {
"type": "string",
- "documentation": "Specify whether MediaConvert generates images for trick play. Keep the default value, None (NONE), to not generate any images. Choose Thumbnail (THUMBNAIL) to generate tiled thumbnails. Choose Thumbnail and full frame (THUMBNAIL_AND_FULLFRAME) to generate tiled thumbnails and full-resolution images of single frames. MediaConvert creates a child manifest for each set of images that you generate and adds corresponding entries to the parent manifest. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md",
+ "documentation": "Specify whether MediaConvert generates images for trick play. Keep the default value, None, to not generate any images. Choose Thumbnail to generate tiled thumbnails. Choose Thumbnail and full frame to generate tiled thumbnails and full-resolution images of single frames. MediaConvert creates a child manifest for each set of images that you generate and adds corresponding entries to the parent manifest. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md",
"enum": [
"NONE",
"THUMBNAIL",
@@ -7106,7 +7106,7 @@
},
"HlsSegmentLengthControl": {
"type": "string",
- "documentation": "Specify how you want MediaConvert to determine the segment length. Choose Exact (EXACT) to have the encoder use the exact length that you specify with the setting Segment length (SegmentLength). This might result in extra I-frames. Choose Multiple of GOP (GOP_MULTIPLE) to have the encoder round up the segment lengths to match the next GOP boundary.",
+ "documentation": "Specify how you want MediaConvert to determine the segment length. Choose Exact to have the encoder use the exact length that you specify with the setting Segment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary.",
"enum": [
"EXACT",
"GOP_MULTIPLE"
@@ -7123,7 +7123,7 @@
"AudioOnlyContainer": {
"shape": "HlsAudioOnlyContainer",
"locationName": "audioOnlyContainer",
- "documentation": "Use this setting only in audio-only outputs. Choose MPEG-2 Transport Stream (M2TS) to create a file in an MPEG2-TS container. Keep the default value Automatic (AUTOMATIC) to create an audio-only file in a raw container. Regardless of the value that you specify here, if this output has video, the service will place the output into an MPEG2-TS container."
+ "documentation": "Use this setting only in audio-only outputs. Choose MPEG-2 Transport Stream (M2TS) to create a file in an MPEG2-TS container. Keep the default value Automatic to create an audio-only file in a raw container. Regardless of the value that you specify here, if this output has video, the service will place the output into an MPEG2-TS container."
},
"AudioRenditionSets": {
"shape": "__string",
@@ -7138,12 +7138,12 @@
"DescriptiveVideoServiceFlag": {
"shape": "HlsDescriptiveVideoServiceFlag",
"locationName": "descriptiveVideoServiceFlag",
- "documentation": "Specify whether to flag this audio track as descriptive video service (DVS) in your HLS parent manifest. When you choose Flag (FLAG), MediaConvert includes the parameter CHARACTERISTICS=\"public.accessibility.describes-video\" in the EXT-X-MEDIA entry for this track. When you keep the default choice, Don't flag (DONT_FLAG), MediaConvert leaves this parameter out. The DVS flag can help with accessibility on Apple devices. For more information, see the Apple documentation."
+ "documentation": "Specify whether to flag this audio track as descriptive video service (DVS) in your HLS parent manifest. When you choose Flag, MediaConvert includes the parameter CHARACTERISTICS=\"public.accessibility.describes-video\" in the EXT-X-MEDIA entry for this track. When you keep the default choice, Don't flag, MediaConvert leaves this parameter out. The DVS flag can help with accessibility on Apple devices. For more information, see the Apple documentation."
},
"IFrameOnlyManifest": {
"shape": "HlsIFrameOnlyManifest",
"locationName": "iFrameOnlyManifest",
- "documentation": "Choose Include (INCLUDE) to have MediaConvert generate a child manifest that lists only the I-frames for this rendition, in addition to your regular manifest for this rendition. You might use this manifest as part of a workflow that creates preview functions for your video. MediaConvert adds both the I-frame only child manifest and the regular child manifest to the parent manifest. When you don't need the I-frame only child manifest, keep the default value Exclude (EXCLUDE)."
+ "documentation": "Choose Include to have MediaConvert generate a child manifest that lists only the I-frames for this rendition, in addition to your regular manifest for this rendition. You might use this manifest as part of a workflow that creates preview functions for your video. MediaConvert adds both the I-frame only child manifest and the regular child manifest to the parent manifest. When you don't need the I-frame only child manifest, keep the default value Exclude."
},
"SegmentModifier": {
"shape": "__string",
@@ -7171,7 +7171,7 @@
},
"HlsTimedMetadataId3Frame": {
"type": "string",
- "documentation": "Specify the type of the ID3 frame (timedMetadataId3Frame) to use for ID3 timestamps (timedMetadataId3Period) in your output. To include ID3 timestamps: Specify PRIV (PRIV) or TDRL (TDRL) and set ID3 metadata (timedMetadata) to Passthrough (PASSTHROUGH). To exclude ID3 timestamps: Set ID3 timestamp frame type to None (NONE).",
+ "documentation": "Specify the type of the ID3 frame to use for ID3 timestamps in your output. To include ID3 timestamps: Specify PRIV or TDRL and set ID3 metadata to Passthrough. To exclude ID3 timestamps: Set ID3 timestamp frame type to None.",
"enum": [
"NONE",
"PRIV",
@@ -7205,15 +7205,15 @@
"Id3": {
"shape": "__stringPatternAZaZ0902",
"locationName": "id3",
- "documentation": "Use ID3 tag (Id3) to provide a fully formed ID3 tag in base64-encode format."
+ "documentation": "Use ID3 tag to provide a fully formed ID3 tag in base64-encode format."
},
"Timecode": {
"shape": "__stringPattern010920405090509092",
"locationName": "timecode",
- "documentation": "Provide a Timecode (TimeCode) in HH:MM:SS:FF or HH:MM:SS;FF format."
+ "documentation": "Provide a Timecode in HH:MM:SS:FF or HH:MM:SS;FF format."
}
},
- "documentation": "To insert ID3 tags in your output, specify two values. Use ID3 tag (Id3) to specify the base 64 encoded string and use Timecode (TimeCode) to specify the time when the tag should be inserted. To insert multiple ID3 tags in your output, create multiple instances of ID3 insertion (Id3Insertion)."
+ "documentation": "To insert ID3 tags in your output, specify two values. Use ID3 tag to specify the base 64 encoded string and use Timecode to specify the time when the tag should be inserted. To insert multiple ID3 tags in your output, create multiple instances of ID3 insertion."
},
"ImageInserter": {
"type": "structure",
@@ -7253,7 +7253,7 @@
"documentation": "Keep this setting enabled to have MediaConvert use the font style and position information from the captions source in the output. This option is available only when your input captions are IMSC, SMPTE-TT, or TTML. Disable this setting for simplified output captions."
}
},
- "documentation": "Settings related to IMSC captions. IMSC is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to IMSC."
+ "documentation": "Settings related to IMSC captions. IMSC is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html."
},
"ImscStylePassthrough": {
"type": "string",
@@ -7279,12 +7279,12 @@
"AudioSelectorGroups": {
"shape": "__mapOfAudioSelectorGroup",
"locationName": "audioSelectorGroups",
- "documentation": "Use audio selector groups to combine multiple sidecar audio inputs so that you can assign them to a single output audio tab (AudioDescription). Note that, if you're working with embedded audio, it's simpler to assign multiple input tracks into a single audio selector rather than use an audio selector group."
+ "documentation": "Use audio selector groups to combine multiple sidecar audio inputs so that you can assign them to a single output audio tab. Note that, if you're working with embedded audio, it's simpler to assign multiple input tracks into a single audio selector rather than use an audio selector group."
},
"AudioSelectors": {
"shape": "__mapOfAudioSelector",
"locationName": "audioSelectors",
- "documentation": "Use Audio selectors (AudioSelectors) to specify a track or set of tracks from the input that you will use in your outputs. You can use multiple Audio selectors per input."
+ "documentation": "Use Audio selectors to specify a track or set of tracks from the input that you will use in your outputs. You can use multiple Audio selectors per input."
},
"CaptionSelectors": {
"shape": "__mapOfCaptionSelector",
@@ -7294,12 +7294,12 @@
"Crop": {
"shape": "Rectangle",
"locationName": "crop",
- "documentation": "Use Cropping selection (crop) to specify the video area that the service will include in the output video frame. If you specify a value here, it will override any value that you specify in the output setting Cropping selection (crop)."
+ "documentation": "Use Cropping selection to specify the video area that the service will include in the output video frame. If you specify a value here, it will override any value that you specify in the output setting Cropping selection."
},
"DeblockFilter": {
"shape": "InputDeblockFilter",
"locationName": "deblockFilter",
- "documentation": "Enable Deblock (InputDeblockFilter) to produce smoother motion in the output. Default is disabled. Only manually controllable for MPEG2 and uncompressed video inputs."
+ "documentation": "Enable Deblock to produce smoother motion in the output. Default is disabled. Only manually controllable for MPEG2 and uncompressed video inputs."
},
"DecryptionSettings": {
"shape": "InputDecryptionSettings",
@@ -7309,7 +7309,7 @@
"DenoiseFilter": {
"shape": "InputDenoiseFilter",
"locationName": "denoiseFilter",
- "documentation": "Enable Denoise (InputDenoiseFilter) to filter noise from the input. Default is disabled. Only applicable to MPEG2, H.264, H.265, and uncompressed video inputs."
+ "documentation": "Enable Denoise to filter noise from the input. Default is disabled. Only applicable to MPEG2, H.264, H.265, and uncompressed video inputs."
},
"DolbyVisionMetadataXml": {
"shape": "__stringMin14PatternS3XmlXMLHttpsXmlXML",
@@ -7319,7 +7319,7 @@
"FileInput": {
"shape": "__stringPatternS3Https",
"locationName": "fileInput",
- "documentation": "Specify the source file for your transcoding job. You can use multiple inputs in a single job. The service concatenates these inputs, in the order that you specify them in the job, to create the outputs. If your input format is IMF, specify your input by providing the path to your CPL. For example, \"s3://bucket/vf/cpl.xml\". If the CPL is in an incomplete IMP, make sure to use *Supplemental IMPs* (SupplementalImps) to specify any supplemental IMPs that contain assets referenced by the CPL."
+ "documentation": "Specify the source file for your transcoding job. You can use multiple inputs in a single job. The service concatenates these inputs, in the order that you specify them in the job, to create the outputs. If your input format is IMF, specify your input by providing the path to your CPL. For example, \"s3://bucket/vf/cpl.xml\". If the CPL is in an incomplete IMP, make sure to use *Supplemental IMPs* to specify any supplemental IMPs that contain assets referenced by the CPL."
},
"FilterEnable": {
"shape": "InputFilterEnable",
@@ -7339,27 +7339,27 @@
"InputClippings": {
"shape": "__listOfInputClipping",
"locationName": "inputClippings",
- "documentation": "(InputClippings) contains sets of start and end times that together specify a portion of the input to be used in the outputs. If you provide only a start time, the clip will be the entire input from that point to the end. If you provide only an end time, it will be the entire input up to that point. When you specify more than one input clip, the transcoding service creates the job outputs by stringing the clips together in the order you specify them."
+ "documentation": "Contains sets of start and end times that together specify a portion of the input to be used in the outputs. If you provide only a start time, the clip will be the entire input from that point to the end. If you provide only an end time, it will be the entire input up to that point. When you specify more than one input clip, the transcoding service creates the job outputs by stringing the clips together in the order you specify them."
},
"InputScanType": {
"shape": "InputScanType",
"locationName": "inputScanType",
- "documentation": "When you have a progressive segmented frame (PsF) input, use this setting to flag the input as PsF. MediaConvert doesn't automatically detect PsF. Therefore, flagging your input as PsF results in better preservation of video quality when you do deinterlacing and frame rate conversion. If you don't specify, the default value is Auto (AUTO). Auto is the correct setting for all inputs that are not PsF. Don't set this value to PsF when your input is interlaced. Doing so creates horizontal interlacing artifacts."
+ "documentation": "When you have a progressive segmented frame (PsF) input, use this setting to flag the input as PsF. MediaConvert doesn't automatically detect PsF. Therefore, flagging your input as PsF results in better preservation of video quality when you do deinterlacing and frame rate conversion. If you don't specify, the default value is Auto. Auto is the correct setting for all inputs that are not PsF. Don't set this value to PsF when your input is interlaced. Doing so creates horizontal interlacing artifacts."
},
"Position": {
"shape": "Rectangle",
"locationName": "position",
- "documentation": "Use Selection placement (position) to define the video area in your output frame. The area outside of the rectangle that you specify here is black. If you specify a value here, it will override any value that you specify in the output setting Selection placement (position). If you specify a value here, this will override any AFD values in your input, even if you set Respond to AFD (RespondToAfd) to Respond (RESPOND). If you specify a value here, this will ignore anything that you specify for the setting Scaling Behavior (scalingBehavior)."
+ "documentation": "Use Selection placement to define the video area in your output frame. The area outside of the rectangle that you specify here is black. If you specify a value here, it will override any value that you specify in the output setting Selection placement. If you specify a value here, this will override any AFD values in your input, even if you set Respond to AFD to Respond. If you specify a value here, this will ignore anything that you specify for the setting Scaling Behavior."
},
"ProgramNumber": {
"shape": "__integerMin1Max2147483647",
"locationName": "programNumber",
- "documentation": "Use Program (programNumber) to select a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported. Default is the first program within the transport stream. If the program you specify doesn't exist, the transcoding service will use this default."
+ "documentation": "Use Program to select a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported. Default is the first program within the transport stream. If the program you specify doesn't exist, the transcoding service will use this default."
},
"PsiControl": {
"shape": "InputPsiControl",
"locationName": "psiControl",
- "documentation": "Set PSI control (InputPsiControl) for transport stream inputs to specify which data the demux process to scans. * Ignore PSI - Scan all PIDs for audio and video. * Use PSI - Scan only PSI data."
+ "documentation": "Set PSI control for transport stream inputs to specify which data the demux process to scans.\n* Ignore PSI - Scan all PIDs for audio and video.\n* Use PSI - Scan only PSI data."
},
"SupplementalImps": {
"shape": "__listOf__stringPatternS3ASSETMAPXml",
@@ -7369,12 +7369,12 @@
"TimecodeSource": {
"shape": "InputTimecodeSource",
"locationName": "timecodeSource",
- "documentation": "Use this Timecode source setting, located under the input settings (InputTimecodeSource), to specify how the service counts input video frames. This input frame count affects only the behavior of features that apply to a single input at a time, such as input clipping and synchronizing some captions formats. Choose Embedded (EMBEDDED) to use the timecodes in your input video. Choose Start at zero (ZEROBASED) to start the first frame at zero. Choose Specified start (SPECIFIEDSTART) to start the first frame at the timecode that you specify in the setting Start timecode (timecodeStart). If you don't specify a value for Timecode source, the service will use Embedded by default. For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode."
+ "documentation": "Use this Timecode source setting, located under the input settings, to specify how the service counts input video frames. This input frame count affects only the behavior of features that apply to a single input at a time, such as input clipping and synchronizing some captions formats. Choose Embedded to use the timecodes in your input video. Choose Start at zero to start the first frame at zero. Choose Specified start to start the first frame at the timecode that you specify in the setting Start timecode. If you don't specify a value for Timecode source, the service will use Embedded by default. For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode."
},
"TimecodeStart": {
"shape": "__stringMin11Max11Pattern01D20305D205D",
"locationName": "timecodeStart",
- "documentation": "Specify the timecode that you want the service to use for this input's initial frame. To use this setting, you must set the Timecode source setting, located under the input settings (InputTimecodeSource), to Specified start (SPECIFIEDSTART). For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode."
+ "documentation": "Specify the timecode that you want the service to use for this input's initial frame. To use this setting, you must set the Timecode source setting, located under the input settings, to Specified start. For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode."
},
"VideoGenerator": {
"shape": "InputVideoGenerator",
@@ -7395,19 +7395,19 @@
"EndTimecode": {
"shape": "__stringPattern010920405090509092",
"locationName": "endTimecode",
- "documentation": "Set End timecode (EndTimecode) to the end of the portion of the input you are clipping. The frame corresponding to the End timecode value is included in the clip. Start timecode or End timecode may be left blank, but not both. Use the format HH:MM:SS:FF or HH:MM:SS;FF, where HH is the hour, MM is the minute, SS is the second, and FF is the frame number. When choosing this value, take into account your setting for timecode source under input settings (InputTimecodeSource). For example, if you have embedded timecodes that start at 01:00:00:00 and you want your clip to end six minutes into the video, use 01:06:00:00."
+ "documentation": "Set End timecode to the end of the portion of the input you are clipping. The frame corresponding to the End timecode value is included in the clip. Start timecode or End timecode may be left blank, but not both. Use the format HH:MM:SS:FF or HH:MM:SS;FF, where HH is the hour, MM is the minute, SS is the second, and FF is the frame number. When choosing this value, take into account your setting for timecode source under input settings. For example, if you have embedded timecodes that start at 01:00:00:00 and you want your clip to end six minutes into the video, use 01:06:00:00."
},
"StartTimecode": {
"shape": "__stringPattern010920405090509092",
"locationName": "startTimecode",
- "documentation": "Set Start timecode (StartTimecode) to the beginning of the portion of the input you are clipping. The frame corresponding to the Start timecode value is included in the clip. Start timecode or End timecode may be left blank, but not both. Use the format HH:MM:SS:FF or HH:MM:SS;FF, where HH is the hour, MM is the minute, SS is the second, and FF is the frame number. When choosing this value, take into account your setting for Input timecode source. For example, if you have embedded timecodes that start at 01:00:00:00 and you want your clip to begin five minutes into the video, use 01:05:00:00."
+ "documentation": "Set Start timecode to the beginning of the portion of the input you are clipping. The frame corresponding to the Start timecode value is included in the clip. Start timecode or End timecode may be left blank, but not both. Use the format HH:MM:SS:FF or HH:MM:SS;FF, where HH is the hour, MM is the minute, SS is the second, and FF is the frame number. When choosing this value, take into account your setting for Input timecode source. For example, if you have embedded timecodes that start at 01:00:00:00 and you want your clip to begin five minutes into the video, use 01:05:00:00."
}
},
"documentation": "To transcode only portions of your input, include one input clip for each part of your input that you want in your output. All input clips that you specify will be included in every output of the job. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/assembling-multiple-inputs-and-input-clips.html."
},
"InputDeblockFilter": {
"type": "string",
- "documentation": "Enable Deblock (InputDeblockFilter) to produce smoother motion in the output. Default is disabled. Only manually controllable for MPEG2 and uncompressed video inputs.",
+ "documentation": "Enable Deblock to produce smoother motion in the output. Default is disabled. Only manually controllable for MPEG2 and uncompressed video inputs.",
"enum": [
"ENABLED",
"DISABLED"
@@ -7441,7 +7441,7 @@
},
"InputDenoiseFilter": {
"type": "string",
- "documentation": "Enable Denoise (InputDenoiseFilter) to filter noise from the input. Default is disabled. Only applicable to MPEG2, H.264, H.265, and uncompressed video inputs.",
+ "documentation": "Enable Denoise to filter noise from the input. Default is disabled. Only applicable to MPEG2, H.264, H.265, and uncompressed video inputs.",
"enum": [
"ENABLED",
"DISABLED"
@@ -7466,7 +7466,7 @@
},
"InputPsiControl": {
"type": "string",
- "documentation": "Set PSI control (InputPsiControl) for transport stream inputs to specify which data the demux process to scans. * Ignore PSI - Scan all PIDs for audio and video. * Use PSI - Scan only PSI data.",
+ "documentation": "Set PSI control for transport stream inputs to specify which data the demux process to scans.\n* Ignore PSI - Scan all PIDs for audio and video.\n* Use PSI - Scan only PSI data.",
"enum": [
"IGNORE_PSI",
"USE_PSI"
@@ -7474,7 +7474,7 @@
},
"InputRotate": {
"type": "string",
- "documentation": "Use Rotate (InputRotate) to specify how the service rotates your video. You can choose automatic rotation or specify a rotation. You can specify a clockwise rotation of 0, 90, 180, or 270 degrees. If your input video container is .mov or .mp4 and your input has rotation metadata, you can choose Automatic to have the service rotate your video according to the rotation specified in the metadata. The rotation must be within one degree of 90, 180, or 270 degrees. If the rotation metadata specifies any other rotation, the service will default to no rotation. By default, the service does no rotation, even if your input video has rotation metadata. The service doesn't pass through rotation metadata.",
+ "documentation": "Use Rotate to specify how the service rotates your video. You can choose automatic rotation or specify a rotation. You can specify a clockwise rotation of 0, 90, 180, or 270 degrees. If your input video container is .mov or .mp4 and your input has rotation metadata, you can choose Automatic to have the service rotate your video according to the rotation specified in the metadata. The rotation must be within one degree of 90, 180, or 270 degrees. If the rotation metadata specifies any other rotation, the service will default to no rotation. By default, the service does no rotation, even if your input video has rotation metadata. The service doesn't pass through rotation metadata.",
"enum": [
"DEGREE_0",
"DEGREES_90",
@@ -7485,7 +7485,7 @@
},
"InputSampleRange": {
"type": "string",
- "documentation": "If the sample range metadata in your input video is accurate, or if you don't know about sample range, keep the default value, Follow (FOLLOW), for this setting. When you do, the service automatically detects your input sample range. If your input video has metadata indicating the wrong sample range, specify the accurate sample range here. When you do, MediaConvert ignores any sample range information in the input metadata. Regardless of whether MediaConvert uses the input sample range or the sample range that you specify, MediaConvert uses the sample range for transcoding and also writes it to the output metadata.",
+ "documentation": "If the sample range metadata in your input video is accurate, or if you don't know about sample range, keep the default value, Follow, for this setting. When you do, the service automatically detects your input sample range. If your input video has metadata indicating the wrong sample range, specify the accurate sample range here. When you do, MediaConvert ignores any sample range information in the input metadata. Regardless of whether MediaConvert uses the input sample range or the sample range that you specify, MediaConvert uses the sample range for transcoding and also writes it to the output metadata.",
"enum": [
"FOLLOW",
"FULL_RANGE",
@@ -7494,7 +7494,7 @@
},
"InputScanType": {
"type": "string",
- "documentation": "When you have a progressive segmented frame (PsF) input, use this setting to flag the input as PsF. MediaConvert doesn't automatically detect PsF. Therefore, flagging your input as PsF results in better preservation of video quality when you do deinterlacing and frame rate conversion. If you don't specify, the default value is Auto (AUTO). Auto is the correct setting for all inputs that are not PsF. Don't set this value to PsF when your input is interlaced. Doing so creates horizontal interlacing artifacts.",
+ "documentation": "When you have a progressive segmented frame (PsF) input, use this setting to flag the input as PsF. MediaConvert doesn't automatically detect PsF. Therefore, flagging your input as PsF results in better preservation of video quality when you do deinterlacing and frame rate conversion. If you don't specify, the default value is Auto. Auto is the correct setting for all inputs that are not PsF. Don't set this value to PsF when your input is interlaced. Doing so creates horizontal interlacing artifacts.",
"enum": [
"AUTO",
"PSF"
@@ -7516,12 +7516,12 @@
"AudioSelectorGroups": {
"shape": "__mapOfAudioSelectorGroup",
"locationName": "audioSelectorGroups",
- "documentation": "Use audio selector groups to combine multiple sidecar audio inputs so that you can assign them to a single output audio tab (AudioDescription). Note that, if you're working with embedded audio, it's simpler to assign multiple input tracks into a single audio selector rather than use an audio selector group."
+ "documentation": "Use audio selector groups to combine multiple sidecar audio inputs so that you can assign them to a single output audio tab. Note that, if you're working with embedded audio, it's simpler to assign multiple input tracks into a single audio selector rather than use an audio selector group."
},
"AudioSelectors": {
"shape": "__mapOfAudioSelector",
"locationName": "audioSelectors",
- "documentation": "Use Audio selectors (AudioSelectors) to specify a track or set of tracks from the input that you will use in your outputs. You can use multiple Audio selectors per input."
+ "documentation": "Use Audio selectors to specify a track or set of tracks from the input that you will use in your outputs. You can use multiple Audio selectors per input."
},
"CaptionSelectors": {
"shape": "__mapOfCaptionSelector",
@@ -7531,17 +7531,17 @@
"Crop": {
"shape": "Rectangle",
"locationName": "crop",
- "documentation": "Use Cropping selection (crop) to specify the video area that the service will include in the output video frame. If you specify a value here, it will override any value that you specify in the output setting Cropping selection (crop)."
+ "documentation": "Use Cropping selection to specify the video area that the service will include in the output video frame. If you specify a value here, it will override any value that you specify in the output setting Cropping selection."
},
"DeblockFilter": {
"shape": "InputDeblockFilter",
"locationName": "deblockFilter",
- "documentation": "Enable Deblock (InputDeblockFilter) to produce smoother motion in the output. Default is disabled. Only manually controllable for MPEG2 and uncompressed video inputs."
+ "documentation": "Enable Deblock to produce smoother motion in the output. Default is disabled. Only manually controllable for MPEG2 and uncompressed video inputs."
},
"DenoiseFilter": {
"shape": "InputDenoiseFilter",
"locationName": "denoiseFilter",
- "documentation": "Enable Denoise (InputDenoiseFilter) to filter noise from the input. Default is disabled. Only applicable to MPEG2, H.264, H.265, and uncompressed video inputs."
+ "documentation": "Enable Denoise to filter noise from the input. Default is disabled. Only applicable to MPEG2, H.264, H.265, and uncompressed video inputs."
},
"DolbyVisionMetadataXml": {
"shape": "__stringMin14PatternS3XmlXMLHttpsXmlXML",
@@ -7566,37 +7566,37 @@
"InputClippings": {
"shape": "__listOfInputClipping",
"locationName": "inputClippings",
- "documentation": "(InputClippings) contains sets of start and end times that together specify a portion of the input to be used in the outputs. If you provide only a start time, the clip will be the entire input from that point to the end. If you provide only an end time, it will be the entire input up to that point. When you specify more than one input clip, the transcoding service creates the job outputs by stringing the clips together in the order you specify them."
+ "documentation": "Contains sets of start and end times that together specify a portion of the input to be used in the outputs. If you provide only a start time, the clip will be the entire input from that point to the end. If you provide only an end time, it will be the entire input up to that point. When you specify more than one input clip, the transcoding service creates the job outputs by stringing the clips together in the order you specify them."
},
"InputScanType": {
"shape": "InputScanType",
"locationName": "inputScanType",
- "documentation": "When you have a progressive segmented frame (PsF) input, use this setting to flag the input as PsF. MediaConvert doesn't automatically detect PsF. Therefore, flagging your input as PsF results in better preservation of video quality when you do deinterlacing and frame rate conversion. If you don't specify, the default value is Auto (AUTO). Auto is the correct setting for all inputs that are not PsF. Don't set this value to PsF when your input is interlaced. Doing so creates horizontal interlacing artifacts."
+ "documentation": "When you have a progressive segmented frame (PsF) input, use this setting to flag the input as PsF. MediaConvert doesn't automatically detect PsF. Therefore, flagging your input as PsF results in better preservation of video quality when you do deinterlacing and frame rate conversion. If you don't specify, the default value is Auto. Auto is the correct setting for all inputs that are not PsF. Don't set this value to PsF when your input is interlaced. Doing so creates horizontal interlacing artifacts."
},
"Position": {
"shape": "Rectangle",
"locationName": "position",
- "documentation": "Use Selection placement (position) to define the video area in your output frame. The area outside of the rectangle that you specify here is black. If you specify a value here, it will override any value that you specify in the output setting Selection placement (position). If you specify a value here, this will override any AFD values in your input, even if you set Respond to AFD (RespondToAfd) to Respond (RESPOND). If you specify a value here, this will ignore anything that you specify for the setting Scaling Behavior (scalingBehavior)."
+ "documentation": "Use Selection placement to define the video area in your output frame. The area outside of the rectangle that you specify here is black. If you specify a value here, it will override any value that you specify in the output setting Selection placement. If you specify a value here, this will override any AFD values in your input, even if you set Respond to AFD to Respond. If you specify a value here, this will ignore anything that you specify for the setting Scaling Behavior."
},
"ProgramNumber": {
"shape": "__integerMin1Max2147483647",
"locationName": "programNumber",
- "documentation": "Use Program (programNumber) to select a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported. Default is the first program within the transport stream. If the program you specify doesn't exist, the transcoding service will use this default."
+ "documentation": "Use Program to select a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported. Default is the first program within the transport stream. If the program you specify doesn't exist, the transcoding service will use this default."
},
"PsiControl": {
"shape": "InputPsiControl",
"locationName": "psiControl",
- "documentation": "Set PSI control (InputPsiControl) for transport stream inputs to specify which data the demux process to scans. * Ignore PSI - Scan all PIDs for audio and video. * Use PSI - Scan only PSI data."
+ "documentation": "Set PSI control for transport stream inputs to specify which data the demux process to scans.\n* Ignore PSI - Scan all PIDs for audio and video.\n* Use PSI - Scan only PSI data."
},
"TimecodeSource": {
"shape": "InputTimecodeSource",
"locationName": "timecodeSource",
- "documentation": "Use this Timecode source setting, located under the input settings (InputTimecodeSource), to specify how the service counts input video frames. This input frame count affects only the behavior of features that apply to a single input at a time, such as input clipping and synchronizing some captions formats. Choose Embedded (EMBEDDED) to use the timecodes in your input video. Choose Start at zero (ZEROBASED) to start the first frame at zero. Choose Specified start (SPECIFIEDSTART) to start the first frame at the timecode that you specify in the setting Start timecode (timecodeStart). If you don't specify a value for Timecode source, the service will use Embedded by default. For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode."
+ "documentation": "Use this Timecode source setting, located under the input settings, to specify how the service counts input video frames. This input frame count affects only the behavior of features that apply to a single input at a time, such as input clipping and synchronizing some captions formats. Choose Embedded to use the timecodes in your input video. Choose Start at zero to start the first frame at zero. Choose Specified start to start the first frame at the timecode that you specify in the setting Start timecode. If you don't specify a value for Timecode source, the service will use Embedded by default. For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode."
},
"TimecodeStart": {
"shape": "__stringMin11Max11Pattern01D20305D205D",
"locationName": "timecodeStart",
- "documentation": "Specify the timecode that you want the service to use for this input's initial frame. To use this setting, you must set the Timecode source setting, located under the input settings (InputTimecodeSource), to Specified start (SPECIFIEDSTART). For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode."
+ "documentation": "Specify the timecode that you want the service to use for this input's initial frame. To use this setting, you must set the Timecode source setting, located under the input settings, to Specified start. For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode."
},
"VideoSelector": {
"shape": "VideoSelector",
@@ -7608,7 +7608,7 @@
},
"InputTimecodeSource": {
"type": "string",
- "documentation": "Use this Timecode source setting, located under the input settings (InputTimecodeSource), to specify how the service counts input video frames. This input frame count affects only the behavior of features that apply to a single input at a time, such as input clipping and synchronizing some captions formats. Choose Embedded (EMBEDDED) to use the timecodes in your input video. Choose Start at zero (ZEROBASED) to start the first frame at zero. Choose Specified start (SPECIFIEDSTART) to start the first frame at the timecode that you specify in the setting Start timecode (timecodeStart). If you don't specify a value for Timecode source, the service will use Embedded by default. For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode.",
+ "documentation": "Use this Timecode source setting, located under the input settings, to specify how the service counts input video frames. This input frame count affects only the behavior of features that apply to a single input at a time, such as input clipping and synchronizing some captions formats. Choose Embedded to use the timecodes in your input video. Choose Start at zero to start the first frame at zero. Choose Specified start to start the first frame at the timecode that you specify in the setting Start timecode. If you don't specify a value for Timecode source, the service will use Embedded by default. For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode.",
"enum": [
"EMBEDDED",
"ZEROBASED",
@@ -7672,7 +7672,7 @@
"Opacity": {
"shape": "__integerMin0Max100",
"locationName": "opacity",
- "documentation": "Use Opacity (Opacity) to specify how much of the underlying video shows through the inserted image. 0 is transparent and 100 is fully opaque. Default is 50."
+ "documentation": "Use Opacity to specify how much of the underlying video shows through the inserted image. 0 is transparent and 100 is fully opaque. Default is 50."
},
"StartTime": {
"shape": "__stringPattern01D20305D205D",
@@ -7897,7 +7897,7 @@
"Inputs": {
"shape": "__listOfInput",
"locationName": "inputs",
- "documentation": "Use Inputs (inputs) to define source file used in the transcode job. There can be multiple inputs add in a job. These inputs will be concantenated together to create the output."
+ "documentation": "Use Inputs to define source file used in the transcode job. There can be multiple inputs add in a job. These inputs will be concantenated together to create the output."
},
"KantarWatermark": {
"shape": "KantarWatermarkSettings",
@@ -7912,7 +7912,7 @@
"NielsenConfiguration": {
"shape": "NielsenConfiguration",
"locationName": "nielsenConfiguration",
- "documentation": "Settings for your Nielsen configuration. If you don't do Nielsen measurement and analytics, ignore these settings. When you enable Nielsen configuration (nielsenConfiguration), MediaConvert enables PCM to ID3 tagging for all outputs in the job. To enable Nielsen configuration programmatically, include an instance of nielsenConfiguration in your JSON job specification. Even if you don't include any children of nielsenConfiguration, you still enable the setting."
+ "documentation": "Settings for your Nielsen configuration. If you don't do Nielsen measurement and analytics, ignore these settings. When you enable Nielsen configuration, MediaConvert enables PCM to ID3 tagging for all outputs in the job."
},
"NielsenNonLinearWatermark": {
"shape": "NielsenNonLinearWatermarkSettings",
@@ -7922,7 +7922,7 @@
"OutputGroups": {
"shape": "__listOfOutputGroup",
"locationName": "outputGroups",
- "documentation": "(OutputGroups) contains one group of settings for each set of outputs that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, MXF, and no container) are grouped in a single output group as well. Required in (OutputGroups) is a group of settings that apply to the whole group. This required object depends on the value you set for (Type) under (OutputGroups)>(OutputGroupSettings). Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings * CMAF_GROUP_SETTINGS, CmafGroupSettings"
+ "documentation": "Contains one group of settings for each set of outputs that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, MXF, and no container) are grouped in a single output group as well. Required in is a group of settings that apply to the whole group. This required object depends on the value you set for Type. Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings * CMAF_GROUP_SETTINGS, CmafGroupSettings"
},
"TimecodeConfig": {
"shape": "TimecodeConfig",
@@ -7932,7 +7932,7 @@
"TimedMetadataInsertion": {
"shape": "TimedMetadataInsertion",
"locationName": "timedMetadataInsertion",
- "documentation": "Insert user-defined custom ID3 metadata (id3) at timecodes (timecode) that you specify. In each output that you want to include this metadata, you must set ID3 metadata (timedMetadata) to Passthrough (PASSTHROUGH)."
+ "documentation": "Insert user-defined custom ID3 metadata at timecodes that you specify. In each output that you want to include this metadata, you must set ID3 metadata to Passthrough."
}
},
"documentation": "JobSettings contains all the transcode settings for a job."
@@ -8058,7 +8058,7 @@
"Inputs": {
"shape": "__listOfInputTemplate",
"locationName": "inputs",
- "documentation": "Use Inputs (inputs) to define the source file used in the transcode job. There can only be one input in a job template. Using the API, you can include multiple inputs when referencing a job template."
+ "documentation": "Use Inputs to define the source file used in the transcode job. There can only be one input in a job template. Using the API, you can include multiple inputs when referencing a job template."
},
"KantarWatermark": {
"shape": "KantarWatermarkSettings",
@@ -8073,7 +8073,7 @@
"NielsenConfiguration": {
"shape": "NielsenConfiguration",
"locationName": "nielsenConfiguration",
- "documentation": "Settings for your Nielsen configuration. If you don't do Nielsen measurement and analytics, ignore these settings. When you enable Nielsen configuration (nielsenConfiguration), MediaConvert enables PCM to ID3 tagging for all outputs in the job. To enable Nielsen configuration programmatically, include an instance of nielsenConfiguration in your JSON job specification. Even if you don't include any children of nielsenConfiguration, you still enable the setting."
+ "documentation": "Settings for your Nielsen configuration. If you don't do Nielsen measurement and analytics, ignore these settings. When you enable Nielsen configuration, MediaConvert enables PCM to ID3 tagging for all outputs in the job."
},
"NielsenNonLinearWatermark": {
"shape": "NielsenNonLinearWatermarkSettings",
@@ -8083,7 +8083,7 @@
"OutputGroups": {
"shape": "__listOfOutputGroup",
"locationName": "outputGroups",
- "documentation": "(OutputGroups) contains one group of settings for each set of outputs that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, MXF, and no container) are grouped in a single output group as well. Required in (OutputGroups) is a group of settings that apply to the whole group. This required object depends on the value you set for (Type) under (OutputGroups)>(OutputGroupSettings). Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings * CMAF_GROUP_SETTINGS, CmafGroupSettings"
+ "documentation": "Contains one group of settings for each set of outputs that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, MXF, and no container) are grouped in a single output group as well. Required in is a group of settings that apply to the whole group. This required object depends on the value you set for Type. Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings * CMAF_GROUP_SETTINGS, CmafGroupSettings"
},
"TimecodeConfig": {
"shape": "TimecodeConfig",
@@ -8093,7 +8093,7 @@
"TimedMetadataInsertion": {
"shape": "TimedMetadataInsertion",
"locationName": "timedMetadataInsertion",
- "documentation": "Insert user-defined custom ID3 metadata (id3) at timecodes (timecode) that you specify. In each output that you want to include this metadata, you must set ID3 metadata (timedMetadata) to Passthrough (PASSTHROUGH)."
+ "documentation": "Insert user-defined custom ID3 metadata at timecodes that you specify. In each output that you want to include this metadata, you must set ID3 metadata to Passthrough."
}
},
"documentation": "JobTemplateSettings contains all the transcode settings saved in the template that will be applied to jobs created from it."
@@ -8595,7 +8595,7 @@
},
"M2tsAudioDuration": {
"type": "string",
- "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration (MATCH_VIDEO_DURATION). In all other cases, keep the default value, Default codec duration (DEFAULT_CODEC_DURATION). When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.",
+ "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.",
"enum": [
"DEFAULT_CODEC_DURATION",
"MATCH_VIDEO_DURATION"
@@ -8611,7 +8611,7 @@
},
"M2tsDataPtsControl": {
"type": "string",
- "documentation": "If you select ALIGN_TO_VIDEO, MediaConvert writes captions and data packets with Presentation Timestamp (PTS) values greater than or equal to the first video packet PTS (MediaConvert drops captions and data packets with lesser PTS values). Keep the default value (AUTO) to allow all PTS values.",
+ "documentation": "If you select ALIGN_TO_VIDEO, MediaConvert writes captions and data packets with Presentation Timestamp (PTS) values greater than or equal to the first video packet PTS (MediaConvert drops captions and data packets with lesser PTS values). Keep the default value to allow all PTS values.",
"enum": [
"AUTO",
"ALIGN_TO_VIDEO"
@@ -8643,7 +8643,7 @@
},
"M2tsForceTsVideoEbpOrder": {
"type": "string",
- "documentation": "Keep the default value (DEFAULT) unless you know that your audio EBP markers are incorrectly appearing before your video EBP markers. To correct this problem, set this value to Force (FORCE).",
+ "documentation": "Keep the default value unless you know that your audio EBP markers are incorrectly appearing before your video EBP markers. To correct this problem, set this value to Force.",
"enum": [
"FORCE",
"DEFAULT"
@@ -8690,11 +8690,11 @@
"documentation": "Packet Identifier (PID) of the SCTE-35 stream in the transport stream generated by ESAM."
}
},
- "documentation": "Settings for SCTE-35 signals from ESAM. Include this in your job settings to put SCTE-35 markers in your HLS and transport stream outputs at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML (sccXml)."
+ "documentation": "Settings for SCTE-35 signals from ESAM. Include this in your job settings to put SCTE-35 markers in your HLS and transport stream outputs at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML."
},
"M2tsScte35Source": {
"type": "string",
- "documentation": "For SCTE-35 markers from your input-- Choose Passthrough (PASSTHROUGH) if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None (NONE) if you don't want SCTE-35 markers in this output. For SCTE-35 markers from an ESAM XML document-- Choose None (NONE). Also provide the ESAM XML as a string in the setting Signal processing notification XML (sccXml). Also enable ESAM SCTE-35 (include the property scte35Esam).",
+ "documentation": "For SCTE-35 markers from your input-- Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want SCTE-35 markers in this output. For SCTE-35 markers from an ESAM XML document-- Choose None. Also provide the ESAM XML as a string in the setting Signal processing notification XML. Also enable ESAM SCTE-35 (include the property scte35Esam).",
"enum": [
"PASSTHROUGH",
"NONE"
@@ -8731,7 +8731,7 @@
"AudioDuration": {
"shape": "M2tsAudioDuration",
"locationName": "audioDuration",
- "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration (MATCH_VIDEO_DURATION). In all other cases, keep the default value, Default codec duration (DEFAULT_CODEC_DURATION). When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec."
+ "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec."
},
"AudioFramesPerPes": {
"shape": "__integerMin0Max2147483647",
@@ -8756,17 +8756,17 @@
"DataPTSControl": {
"shape": "M2tsDataPtsControl",
"locationName": "dataPTSControl",
- "documentation": "If you select ALIGN_TO_VIDEO, MediaConvert writes captions and data packets with Presentation Timestamp (PTS) values greater than or equal to the first video packet PTS (MediaConvert drops captions and data packets with lesser PTS values). Keep the default value (AUTO) to allow all PTS values."
+ "documentation": "If you select ALIGN_TO_VIDEO, MediaConvert writes captions and data packets with Presentation Timestamp (PTS) values greater than or equal to the first video packet PTS (MediaConvert drops captions and data packets with lesser PTS values). Keep the default value to allow all PTS values."
},
"DvbNitSettings": {
"shape": "DvbNitSettings",
"locationName": "dvbNitSettings",
- "documentation": "Use these settings to insert a DVB Network Information Table (NIT) in the transport stream of this output. When you work directly in your JSON job specification, include this object only when your job has a transport stream output and the container settings contain the object M2tsSettings."
+ "documentation": "Use these settings to insert a DVB Network Information Table (NIT) in the transport stream of this output."
},
"DvbSdtSettings": {
"shape": "DvbSdtSettings",
"locationName": "dvbSdtSettings",
- "documentation": "Use these settings to insert a DVB Service Description Table (SDT) in the transport stream of this output. When you work directly in your JSON job specification, include this object only when your job has a transport stream output and the container settings contain the object M2tsSettings."
+ "documentation": "Use these settings to insert a DVB Service Description Table (SDT) in the transport stream of this output."
},
"DvbSubPids": {
"shape": "__listOf__integerMin32Max8182",
@@ -8776,7 +8776,7 @@
"DvbTdtSettings": {
"shape": "DvbTdtSettings",
"locationName": "dvbTdtSettings",
- "documentation": "Use these settings to insert a DVB Time and Date Table (TDT) in the transport stream of this output. When you work directly in your JSON job specification, include this object only when your job has a transport stream output and the container settings contain the object M2tsSettings."
+ "documentation": "Use these settings to insert a DVB Time and Date Table (TDT) in the transport stream of this output."
},
"DvbTeletextPid": {
"shape": "__integerMin32Max8182",
@@ -8801,7 +8801,7 @@
"ForceTsVideoEbpOrder": {
"shape": "M2tsForceTsVideoEbpOrder",
"locationName": "forceTsVideoEbpOrder",
- "documentation": "Keep the default value (DEFAULT) unless you know that your audio EBP markers are incorrectly appearing before your video EBP markers. To correct this problem, set this value to Force (FORCE)."
+ "documentation": "Keep the default value unless you know that your audio EBP markers are incorrectly appearing before your video EBP markers. To correct this problem, set this value to Force."
},
"FragmentTime": {
"shape": "__doubleMin0",
@@ -8846,7 +8846,7 @@
"PcrPid": {
"shape": "__integerMin32Max8182",
"locationName": "pcrPid",
- "documentation": "Specify the packet identifier (PID) for the program clock reference (PCR) in this output. If you do not specify a value, the service will use the value for Video PID (VideoPid)."
+ "documentation": "Specify the packet identifier (PID) for the program clock reference (PCR) in this output. If you do not specify a value, the service will use the value for Video PID."
},
"PmtInterval": {
"shape": "__integerMin0Max1000",
@@ -8866,7 +8866,7 @@
"ProgramNumber": {
"shape": "__integerMin0Max65535",
"locationName": "programNumber",
- "documentation": "Use Program number (programNumber) to specify the program number used in the program map table (PMT) for this output. Default is 1. Program numbers and program map tables are parts of MPEG-2 transport stream containers, used for organizing data."
+ "documentation": "Use Program number to specify the program number used in the program map table (PMT) for this output. Default is 1. Program numbers and program map tables are parts of MPEG-2 transport stream containers, used for organizing data."
},
"RateMode": {
"shape": "M2tsRateMode",
@@ -8876,7 +8876,7 @@
"Scte35Esam": {
"shape": "M2tsScte35Esam",
"locationName": "scte35Esam",
- "documentation": "Include this in your job settings to put SCTE-35 markers in your HLS and transport stream outputs at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML (sccXml)."
+ "documentation": "Include this in your job settings to put SCTE-35 markers in your HLS and transport stream outputs at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML."
},
"Scte35Pid": {
"shape": "__integerMin32Max8182",
@@ -8886,7 +8886,7 @@
"Scte35Source": {
"shape": "M2tsScte35Source",
"locationName": "scte35Source",
- "documentation": "For SCTE-35 markers from your input-- Choose Passthrough (PASSTHROUGH) if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None (NONE) if you don't want SCTE-35 markers in this output. For SCTE-35 markers from an ESAM XML document-- Choose None (NONE). Also provide the ESAM XML as a string in the setting Signal processing notification XML (sccXml). Also enable ESAM SCTE-35 (include the property scte35Esam)."
+ "documentation": "For SCTE-35 markers from your input-- Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want SCTE-35 markers in this output. For SCTE-35 markers from an ESAM XML document-- Choose None. Also provide the ESAM XML as a string in the setting Signal processing notification XML. Also enable ESAM SCTE-35 (include the property scte35Esam)."
},
"SegmentationMarkers": {
"shape": "M2tsSegmentationMarkers",
@@ -8919,11 +8919,11 @@
"documentation": "Specify the packet identifier (PID) of the elementary video stream in the transport stream."
}
},
- "documentation": "MPEG-2 TS container settings. These apply to outputs in a File output group when the output's container (ContainerType) is MPEG-2 Transport Stream (M2TS). In these assets, data is organized by the program map table (PMT). Each transport stream program contains subsets of data, including audio, video, and metadata. Each of these subsets of data has a numerical label called a packet identifier (PID). Each transport stream program corresponds to one MediaConvert output. The PMT lists the types of data in a program along with their PID. Downstream systems and players use the program map table to look up the PID for each type of data it accesses and then uses the PIDs to locate specific data within the asset."
+ "documentation": "MPEG-2 TS container settings. These apply to outputs in a File output group when the output's container is MPEG-2 Transport Stream (M2TS). In these assets, data is organized by the program map table (PMT). Each transport stream program contains subsets of data, including audio, video, and metadata. Each of these subsets of data has a numerical label called a packet identifier (PID). Each transport stream program corresponds to one MediaConvert output. The PMT lists the types of data in a program along with their PID. Downstream systems and players use the program map table to look up the PID for each type of data it accesses and then uses the PIDs to locate specific data within the asset."
},
"M3u8AudioDuration": {
"type": "string",
- "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration (MATCH_VIDEO_DURATION). In all other cases, keep the default value, Default codec duration (DEFAULT_CODEC_DURATION). When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.",
+ "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.",
"enum": [
"DEFAULT_CODEC_DURATION",
"MATCH_VIDEO_DURATION"
@@ -8931,7 +8931,7 @@
},
"M3u8DataPtsControl": {
"type": "string",
- "documentation": "If you select ALIGN_TO_VIDEO, MediaConvert writes captions and data packets with Presentation Timestamp (PTS) values greater than or equal to the first video packet PTS (MediaConvert drops captions and data packets with lesser PTS values). Keep the default value (AUTO) to allow all PTS values.",
+ "documentation": "If you select ALIGN_TO_VIDEO, MediaConvert writes captions and data packets with Presentation Timestamp (PTS) values greater than or equal to the first video packet PTS (MediaConvert drops captions and data packets with lesser PTS values). Keep the default value AUTO to allow all PTS values.",
"enum": [
"AUTO",
"ALIGN_TO_VIDEO"
@@ -8955,7 +8955,7 @@
},
"M3u8Scte35Source": {
"type": "string",
- "documentation": "For SCTE-35 markers from your input-- Choose Passthrough (PASSTHROUGH) if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None (NONE) if you don't want SCTE-35 markers in this output. For SCTE-35 markers from an ESAM XML document-- Choose None (NONE) if you don't want manifest conditioning. Choose Passthrough (PASSTHROUGH) and choose Ad markers (adMarkers) if you do want manifest conditioning. In both cases, also provide the ESAM XML as a string in the setting Signal processing notification XML (sccXml).",
+ "documentation": "For SCTE-35 markers from your input-- Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want SCTE-35 markers in this output. For SCTE-35 markers from an ESAM XML document-- Choose None if you don't want manifest conditioning. Choose Passthrough and choose Ad markers if you do want manifest conditioning. In both cases, also provide the ESAM XML as a string in the setting Signal processing notification XML.",
"enum": [
"PASSTHROUGH",
"NONE"
@@ -8967,7 +8967,7 @@
"AudioDuration": {
"shape": "M3u8AudioDuration",
"locationName": "audioDuration",
- "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration (MATCH_VIDEO_DURATION). In all other cases, keep the default value, Default codec duration (DEFAULT_CODEC_DURATION). When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec."
+ "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec."
},
"AudioFramesPerPes": {
"shape": "__integerMin0Max2147483647",
@@ -8982,7 +8982,7 @@
"DataPTSControl": {
"shape": "M3u8DataPtsControl",
"locationName": "dataPTSControl",
- "documentation": "If you select ALIGN_TO_VIDEO, MediaConvert writes captions and data packets with Presentation Timestamp (PTS) values greater than or equal to the first video packet PTS (MediaConvert drops captions and data packets with lesser PTS values). Keep the default value (AUTO) to allow all PTS values."
+ "documentation": "If you select ALIGN_TO_VIDEO, MediaConvert writes captions and data packets with Presentation Timestamp (PTS) values greater than or equal to the first video packet PTS (MediaConvert drops captions and data packets with lesser PTS values). Keep the default value AUTO to allow all PTS values."
},
"MaxPcrInterval": {
"shape": "__integerMin0Max500",
@@ -9037,12 +9037,12 @@
"Scte35Source": {
"shape": "M3u8Scte35Source",
"locationName": "scte35Source",
- "documentation": "For SCTE-35 markers from your input-- Choose Passthrough (PASSTHROUGH) if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None (NONE) if you don't want SCTE-35 markers in this output. For SCTE-35 markers from an ESAM XML document-- Choose None (NONE) if you don't want manifest conditioning. Choose Passthrough (PASSTHROUGH) and choose Ad markers (adMarkers) if you do want manifest conditioning. In both cases, also provide the ESAM XML as a string in the setting Signal processing notification XML (sccXml)."
+ "documentation": "For SCTE-35 markers from your input-- Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want SCTE-35 markers in this output. For SCTE-35 markers from an ESAM XML document-- Choose None if you don't want manifest conditioning. Choose Passthrough and choose Ad markers if you do want manifest conditioning. In both cases, also provide the ESAM XML as a string in the setting Signal processing notification XML."
},
"TimedMetadata": {
"shape": "TimedMetadata",
"locationName": "timedMetadata",
- "documentation": "Set ID3 metadata (timedMetadata) to Passthrough (PASSTHROUGH) to include ID3 metadata in this output. This includes ID3 metadata from the following features: ID3 timestamp period (timedMetadataId3Period), and Custom ID3 metadata inserter (timedMetadataInsertion). To exclude this ID3 metadata in this output: set ID3 metadata to None (NONE) or leave blank."
+ "documentation": "Set ID3 metadata to Passthrough to include ID3 metadata in this output. This includes ID3 metadata from the following features: ID3 timestamp period, and Custom ID3 metadata inserter. To exclude this ID3 metadata in this output: set ID3 metadata to None or leave blank."
},
"TimedMetadataPid": {
"shape": "__integerMin32Max8182",
@@ -9125,7 +9125,7 @@
"StartTime": {
"shape": "__stringMin11Max11Pattern01D20305D205D",
"locationName": "startTime",
- "documentation": "Specify when the motion overlay begins. Use timecode format (HH:MM:SS:FF or HH:MM:SS;FF). Make sure that the timecode you provide here takes into account how you have set up your timecode configuration under both job settings and input settings. The simplest way to do that is to set both to start at 0. If you need to set up your job to follow timecodes embedded in your source that don't start at zero, make sure that you specify a start time that is after the first embedded timecode. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/setting-up-timecode.html Find job-wide and input timecode configuration settings in your JSON job settings specification at settings>timecodeConfig>source and settings>inputs>timecodeSource."
+ "documentation": "Specify when the motion overlay begins. Use timecode format (HH:MM:SS:FF or HH:MM:SS;FF). Make sure that the timecode you provide here takes into account how you have set up your timecode configuration under both job settings and input settings. The simplest way to do that is to set both to start at 0. If you need to set up your job to follow timecodes embedded in your source that don't start at zero, make sure that you specify a start time that is after the first embedded timecode. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/setting-up-timecode.html"
}
},
"documentation": "Overlay motion graphics on top of your video. The motion graphics that you specify here appear on all outputs in all output groups. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/motion-graphic-overlay.html."
@@ -9260,7 +9260,7 @@
"Channels": {
"shape": "__integerMin1Max2",
"locationName": "channels",
- "documentation": "Set Channels to specify the number of channels in this output audio track. Choosing Mono in the console will give you 1 output channel; choosing Stereo will give you 2. In the API, valid values are 1 and 2."
+ "documentation": "Set Channels to specify the number of channels in this output audio track. Choosing Mono in will give you 1 output channel; choosing Stereo will give you 2. In the API, valid values are 1 and 2."
},
"SampleRate": {
"shape": "__integerMin32000Max48000",
@@ -9268,7 +9268,7 @@
"documentation": "Sample rate in hz."
}
},
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value MP2."
+ "documentation": "Required when you set Codec to the value MP2."
},
"Mp3RateControlMode": {
"type": "string",
@@ -9289,7 +9289,7 @@
"Channels": {
"shape": "__integerMin1Max2",
"locationName": "channels",
- "documentation": "Specify the number of channels in this output audio track. Choosing Mono on the console gives you 1 output channel; choosing Stereo gives you 2. In the API, valid values are 1 and 2."
+ "documentation": "Specify the number of channels in this output audio track. Choosing Mono gives you 1 output channel; choosing Stereo gives you 2. In the API, valid values are 1 and 2."
},
"RateControlMode": {
"shape": "Mp3RateControlMode",
@@ -9304,7 +9304,7 @@
"VbrQuality": {
"shape": "__integerMin0Max9",
"locationName": "vbrQuality",
- "documentation": "Required when you set Bitrate control mode (rateControlMode) to VBR. Specify the audio quality of this MP3 output from 0 (highest quality) to 9 (lowest quality)."
+ "documentation": "Required when you set Bitrate control mode to VBR. Specify the audio quality of this MP3 output from 0 (highest quality) to 9 (lowest quality)."
}
},
"documentation": "Required when you set Codec, under AudioDescriptions>CodecSettings, to the value MP3."
@@ -9339,7 +9339,7 @@
"AudioDuration": {
"shape": "CmfcAudioDuration",
"locationName": "audioDuration",
- "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration (MATCH_VIDEO_DURATION). In all other cases, keep the default value, Default codec duration (DEFAULT_CODEC_DURATION). When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec."
+ "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec."
},
"CslgAtom": {
"shape": "Mp4CslgAtom",
@@ -9349,7 +9349,7 @@
"CttsVersion": {
"shape": "__integerMin0Max1",
"locationName": "cttsVersion",
- "documentation": "Ignore this setting unless compliance to the CTTS box version specification matters in your workflow. Specify a value of 1 to set your CTTS box version to 1 and make your output compliant with the specification. When you specify a value of 1, you must also set CSLG atom (cslgAtom) to the value INCLUDE. Keep the default value 0 to set your CTTS box version to 0. This can provide backward compatibility for some players and packagers."
+ "documentation": "Ignore this setting unless compliance to the CTTS box version specification matters in your workflow. Specify a value of 1 to set your CTTS box version to 1 and make your output compliant with the specification. When you specify a value of 1, you must also set CSLG atom to the value INCLUDE. Keep the default value 0 to set your CTTS box version to 0. This can provide backward compatibility for some players and packagers."
},
"FreeSpaceBox": {
"shape": "Mp4FreeSpaceBox",
@@ -9371,7 +9371,7 @@
},
"MpdAccessibilityCaptionHints": {
"type": "string",
- "documentation": "Optional. Choose Include (INCLUDE) to have MediaConvert mark up your DASH manifest with elements for embedded 608 captions. This markup isn't generally required, but some video players require it to discover and play embedded 608 captions. Keep the default value, Exclude (EXCLUDE), to leave these elements out. When you enable this setting, this is the markup that MediaConvert includes in your manifest: ",
+ "documentation": "Optional. Choose Include to have MediaConvert mark up your DASH manifest with elements for embedded 608 captions. This markup isn't generally required, but some video players require it to discover and play embedded 608 captions. Keep the default value, Exclude, to leave these elements out. When you enable this setting, this is the markup that MediaConvert includes in your manifest: ",
"enum": [
"INCLUDE",
"EXCLUDE"
@@ -9379,7 +9379,7 @@
},
"MpdAudioDuration": {
"type": "string",
- "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration (MATCH_VIDEO_DURATION). In all other cases, keep the default value, Default codec duration (DEFAULT_CODEC_DURATION). When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.",
+ "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.",
"enum": [
"DEFAULT_CODEC_DURATION",
"MATCH_VIDEO_DURATION"
@@ -9387,7 +9387,7 @@
},
"MpdCaptionContainerType": {
"type": "string",
- "documentation": "Use this setting only in DASH output groups that include sidecar TTML or IMSC captions. You specify sidecar captions in a separate output from your audio and video. Choose Raw (RAW) for captions in a single XML file in a raw container. Choose Fragmented MPEG-4 (FRAGMENTED_MP4) for captions in XML format contained within fragmented MP4 files. This set of fragmented MP4 files is separate from your video and audio fragmented MP4 files.",
+ "documentation": "Use this setting only in DASH output groups that include sidecar TTML or IMSC captions. You specify sidecar captions in a separate output from your audio and video. Choose Raw for captions in a single XML file in a raw container. Choose Fragmented MPEG-4 for captions in XML format contained within fragmented MP4 files. This set of fragmented MP4 files is separate from your video and audio fragmented MP4 files.",
"enum": [
"RAW",
"FRAGMENTED_MP4"
@@ -9403,7 +9403,7 @@
},
"MpdManifestMetadataSignaling": {
"type": "string",
- "documentation": "To add an InbandEventStream element in your output MPD manifest for each type of event message, set Manifest metadata signaling to Enabled. For ID3 event messages, the InbandEventStream element schemeIdUri will be same value that you specify for ID3 metadata scheme ID URI. For SCTE35 event messages, the InbandEventStream element schemeIdUri will be \"urn:scte:scte35:2013:bin\". To leave these elements out of your output MPD manifest, set Manifest metadata signaling to Disabled. To enable Manifest metadata signaling, you must also set SCTE-35 source to Passthrough, ESAM SCTE-35 to insert, or ID3 metadata (TimedMetadata) to Passthrough.",
+ "documentation": "To add an InbandEventStream element in your output MPD manifest for each type of event message, set Manifest metadata signaling to Enabled. For ID3 event messages, the InbandEventStream element schemeIdUri will be same value that you specify for ID3 metadata scheme ID URI. For SCTE35 event messages, the InbandEventStream element schemeIdUri will be \"urn:scte:scte35:2013:bin\". To leave these elements out of your output MPD manifest, set Manifest metadata signaling to Disabled. To enable Manifest metadata signaling, you must also set SCTE-35 source to Passthrough, ESAM SCTE-35 to insert, or ID3 metadata to Passthrough.",
"enum": [
"ENABLED",
"DISABLED"
@@ -9411,7 +9411,7 @@
},
"MpdScte35Esam": {
"type": "string",
- "documentation": "Use this setting only when you specify SCTE-35 markers from ESAM. Choose INSERT to put SCTE-35 markers in this output at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML (sccXml).",
+ "documentation": "Use this setting only when you specify SCTE-35 markers from ESAM. Choose INSERT to put SCTE-35 markers in this output at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML.",
"enum": [
"INSERT",
"NONE"
@@ -9419,7 +9419,7 @@
},
"MpdScte35Source": {
"type": "string",
- "documentation": "Ignore this setting unless you have SCTE-35 markers in your input video file. Choose Passthrough (PASSTHROUGH) if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None (NONE) if you don't want those SCTE-35 markers in this output.",
+ "documentation": "Ignore this setting unless you have SCTE-35 markers in your input video file. Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want those SCTE-35 markers in this output.",
"enum": [
"PASSTHROUGH",
"NONE"
@@ -9431,17 +9431,17 @@
"AccessibilityCaptionHints": {
"shape": "MpdAccessibilityCaptionHints",
"locationName": "accessibilityCaptionHints",
- "documentation": "Optional. Choose Include (INCLUDE) to have MediaConvert mark up your DASH manifest with elements for embedded 608 captions. This markup isn't generally required, but some video players require it to discover and play embedded 608 captions. Keep the default value, Exclude (EXCLUDE), to leave these elements out. When you enable this setting, this is the markup that MediaConvert includes in your manifest: "
+ "documentation": "Optional. Choose Include to have MediaConvert mark up your DASH manifest with elements for embedded 608 captions. This markup isn't generally required, but some video players require it to discover and play embedded 608 captions. Keep the default value, Exclude, to leave these elements out. When you enable this setting, this is the markup that MediaConvert includes in your manifest: "
},
"AudioDuration": {
"shape": "MpdAudioDuration",
"locationName": "audioDuration",
- "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration (MATCH_VIDEO_DURATION). In all other cases, keep the default value, Default codec duration (DEFAULT_CODEC_DURATION). When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec."
+ "documentation": "Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec."
},
"CaptionContainerType": {
"shape": "MpdCaptionContainerType",
"locationName": "captionContainerType",
- "documentation": "Use this setting only in DASH output groups that include sidecar TTML or IMSC captions. You specify sidecar captions in a separate output from your audio and video. Choose Raw (RAW) for captions in a single XML file in a raw container. Choose Fragmented MPEG-4 (FRAGMENTED_MP4) for captions in XML format contained within fragmented MP4 files. This set of fragmented MP4 files is separate from your video and audio fragmented MP4 files."
+ "documentation": "Use this setting only in DASH output groups that include sidecar TTML or IMSC captions. You specify sidecar captions in a separate output from your audio and video. Choose Raw for captions in a single XML file in a raw container. Choose Fragmented MPEG-4 for captions in XML format contained within fragmented MP4 files. This set of fragmented MP4 files is separate from your video and audio fragmented MP4 files."
},
"KlvMetadata": {
"shape": "MpdKlvMetadata",
@@ -9451,44 +9451,44 @@
"ManifestMetadataSignaling": {
"shape": "MpdManifestMetadataSignaling",
"locationName": "manifestMetadataSignaling",
- "documentation": "To add an InbandEventStream element in your output MPD manifest for each type of event message, set Manifest metadata signaling to Enabled. For ID3 event messages, the InbandEventStream element schemeIdUri will be same value that you specify for ID3 metadata scheme ID URI. For SCTE35 event messages, the InbandEventStream element schemeIdUri will be \"urn:scte:scte35:2013:bin\". To leave these elements out of your output MPD manifest, set Manifest metadata signaling to Disabled. To enable Manifest metadata signaling, you must also set SCTE-35 source to Passthrough, ESAM SCTE-35 to insert, or ID3 metadata (TimedMetadata) to Passthrough."
+ "documentation": "To add an InbandEventStream element in your output MPD manifest for each type of event message, set Manifest metadata signaling to Enabled. For ID3 event messages, the InbandEventStream element schemeIdUri will be same value that you specify for ID3 metadata scheme ID URI. For SCTE35 event messages, the InbandEventStream element schemeIdUri will be \"urn:scte:scte35:2013:bin\". To leave these elements out of your output MPD manifest, set Manifest metadata signaling to Disabled. To enable Manifest metadata signaling, you must also set SCTE-35 source to Passthrough, ESAM SCTE-35 to insert, or ID3 metadata to Passthrough."
},
"Scte35Esam": {
"shape": "MpdScte35Esam",
"locationName": "scte35Esam",
- "documentation": "Use this setting only when you specify SCTE-35 markers from ESAM. Choose INSERT to put SCTE-35 markers in this output at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML (sccXml)."
+ "documentation": "Use this setting only when you specify SCTE-35 markers from ESAM. Choose INSERT to put SCTE-35 markers in this output at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML."
},
"Scte35Source": {
"shape": "MpdScte35Source",
"locationName": "scte35Source",
- "documentation": "Ignore this setting unless you have SCTE-35 markers in your input video file. Choose Passthrough (PASSTHROUGH) if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None (NONE) if you don't want those SCTE-35 markers in this output."
+ "documentation": "Ignore this setting unless you have SCTE-35 markers in your input video file. Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want those SCTE-35 markers in this output."
},
"TimedMetadata": {
"shape": "MpdTimedMetadata",
"locationName": "timedMetadata",
- "documentation": "To include ID3 metadata in this output: Set ID3 metadata (timedMetadata) to Passthrough (PASSTHROUGH). Specify this ID3 metadata in Custom ID3 metadata inserter (timedMetadataInsertion). MediaConvert writes each instance of ID3 metadata in a separate Event Message (eMSG) box. To exclude this ID3 metadata: Set ID3 metadata to None (NONE) or leave blank."
+ "documentation": "To include ID3 metadata in this output: Set ID3 metadata to Passthrough. Specify this ID3 metadata in Custom ID3 metadata inserter. MediaConvert writes each instance of ID3 metadata in a separate Event Message (eMSG) box. To exclude this ID3 metadata: Set ID3 metadata to None or leave blank."
},
"TimedMetadataBoxVersion": {
"shape": "MpdTimedMetadataBoxVersion",
"locationName": "timedMetadataBoxVersion",
- "documentation": "Specify the event message box (eMSG) version for ID3 timed metadata in your output.\nFor more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.3 Syntax.\nLeave blank to use the default value Version 0.\nWhen you specify Version 1, you must also set ID3 metadata (timedMetadata) to Passthrough."
+ "documentation": "Specify the event message box (eMSG) version for ID3 timed metadata in your output.\nFor more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.3 Syntax.\nLeave blank to use the default value Version 0.\nWhen you specify Version 1, you must also set ID3 metadata to Passthrough."
},
"TimedMetadataSchemeIdUri": {
"shape": "__stringMax1000",
"locationName": "timedMetadataSchemeIdUri",
- "documentation": "Specify the event message box (eMSG) scheme ID URI (scheme_id_uri) for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. Leave blank to use the default value: https://aomedia.org/emsg/ID3 When you specify a value for ID3 metadata scheme ID URI, you must also set ID3 metadata (timedMetadata) to Passthrough."
+ "documentation": "Specify the event message box (eMSG) scheme ID URI for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. Leave blank to use the default value: https://aomedia.org/emsg/ID3 When you specify a value for ID3 metadata scheme ID URI, you must also set ID3 metadata to Passthrough."
},
"TimedMetadataValue": {
"shape": "__stringMax1000",
"locationName": "timedMetadataValue",
- "documentation": "Specify the event message box (eMSG) value for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. When you specify a value for ID3 Metadata Value, you must also set ID3 metadata (timedMetadata) to Passthrough."
+ "documentation": "Specify the event message box (eMSG) value for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. When you specify a value for ID3 Metadata Value, you must also set ID3 metadata to Passthrough."
}
},
"documentation": "These settings relate to the fragmented MP4 container for the segments in your DASH outputs."
},
"MpdTimedMetadata": {
"type": "string",
- "documentation": "To include ID3 metadata in this output: Set ID3 metadata (timedMetadata) to Passthrough (PASSTHROUGH). Specify this ID3 metadata in Custom ID3 metadata inserter (timedMetadataInsertion). MediaConvert writes each instance of ID3 metadata in a separate Event Message (eMSG) box. To exclude this ID3 metadata: Set ID3 metadata to None (NONE) or leave blank.",
+ "documentation": "To include ID3 metadata in this output: Set ID3 metadata to Passthrough. Specify this ID3 metadata in Custom ID3 metadata inserter. MediaConvert writes each instance of ID3 metadata in a separate Event Message (eMSG) box. To exclude this ID3 metadata: Set ID3 metadata to None or leave blank.",
"enum": [
"PASSTHROUGH",
"NONE"
@@ -9496,7 +9496,7 @@
},
"MpdTimedMetadataBoxVersion": {
"type": "string",
- "documentation": "Specify the event message box (eMSG) version for ID3 timed metadata in your output.\nFor more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.3 Syntax.\nLeave blank to use the default value Version 0.\nWhen you specify Version 1, you must also set ID3 metadata (timedMetadata) to Passthrough.",
+ "documentation": "Specify the event message box (eMSG) version for ID3 timed metadata in your output.\nFor more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.3 Syntax.\nLeave blank to use the default value Version 0.\nWhen you specify Version 1, you must also set ID3 metadata to Passthrough.",
"enum": [
"VERSION_0",
"VERSION_1"
@@ -9504,7 +9504,7 @@
},
"Mpeg2AdaptiveQuantization": {
"type": "string",
- "documentation": "Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Spatial adaptive quantization (spatialAdaptiveQuantization), and Temporal adaptive quantization (temporalAdaptiveQuantization).",
+ "documentation": "Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Spatial adaptive quantization, and Temporal adaptive quantization.",
"enum": [
"OFF",
"LOW",
@@ -9514,7 +9514,7 @@
},
"Mpeg2CodecLevel": {
"type": "string",
- "documentation": "Use Level (Mpeg2CodecLevel) to set the MPEG-2 level for the video output.",
+ "documentation": "Use Level to set the MPEG-2 level for the video output.",
"enum": [
"AUTO",
"LOW",
@@ -9525,7 +9525,7 @@
},
"Mpeg2CodecProfile": {
"type": "string",
- "documentation": "Use Profile (Mpeg2CodecProfile) to set the MPEG-2 profile for the video output.",
+ "documentation": "Use Profile to set the MPEG-2 profile for the video output.",
"enum": [
"MAIN",
"PROFILE_422"
@@ -9533,7 +9533,7 @@
},
"Mpeg2DynamicSubGop": {
"type": "string",
- "documentation": "Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames (numberBFramesBetweenReferenceFrames).",
+ "documentation": "Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames.",
"enum": [
"ADAPTIVE",
"STATIC"
@@ -9541,7 +9541,7 @@
},
"Mpeg2FramerateControl": {
"type": "string",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator.",
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -9558,7 +9558,7 @@
},
"Mpeg2GopSizeUnits": {
"type": "string",
- "documentation": "Specify the units for GOP size (GopSize). If you don't specify a value here, by default the encoder measures GOP size in frames.",
+ "documentation": "Specify the units for GOP size. If you don't specify a value here, by default the encoder measures GOP size in frames.",
"enum": [
"FRAMES",
"SECONDS"
@@ -9566,7 +9566,7 @@
},
"Mpeg2InterlaceMode": {
"type": "string",
- "documentation": "Choose the scan line type for the output. Keep the default value, Progressive (PROGRESSIVE) to create a progressive output, regardless of the scan type of your input. Use Top field first (TOP_FIELD) or Bottom field first (BOTTOM_FIELD) to create an output that's interlaced with the same field polarity throughout. Use Follow, default top (FOLLOW_TOP_FIELD) or Follow, default bottom (FOLLOW_BOTTOM_FIELD) to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.",
+ "documentation": "Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.",
"enum": [
"PROGRESSIVE",
"TOP_FIELD",
@@ -9577,7 +9577,7 @@
},
"Mpeg2IntraDcPrecision": {
"type": "string",
- "documentation": "Use Intra DC precision (Mpeg2IntraDcPrecision) to set quantization precision for intra-block DC coefficients. If you choose the value auto, the service will automatically select the precision based on the per-frame compression ratio.",
+ "documentation": "Use Intra DC precision to set quantization precision for intra-block DC coefficients. If you choose the value auto, the service will automatically select the precision based on the per-frame compression ratio.",
"enum": [
"AUTO",
"INTRA_DC_PRECISION_8",
@@ -9588,7 +9588,7 @@
},
"Mpeg2ParControl": {
"type": "string",
- "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source (INITIALIZE_FROM_SOURCE), uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. To specify a different PAR by editing the JSON job specification, choose SPECIFIED. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.",
+ "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -9596,7 +9596,7 @@
},
"Mpeg2QualityTuningLevel": {
"type": "string",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.",
+ "documentation": "Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.",
"enum": [
"SINGLE_PASS",
"MULTI_PASS"
@@ -9604,7 +9604,7 @@
},
"Mpeg2RateControlMode": {
"type": "string",
- "documentation": "Use Rate control mode (Mpeg2RateControlMode) to specify whether the bitrate is variable (vbr) or constant (cbr).",
+ "documentation": "Use Rate control mode to specify whether the bitrate is variable (vbr) or constant (cbr).",
"enum": [
"VBR",
"CBR"
@@ -9612,7 +9612,7 @@
},
"Mpeg2ScanTypeConversionMode": {
"type": "string",
- "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing (INTERLACED_OPTIMIZE) to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing (INTERLACED), for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine (telecine) to None (NONE) or Soft (SOFT). You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode (interlaceMode) to a value other than Progressive (PROGRESSIVE).",
+ "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.",
"enum": [
"INTERLACED",
"INTERLACED_OPTIMIZE"
@@ -9632,7 +9632,7 @@
"AdaptiveQuantization": {
"shape": "Mpeg2AdaptiveQuantization",
"locationName": "adaptiveQuantization",
- "documentation": "Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Spatial adaptive quantization (spatialAdaptiveQuantization), and Temporal adaptive quantization (temporalAdaptiveQuantization)."
+ "documentation": "Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Spatial adaptive quantization, and Temporal adaptive quantization."
},
"Bitrate": {
"shape": "__integerMin1000Max288000000",
@@ -9642,22 +9642,22 @@
"CodecLevel": {
"shape": "Mpeg2CodecLevel",
"locationName": "codecLevel",
- "documentation": "Use Level (Mpeg2CodecLevel) to set the MPEG-2 level for the video output."
+ "documentation": "Use Level to set the MPEG-2 level for the video output."
},
"CodecProfile": {
"shape": "Mpeg2CodecProfile",
"locationName": "codecProfile",
- "documentation": "Use Profile (Mpeg2CodecProfile) to set the MPEG-2 profile for the video output."
+ "documentation": "Use Profile to set the MPEG-2 profile for the video output."
},
"DynamicSubGop": {
"shape": "Mpeg2DynamicSubGop",
"locationName": "dynamicSubGop",
- "documentation": "Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames (numberBFramesBetweenReferenceFrames)."
+ "documentation": "Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames."
},
"FramerateControl": {
"shape": "Mpeg2FramerateControl",
"locationName": "framerateControl",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator."
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction."
},
"FramerateConversionAlgorithm": {
"shape": "Mpeg2FramerateConversionAlgorithm",
@@ -9682,12 +9682,12 @@
"GopSize": {
"shape": "__doubleMin0",
"locationName": "gopSize",
- "documentation": "Specify the interval between keyframes, in seconds or frames, for this output. Default: 12 Related settings: When you specify the GOP size in seconds, set GOP mode control (GopSizeUnits) to Specified, seconds (SECONDS). The default value for GOP mode control (GopSizeUnits) is Frames (FRAMES)."
+ "documentation": "Specify the interval between keyframes, in seconds or frames, for this output. Default: 12 Related settings: When you specify the GOP size in seconds, set GOP mode control to Specified, seconds. The default value for GOP mode control is Frames."
},
"GopSizeUnits": {
"shape": "Mpeg2GopSizeUnits",
"locationName": "gopSizeUnits",
- "documentation": "Specify the units for GOP size (GopSize). If you don't specify a value here, by default the encoder measures GOP size in frames."
+ "documentation": "Specify the units for GOP size. If you don't specify a value here, by default the encoder measures GOP size in frames."
},
"HrdBufferFinalFillPercentage": {
"shape": "__integerMin0Max100",
@@ -9707,12 +9707,12 @@
"InterlaceMode": {
"shape": "Mpeg2InterlaceMode",
"locationName": "interlaceMode",
- "documentation": "Choose the scan line type for the output. Keep the default value, Progressive (PROGRESSIVE) to create a progressive output, regardless of the scan type of your input. Use Top field first (TOP_FIELD) or Bottom field first (BOTTOM_FIELD) to create an output that's interlaced with the same field polarity throughout. Use Follow, default top (FOLLOW_TOP_FIELD) or Follow, default bottom (FOLLOW_BOTTOM_FIELD) to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose."
+ "documentation": "Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose."
},
"IntraDcPrecision": {
"shape": "Mpeg2IntraDcPrecision",
"locationName": "intraDcPrecision",
- "documentation": "Use Intra DC precision (Mpeg2IntraDcPrecision) to set quantization precision for intra-block DC coefficients. If you choose the value auto, the service will automatically select the precision based on the per-frame compression ratio."
+ "documentation": "Use Intra DC precision to set quantization precision for intra-block DC coefficients. If you choose the value auto, the service will automatically select the precision based on the per-frame compression ratio."
},
"MaxBitrate": {
"shape": "__integerMin1000Max300000000",
@@ -9722,7 +9722,7 @@
"MinIInterval": {
"shape": "__integerMin0Max30",
"locationName": "minIInterval",
- "documentation": "Use this setting only when you also enable Scene change detection (SceneChangeDetect). This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. When you specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval (minIInterval) to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs."
+ "documentation": "Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. When you specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs."
},
"NumberBFramesBetweenReferenceFrames": {
"shape": "__integerMin0Max7",
@@ -9732,32 +9732,32 @@
"ParControl": {
"shape": "Mpeg2ParControl",
"locationName": "parControl",
- "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source (INITIALIZE_FROM_SOURCE), uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. To specify a different PAR by editing the JSON job specification, choose SPECIFIED. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings."
+ "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings."
},
"ParDenominator": {
"shape": "__integerMin1Max2147483647",
"locationName": "parDenominator",
- "documentation": "Required when you set Pixel aspect ratio (parControl) to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33."
+ "documentation": "Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33."
},
"ParNumerator": {
"shape": "__integerMin1Max2147483647",
"locationName": "parNumerator",
- "documentation": "Required when you set Pixel aspect ratio (parControl) to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40."
+ "documentation": "Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40."
},
"QualityTuningLevel": {
"shape": "Mpeg2QualityTuningLevel",
"locationName": "qualityTuningLevel",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding."
+ "documentation": "Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding."
},
"RateControlMode": {
"shape": "Mpeg2RateControlMode",
"locationName": "rateControlMode",
- "documentation": "Use Rate control mode (Mpeg2RateControlMode) to specify whether the bitrate is variable (vbr) or constant (cbr)."
+ "documentation": "Use Rate control mode to specify whether the bitrate is variable (vbr) or constant (cbr)."
},
"ScanTypeConversionMode": {
"shape": "Mpeg2ScanTypeConversionMode",
"locationName": "scanTypeConversionMode",
- "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing (INTERLACED_OPTIMIZE) to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing (INTERLACED), for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine (telecine) to None (NONE) or Soft (SOFT). You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode (interlaceMode) to a value other than Progressive (PROGRESSIVE)."
+ "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive."
},
"SceneChangeDetect": {
"shape": "Mpeg2SceneChangeDetect",
@@ -9767,39 +9767,39 @@
"SlowPal": {
"shape": "Mpeg2SlowPal",
"locationName": "slowPal",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1."
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25."
},
"Softness": {
"shape": "__integerMin0Max128",
"locationName": "softness",
- "documentation": "Ignore this setting unless you need to comply with a specification that requires a specific value. If you don't have a specification requirement, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness (sharpness) or by enabling a noise reducer filter (noiseReducerFilter). The Softness (softness) setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, to use the AWS Elemental default matrices. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video."
+ "documentation": "Ignore this setting unless you need to comply with a specification that requires a specific value. If you don't have a specification requirement, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness or by enabling a noise reducer filter. The Softness setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, to use the AWS Elemental default matrices. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video."
},
"SpatialAdaptiveQuantization": {
"shape": "Mpeg2SpatialAdaptiveQuantization",
"locationName": "spatialAdaptiveQuantization",
- "documentation": "Keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization (adaptiveQuantization) depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher."
+ "documentation": "Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher."
},
"Syntax": {
"shape": "Mpeg2Syntax",
"locationName": "syntax",
- "documentation": "Specify whether this output's video uses the D10 syntax. Keep the default value to not use the syntax. Related settings: When you choose D10 (D_10) for your MXF profile (profile), you must also set this value to D10 (D_10)."
+ "documentation": "Specify whether this output's video uses the D10 syntax. Keep the default value to not use the syntax. Related settings: When you choose D10 for your MXF profile, you must also set this value to D10."
},
"Telecine": {
"shape": "Mpeg2Telecine",
"locationName": "telecine",
- "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine (HARD) produces a 29.97i output. Soft telecine (SOFT) produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None (NONE), MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture."
+ "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine produces a 29.97i output. Soft telecine produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture."
},
"TemporalAdaptiveQuantization": {
"shape": "Mpeg2TemporalAdaptiveQuantization",
"locationName": "temporalAdaptiveQuantization",
- "documentation": "Keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization (adaptiveQuantization)."
+ "documentation": "Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization."
}
},
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value MPEG2."
+ "documentation": "Required when you set Codec to the value MPEG2."
},
"Mpeg2SlowPal": {
"type": "string",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1.",
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.",
"enum": [
"DISABLED",
"ENABLED"
@@ -9807,7 +9807,7 @@
},
"Mpeg2SpatialAdaptiveQuantization": {
"type": "string",
- "documentation": "Keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization (adaptiveQuantization) depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.",
+ "documentation": "Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.",
"enum": [
"DISABLED",
"ENABLED"
@@ -9815,7 +9815,7 @@
},
"Mpeg2Syntax": {
"type": "string",
- "documentation": "Specify whether this output's video uses the D10 syntax. Keep the default value to not use the syntax. Related settings: When you choose D10 (D_10) for your MXF profile (profile), you must also set this value to D10 (D_10).",
+ "documentation": "Specify whether this output's video uses the D10 syntax. Keep the default value to not use the syntax. Related settings: When you choose D10 for your MXF profile, you must also set this value to D10.",
"enum": [
"DEFAULT",
"D_10"
@@ -9823,7 +9823,7 @@
},
"Mpeg2Telecine": {
"type": "string",
- "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine (HARD) produces a 29.97i output. Soft telecine (SOFT) produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None (NONE), MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.",
+ "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine produces a 29.97i output. Soft telecine produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.",
"enum": [
"NONE",
"SOFT",
@@ -9832,7 +9832,7 @@
},
"Mpeg2TemporalAdaptiveQuantization": {
"type": "string",
- "documentation": "Keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization (adaptiveQuantization).",
+ "documentation": "Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.",
"enum": [
"DISABLED",
"ENABLED"
@@ -9871,11 +9871,11 @@
"documentation": "If your output group type is HLS, DASH, or Microsoft Smooth, use these settings when doing DRM encryption with a SPEKE-compliant key provider. If your output group type is CMAF, use the SpekeKeyProviderCmaf settings instead."
}
},
- "documentation": "If you are using DRM, set DRM System (MsSmoothEncryptionSettings) to specify the value SpekeKeyProvider."
+ "documentation": "If you are using DRM, set DRM System to specify the value SpekeKeyProvider."
},
"MsSmoothFragmentLengthControl": {
"type": "string",
- "documentation": "Specify how you want MediaConvert to determine the fragment length. Choose Exact (EXACT) to have the encoder use the exact length that you specify with the setting Fragment length (FragmentLength). This might result in extra I-frames. Choose Multiple of GOP (GOP_MULTIPLE) to have the encoder round up the segment lengths to match the next GOP boundary.",
+ "documentation": "Specify how you want MediaConvert to determine the fragment length. Choose Exact to have the encoder use the exact length that you specify with the setting Fragment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary.",
"enum": [
"EXACT",
"GOP_MULTIPLE"
@@ -9897,7 +9897,7 @@
"Destination": {
"shape": "__stringPatternS3",
"locationName": "destination",
- "documentation": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file."
+ "documentation": "Use Destination to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file."
},
"DestinationSettings": {
"shape": "DestinationSettings",
@@ -9907,29 +9907,29 @@
"Encryption": {
"shape": "MsSmoothEncryptionSettings",
"locationName": "encryption",
- "documentation": "If you are using DRM, set DRM System (MsSmoothEncryptionSettings) to specify the value SpekeKeyProvider."
+ "documentation": "If you are using DRM, set DRM System to specify the value SpekeKeyProvider."
},
"FragmentLength": {
"shape": "__integerMin1Max2147483647",
"locationName": "fragmentLength",
- "documentation": "Specify how you want MediaConvert to determine the fragment length. Choose Exact (EXACT) to have the encoder use the exact length that you specify with the setting Fragment length (FragmentLength). This might result in extra I-frames. Choose Multiple of GOP (GOP_MULTIPLE) to have the encoder round up the segment lengths to match the next GOP boundary."
+ "documentation": "Specify how you want MediaConvert to determine the fragment length. Choose Exact to have the encoder use the exact length that you specify with the setting Fragment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary."
},
"FragmentLengthControl": {
"shape": "MsSmoothFragmentLengthControl",
"locationName": "fragmentLengthControl",
- "documentation": "Specify how you want MediaConvert to determine the fragment length. Choose Exact (EXACT) to have the encoder use the exact length that you specify with the setting Fragment length (FragmentLength). This might result in extra I-frames. Choose Multiple of GOP (GOP_MULTIPLE) to have the encoder round up the segment lengths to match the next GOP boundary."
+ "documentation": "Specify how you want MediaConvert to determine the fragment length. Choose Exact to have the encoder use the exact length that you specify with the setting Fragment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary."
},
"ManifestEncoding": {
"shape": "MsSmoothManifestEncoding",
"locationName": "manifestEncoding",
- "documentation": "Use Manifest encoding (MsSmoothManifestEncoding) to specify the encoding format for the server and client manifest. Valid options are utf8 and utf16."
+ "documentation": "Use Manifest encoding to specify the encoding format for the server and client manifest. Valid options are utf8 and utf16."
}
},
- "documentation": "Settings related to your Microsoft Smooth Streaming output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html. When you work directly in your JSON job specification, include this object and any required children when you set Type, under OutputGroupSettings, to MS_SMOOTH_GROUP_SETTINGS."
+ "documentation": "Settings related to your Microsoft Smooth Streaming output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html."
},
"MsSmoothManifestEncoding": {
"type": "string",
- "documentation": "Use Manifest encoding (MsSmoothManifestEncoding) to specify the encoding format for the server and client manifest. Valid options are utf8 and utf16.",
+ "documentation": "Use Manifest encoding to specify the encoding format for the server and client manifest. Valid options are utf8 and utf16.",
"enum": [
"UTF8",
"UTF16"
@@ -9937,7 +9937,7 @@
},
"MxfAfdSignaling": {
"type": "string",
- "documentation": "Optional. When you have AFD signaling set up in your output video stream, use this setting to choose whether to also include it in the MXF wrapper. Choose Don't copy (NO_COPY) to exclude AFD signaling from the MXF wrapper. Choose Copy from video stream (COPY_FROM_VIDEO) to copy the AFD values from the video stream for this output to the MXF wrapper. Regardless of which option you choose, the AFD values remain in the video stream. Related settings: To set up your output to include or exclude AFD values, see AfdSignaling, under VideoDescription. On the console, find AFD signaling under the output's video encoding settings.",
+ "documentation": "Optional. When you have AFD signaling set up in your output video stream, use this setting to choose whether to also include it in the MXF wrapper. Choose Don't copy to exclude AFD signaling from the MXF wrapper. Choose Copy from video stream to copy the AFD values from the video stream for this output to the MXF wrapper. Regardless of which option you choose, the AFD values remain in the video stream. Related settings: To set up your output to include or exclude AFD values, see AfdSignaling, under VideoDescription. On the console, find AFD signaling under the output's video encoding settings.",
"enum": [
"NO_COPY",
"COPY_FROM_VIDEO"
@@ -9960,7 +9960,7 @@
"AfdSignaling": {
"shape": "MxfAfdSignaling",
"locationName": "afdSignaling",
- "documentation": "Optional. When you have AFD signaling set up in your output video stream, use this setting to choose whether to also include it in the MXF wrapper. Choose Don't copy (NO_COPY) to exclude AFD signaling from the MXF wrapper. Choose Copy from video stream (COPY_FROM_VIDEO) to copy the AFD values from the video stream for this output to the MXF wrapper. Regardless of which option you choose, the AFD values remain in the video stream. Related settings: To set up your output to include or exclude AFD values, see AfdSignaling, under VideoDescription. On the console, find AFD signaling under the output's video encoding settings."
+ "documentation": "Optional. When you have AFD signaling set up in your output video stream, use this setting to choose whether to also include it in the MXF wrapper. Choose Don't copy to exclude AFD signaling from the MXF wrapper. Choose Copy from video stream to copy the AFD values from the video stream for this output to the MXF wrapper. Regardless of which option you choose, the AFD values remain in the video stream. Related settings: To set up your output to include or exclude AFD values, see AfdSignaling, under VideoDescription. On the console, find AFD signaling under the output's video encoding settings."
},
"Profile": {
"shape": "MxfProfile",
@@ -9977,7 +9977,7 @@
},
"MxfXavcDurationMode": {
"type": "string",
- "documentation": "To create an output that complies with the XAVC file format guidelines for interoperability, keep the default value, Drop frames for compliance (DROP_FRAMES_FOR_COMPLIANCE). To include all frames from your input in this output, keep the default setting, Allow any duration (ALLOW_ANY_DURATION). The number of frames that MediaConvert excludes when you set this to Drop frames for compliance depends on the output frame rate and duration.",
+ "documentation": "To create an output that complies with the XAVC file format guidelines for interoperability, keep the default value, Drop frames for compliance. To include all frames from your input in this output, keep the default setting, Allow any duration. The number of frames that MediaConvert excludes when you set this to Drop frames for compliance depends on the output frame rate and duration.",
"enum": [
"ALLOW_ANY_DURATION",
"DROP_FRAMES_FOR_COMPLIANCE"
@@ -9989,12 +9989,12 @@
"DurationMode": {
"shape": "MxfXavcDurationMode",
"locationName": "durationMode",
- "documentation": "To create an output that complies with the XAVC file format guidelines for interoperability, keep the default value, Drop frames for compliance (DROP_FRAMES_FOR_COMPLIANCE). To include all frames from your input in this output, keep the default setting, Allow any duration (ALLOW_ANY_DURATION). The number of frames that MediaConvert excludes when you set this to Drop frames for compliance depends on the output frame rate and duration."
+ "documentation": "To create an output that complies with the XAVC file format guidelines for interoperability, keep the default value, Drop frames for compliance. To include all frames from your input in this output, keep the default setting, Allow any duration. The number of frames that MediaConvert excludes when you set this to Drop frames for compliance depends on the output frame rate and duration."
},
"MaxAncDataSize": {
"shape": "__integerMin0Max2147483647",
"locationName": "maxAncDataSize",
- "documentation": "Specify a value for this setting only for outputs that you set up with one of these two XAVC profiles: XAVC HD Intra CBG (XAVC_HD_INTRA_CBG) or XAVC 4K Intra CBG (XAVC_4K_INTRA_CBG). Specify the amount of space in each frame that the service reserves for ancillary data, such as teletext captions. The default value for this setting is 1492 bytes per frame. This should be sufficient to prevent overflow unless you have multiple pages of teletext captions data. If you have a large amount of teletext data, specify a larger number."
+ "documentation": "Specify a value for this setting only for outputs that you set up with one of these two XAVC profiles: XAVC HD Intra CBG or XAVC 4K Intra CBG. Specify the amount of space in each frame that the service reserves for ancillary data, such as teletext captions. The default value for this setting is 1492 bytes per frame. This should be sufficient to prevent overflow unless you have multiple pages of teletext captions data. If you have a large amount of teletext data, specify a larger number."
}
},
"documentation": "Specify the XAVC profile settings for MXF outputs when you set your MXF profile to XAVC."
@@ -10005,17 +10005,17 @@
"License": {
"shape": "__stringMin1Max100000",
"locationName": "license",
- "documentation": "Use the base64 license string that Nagra provides you. Enter it directly in your JSON job specification or in the console. Required when you include Nagra NexGuard File Marker watermarking (NexGuardWatermarkingSettings) in your job."
+ "documentation": "Use the base64 license string that Nagra provides you. Enter it directly in your JSON job specification or in the console. Required when you include Nagra NexGuard File Marker watermarking in your job."
},
"Payload": {
"shape": "__integerMin0Max4194303",
"locationName": "payload",
- "documentation": "Specify the payload ID that you want associated with this output. Valid values vary depending on your Nagra NexGuard forensic watermarking workflow. Required when you include Nagra NexGuard File Marker watermarking (NexGuardWatermarkingSettings) in your job. For PreRelease Content (NGPR/G2), specify an integer from 1 through 4,194,303. You must generate a unique ID for each asset you watermark, and keep a record of which ID you have assigned to each asset. Neither Nagra nor MediaConvert keep track of the relationship between output files and your IDs. For OTT Streaming, create two adaptive bitrate (ABR) stacks for each asset. Do this by setting up two output groups. For one output group, set the value of Payload ID (payload) to 0 in every output. For the other output group, set Payload ID (payload) to 1 in every output."
+ "documentation": "Specify the payload ID that you want associated with this output. Valid values vary depending on your Nagra NexGuard forensic watermarking workflow. Required when you include Nagra NexGuard File Marker watermarking in your job. For PreRelease Content (NGPR/G2), specify an integer from 1 through 4,194,303. You must generate a unique ID for each asset you watermark, and keep a record of which ID you have assigned to each asset. Neither Nagra nor MediaConvert keep track of the relationship between output files and your IDs. For OTT Streaming, create two adaptive bitrate (ABR) stacks for each asset. Do this by setting up two output groups. For one output group, set the value of Payload ID to 0 in every output. For the other output group, set Payload ID to 1 in every output."
},
"Preset": {
"shape": "__stringMin1Max256",
"locationName": "preset",
- "documentation": "Enter one of the watermarking preset strings that Nagra provides you. Required when you include Nagra NexGuard File Marker watermarking (NexGuardWatermarkingSettings) in your job."
+ "documentation": "Enter one of the watermarking preset strings that Nagra provides you. Required when you include Nagra NexGuard File Marker watermarking in your job."
},
"Strength": {
"shape": "WatermarkingStrength",
@@ -10027,7 +10027,7 @@
},
"NielsenActiveWatermarkProcessType": {
"type": "string",
- "documentation": "Choose the type of Nielsen watermarks that you want in your outputs. When you choose NAES 2 and NW (NAES2_AND_NW), you must provide a value for the setting SID (sourceId). When you choose CBET (CBET), you must provide a value for the setting CSID (cbetSourceId). When you choose NAES 2, NW, and CBET (NAES2_AND_NW_AND_CBET), you must provide values for both of these settings.",
+ "documentation": "Choose the type of Nielsen watermarks that you want in your outputs. When you choose NAES 2 and NW, you must provide a value for the setting SID. When you choose CBET, you must provide a value for the setting CSID. When you choose NAES 2, NW, and CBET, you must provide values for both of these settings.",
"enum": [
"NAES2_AND_NW",
"CBET",
@@ -10045,10 +10045,10 @@
"DistributorId": {
"shape": "__string",
"locationName": "distributorId",
- "documentation": "Use Distributor ID (DistributorID) to specify the distributor ID that is assigned to your organization by Neilsen."
+ "documentation": "Use Distributor ID to specify the distributor ID that is assigned to your organization by Nielsen."
}
},
- "documentation": "Settings for your Nielsen configuration. If you don't do Nielsen measurement and analytics, ignore these settings. When you enable Nielsen configuration (nielsenConfiguration), MediaConvert enables PCM to ID3 tagging for all outputs in the job. To enable Nielsen configuration programmatically, include an instance of nielsenConfiguration in your JSON job specification. Even if you don't include any children of nielsenConfiguration, you still enable the setting."
+ "documentation": "Settings for your Nielsen configuration. If you don't do Nielsen measurement and analytics, ignore these settings. When you enable Nielsen configuration, MediaConvert enables PCM to ID3 tagging for all outputs in the job."
},
"NielsenNonLinearWatermarkSettings": {
"type": "structure",
@@ -10056,12 +10056,12 @@
"ActiveWatermarkProcess": {
"shape": "NielsenActiveWatermarkProcessType",
"locationName": "activeWatermarkProcess",
- "documentation": "Choose the type of Nielsen watermarks that you want in your outputs. When you choose NAES 2 and NW (NAES2_AND_NW), you must provide a value for the setting SID (sourceId). When you choose CBET (CBET), you must provide a value for the setting CSID (cbetSourceId). When you choose NAES 2, NW, and CBET (NAES2_AND_NW_AND_CBET), you must provide values for both of these settings."
+ "documentation": "Choose the type of Nielsen watermarks that you want in your outputs. When you choose NAES 2 and NW, you must provide a value for the setting SID. When you choose CBET, you must provide a value for the setting CSID. When you choose NAES 2, NW, and CBET, you must provide values for both of these settings."
},
"AdiFilename": {
"shape": "__stringPatternS3",
"locationName": "adiFilename",
- "documentation": "Optional. Use this setting when you want the service to include an ADI file in the Nielsen metadata .zip file. To provide an ADI file, store it in Amazon S3 and provide a URL to it here. The URL should be in the following format: S3://bucket/path/ADI-file. For more information about the metadata .zip file, see the setting Metadata destination (metadataDestination)."
+ "documentation": "Optional. Use this setting when you want the service to include an ADI file in the Nielsen metadata .zip file. To provide an ADI file, store it in Amazon S3 and provide a URL to it here. The URL should be in the following format: S3://bucket/path/ADI-file. For more information about the metadata .zip file, see the setting Metadata destination."
},
"AssetId": {
"shape": "__stringMin1Max20",
@@ -10076,7 +10076,7 @@
"CbetSourceId": {
"shape": "__stringPattern0xAFaF0908190908",
"locationName": "cbetSourceId",
- "documentation": "Use the CSID that Nielsen provides to you. This CBET source ID should be unique to your Nielsen account but common to all of your output assets that have CBET watermarking. Required when you choose a value for the setting Watermark types (ActiveWatermarkProcess) that includes CBET."
+ "documentation": "Use the CSID that Nielsen provides to you. This CBET source ID should be unique to your Nielsen account but common to all of your output assets that have CBET watermarking. Required when you choose a value for the setting Watermark types that includes CBET."
},
"EpisodeId": {
"shape": "__stringMin1Max20",
@@ -10086,7 +10086,7 @@
"MetadataDestination": {
"shape": "__stringPatternS3",
"locationName": "metadataDestination",
- "documentation": "Specify the Amazon S3 location where you want MediaConvert to save your Nielsen non-linear metadata .zip file. This Amazon S3 bucket must be in the same Region as the one where you do your MediaConvert transcoding. If you want to include an ADI file in this .zip file, use the setting ADI file (adiFilename) to specify it. MediaConvert delivers the Nielsen metadata .zip files only to your metadata destination Amazon S3 bucket. It doesn't deliver the .zip files to Nielsen. You are responsible for delivering the metadata .zip files to Nielsen."
+ "documentation": "Specify the Amazon S3 location where you want MediaConvert to save your Nielsen non-linear metadata .zip file. This Amazon S3 bucket must be in the same Region as the one where you do your MediaConvert transcoding. If you want to include an ADI file in this .zip file, use the setting ADI file to specify it. MediaConvert delivers the Nielsen metadata .zip files only to your metadata destination Amazon S3 bucket. It doesn't deliver the .zip files to Nielsen. You are responsible for delivering the metadata .zip files to Nielsen."
},
"SourceId": {
"shape": "__integerMin0Max65534",
@@ -10096,7 +10096,7 @@
"SourceWatermarkStatus": {
"shape": "NielsenSourceWatermarkStatusType",
"locationName": "sourceWatermarkStatus",
- "documentation": "Required. Specify whether your source content already contains Nielsen non-linear watermarks. When you set this value to Watermarked (WATERMARKED), the service fails the job. Nielsen requires that you add non-linear watermarking to only clean content that doesn't already have non-linear Nielsen watermarks."
+ "documentation": "Required. Specify whether your source content already contains Nielsen non-linear watermarks. When you set this value to Watermarked, the service fails the job. Nielsen requires that you add non-linear watermarking to only clean content that doesn't already have non-linear Nielsen watermarks."
},
"TicServerUrl": {
"shape": "__stringPatternHttps",
@@ -10106,14 +10106,14 @@
"UniqueTicPerAudioTrack": {
"shape": "NielsenUniqueTicPerAudioTrackType",
"locationName": "uniqueTicPerAudioTrack",
- "documentation": "To create assets that have the same TIC values in each audio track, keep the default value Share TICs (SAME_TICS_PER_TRACK). To create assets that have unique TIC values for each audio track, choose Use unique TICs (RESERVE_UNIQUE_TICS_PER_TRACK)."
+ "documentation": "To create assets that have the same TIC values in each audio track, keep the default value Share TICs. To create assets that have unique TIC values for each audio track, choose Use unique TICs."
}
},
"documentation": "Ignore these settings unless you are using Nielsen non-linear watermarking. Specify the values that MediaConvert uses to generate and place Nielsen watermarks in your output audio. In addition to specifying these values, you also need to set up your cloud TIC server. These settings apply to every output in your job. The MediaConvert implementation is currently with the following Nielsen versions: Nielsen Watermark SDK Version 5.2.1 Nielsen NLM Watermark Engine Version 1.2.7 Nielsen Watermark Authenticator [SID_TIC] Version [5.0.0]"
},
"NielsenSourceWatermarkStatusType": {
"type": "string",
- "documentation": "Required. Specify whether your source content already contains Nielsen non-linear watermarks. When you set this value to Watermarked (WATERMARKED), the service fails the job. Nielsen requires that you add non-linear watermarking to only clean content that doesn't already have non-linear Nielsen watermarks.",
+ "documentation": "Required. Specify whether your source content already contains Nielsen non-linear watermarks. When you set this value to Watermarked, the service fails the job. Nielsen requires that you add non-linear watermarking to only clean content that doesn't already have non-linear Nielsen watermarks.",
"enum": [
"CLEAN",
"WATERMARKED"
@@ -10121,7 +10121,7 @@
},
"NielsenUniqueTicPerAudioTrackType": {
"type": "string",
- "documentation": "To create assets that have the same TIC values in each audio track, keep the default value Share TICs (SAME_TICS_PER_TRACK). To create assets that have unique TIC values for each audio track, choose Use unique TICs (RESERVE_UNIQUE_TICS_PER_TRACK).",
+ "documentation": "To create assets that have the same TIC values in each audio track, keep the default value Share TICs. To create assets that have unique TIC values for each audio track, choose Use unique TICs.",
"enum": [
"RESERVE_UNIQUE_TICS_PER_TRACK",
"SAME_TICS_PER_TRACK"
@@ -10129,7 +10129,7 @@
},
"NoiseFilterPostTemporalSharpening": {
"type": "string",
- "documentation": "When you set Noise reducer (noiseReducer) to Temporal (TEMPORAL), the bandwidth and sharpness of your output is reduced. You can optionally use Post temporal sharpening (postTemporalSharpening) to apply sharpening to the edges of your output. Note that Post temporal sharpening will also make the bandwidth reduction from the Noise reducer smaller. The default behavior, Auto (AUTO), allows the transcoder to determine whether to apply sharpening, depending on your input type and quality. When you set Post temporal sharpening to Enabled (ENABLED), specify how much sharpening is applied using Post temporal sharpening strength (postTemporalSharpeningStrength). Set Post temporal sharpening to Disabled (DISABLED) to not apply sharpening.",
+ "documentation": "When you set Noise reducer to Temporal, the bandwidth and sharpness of your output is reduced. You can optionally use Post temporal sharpening to apply sharpening to the edges of your output. Note that Post temporal sharpening will also make the bandwidth reduction from the Noise reducer smaller. The default behavior, Auto, allows the transcoder to determine whether to apply sharpening, depending on your input type and quality. When you set Post temporal sharpening to Enabled, specify how much sharpening is applied using Post temporal sharpening strength. Set Post temporal sharpening to Disabled to not apply sharpening.",
"enum": [
"DISABLED",
"ENABLED",
@@ -10138,7 +10138,7 @@
},
"NoiseFilterPostTemporalSharpeningStrength": {
"type": "string",
- "documentation": "Use Post temporal sharpening strength (postTemporalSharpeningStrength) to define the amount of sharpening the transcoder applies to your output. Set Post temporal sharpening strength to Low (LOW), Medium (MEDIUM), or High (HIGH) to indicate the amount of sharpening.",
+ "documentation": "Use Post temporal sharpening strength to define the amount of sharpening the transcoder applies to your output. Set Post temporal sharpening strength to Low, Medium, or High to indicate the amount of sharpening.",
"enum": [
"LOW",
"MEDIUM",
@@ -10151,7 +10151,7 @@
"Filter": {
"shape": "NoiseReducerFilter",
"locationName": "filter",
- "documentation": "Use Noise reducer filter (NoiseReducerFilter) to select one of the following spatial image filtering functions. To use this setting, you must also enable Noise reducer (NoiseReducer). * Bilateral preserves edges while reducing noise. * Mean (softest), Gaussian, Lanczos, and Sharpen (sharpest) do convolution filtering. * Conserve does min/max noise reduction. * Spatial does frequency-domain filtering based on JND principles. * Temporal optimizes video quality for complex motion."
+ "documentation": "Use Noise reducer filter to select one of the following spatial image filtering functions. To use this setting, you must also enable Noise reducer. * Bilateral preserves edges while reducing noise. * Mean (softest), Gaussian, Lanczos, and Sharpen (sharpest) do convolution filtering. * Conserve does min/max noise reduction. * Spatial does frequency-domain filtering based on JND principles. * Temporal optimizes video quality for complex motion."
},
"FilterSettings": {
"shape": "NoiseReducerFilterSettings",
@@ -10173,7 +10173,7 @@
},
"NoiseReducerFilter": {
"type": "string",
- "documentation": "Use Noise reducer filter (NoiseReducerFilter) to select one of the following spatial image filtering functions. To use this setting, you must also enable Noise reducer (NoiseReducer). * Bilateral preserves edges while reducing noise. * Mean (softest), Gaussian, Lanczos, and Sharpen (sharpest) do convolution filtering. * Conserve does min/max noise reduction. * Spatial does frequency-domain filtering based on JND principles. * Temporal optimizes video quality for complex motion.",
+ "documentation": "Use Noise reducer filter to select one of the following spatial image filtering functions. To use this setting, you must also enable Noise reducer. * Bilateral preserves edges while reducing noise. * Mean (softest), Gaussian, Lanczos, and Sharpen (sharpest) do convolution filtering. * Conserve does min/max noise reduction. * Spatial does frequency-domain filtering based on JND principles. * Temporal optimizes video quality for complex motion.",
"enum": [
"BILATERAL",
"MEAN",
@@ -10228,12 +10228,12 @@
"PostTemporalSharpening": {
"shape": "NoiseFilterPostTemporalSharpening",
"locationName": "postTemporalSharpening",
- "documentation": "When you set Noise reducer (noiseReducer) to Temporal (TEMPORAL), the bandwidth and sharpness of your output is reduced. You can optionally use Post temporal sharpening (postTemporalSharpening) to apply sharpening to the edges of your output. Note that Post temporal sharpening will also make the bandwidth reduction from the Noise reducer smaller. The default behavior, Auto (AUTO), allows the transcoder to determine whether to apply sharpening, depending on your input type and quality. When you set Post temporal sharpening to Enabled (ENABLED), specify how much sharpening is applied using Post temporal sharpening strength (postTemporalSharpeningStrength). Set Post temporal sharpening to Disabled (DISABLED) to not apply sharpening."
+ "documentation": "When you set Noise reducer to Temporal, the bandwidth and sharpness of your output is reduced. You can optionally use Post temporal sharpening to apply sharpening to the edges of your output. Note that Post temporal sharpening will also make the bandwidth reduction from the Noise reducer smaller. The default behavior, Auto, allows the transcoder to determine whether to apply sharpening, depending on your input type and quality. When you set Post temporal sharpening to Enabled, specify how much sharpening is applied using Post temporal sharpening strength. Set Post temporal sharpening to Disabled to not apply sharpening."
},
"PostTemporalSharpeningStrength": {
"shape": "NoiseFilterPostTemporalSharpeningStrength",
"locationName": "postTemporalSharpeningStrength",
- "documentation": "Use Post temporal sharpening strength (postTemporalSharpeningStrength) to define the amount of sharpening the transcoder applies to your output. Set Post temporal sharpening strength to Low (LOW), Medium (MEDIUM), or High (HIGH) to indicate the amount of sharpening."
+ "documentation": "Use Post temporal sharpening strength to define the amount of sharpening the transcoder applies to your output. Set Post temporal sharpening strength to Low, Medium, or High to indicate the amount of sharpening."
},
"Speed": {
"shape": "__integerMinNegative1Max3",
@@ -10273,7 +10273,7 @@
"Channels": {
"shape": "__integerMin1Max2",
"locationName": "channels",
- "documentation": "Specify the number of channels in this output audio track. Choosing Mono on the console gives you 1 output channel; choosing Stereo gives you 2. In the API, valid values are 1 and 2."
+ "documentation": "Specify the number of channels in this output audio track. Choosing Mono on gives you 1 output channel; choosing Stereo gives you 2. In the API, valid values are 1 and 2."
},
"SampleRate": {
"shape": "__integerMin16000Max48000",
@@ -10297,12 +10297,12 @@
"AudioDescriptions": {
"shape": "__listOfAudioDescription",
"locationName": "audioDescriptions",
- "documentation": "(AudioDescriptions) contains groups of audio encoding settings organized by audio codec. Include one instance of (AudioDescriptions) per output. (AudioDescriptions) can contain multiple groups of encoding settings."
+ "documentation": "Contains groups of audio encoding settings organized by audio codec. Include one instance of per output. Can contain multiple groups of encoding settings."
},
"CaptionDescriptions": {
"shape": "__listOfCaptionDescription",
"locationName": "captionDescriptions",
- "documentation": "(CaptionDescriptions) contains groups of captions settings. For each output that has captions, include one instance of (CaptionDescriptions). (CaptionDescriptions) can contain multiple groups of captions settings."
+ "documentation": "Contains groups of captions settings. For each output that has captions, include one instance of CaptionDescriptions. Can contain multiple groups of captions settings."
},
"ContainerSettings": {
"shape": "ContainerSettings",
@@ -10312,12 +10312,12 @@
"Extension": {
"shape": "__string",
"locationName": "extension",
- "documentation": "Use Extension (Extension) to specify the file extension for outputs in File output groups. If you do not specify a value, the service will use default extensions by container type as follows * MPEG-2 transport stream, m2ts * Quicktime, mov * MXF container, mxf * MPEG-4 container, mp4 * WebM container, webm * No Container, the service will use codec extensions (e.g. AAC, H265, H265, AC3)"
+ "documentation": "Use Extension to specify the file extension for outputs in File output groups. If you do not specify a value, the service will use default extensions by container type as follows * MPEG-2 transport stream, m2ts * Quicktime, mov * MXF container, mxf * MPEG-4 container, mp4 * WebM container, webm * No Container, the service will use codec extensions (e.g. AAC, H265, H265, AC3)"
},
"NameModifier": {
"shape": "__stringMin1",
"locationName": "nameModifier",
- "documentation": "Use Name modifier (NameModifier) to have the service add a string to the end of each output filename. You specify the base filename as part of your destination URI. When you create multiple outputs in the same output group, Name modifier (NameModifier) is required. Name modifier also accepts format identifiers. For DASH ISO outputs, if you use the format identifiers $Number$ or $Time$ in one output, you must use them in the same way in all outputs of the output group."
+ "documentation": "Use Name modifier to have the service add a string to the end of each output filename. You specify the base filename as part of your destination URI. When you create multiple outputs in the same output group, Name modifier is required. Name modifier also accepts format identifiers. For DASH ISO outputs, if you use the format identifiers $Number$ or $Time$ in one output, you must use them in the same way in all outputs of the output group."
},
"OutputSettings": {
"shape": "OutputSettings",
@@ -10327,7 +10327,7 @@
"Preset": {
"shape": "__stringMin0",
"locationName": "preset",
- "documentation": "Use Preset (Preset) to specify a preset for your transcoding settings. Provide the system or custom preset name. You can specify either Preset (Preset) or Container settings (ContainerSettings), but not both."
+ "documentation": "Use Preset to specify a preset for your transcoding settings. Provide the system or custom preset name. You can specify either Preset or Container settings, but not both."
},
"VideoDescription": {
"shape": "VideoDescription",
@@ -10380,7 +10380,7 @@
"CustomName": {
"shape": "__string",
"locationName": "customName",
- "documentation": "Use Custom Group Name (CustomName) to specify a name for the output group. This value is displayed on the console and can make your job settings JSON more human-readable. It does not affect your outputs. Use up to twelve characters that are either letters, numbers, spaces, or underscores."
+ "documentation": "Use Custom Group Name to specify a name for the output group. This value is displayed on the console and can make your job settings JSON more human-readable. It does not affect your outputs. Use up to twelve characters that are either letters, numbers, spaces, or underscores."
},
"Name": {
"shape": "__string",
@@ -10417,27 +10417,27 @@
"CmafGroupSettings": {
"shape": "CmafGroupSettings",
"locationName": "cmafGroupSettings",
- "documentation": "Settings related to your CMAF output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html. When you work directly in your JSON job specification, include this object and any required children when you set Type, under OutputGroupSettings, to CMAF_GROUP_SETTINGS."
+ "documentation": "Settings related to your CMAF output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html."
},
"DashIsoGroupSettings": {
"shape": "DashIsoGroupSettings",
"locationName": "dashIsoGroupSettings",
- "documentation": "Settings related to your DASH output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html. When you work directly in your JSON job specification, include this object and any required children when you set Type, under OutputGroupSettings, to DASH_ISO_GROUP_SETTINGS."
+ "documentation": "Settings related to your DASH output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html."
},
"FileGroupSettings": {
"shape": "FileGroupSettings",
"locationName": "fileGroupSettings",
- "documentation": "Settings related to your File output group. MediaConvert uses this group of settings to generate a single standalone file, rather than a streaming package. When you work directly in your JSON job specification, include this object and any required children when you set Type, under OutputGroupSettings, to FILE_GROUP_SETTINGS."
+ "documentation": "Settings related to your File output group. MediaConvert uses this group of settings to generate a single standalone file, rather than a streaming package."
},
"HlsGroupSettings": {
"shape": "HlsGroupSettings",
"locationName": "hlsGroupSettings",
- "documentation": "Settings related to your HLS output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html. When you work directly in your JSON job specification, include this object and any required children when you set Type, under OutputGroupSettings, to HLS_GROUP_SETTINGS."
+ "documentation": "Settings related to your HLS output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html."
},
"MsSmoothGroupSettings": {
"shape": "MsSmoothGroupSettings",
"locationName": "msSmoothGroupSettings",
- "documentation": "Settings related to your Microsoft Smooth Streaming output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html. When you work directly in your JSON job specification, include this object and any required children when you set Type, under OutputGroupSettings, to MS_SMOOTH_GROUP_SETTINGS."
+ "documentation": "Settings related to your Microsoft Smooth Streaming output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html."
},
"Type": {
"shape": "OutputGroupType",
@@ -10481,7 +10481,7 @@
},
"PadVideo": {
"type": "string",
- "documentation": "Use this setting if your input has video and audio durations that don't align, and your output or player has strict alignment requirements. Examples: Input audio track has a delayed start. Input video track ends before audio ends. When you set Pad video (padVideo) to Black (BLACK), MediaConvert generates black video frames so that output video and audio durations match. Black video frames are added at the beginning or end, depending on your input. To keep the default behavior and not generate black video, set Pad video to Disabled (DISABLED) or leave blank.",
+ "documentation": "Use this setting if your input has video and audio durations that don't align, and your output or player has strict alignment requirements. Examples: Input audio track has a delayed start. Input video track ends before audio ends. When you set Pad video to Black, MediaConvert generates black video frames so that output video and audio durations match. Black video frames are added at the beginning or end, depending on your input. To keep the default behavior and not generate black video, set Pad video to Disabled or leave blank.",
"enum": [
"DISABLED",
"BLACK"
@@ -10584,7 +10584,7 @@
"AudioDescriptions": {
"shape": "__listOfAudioDescription",
"locationName": "audioDescriptions",
- "documentation": "(AudioDescriptions) contains groups of audio encoding settings organized by audio codec. Include one instance of (AudioDescriptions) per output. (AudioDescriptions) can contain multiple groups of encoding settings."
+ "documentation": "Contains groups of audio encoding settings organized by audio codec. Include one instance of per output. Can contain multiple groups of encoding settings."
},
"CaptionDescriptions": {
"shape": "__listOfCaptionDescriptionPreset",
@@ -10614,7 +10614,7 @@
},
"ProresChromaSampling": {
"type": "string",
- "documentation": "This setting applies only to ProRes 4444 and ProRes 4444 XQ outputs that you create from inputs that use 4:4:4 chroma sampling. Set Preserve 4:4:4 sampling (PRESERVE_444_SAMPLING) to allow outputs to also use 4:4:4 chroma sampling. You must specify a value for this setting when your output codec profile supports 4:4:4 chroma sampling. Related Settings: When you set Chroma sampling to Preserve 4:4:4 sampling (PRESERVE_444_SAMPLING), you must choose an output codec profile that supports 4:4:4 chroma sampling. These values for Profile (CodecProfile) support 4:4:4 chroma sampling: Apple ProRes 4444 (APPLE_PRORES_4444) or Apple ProRes 4444 XQ (APPLE_PRORES_4444_XQ). When you set Chroma sampling to Preserve 4:4:4 sampling, you must disable all video preprocessors except for Nexguard file marker (PartnerWatermarking). When you set Chroma sampling to Preserve 4:4:4 sampling and use framerate conversion, you must set Frame rate conversion algorithm (FramerateConversionAlgorithm) to Drop duplicate (DUPLICATE_DROP).",
+ "documentation": "This setting applies only to ProRes 4444 and ProRes 4444 XQ outputs that you create from inputs that use 4:4:4 chroma sampling. Set Preserve 4:4:4 sampling to allow outputs to also use 4:4:4 chroma sampling. You must specify a value for this setting when your output codec profile supports 4:4:4 chroma sampling. Related Settings: For Apple ProRes outputs with 4:4:4 chroma sampling: Choose Preserve 4:4:4 sampling. Use when your input has 4:4:4 chroma sampling and your output codec Profile is Apple ProRes 4444 or 4444 XQ. Note that when you choose Preserve 4:4:4 sampling, you cannot include any of the following Preprocessors: Dolby Vision, HDR10+, or Noise reducer.",
"enum": [
"PRESERVE_444_SAMPLING",
"SUBSAMPLE_TO_422"
@@ -10622,7 +10622,7 @@
},
"ProresCodecProfile": {
"type": "string",
- "documentation": "Use Profile (ProResCodecProfile) to specify the type of Apple ProRes codec to use for this output.",
+ "documentation": "Use Profile to specify the type of Apple ProRes codec to use for this output.",
"enum": [
"APPLE_PRORES_422",
"APPLE_PRORES_422_HQ",
@@ -10634,7 +10634,7 @@
},
"ProresFramerateControl": {
"type": "string",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator.",
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -10651,7 +10651,7 @@
},
"ProresInterlaceMode": {
"type": "string",
- "documentation": "Choose the scan line type for the output. Keep the default value, Progressive (PROGRESSIVE) to create a progressive output, regardless of the scan type of your input. Use Top field first (TOP_FIELD) or Bottom field first (BOTTOM_FIELD) to create an output that's interlaced with the same field polarity throughout. Use Follow, default top (FOLLOW_TOP_FIELD) or Follow, default bottom (FOLLOW_BOTTOM_FIELD) to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.",
+ "documentation": "Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.",
"enum": [
"PROGRESSIVE",
"TOP_FIELD",
@@ -10662,7 +10662,7 @@
},
"ProresParControl": {
"type": "string",
- "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source (INITIALIZE_FROM_SOURCE), uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. To specify a different PAR by editing the JSON job specification, choose SPECIFIED. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.",
+ "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -10670,7 +10670,7 @@
},
"ProresScanTypeConversionMode": {
"type": "string",
- "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing (INTERLACED_OPTIMIZE) to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing (INTERLACED), for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine (telecine) to None (NONE) or Soft (SOFT). You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode (interlaceMode) to a value other than Progressive (PROGRESSIVE).",
+ "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.",
"enum": [
"INTERLACED",
"INTERLACED_OPTIMIZE"
@@ -10682,17 +10682,17 @@
"ChromaSampling": {
"shape": "ProresChromaSampling",
"locationName": "chromaSampling",
- "documentation": "This setting applies only to ProRes 4444 and ProRes 4444 XQ outputs that you create from inputs that use 4:4:4 chroma sampling. Set Preserve 4:4:4 sampling (PRESERVE_444_SAMPLING) to allow outputs to also use 4:4:4 chroma sampling. You must specify a value for this setting when your output codec profile supports 4:4:4 chroma sampling. Related Settings: When you set Chroma sampling to Preserve 4:4:4 sampling (PRESERVE_444_SAMPLING), you must choose an output codec profile that supports 4:4:4 chroma sampling. These values for Profile (CodecProfile) support 4:4:4 chroma sampling: Apple ProRes 4444 (APPLE_PRORES_4444) or Apple ProRes 4444 XQ (APPLE_PRORES_4444_XQ). When you set Chroma sampling to Preserve 4:4:4 sampling, you must disable all video preprocessors except for Nexguard file marker (PartnerWatermarking). When you set Chroma sampling to Preserve 4:4:4 sampling and use framerate conversion, you must set Frame rate conversion algorithm (FramerateConversionAlgorithm) to Drop duplicate (DUPLICATE_DROP)."
+ "documentation": "This setting applies only to ProRes 4444 and ProRes 4444 XQ outputs that you create from inputs that use 4:4:4 chroma sampling. Set Preserve 4:4:4 sampling to allow outputs to also use 4:4:4 chroma sampling. You must specify a value for this setting when your output codec profile supports 4:4:4 chroma sampling. Related Settings: For Apple ProRes outputs with 4:4:4 chroma sampling: Choose Preserve 4:4:4 sampling. Use when your input has 4:4:4 chroma sampling and your output codec Profile is Apple ProRes 4444 or 4444 XQ. Note that when you choose Preserve 4:4:4 sampling, you cannot include any of the following Preprocessors: Dolby Vision, HDR10+, or Noise reducer."
},
"CodecProfile": {
"shape": "ProresCodecProfile",
"locationName": "codecProfile",
- "documentation": "Use Profile (ProResCodecProfile) to specify the type of Apple ProRes codec to use for this output."
+ "documentation": "Use Profile to specify the type of Apple ProRes codec to use for this output."
},
"FramerateControl": {
"shape": "ProresFramerateControl",
"locationName": "framerateControl",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator."
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction."
},
"FramerateConversionAlgorithm": {
"shape": "ProresFramerateConversionAlgorithm",
@@ -10712,44 +10712,44 @@
"InterlaceMode": {
"shape": "ProresInterlaceMode",
"locationName": "interlaceMode",
- "documentation": "Choose the scan line type for the output. Keep the default value, Progressive (PROGRESSIVE) to create a progressive output, regardless of the scan type of your input. Use Top field first (TOP_FIELD) or Bottom field first (BOTTOM_FIELD) to create an output that's interlaced with the same field polarity throughout. Use Follow, default top (FOLLOW_TOP_FIELD) or Follow, default bottom (FOLLOW_BOTTOM_FIELD) to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose."
+ "documentation": "Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose."
},
"ParControl": {
"shape": "ProresParControl",
"locationName": "parControl",
- "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source (INITIALIZE_FROM_SOURCE), uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. To specify a different PAR by editing the JSON job specification, choose SPECIFIED. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings."
+ "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings."
},
"ParDenominator": {
"shape": "__integerMin1Max2147483647",
"locationName": "parDenominator",
- "documentation": "Required when you set Pixel aspect ratio (parControl) to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33."
+ "documentation": "Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33."
},
"ParNumerator": {
"shape": "__integerMin1Max2147483647",
"locationName": "parNumerator",
- "documentation": "Required when you set Pixel aspect ratio (parControl) to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40."
+ "documentation": "Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40."
},
"ScanTypeConversionMode": {
"shape": "ProresScanTypeConversionMode",
"locationName": "scanTypeConversionMode",
- "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing (INTERLACED_OPTIMIZE) to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing (INTERLACED), for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine (telecine) to None (NONE) or Soft (SOFT). You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode (interlaceMode) to a value other than Progressive (PROGRESSIVE)."
+ "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive."
},
"SlowPal": {
"shape": "ProresSlowPal",
"locationName": "slowPal",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1."
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25."
},
"Telecine": {
"shape": "ProresTelecine",
"locationName": "telecine",
- "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine (HARD) to create a smoother picture. When you keep the default value, None (NONE), MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture."
+ "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture."
}
},
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value PRORES."
+ "documentation": "Required when you set Codec to the value PRORES."
},
"ProresSlowPal": {
"type": "string",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1.",
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.",
"enum": [
"DISABLED",
"ENABLED"
@@ -10757,7 +10757,7 @@
},
"ProresTelecine": {
"type": "string",
- "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine (HARD) to create a smoother picture. When you keep the default value, None (NONE), MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.",
+ "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.",
"enum": [
"NONE",
"HARD"
@@ -10919,7 +10919,7 @@
"ChannelMapping": {
"shape": "ChannelMapping",
"locationName": "channelMapping",
- "documentation": "Channel mapping (ChannelMapping) contains the group of fields that hold the remixing value for each channel, in dB. Specify remix values to indicate how much of the content from your input audio channel you want in your output audio channels. Each instance of the InputChannels or InputChannelsFineTune array specifies these values for one output channel. Use one instance of this array for each output channel. In the console, each array corresponds to a column in the graphical depiction of the mapping matrix. The rows of the graphical matrix correspond to input channels. Valid values are within the range from -60 (mute) through 6. A setting of 0 passes the input channel unchanged to the output channel (no attenuation or amplification). Use InputChannels or InputChannelsFineTune to specify your remix values. Don't use both."
+ "documentation": "Channel mapping contains the group of fields that hold the remixing value for each channel, in dB. Specify remix values to indicate how much of the content from your input audio channel you want in your output audio channels. Each instance of the InputChannels or InputChannelsFineTune array specifies these values for one output channel. Use one instance of this array for each output channel. In the console, each array corresponds to a column in the graphical depiction of the mapping matrix. The rows of the graphical matrix correspond to input channels. Valid values are within the range from -60 (mute) through 6. A setting of 0 passes the input channel unchanged to the output channel (no attenuation or amplification). Use InputChannels or InputChannelsFineTune to specify your remix values. Don't use both."
},
"ChannelsIn": {
"shape": "__integerMin1Max64",
@@ -10932,7 +10932,7 @@
"documentation": "Specify the number of channels in this output after remixing. Valid values: 1, 2, 4, 6, 8... 64. (1 and even numbers to 64.) If you are doing both input channel mapping and output channel mapping, the number of output channels in your input mapping must be the same as the number of input channels in your output mapping."
}
},
- "documentation": "Use Manual audio remixing (RemixSettings) to adjust audio levels for each audio channel in each output of your job. With audio remixing, you can output more or fewer audio channels than your input audio source provides."
+ "documentation": "Use Manual audio remixing to adjust audio levels for each audio channel in each output of your job. With audio remixing, you can output more or fewer audio channels than your input audio source provides."
},
"RenewalType": {
"type": "string",
@@ -11038,7 +11038,7 @@
},
"RespondToAfd": {
"type": "string",
- "documentation": "Use Respond to AFD (RespondToAfd) to specify how the service changes the video itself in response to AFD values in the input. * Choose Respond to clip the input video frame according to the AFD value, input display aspect ratio, and output display aspect ratio. * Choose Passthrough to include the input AFD values. Do not choose this when AfdSignaling is set to (NONE). A preferred implementation of this workflow is to set RespondToAfd to (NONE) and set AfdSignaling to (AUTO). * Choose None to remove all input AFD values from this output.",
+ "documentation": "Use Respond to AFD to specify how the service changes the video itself in response to AFD values in the input. * Choose Respond to clip the input video frame according to the AFD value, input display aspect ratio, and output display aspect ratio. * Choose Passthrough to include the input AFD values. Do not choose this when AfdSignaling is set to NONE. A preferred implementation of this workflow is to set RespondToAfd to and set AfdSignaling to AUTO. * Choose None to remove all input AFD values from this output.",
"enum": [
"NONE",
"RESPOND",
@@ -11088,17 +11088,17 @@
"EncryptionType": {
"shape": "S3ServerSideEncryptionType",
"locationName": "encryptionType",
- "documentation": "Specify how you want your data keys managed. AWS uses data keys to encrypt your content. AWS also encrypts the data keys themselves, using a customer master key (CMK), and then stores the encrypted data keys alongside your encrypted content. Use this setting to specify which AWS service manages the CMK. For simplest set up, choose Amazon S3 (SERVER_SIDE_ENCRYPTION_S3). If you want your master key to be managed by AWS Key Management Service (KMS), choose AWS KMS (SERVER_SIDE_ENCRYPTION_KMS). By default, when you choose AWS KMS, KMS uses the AWS managed customer master key (CMK) associated with Amazon S3 to encrypt your data keys. You can optionally choose to specify a different, customer managed CMK. Do so by specifying the Amazon Resource Name (ARN) of the key for the setting KMS ARN (kmsKeyArn)."
+ "documentation": "Specify how you want your data keys managed. AWS uses data keys to encrypt your content. AWS also encrypts the data keys themselves, using a customer master key (CMK), and then stores the encrypted data keys alongside your encrypted content. Use this setting to specify which AWS service manages the CMK. For simplest set up, choose Amazon S3. If you want your master key to be managed by AWS Key Management Service (KMS), choose AWS KMS. By default, when you choose AWS KMS, KMS uses the AWS managed customer master key (CMK) associated with Amazon S3 to encrypt your data keys. You can optionally choose to specify a different, customer managed CMK. Do so by specifying the Amazon Resource Name (ARN) of the key for the setting KMS ARN."
},
"KmsEncryptionContext": {
"shape": "__stringPatternAZaZ0902",
"locationName": "kmsEncryptionContext",
- "documentation": "Optionally, specify the encryption context that you want to use alongside your KMS key. AWS KMS uses this encryption context as additional authenticated data (AAD) to support authenticated encryption. This value must be a base64-encoded UTF-8 string holding JSON which represents a string-string map. To use this setting, you must also set Server-side encryption (S3ServerSideEncryptionType) to AWS KMS (SERVER_SIDE_ENCRYPTION_KMS). For more information about encryption context, see: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#encrypt_context."
+ "documentation": "Optionally, specify the encryption context that you want to use alongside your KMS key. AWS KMS uses this encryption context as additional authenticated data (AAD) to support authenticated encryption. This value must be a base64-encoded UTF-8 string holding JSON which represents a string-string map. To use this setting, you must also set Server-side encryption to AWS KMS. For more information about encryption context, see: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#encrypt_context."
},
"KmsKeyArn": {
"shape": "__stringPatternArnAwsUsGovCnKmsAZ26EastWestCentralNorthSouthEastWest1912D12KeyAFAF098AFAF094AFAF094AFAF094AFAF0912MrkAFAF0932",
"locationName": "kmsKeyArn",
- "documentation": "Optionally, specify the customer master key (CMK) that you want to use to encrypt the data key that AWS uses to encrypt your output content. Enter the Amazon Resource Name (ARN) of the CMK. To use this setting, you must also set Server-side encryption (S3ServerSideEncryptionType) to AWS KMS (SERVER_SIDE_ENCRYPTION_KMS). If you set Server-side encryption to AWS KMS but don't specify a CMK here, AWS uses the AWS managed CMK associated with Amazon S3."
+ "documentation": "Optionally, specify the customer master key (CMK) that you want to use to encrypt the data key that AWS uses to encrypt your output content. Enter the Amazon Resource Name (ARN) of the CMK. To use this setting, you must also set Server-side encryption to AWS KMS. If you set Server-side encryption to AWS KMS but don't specify a CMK here, AWS uses the AWS managed CMK associated with Amazon S3."
}
},
"documentation": "Settings for how your job outputs are encrypted as they are uploaded to Amazon S3."
@@ -11115,7 +11115,7 @@
},
"S3ServerSideEncryptionType": {
"type": "string",
- "documentation": "Specify how you want your data keys managed. AWS uses data keys to encrypt your content. AWS also encrypts the data keys themselves, using a customer master key (CMK), and then stores the encrypted data keys alongside your encrypted content. Use this setting to specify which AWS service manages the CMK. For simplest set up, choose Amazon S3 (SERVER_SIDE_ENCRYPTION_S3). If you want your master key to be managed by AWS Key Management Service (KMS), choose AWS KMS (SERVER_SIDE_ENCRYPTION_KMS). By default, when you choose AWS KMS, KMS uses the AWS managed customer master key (CMK) associated with Amazon S3 to encrypt your data keys. You can optionally choose to specify a different, customer managed CMK. Do so by specifying the Amazon Resource Name (ARN) of the key for the setting KMS ARN (kmsKeyArn).",
+ "documentation": "Specify how you want your data keys managed. AWS uses data keys to encrypt your content. AWS also encrypts the data keys themselves, using a customer master key (CMK), and then stores the encrypted data keys alongside your encrypted content. Use this setting to specify which AWS service manages the CMK. For simplest set up, choose Amazon S3. If you want your master key to be managed by AWS Key Management Service (KMS), choose AWS KMS. By default, when you choose AWS KMS, KMS uses the AWS managed customer master key (CMK) associated with Amazon S3 to encrypt your data keys. You can optionally choose to specify a different, customer managed CMK. Do so by specifying the Amazon Resource Name (ARN) of the key for the setting KMS ARN.",
"enum": [
"SERVER_SIDE_ENCRYPTION_S3",
"SERVER_SIDE_ENCRYPTION_KMS"
@@ -11132,7 +11132,7 @@
},
"ScalingBehavior": {
"type": "string",
- "documentation": "Specify how the service handles outputs that have a different aspect ratio from the input aspect ratio. Choose Stretch to output (STRETCH_TO_OUTPUT) to have the service stretch your video image to fit. Keep the setting Default (DEFAULT) to have the service letterbox your video instead. This setting overrides any value that you specify for the setting Selection placement (position) in this output.",
+ "documentation": "Specify how the service handles outputs that have a different aspect ratio from the input aspect ratio. Choose Stretch to output to have the service stretch your video image to fit. Keep the setting Default to have the service letterbox your video instead. This setting overrides any value that you specify for the setting Selection placement in this output.",
"enum": [
"DEFAULT",
"STRETCH_TO_OUTPUT"
@@ -11140,7 +11140,7 @@
},
"SccDestinationFramerate": {
"type": "string",
- "documentation": "Set Framerate (SccDestinationFramerate) to make sure that the captions and the video are synchronized in the output. Specify a frame rate that matches the frame rate of the associated video. If the video frame rate is 29.97, choose 29.97 dropframe (FRAMERATE_29_97_DROPFRAME) only if the video has video_insertion=true and drop_frame_timecode=true; otherwise, choose 29.97 non-dropframe (FRAMERATE_29_97_NON_DROPFRAME).",
+ "documentation": "Set Framerate to make sure that the captions and the video are synchronized in the output. Specify a frame rate that matches the frame rate of the associated video. If the video frame rate is 29.97, choose 29.97 dropframe only if the video has video_insertion=true and drop_frame_timecode=true; otherwise, choose 29.97 non-dropframe.",
"enum": [
"FRAMERATE_23_97",
"FRAMERATE_24",
@@ -11155,10 +11155,10 @@
"Framerate": {
"shape": "SccDestinationFramerate",
"locationName": "framerate",
- "documentation": "Set Framerate (SccDestinationFramerate) to make sure that the captions and the video are synchronized in the output. Specify a frame rate that matches the frame rate of the associated video. If the video frame rate is 29.97, choose 29.97 dropframe (FRAMERATE_29_97_DROPFRAME) only if the video has video_insertion=true and drop_frame_timecode=true; otherwise, choose 29.97 non-dropframe (FRAMERATE_29_97_NON_DROPFRAME)."
+ "documentation": "Set Framerate to make sure that the captions and the video are synchronized in the output. Specify a frame rate that matches the frame rate of the associated video. If the video frame rate is 29.97, choose 29.97 dropframe only if the video has video_insertion=true and drop_frame_timecode=true; otherwise, choose 29.97 non-dropframe."
}
},
- "documentation": "Settings related to SCC captions. SCC is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/scc-srt-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to SCC."
+ "documentation": "Settings related to SCC captions. SCC is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/scc-srt-output-captions.html."
},
"SimulateReservedQueue": {
"type": "string",
@@ -11231,14 +11231,14 @@
"StylePassthrough": {
"shape": "SrtStylePassthrough",
"locationName": "stylePassthrough",
- "documentation": "Set Style passthrough (StylePassthrough) to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use simplified output captions."
+ "documentation": "Set Style passthrough to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use simplified output captions."
}
},
- "documentation": "Settings related to SRT captions. SRT is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to SRT."
+ "documentation": "Settings related to SRT captions. SRT is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video."
},
"SrtStylePassthrough": {
"type": "string",
- "documentation": "Set Style passthrough (StylePassthrough) to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use simplified output captions.",
+ "documentation": "Set Style passthrough to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use simplified output captions.",
"enum": [
"ENABLED",
"DISABLED"
@@ -11260,7 +11260,7 @@
"StaticKeyValue": {
"shape": "__stringPatternAZaZ0932",
"locationName": "staticKeyValue",
- "documentation": "Relates to DRM implementation. Use a 32-character hexidecimal string to specify Key Value (StaticKeyValue)."
+ "documentation": "Relates to DRM implementation. Use a 32-character hexidecimal string to specify Key Value."
},
"Url": {
"shape": "__string",
@@ -11326,10 +11326,10 @@
"PageTypes": {
"shape": "__listOfTeletextPageType",
"locationName": "pageTypes",
- "documentation": "Specify the page types for this Teletext page. If you don't specify a value here, the service sets the page type to the default value Subtitle (PAGE_TYPE_SUBTITLE). If you pass through the entire set of Teletext data, don't use this field. When you pass through a set of Teletext pages, your output has the same page types as your input."
+ "documentation": "Specify the page types for this Teletext page. If you don't specify a value here, the service sets the page type to the default value Subtitle. If you pass through the entire set of Teletext data, don't use this field. When you pass through a set of Teletext pages, your output has the same page types as your input."
}
},
- "documentation": "Settings related to teletext captions. Set up teletext captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/teletext-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to TELETEXT."
+ "documentation": "Settings related to teletext captions. Set up teletext captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/teletext-output-captions.html."
},
"TeletextPageType": {
"type": "string",
@@ -11348,7 +11348,7 @@
"PageNumber": {
"shape": "__stringMin3Max3Pattern1809aFAF09aEAE",
"locationName": "pageNumber",
- "documentation": "Use Page Number (PageNumber) to specify the three-digit hexadecimal page number that will be used for Teletext captions. Do not use this setting if you are passing through teletext from the input source to output."
+ "documentation": "Use Page Number to specify the three-digit hexadecimal page number that will be used for Teletext captions. Do not use this setting if you are passing through teletext from the input source to output."
}
},
"documentation": "Settings specific to Teletext caption sources, including Page number."
@@ -11359,24 +11359,24 @@
"FontSize": {
"shape": "__integerMin10Max48",
"locationName": "fontSize",
- "documentation": "Use Font Size (FontSize) to set the font size of any burned-in timecode. Valid values are 10, 16, 32, 48."
+ "documentation": "Use Font size to set the font size of any burned-in timecode. Valid values are 10, 16, 32, 48."
},
"Position": {
"shape": "TimecodeBurninPosition",
"locationName": "position",
- "documentation": "Use Position (Position) under under Timecode burn-in (TimecodeBurnIn) to specify the location the burned-in timecode on output video."
+ "documentation": "Use Position under Timecode burn-in to specify the location the burned-in timecode on output video."
},
"Prefix": {
"shape": "__stringPattern",
"locationName": "prefix",
- "documentation": "Use Prefix (Prefix) to place ASCII characters before any burned-in timecode. For example, a prefix of \"EZ-\" will result in the timecode \"EZ-00:00:00:00\". Provide either the characters themselves or the ASCII code equivalents. The supported range of characters is 0x20 through 0x7e. This includes letters, numbers, and all special characters represented on a standard English keyboard."
+ "documentation": "Use Prefix to place ASCII characters before any burned-in timecode. For example, a prefix of \"EZ-\" will result in the timecode \"EZ-00:00:00:00\". Provide either the characters themselves or the ASCII code equivalents. The supported range of characters is 0x20 through 0x7e. This includes letters, numbers, and all special characters represented on a standard English keyboard."
}
},
"documentation": "Settings for burning the output timecode and specified prefix into the output."
},
"TimecodeBurninPosition": {
"type": "string",
- "documentation": "Use Position (Position) under under Timecode burn-in (TimecodeBurnIn) to specify the location the burned-in timecode on output video.",
+ "documentation": "Use Position under Timecode burn-in to specify the location the burned-in timecode on output video.",
"enum": [
"TOP_CENTER",
"TOP_LEFT",
@@ -11395,29 +11395,29 @@
"Anchor": {
"shape": "__stringPattern010920405090509092",
"locationName": "anchor",
- "documentation": "If you use an editing platform that relies on an anchor timecode, use Anchor Timecode (Anchor) to specify a timecode that will match the input video frame to the output video frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF). This setting ignores frame rate conversion. System behavior for Anchor Timecode varies depending on your setting for Source (TimecodeSource). * If Source (TimecodeSource) is set to Specified Start (SPECIFIEDSTART), the first input frame is the specified value in Start Timecode (Start). Anchor Timecode (Anchor) and Start Timecode (Start) are used calculate output timecode. * If Source (TimecodeSource) is set to Start at 0 (ZEROBASED) the first frame is 00:00:00:00. * If Source (TimecodeSource) is set to Embedded (EMBEDDED), the first frame is the timecode value on the first input frame of the input."
+ "documentation": "If you use an editing platform that relies on an anchor timecode, use Anchor Timecode to specify a timecode that will match the input video frame to the output video frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF). This setting ignores frame rate conversion. System behavior for Anchor Timecode varies depending on your setting for Source. * If Source is set to Specified Start, the first input frame is the specified value in Start Timecode. Anchor Timecode and Start Timecode are used calculate output timecode. * If Source is set to Start at 0 the first frame is 00:00:00:00. * If Source is set to Embedded, the first frame is the timecode value on the first input frame of the input."
},
"Source": {
"shape": "TimecodeSource",
"locationName": "source",
- "documentation": "Use Source (TimecodeSource) to set how timecodes are handled within this job. To make sure that your video, audio, captions, and markers are synchronized and that time-based features, such as image inserter, work correctly, choose the Timecode source option that matches your assets. All timecodes are in a 24-hour format with frame number (HH:MM:SS:FF). * Embedded (EMBEDDED) - Use the timecode that is in the input video. If no embedded timecode is in the source, the service will use Start at 0 (ZEROBASED) instead. * Start at 0 (ZEROBASED) - Set the timecode of the initial frame to 00:00:00:00. * Specified Start (SPECIFIEDSTART) - Set the timecode of the initial frame to a value other than zero. You use Start timecode (Start) to provide this value."
+ "documentation": "Use Source to set how timecodes are handled within this job. To make sure that your video, audio, captions, and markers are synchronized and that time-based features, such as image inserter, work correctly, choose the Timecode source option that matches your assets. All timecodes are in a 24-hour format with frame number (HH:MM:SS:FF). * Embedded - Use the timecode that is in the input video. If no embedded timecode is in the source, the service will use Start at 0 instead. * Start at 0 - Set the timecode of the initial frame to 00:00:00:00. * Specified Start - Set the timecode of the initial frame to a value other than zero. You use Start timecode to provide this value."
},
"Start": {
"shape": "__stringPattern010920405090509092",
"locationName": "start",
- "documentation": "Only use when you set Source (TimecodeSource) to Specified start (SPECIFIEDSTART). Use Start timecode (Start) to specify the timecode for the initial frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF)."
+ "documentation": "Only use when you set Source to Specified start. Use Start timecode to specify the timecode for the initial frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF)."
},
"TimestampOffset": {
"shape": "__stringPattern0940191020191209301",
"locationName": "timestampOffset",
- "documentation": "Only applies to outputs that support program-date-time stamp. Use Timestamp offset (TimestampOffset) to overwrite the timecode date without affecting the time and frame number. Provide the new date as a string in the format \"yyyy-mm-dd\". To use Time stamp offset, you must also enable Insert program-date-time (InsertProgramDateTime) in the output settings. For example, if the date part of your timecodes is 2002-1-25 and you want to change it to one year later, set Timestamp offset (TimestampOffset) to 2003-1-25."
+ "documentation": "Only applies to outputs that support program-date-time stamp. Use Timestamp offset to overwrite the timecode date without affecting the time and frame number. Provide the new date as a string in the format \"yyyy-mm-dd\". To use Timestamp offset, you must also enable Insert program-date-time in the output settings. For example, if the date part of your timecodes is 2002-1-25 and you want to change it to one year later, set Timestamp offset to 2003-1-25."
}
},
"documentation": "These settings control how the service handles timecodes throughout the job. These settings don't affect input clipping."
},
"TimecodeSource": {
"type": "string",
- "documentation": "Use Source (TimecodeSource) to set how timecodes are handled within this job. To make sure that your video, audio, captions, and markers are synchronized and that time-based features, such as image inserter, work correctly, choose the Timecode source option that matches your assets. All timecodes are in a 24-hour format with frame number (HH:MM:SS:FF). * Embedded (EMBEDDED) - Use the timecode that is in the input video. If no embedded timecode is in the source, the service will use Start at 0 (ZEROBASED) instead. * Start at 0 (ZEROBASED) - Set the timecode of the initial frame to 00:00:00:00. * Specified Start (SPECIFIEDSTART) - Set the timecode of the initial frame to a value other than zero. You use Start timecode (Start) to provide this value.",
+ "documentation": "Use Source to set how timecodes are handled within this job. To make sure that your video, audio, captions, and markers are synchronized and that time-based features, such as image inserter, work correctly, choose the Timecode source option that matches your assets. All timecodes are in a 24-hour format with frame number (HH:MM:SS:FF). * Embedded - Use the timecode that is in the input video. If no embedded timecode is in the source, the service will use Start at 0 instead. * Start at 0 - Set the timecode of the initial frame to 00:00:00:00. * Specified Start - Set the timecode of the initial frame to a value other than zero. You use Start timecode to provide this value.",
"enum": [
"EMBEDDED",
"ZEROBASED",
@@ -11426,7 +11426,7 @@
},
"TimedMetadata": {
"type": "string",
- "documentation": "Set ID3 metadata (timedMetadata) to Passthrough (PASSTHROUGH) to include ID3 metadata in this output. This includes ID3 metadata from the following features: ID3 timestamp period (timedMetadataId3Period), and Custom ID3 metadata inserter (timedMetadataInsertion). To exclude this ID3 metadata in this output: set ID3 metadata to None (NONE) or leave blank.",
+ "documentation": "Set ID3 metadata to Passthrough to include ID3 metadata in this output. This includes ID3 metadata from the following features: ID3 timestamp period, and Custom ID3 metadata inserter. To exclude this ID3 metadata in this output: set ID3 metadata to None or leave blank.",
"enum": [
"PASSTHROUGH",
"NONE"
@@ -11441,7 +11441,7 @@
"documentation": "Id3Insertions contains the array of Id3Insertion instances."
}
},
- "documentation": "Insert user-defined custom ID3 metadata (id3) at timecodes (timecode) that you specify. In each output that you want to include this metadata, you must set ID3 metadata (timedMetadata) to Passthrough (PASSTHROUGH)."
+ "documentation": "Insert user-defined custom ID3 metadata at timecodes that you specify. In each output that you want to include this metadata, you must set ID3 metadata to Passthrough."
},
"Timing": {
"type": "structure",
@@ -11498,7 +11498,7 @@
"documentation": "Pass through style and position information from a TTML-like input source (TTML, IMSC, SMPTE-TT) to the TTML output."
}
},
- "documentation": "Settings related to TTML captions. TTML is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to TTML."
+ "documentation": "Settings related to TTML captions. TTML is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html."
},
"TtmlStylePassthrough": {
"type": "string",
@@ -11683,7 +11683,7 @@
},
"Vc3Class": {
"type": "string",
- "documentation": "Specify the VC3 class to choose the quality characteristics for this output. VC3 class, together with the settings Framerate (framerateNumerator and framerateDenominator) and Resolution (height and width), determine your output bitrate. For example, say that your video resolution is 1920x1080 and your framerate is 29.97. Then Class 145 (CLASS_145) gives you an output with a bitrate of approximately 145 Mbps and Class 220 (CLASS_220) gives you and output with a bitrate of approximately 220 Mbps. VC3 class also specifies the color bit depth of your output.",
+ "documentation": "Specify the VC3 class to choose the quality characteristics for this output. VC3 class, together with the settings Framerate (framerateNumerator and framerateDenominator) and Resolution (height and width), determine your output bitrate. For example, say that your video resolution is 1920x1080 and your framerate is 29.97. Then Class 145 gives you an output with a bitrate of approximately 145 Mbps and Class 220 gives you and output with a bitrate of approximately 220 Mbps. VC3 class also specifies the color bit depth of your output.",
"enum": [
"CLASS_145_8BIT",
"CLASS_220_8BIT",
@@ -11692,7 +11692,7 @@
},
"Vc3FramerateControl": {
"type": "string",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator.",
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -11717,7 +11717,7 @@
},
"Vc3ScanTypeConversionMode": {
"type": "string",
- "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing (INTERLACED_OPTIMIZE) to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing (INTERLACED), for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine (telecine) to None (NONE) or Soft (SOFT). You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode (interlaceMode) to a value other than Progressive (PROGRESSIVE).",
+ "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.",
"enum": [
"INTERLACED",
"INTERLACED_OPTIMIZE"
@@ -11729,7 +11729,7 @@
"FramerateControl": {
"shape": "Vc3FramerateControl",
"locationName": "framerateControl",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator."
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction."
},
"FramerateConversionAlgorithm": {
"shape": "Vc3FramerateConversionAlgorithm",
@@ -11754,29 +11754,29 @@
"ScanTypeConversionMode": {
"shape": "Vc3ScanTypeConversionMode",
"locationName": "scanTypeConversionMode",
- "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing (INTERLACED_OPTIMIZE) to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing (INTERLACED), for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine (telecine) to None (NONE) or Soft (SOFT). You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode (interlaceMode) to a value other than Progressive (PROGRESSIVE)."
+ "documentation": "Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive."
},
"SlowPal": {
"shape": "Vc3SlowPal",
"locationName": "slowPal",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Framerate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1."
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Framerate to 25."
},
"Telecine": {
"shape": "Vc3Telecine",
"locationName": "telecine",
- "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine (HARD) to create a smoother picture. When you keep the default value, None (NONE), MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture."
+ "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture."
},
"Vc3Class": {
"shape": "Vc3Class",
"locationName": "vc3Class",
- "documentation": "Specify the VC3 class to choose the quality characteristics for this output. VC3 class, together with the settings Framerate (framerateNumerator and framerateDenominator) and Resolution (height and width), determine your output bitrate. For example, say that your video resolution is 1920x1080 and your framerate is 29.97. Then Class 145 (CLASS_145) gives you an output with a bitrate of approximately 145 Mbps and Class 220 (CLASS_220) gives you and output with a bitrate of approximately 220 Mbps. VC3 class also specifies the color bit depth of your output."
+ "documentation": "Specify the VC3 class to choose the quality characteristics for this output. VC3 class, together with the settings Framerate (framerateNumerator and framerateDenominator) and Resolution (height and width), determine your output bitrate. For example, say that your video resolution is 1920x1080 and your framerate is 29.97. Then Class 145 gives you an output with a bitrate of approximately 145 Mbps and Class 220 gives you and output with a bitrate of approximately 220 Mbps. VC3 class also specifies the color bit depth of your output."
}
},
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value VC3"
+ "documentation": "Required when you set Codec to the value VC3"
},
"Vc3SlowPal": {
"type": "string",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Framerate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1.",
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Framerate to 25.",
"enum": [
"DISABLED",
"ENABLED"
@@ -11784,7 +11784,7 @@
},
"Vc3Telecine": {
"type": "string",
- "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine (HARD) to create a smoother picture. When you keep the default value, None (NONE), MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.",
+ "documentation": "When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.",
"enum": [
"NONE",
"HARD"
@@ -11837,12 +11837,12 @@
"FrameCaptureSettings": {
"shape": "FrameCaptureSettings",
"locationName": "frameCaptureSettings",
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value FRAME_CAPTURE."
+ "documentation": "Required when you set Codec to the value FRAME_CAPTURE."
},
"H264Settings": {
"shape": "H264Settings",
"locationName": "h264Settings",
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value H_264."
+ "documentation": "Required when you set Codec to the value H_264."
},
"H265Settings": {
"shape": "H265Settings",
@@ -11852,35 +11852,35 @@
"Mpeg2Settings": {
"shape": "Mpeg2Settings",
"locationName": "mpeg2Settings",
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value MPEG2."
+ "documentation": "Required when you set Codec to the value MPEG2."
},
"ProresSettings": {
"shape": "ProresSettings",
"locationName": "proresSettings",
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value PRORES."
+ "documentation": "Required when you set Codec to the value PRORES."
},
"Vc3Settings": {
"shape": "Vc3Settings",
"locationName": "vc3Settings",
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value VC3"
+ "documentation": "Required when you set Codec to the value VC3"
},
"Vp8Settings": {
"shape": "Vp8Settings",
"locationName": "vp8Settings",
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value VP8."
+ "documentation": "Required when you set Codec to the value VP8."
},
"Vp9Settings": {
"shape": "Vp9Settings",
"locationName": "vp9Settings",
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value VP9."
+ "documentation": "Required when you set Codec to the value VP9."
},
"XavcSettings": {
"shape": "XavcSettings",
"locationName": "xavcSettings",
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value XAVC."
+ "documentation": "Required when you set Codec to the value XAVC."
}
},
- "documentation": "Video codec settings, (CodecSettings) under (VideoDescription), contains the group of settings related to video encoding. The settings in this group vary depending on the value that you choose for Video codec (Codec). For each codec enum that you choose, define the corresponding settings object. The following lists the codec enum, settings object pairs. * AV1, Av1Settings * AVC_INTRA, AvcIntraSettings * FRAME_CAPTURE, FrameCaptureSettings * H_264, H264Settings * H_265, H265Settings * MPEG2, Mpeg2Settings * PRORES, ProresSettings * VC3, Vc3Settings * VP8, Vp8Settings * VP9, Vp9Settings * XAVC, XavcSettings"
+ "documentation": "Video codec settings contains the group of settings related to video encoding. The settings in this group vary depending on the value that you choose for Video codec. For each codec enum that you choose, define the corresponding settings object. The following lists the codec enum, settings object pairs. * AV1, Av1Settings * AVC_INTRA, AvcIntraSettings * FRAME_CAPTURE, FrameCaptureSettings * H_264, H264Settings * H_265, H265Settings * MPEG2, Mpeg2Settings * PRORES, ProresSettings * VC3, Vc3Settings * VP8, Vp8Settings * VP9, Vp9Settings * XAVC, XavcSettings"
},
"VideoDescription": {
"type": "structure",
@@ -11888,7 +11888,7 @@
"AfdSignaling": {
"shape": "AfdSignaling",
"locationName": "afdSignaling",
- "documentation": "This setting only applies to H.264, H.265, and MPEG2 outputs. Use Insert AFD signaling (AfdSignaling) to specify whether the service includes AFD values in the output video data and what those values are. * Choose None to remove all AFD values from this output. * Choose Fixed to ignore input AFD values and instead encode the value specified in the job. * Choose Auto to calculate output AFD values based on the input AFD scaler data."
+ "documentation": "This setting only applies to H.264, H.265, and MPEG2 outputs. Use Insert AFD signaling to specify whether the service includes AFD values in the output video data and what those values are. * Choose None to remove all AFD values from this output. * Choose Fixed to ignore input AFD values and instead encode the value specified in the job. * Choose Auto to calculate output AFD values based on the input AFD scaler data."
},
"AntiAlias": {
"shape": "AntiAlias",
@@ -11898,27 +11898,27 @@
"CodecSettings": {
"shape": "VideoCodecSettings",
"locationName": "codecSettings",
- "documentation": "Video codec settings, (CodecSettings) under (VideoDescription), contains the group of settings related to video encoding. The settings in this group vary depending on the value that you choose for Video codec (Codec). For each codec enum that you choose, define the corresponding settings object. The following lists the codec enum, settings object pairs. * AV1, Av1Settings * AVC_INTRA, AvcIntraSettings * FRAME_CAPTURE, FrameCaptureSettings * H_264, H264Settings * H_265, H265Settings * MPEG2, Mpeg2Settings * PRORES, ProresSettings * VC3, Vc3Settings * VP8, Vp8Settings * VP9, Vp9Settings * XAVC, XavcSettings"
+ "documentation": "Video codec settings contains the group of settings related to video encoding. The settings in this group vary depending on the value that you choose for Video codec. For each codec enum that you choose, define the corresponding settings object. The following lists the codec enum, settings object pairs. * AV1, Av1Settings * AVC_INTRA, AvcIntraSettings * FRAME_CAPTURE, FrameCaptureSettings * H_264, H264Settings * H_265, H265Settings * MPEG2, Mpeg2Settings * PRORES, ProresSettings * VC3, Vc3Settings * VP8, Vp8Settings * VP9, Vp9Settings * XAVC, XavcSettings"
},
"ColorMetadata": {
"shape": "ColorMetadata",
"locationName": "colorMetadata",
- "documentation": "Choose Insert (INSERT) for this setting to include color metadata in this output. Choose Ignore (IGNORE) to exclude color metadata from this output. If you don't specify a value, the service sets this to Insert by default."
+ "documentation": "Choose Insert for this setting to include color metadata in this output. Choose Ignore to exclude color metadata from this output. If you don't specify a value, the service sets this to Insert by default."
},
"Crop": {
"shape": "Rectangle",
"locationName": "crop",
- "documentation": "Use Cropping selection (crop) to specify the video area that the service will include in the output video frame."
+ "documentation": "Use Cropping selection to specify the video area that the service will include in the output video frame."
},
"DropFrameTimecode": {
"shape": "DropFrameTimecode",
"locationName": "dropFrameTimecode",
- "documentation": "Applies only to 29.97 fps outputs. When this feature is enabled, the service will use drop-frame timecode on outputs. If it is not possible to use drop-frame timecode, the system will fall back to non-drop-frame. This setting is enabled by default when Timecode insertion (TimecodeInsertion) is enabled."
+ "documentation": "Applies only to 29.97 fps outputs. When this feature is enabled, the service will use drop-frame timecode on outputs. If it is not possible to use drop-frame timecode, the system will fall back to non-drop-frame. This setting is enabled by default when Timecode insertion is enabled."
},
"FixedAfd": {
"shape": "__integerMin0Max15",
"locationName": "fixedAfd",
- "documentation": "Applies only if you set AFD Signaling(AfdSignaling) to Fixed (FIXED). Use Fixed (FixedAfd) to specify a four-bit AFD value which the service will write on all frames of this video output."
+ "documentation": "Applies only if you set AFD Signaling to Fixed. Use Fixed to specify a four-bit AFD value which the service will write on all frames of this video output."
},
"Height": {
"shape": "__integerMin32Max8192",
@@ -11928,32 +11928,32 @@
"Position": {
"shape": "Rectangle",
"locationName": "position",
- "documentation": "Use Selection placement (position) to define the video area in your output frame. The area outside of the rectangle that you specify here is black."
+ "documentation": "Use Selection placement to define the video area in your output frame. The area outside of the rectangle that you specify here is black."
},
"RespondToAfd": {
"shape": "RespondToAfd",
"locationName": "respondToAfd",
- "documentation": "Use Respond to AFD (RespondToAfd) to specify how the service changes the video itself in response to AFD values in the input. * Choose Respond to clip the input video frame according to the AFD value, input display aspect ratio, and output display aspect ratio. * Choose Passthrough to include the input AFD values. Do not choose this when AfdSignaling is set to (NONE). A preferred implementation of this workflow is to set RespondToAfd to (NONE) and set AfdSignaling to (AUTO). * Choose None to remove all input AFD values from this output."
+ "documentation": "Use Respond to AFD to specify how the service changes the video itself in response to AFD values in the input. * Choose Respond to clip the input video frame according to the AFD value, input display aspect ratio, and output display aspect ratio. * Choose Passthrough to include the input AFD values. Do not choose this when AfdSignaling is set to NONE. A preferred implementation of this workflow is to set RespondToAfd to and set AfdSignaling to AUTO. * Choose None to remove all input AFD values from this output."
},
"ScalingBehavior": {
"shape": "ScalingBehavior",
"locationName": "scalingBehavior",
- "documentation": "Specify how the service handles outputs that have a different aspect ratio from the input aspect ratio. Choose Stretch to output (STRETCH_TO_OUTPUT) to have the service stretch your video image to fit. Keep the setting Default (DEFAULT) to have the service letterbox your video instead. This setting overrides any value that you specify for the setting Selection placement (position) in this output."
+ "documentation": "Specify how the service handles outputs that have a different aspect ratio from the input aspect ratio. Choose Stretch to output to have the service stretch your video image to fit. Keep the setting Default to have the service letterbox your video instead. This setting overrides any value that you specify for the setting Selection placement in this output."
},
"Sharpness": {
"shape": "__integerMin0Max100",
"locationName": "sharpness",
- "documentation": "Use Sharpness (Sharpness) setting to specify the strength of anti-aliasing. This setting changes the width of the anti-alias filter kernel used for scaling. Sharpness only applies if your output resolution is different from your input resolution. 0 is the softest setting, 100 the sharpest, and 50 recommended for most content."
+ "documentation": "Use Sharpness setting to specify the strength of anti-aliasing. This setting changes the width of the anti-alias filter kernel used for scaling. Sharpness only applies if your output resolution is different from your input resolution. 0 is the softest setting, 100 the sharpest, and 50 recommended for most content."
},
"TimecodeInsertion": {
"shape": "VideoTimecodeInsertion",
"locationName": "timecodeInsertion",
- "documentation": "Applies only to H.264, H.265, MPEG2, and ProRes outputs. Only enable Timecode insertion when the input frame rate is identical to the output frame rate. To include timecodes in this output, set Timecode insertion (VideoTimecodeInsertion) to PIC_TIMING_SEI. To leave them out, set it to DISABLED. Default is DISABLED. When the service inserts timecodes in an output, by default, it uses any embedded timecodes from the input. If none are present, the service will set the timecode for the first output frame to zero. To change this default behavior, adjust the settings under Timecode configuration (TimecodeConfig). In the console, these settings are located under Job > Job settings > Timecode configuration. Note - Timecode source under input settings (InputTimecodeSource) does not affect the timecodes that are inserted in the output. Source under Job settings > Timecode configuration (TimecodeSource) does."
+ "documentation": "Applies only to H.264, H.265, MPEG2, and ProRes outputs. Only enable Timecode insertion when the input frame rate is identical to the output frame rate. To include timecodes in this output, set Timecode insertion to PIC_TIMING_SEI. To leave them out, set it to DISABLED. Default is DISABLED. When the service inserts timecodes in an output, by default, it uses any embedded timecodes from the input. If none are present, the service will set the timecode for the first output frame to zero. To change this default behavior, adjust the settings under Timecode configuration. In the console, these settings are located under Job > Job settings > Timecode configuration. Note - Timecode source under input settings does not affect the timecodes that are inserted in the output. Source under Job settings > Timecode configuration does."
},
"VideoPreprocessors": {
"shape": "VideoPreprocessor",
"locationName": "videoPreprocessors",
- "documentation": "Find additional transcoding features under Preprocessors (VideoPreprocessors). Enable the features at each output individually. These features are disabled by default."
+ "documentation": "Find additional transcoding features under Preprocessors. Enable the features at each output individually. These features are disabled by default."
},
"Width": {
"shape": "__integerMin32Max8192",
@@ -11961,7 +11961,7 @@
"documentation": "Use Width to define the video resolution width, in pixels, for this output. To use the same resolution as your input: Leave both Width and Height blank. To evenly scale from your input resolution: Leave Width blank and enter a value for Height. For example, if your input is 1920x1080 and you set Height to 720, your output will be 1280x720."
}
},
- "documentation": "Settings related to video encoding of your output. The specific video settings depend on the video codec that you choose. When you work directly in your JSON job specification, include one instance of Video description (VideoDescription) per output."
+ "documentation": "Settings related to video encoding of your output. The specific video settings depend on the video codec that you choose."
},
"VideoDetail": {
"type": "structure",
@@ -12005,7 +12005,7 @@
"ImageInserter": {
"shape": "ImageInserter",
"locationName": "imageInserter",
- "documentation": "Enable the Image inserter (ImageInserter) feature to include a graphic overlay on your video. Enable or disable this feature for each output individually. This setting is disabled by default."
+ "documentation": "Enable the Image inserter feature to include a graphic overlay on your video. Enable or disable this feature for each output individually. This setting is disabled by default."
},
"NoiseReducer": {
"shape": "NoiseReducer",
@@ -12023,7 +12023,7 @@
"documentation": "Settings for burning the output timecode and specified prefix into the output."
}
},
- "documentation": "Find additional transcoding features under Preprocessors (VideoPreprocessors). Enable the features at each output individually. These features are disabled by default."
+ "documentation": "Find additional transcoding features under Preprocessors. Enable the features at each output individually. These features are disabled by default."
},
"VideoSelector": {
"type": "structure",
@@ -12041,27 +12041,27 @@
"ColorSpaceUsage": {
"shape": "ColorSpaceUsage",
"locationName": "colorSpaceUsage",
- "documentation": "There are two sources for color metadata, the input file and the job input settings Color space (ColorSpace) and HDR master display information settings(Hdr10Metadata). The Color space usage setting determines which takes precedence. Choose Force (FORCE) to use color metadata from the input job settings. If you don't specify values for those settings, the service defaults to using metadata from your input. FALLBACK - Choose Fallback (FALLBACK) to use color metadata from the source when it is present. If there's no color metadata in your input file, the service defaults to using values you specify in the input settings."
+ "documentation": "There are two sources for color metadata, the input file and the job input settings Color space and HDR master display information settings. The Color space usage setting determines which takes precedence. Choose Force to use color metadata from the input job settings. If you don't specify values for those settings, the service defaults to using metadata from your input. FALLBACK - Choose Fallback to use color metadata from the source when it is present. If there's no color metadata in your input file, the service defaults to using values you specify in the input settings."
},
"EmbeddedTimecodeOverride": {
"shape": "EmbeddedTimecodeOverride",
"locationName": "embeddedTimecodeOverride",
- "documentation": "Set Embedded timecode override (embeddedTimecodeOverride) to Use MDPM (USE_MDPM) when your AVCHD input contains timecode tag data in the Modified Digital Video Pack Metadata (MDPM). When you do, we recommend you also set Timecode source (inputTimecodeSource) to Embedded (EMBEDDED). Leave Embedded timecode override blank, or set to None (NONE), when your input does not contain MDPM timecode."
+ "documentation": "Set Embedded timecode override to Use MDPM when your AVCHD input contains timecode tag data in the Modified Digital Video Pack Metadata. When you do, we recommend you also set Timecode source to Embedded. Leave Embedded timecode override blank, or set to None, when your input does not contain MDPM timecode."
},
"Hdr10Metadata": {
"shape": "Hdr10Metadata",
"locationName": "hdr10Metadata",
- "documentation": "Use these settings to provide HDR 10 metadata that is missing or inaccurate in your input video. Appropriate values vary depending on the input video and must be provided by a color grader. The color grader generates these values during the HDR 10 mastering process. The valid range for each of these settings is 0 to 50,000. Each increment represents 0.00002 in CIE1931 color coordinate. Related settings - When you specify these values, you must also set Color space (ColorSpace) to HDR 10 (HDR10). To specify whether the the values you specify here take precedence over the values in the metadata of your input file, set Color space usage (ColorSpaceUsage). To specify whether color metadata is included in an output, set Color metadata (ColorMetadata). For more information about MediaConvert HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr."
+ "documentation": "Use these settings to provide HDR 10 metadata that is missing or inaccurate in your input video. Appropriate values vary depending on the input video and must be provided by a color grader. The color grader generates these values during the HDR 10 mastering process. The valid range for each of these settings is 0 to 50,000. Each increment represents 0.00002 in CIE1931 color coordinate. Related settings - When you specify these values, you must also set Color space to HDR 10. To specify whether the the values you specify here take precedence over the values in the metadata of your input file, set Color space usage. To specify whether color metadata is included in an output, set Color metadata. For more information about MediaConvert HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr."
},
"PadVideo": {
"shape": "PadVideo",
"locationName": "padVideo",
- "documentation": "Use this setting if your input has video and audio durations that don't align, and your output or player has strict alignment requirements. Examples: Input audio track has a delayed start. Input video track ends before audio ends. When you set Pad video (padVideo) to Black (BLACK), MediaConvert generates black video frames so that output video and audio durations match. Black video frames are added at the beginning or end, depending on your input. To keep the default behavior and not generate black video, set Pad video to Disabled (DISABLED) or leave blank."
+ "documentation": "Use this setting if your input has video and audio durations that don't align, and your output or player has strict alignment requirements. Examples: Input audio track has a delayed start. Input video track ends before audio ends. When you set Pad video to Black, MediaConvert generates black video frames so that output video and audio durations match. Black video frames are added at the beginning or end, depending on your input. To keep the default behavior and not generate black video, set Pad video to Disabled or leave blank."
},
"Pid": {
"shape": "__integerMin1Max2147483647",
"locationName": "pid",
- "documentation": "Use PID (Pid) to select specific video data from an input file. Specify this value as an integer; the system automatically converts it to the hexidecimal value. For example, 257 selects PID 0x101. A PID, or packet identifier, is an identifier for a set of data in an MPEG-2 transport stream container."
+ "documentation": "Use PID to select specific video data from an input file. Specify this value as an integer; the system automatically converts it to the hexidecimal value. For example, 257 selects PID 0x101. A PID, or packet identifier, is an identifier for a set of data in an MPEG-2 transport stream container."
},
"ProgramNumber": {
"shape": "__integerMinNegative2147483648Max2147483647",
@@ -12071,19 +12071,19 @@
"Rotate": {
"shape": "InputRotate",
"locationName": "rotate",
- "documentation": "Use Rotate (InputRotate) to specify how the service rotates your video. You can choose automatic rotation or specify a rotation. You can specify a clockwise rotation of 0, 90, 180, or 270 degrees. If your input video container is .mov or .mp4 and your input has rotation metadata, you can choose Automatic to have the service rotate your video according to the rotation specified in the metadata. The rotation must be within one degree of 90, 180, or 270 degrees. If the rotation metadata specifies any other rotation, the service will default to no rotation. By default, the service does no rotation, even if your input video has rotation metadata. The service doesn't pass through rotation metadata."
+ "documentation": "Use Rotate to specify how the service rotates your video. You can choose automatic rotation or specify a rotation. You can specify a clockwise rotation of 0, 90, 180, or 270 degrees. If your input video container is .mov or .mp4 and your input has rotation metadata, you can choose Automatic to have the service rotate your video according to the rotation specified in the metadata. The rotation must be within one degree of 90, 180, or 270 degrees. If the rotation metadata specifies any other rotation, the service will default to no rotation. By default, the service does no rotation, even if your input video has rotation metadata. The service doesn't pass through rotation metadata."
},
"SampleRange": {
"shape": "InputSampleRange",
"locationName": "sampleRange",
- "documentation": "If the sample range metadata in your input video is accurate, or if you don't know about sample range, keep the default value, Follow (FOLLOW), for this setting. When you do, the service automatically detects your input sample range. If your input video has metadata indicating the wrong sample range, specify the accurate sample range here. When you do, MediaConvert ignores any sample range information in the input metadata. Regardless of whether MediaConvert uses the input sample range or the sample range that you specify, MediaConvert uses the sample range for transcoding and also writes it to the output metadata."
+ "documentation": "If the sample range metadata in your input video is accurate, or if you don't know about sample range, keep the default value, Follow, for this setting. When you do, the service automatically detects your input sample range. If your input video has metadata indicating the wrong sample range, specify the accurate sample range here. When you do, MediaConvert ignores any sample range information in the input metadata. Regardless of whether MediaConvert uses the input sample range or the sample range that you specify, MediaConvert uses the sample range for transcoding and also writes it to the output metadata."
}
},
"documentation": "Input video selectors contain the video settings for the input. Each of your inputs can have up to one video selector."
},
"VideoTimecodeInsertion": {
"type": "string",
- "documentation": "Applies only to H.264, H.265, MPEG2, and ProRes outputs. Only enable Timecode insertion when the input frame rate is identical to the output frame rate. To include timecodes in this output, set Timecode insertion (VideoTimecodeInsertion) to PIC_TIMING_SEI. To leave them out, set it to DISABLED. Default is DISABLED. When the service inserts timecodes in an output, by default, it uses any embedded timecodes from the input. If none are present, the service will set the timecode for the first output frame to zero. To change this default behavior, adjust the settings under Timecode configuration (TimecodeConfig). In the console, these settings are located under Job > Job settings > Timecode configuration. Note - Timecode source under input settings (InputTimecodeSource) does not affect the timecodes that are inserted in the output. Source under Job settings > Timecode configuration (TimecodeSource) does.",
+ "documentation": "Applies only to H.264, H.265, MPEG2, and ProRes outputs. Only enable Timecode insertion when the input frame rate is identical to the output frame rate. To include timecodes in this output, set Timecode insertion to PIC_TIMING_SEI. To leave them out, set it to DISABLED. Default is DISABLED. When the service inserts timecodes in an output, by default, it uses any embedded timecodes from the input. If none are present, the service will set the timecode for the first output frame to zero. To change this default behavior, adjust the settings under Timecode configuration. In the console, these settings are located under Job > Job settings > Timecode configuration. Note - Timecode source under input settings does not affect the timecodes that are inserted in the output. Source under Job settings > Timecode configuration does.",
"enum": [
"DISABLED",
"PIC_TIMING_SEI"
@@ -12112,7 +12112,7 @@
},
"Vp8FramerateControl": {
"type": "string",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator.",
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -12129,7 +12129,7 @@
},
"Vp8ParControl": {
"type": "string",
- "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source (INITIALIZE_FROM_SOURCE), uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. To specify a different PAR by editing the JSON job specification, choose SPECIFIED. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.",
+ "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -12137,7 +12137,7 @@
},
"Vp8QualityTuningLevel": {
"type": "string",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, multi-pass encoding.",
+ "documentation": "Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, multi-pass encoding.",
"enum": [
"MULTI_PASS",
"MULTI_PASS_HQ"
@@ -12161,7 +12161,7 @@
"FramerateControl": {
"shape": "Vp8FramerateControl",
"locationName": "framerateControl",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator."
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction."
},
"FramerateConversionAlgorithm": {
"shape": "Vp8FramerateConversionAlgorithm",
@@ -12196,22 +12196,22 @@
"ParControl": {
"shape": "Vp8ParControl",
"locationName": "parControl",
- "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source (INITIALIZE_FROM_SOURCE), uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. To specify a different PAR by editing the JSON job specification, choose SPECIFIED. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings."
+ "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings."
},
"ParDenominator": {
"shape": "__integerMin1Max2147483647",
"locationName": "parDenominator",
- "documentation": "Required when you set Pixel aspect ratio (parControl) to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33."
+ "documentation": "Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33."
},
"ParNumerator": {
"shape": "__integerMin1Max2147483647",
"locationName": "parNumerator",
- "documentation": "Required when you set Pixel aspect ratio (parControl) to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40."
+ "documentation": "Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40."
},
"QualityTuningLevel": {
"shape": "Vp8QualityTuningLevel",
"locationName": "qualityTuningLevel",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, multi-pass encoding."
+ "documentation": "Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, multi-pass encoding."
},
"RateControlMode": {
"shape": "Vp8RateControlMode",
@@ -12219,11 +12219,11 @@
"documentation": "With the VP8 codec, you can use only the variable bitrate (VBR) rate control mode."
}
},
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value VP8."
+ "documentation": "Required when you set Codec to the value VP8."
},
"Vp9FramerateControl": {
"type": "string",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator.",
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -12240,7 +12240,7 @@
},
"Vp9ParControl": {
"type": "string",
- "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source (INITIALIZE_FROM_SOURCE), uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. To specify a different PAR by editing the JSON job specification, choose SPECIFIED. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.",
+ "documentation": "Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -12248,7 +12248,7 @@
},
"Vp9QualityTuningLevel": {
"type": "string",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, multi-pass encoding.",
+ "documentation": "Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, multi-pass encoding.",
"enum": [
"MULTI_PASS",
"MULTI_PASS_HQ"
@@ -12272,7 +12272,7 @@
"FramerateControl": {
"shape": "Vp9FramerateControl",
"locationName": "framerateControl",
- "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator."
+ "documentation": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction."
},
"FramerateConversionAlgorithm": {
"shape": "Vp9FramerateConversionAlgorithm",
@@ -12312,17 +12312,17 @@
"ParDenominator": {
"shape": "__integerMin1Max2147483647",
"locationName": "parDenominator",
- "documentation": "Required when you set Pixel aspect ratio (parControl) to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33."
+ "documentation": "Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33."
},
"ParNumerator": {
"shape": "__integerMin1Max2147483647",
"locationName": "parNumerator",
- "documentation": "Required when you set Pixel aspect ratio (parControl) to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40."
+ "documentation": "Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40."
},
"QualityTuningLevel": {
"shape": "Vp9QualityTuningLevel",
"locationName": "qualityTuningLevel",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, multi-pass encoding."
+ "documentation": "Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, multi-pass encoding."
},
"RateControlMode": {
"shape": "Vp9RateControlMode",
@@ -12330,7 +12330,7 @@
"documentation": "With the VP9 codec, you can use only the variable bitrate (VBR) rate control mode."
}
},
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value VP9."
+ "documentation": "Required when you set Codec to the value VP9."
},
"WarningGroup": {
"type": "structure",
@@ -12377,7 +12377,7 @@
"BitDepth": {
"shape": "__integerMin16Max24",
"locationName": "bitDepth",
- "documentation": "Specify Bit depth (BitDepth), in bits per sample, to choose the encoding quality for this audio track."
+ "documentation": "Specify Bit depth, in bits per sample, to choose the encoding quality for this audio track."
},
"Channels": {
"shape": "__integerMin1Max64",
@@ -12395,7 +12395,7 @@
"documentation": "Sample rate in Hz."
}
},
- "documentation": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value WAV."
+ "documentation": "Required when you set Codec to the value WAV."
},
"WebvttAccessibilitySubs": {
"type": "string",
@@ -12416,10 +12416,10 @@
"StylePassthrough": {
"shape": "WebvttStylePassthrough",
"locationName": "stylePassthrough",
- "documentation": "To use the available style, color, and position information from your input captions: Set Style passthrough (stylePassthrough) to Enabled (ENABLED). MediaConvert uses default settings when style and position information is missing from your input captions. To recreate the input captions exactly: Set Style passthrough to Strict (STRICT). MediaConvert automatically applies timing adjustments, including adjustments for frame rate conversion, ad avails, and input clipping. Your input captions format must be WebVTT. To ignore the style and position information from your input captions and use simplified output captions: Set Style passthrough to Disabled (DISABLED), or leave blank."
+ "documentation": "To use the available style, color, and position information from your input captions: Set Style passthrough to Enabled. MediaConvert uses default settings when style and position information is missing from your input captions. To recreate the input captions exactly: Set Style passthrough to Strict. MediaConvert automatically applies timing adjustments, including adjustments for frame rate conversion, ad avails, and input clipping. Your input captions format must be WebVTT. To ignore the style and position information from your input captions and use simplified output captions: Set Style passthrough to Disabled, or leave blank."
}
},
- "documentation": "Settings related to WebVTT captions. WebVTT is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to WebVTT."
+ "documentation": "Settings related to WebVTT captions. WebVTT is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html."
},
"WebvttHlsSourceSettings": {
"type": "structure",
@@ -12444,7 +12444,7 @@
},
"WebvttStylePassthrough": {
"type": "string",
- "documentation": "To use the available style, color, and position information from your input captions: Set Style passthrough (stylePassthrough) to Enabled (ENABLED). MediaConvert uses default settings when style and position information is missing from your input captions. To recreate the input captions exactly: Set Style passthrough to Strict (STRICT). MediaConvert automatically applies timing adjustments, including adjustments for frame rate conversion, ad avails, and input clipping. Your input captions format must be WebVTT. To ignore the style and position information from your input captions and use simplified output captions: Set Style passthrough to Disabled (DISABLED), or leave blank.",
+ "documentation": "To use the available style, color, and position information from your input captions: Set Style passthrough to Enabled. MediaConvert uses default settings when style and position information is missing from your input captions. To recreate the input captions exactly: Set Style passthrough to Strict. MediaConvert automatically applies timing adjustments, including adjustments for frame rate conversion, ad avails, and input clipping. Your input captions format must be WebVTT. To ignore the style and position information from your input captions and use simplified output captions: Set Style passthrough to Disabled, or leave blank.",
"enum": [
"ENABLED",
"DISABLED",
@@ -12469,7 +12469,7 @@
"documentation": "Specify the XAVC Intra 4k (CBG) Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class."
}
},
- "documentation": "Required when you set (Profile) under (VideoDescription)>(CodecSettings)>(XavcSettings) to the value XAVC_4K_INTRA_CBG."
+ "documentation": "Required when you set Profile to the value XAVC_4K_INTRA_CBG."
},
"Xavc4kIntraVbrProfileClass": {
"type": "string",
@@ -12489,7 +12489,7 @@
"documentation": "Specify the XAVC Intra 4k (VBR) Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class."
}
},
- "documentation": "Required when you set (Profile) under (VideoDescription)>(CodecSettings)>(XavcSettings) to the value XAVC_4K_INTRA_VBR."
+ "documentation": "Required when you set Profile to the value XAVC_4K_INTRA_VBR."
},
"Xavc4kProfileBitrateClass": {
"type": "string",
@@ -12510,7 +12510,7 @@
},
"Xavc4kProfileQualityTuningLevel": {
"type": "string",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.",
+ "documentation": "Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.",
"enum": [
"SINGLE_PASS",
"SINGLE_PASS_HQ",
@@ -12533,12 +12533,12 @@
"FlickerAdaptiveQuantization": {
"shape": "XavcFlickerAdaptiveQuantization",
"locationName": "flickerAdaptiveQuantization",
- "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto (AUTO), for the setting Adaptive quantization (XavcAdaptiveQuantization). When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set Adaptive quantization (adaptiveQuantization) to a value other than Off (OFF) or Auto (AUTO). Use Adaptive quantization to adjust the degree of smoothing that Flicker adaptive quantization provides."
+ "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set Adaptive quantization to a value other than Off or Auto. Use Adaptive quantization to adjust the degree of smoothing that Flicker adaptive quantization provides."
},
"GopBReference": {
"shape": "XavcGopBReference",
"locationName": "gopBReference",
- "documentation": "Specify whether the encoder uses B-frames as reference frames for other pictures in the same GOP. Choose Allow (ENABLED) to allow the encoder to use B-frames as reference frames. Choose Don't allow (DISABLED) to prevent the encoder from using B-frames as reference frames."
+ "documentation": "Specify whether the encoder uses B-frames as reference frames for other pictures in the same GOP. Choose Allow to allow the encoder to use B-frames as reference frames. Choose Don't allow to prevent the encoder from using B-frames as reference frames."
},
"GopClosedCadence": {
"shape": "__integerMin0Max2147483647",
@@ -12553,7 +12553,7 @@
"QualityTuningLevel": {
"shape": "Xavc4kProfileQualityTuningLevel",
"locationName": "qualityTuningLevel",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding."
+ "documentation": "Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding."
},
"Slices": {
"shape": "__integerMin8Max12",
@@ -12561,11 +12561,11 @@
"documentation": "Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures."
}
},
- "documentation": "Required when you set (Profile) under (VideoDescription)>(CodecSettings)>(XavcSettings) to the value XAVC_4K."
+ "documentation": "Required when you set Profile to the value XAVC_4K."
},
"XavcAdaptiveQuantization": {
"type": "string",
- "documentation": "Keep the default value, Auto (AUTO), for this setting to have MediaConvert automatically apply the best types of quantization for your video content. When you want to apply your quantization settings manually, you must set Adaptive quantization (adaptiveQuantization) to a value other than Auto (AUTO). Use this setting to specify the strength of any adaptive quantization filters that you enable. If you don't want MediaConvert to do any adaptive quantization in this transcode, set Adaptive quantization to Off (OFF). Related settings: The value that you choose here applies to the following settings: Flicker adaptive quantization (flickerAdaptiveQuantization), Spatial adaptive quantization (spatialAdaptiveQuantization), and Temporal adaptive quantization (temporalAdaptiveQuantization).",
+ "documentation": "Keep the default value, Auto, for this setting to have MediaConvert automatically apply the best types of quantization for your video content. When you want to apply your quantization settings manually, you must set Adaptive quantization to a value other than Auto. Use this setting to specify the strength of any adaptive quantization filters that you enable. If you don't want MediaConvert to do any adaptive quantization in this transcode, set Adaptive quantization to Off. Related settings: The value that you choose here applies to the following settings: Flicker adaptive quantization (flickerAdaptiveQuantization), Spatial adaptive quantization, and Temporal adaptive quantization.",
"enum": [
"OFF",
"AUTO",
@@ -12587,7 +12587,7 @@
},
"XavcFlickerAdaptiveQuantization": {
"type": "string",
- "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto (AUTO), for the setting Adaptive quantization (XavcAdaptiveQuantization). When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set Adaptive quantization (adaptiveQuantization) to a value other than Off (OFF) or Auto (AUTO). Use Adaptive quantization to adjust the degree of smoothing that Flicker adaptive quantization provides.",
+ "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set Adaptive quantization to a value other than Off or Auto. Use Adaptive quantization to adjust the degree of smoothing that Flicker adaptive quantization provides.",
"enum": [
"DISABLED",
"ENABLED"
@@ -12595,7 +12595,7 @@
},
"XavcFramerateControl": {
"type": "string",
- "documentation": "If you are using the console, use the Frame rate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list. The framerates shown in the dropdown list are decimal approximations of fractions. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate that you specify in the settings FramerateNumerator and FramerateDenominator.",
+ "documentation": "If you are using the console, use the Frame rate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list. The framerates shown in the dropdown list are decimal approximations of fractions.",
"enum": [
"INITIALIZE_FROM_SOURCE",
"SPECIFIED"
@@ -12612,7 +12612,7 @@
},
"XavcGopBReference": {
"type": "string",
- "documentation": "Specify whether the encoder uses B-frames as reference frames for other pictures in the same GOP. Choose Allow (ENABLED) to allow the encoder to use B-frames as reference frames. Choose Don't allow (DISABLED) to prevent the encoder from using B-frames as reference frames.",
+ "documentation": "Specify whether the encoder uses B-frames as reference frames for other pictures in the same GOP. Choose Allow to allow the encoder to use B-frames as reference frames. Choose Don't allow to prevent the encoder from using B-frames as reference frames.",
"enum": [
"DISABLED",
"ENABLED"
@@ -12636,7 +12636,7 @@
"documentation": "Specify the XAVC Intra HD (CBG) Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class."
}
},
- "documentation": "Required when you set (Profile) under (VideoDescription)>(CodecSettings)>(XavcSettings) to the value XAVC_HD_INTRA_CBG."
+ "documentation": "Required when you set Profile to the value XAVC_HD_INTRA_CBG."
},
"XavcHdProfileBitrateClass": {
"type": "string",
@@ -12649,7 +12649,7 @@
},
"XavcHdProfileQualityTuningLevel": {
"type": "string",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.",
+ "documentation": "Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.",
"enum": [
"SINGLE_PASS",
"SINGLE_PASS_HQ",
@@ -12667,12 +12667,12 @@
"FlickerAdaptiveQuantization": {
"shape": "XavcFlickerAdaptiveQuantization",
"locationName": "flickerAdaptiveQuantization",
- "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto (AUTO), for the setting Adaptive quantization (XavcAdaptiveQuantization). When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set Adaptive quantization (adaptiveQuantization) to a value other than Off (OFF) or Auto (AUTO). Use Adaptive quantization to adjust the degree of smoothing that Flicker adaptive quantization provides."
+ "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set Adaptive quantization to a value other than Off or Auto. Use Adaptive quantization to adjust the degree of smoothing that Flicker adaptive quantization provides."
},
"GopBReference": {
"shape": "XavcGopBReference",
"locationName": "gopBReference",
- "documentation": "Specify whether the encoder uses B-frames as reference frames for other pictures in the same GOP. Choose Allow (ENABLED) to allow the encoder to use B-frames as reference frames. Choose Don't allow (DISABLED) to prevent the encoder from using B-frames as reference frames."
+ "documentation": "Specify whether the encoder uses B-frames as reference frames for other pictures in the same GOP. Choose Allow to allow the encoder to use B-frames as reference frames. Choose Don't allow to prevent the encoder from using B-frames as reference frames."
},
"GopClosedCadence": {
"shape": "__integerMin0Max2147483647",
@@ -12687,12 +12687,12 @@
"InterlaceMode": {
"shape": "XavcInterlaceMode",
"locationName": "interlaceMode",
- "documentation": "Choose the scan line type for the output. Keep the default value, Progressive (PROGRESSIVE) to create a progressive output, regardless of the scan type of your input. Use Top field first (TOP_FIELD) or Bottom field first (BOTTOM_FIELD) to create an output that's interlaced with the same field polarity throughout. Use Follow, default top (FOLLOW_TOP_FIELD) or Follow, default bottom (FOLLOW_BOTTOM_FIELD) to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose."
+ "documentation": "Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose."
},
"QualityTuningLevel": {
"shape": "XavcHdProfileQualityTuningLevel",
"locationName": "qualityTuningLevel",
- "documentation": "Optional. Use Quality tuning level (qualityTuningLevel) to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding."
+ "documentation": "Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding."
},
"Slices": {
"shape": "__integerMin4Max12",
@@ -12702,14 +12702,14 @@
"Telecine": {
"shape": "XavcHdProfileTelecine",
"locationName": "telecine",
- "documentation": "Ignore this setting unless you set Frame rate (framerateNumerator divided by framerateDenominator) to 29.970. If your input framerate is 23.976, choose Hard (HARD). Otherwise, keep the default value None (NONE). For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/working-with-telecine-and-inverse-telecine.html."
+ "documentation": "Ignore this setting unless you set Frame rate (framerateNumerator divided by framerateDenominator) to 29.970. If your input framerate is 23.976, choose Hard. Otherwise, keep the default value None. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/working-with-telecine-and-inverse-telecine.html."
}
},
- "documentation": "Required when you set (Profile) under (VideoDescription)>(CodecSettings)>(XavcSettings) to the value XAVC_HD."
+ "documentation": "Required when you set Profile to the value XAVC_HD."
},
"XavcHdProfileTelecine": {
"type": "string",
- "documentation": "Ignore this setting unless you set Frame rate (framerateNumerator divided by framerateDenominator) to 29.970. If your input framerate is 23.976, choose Hard (HARD). Otherwise, keep the default value None (NONE). For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/working-with-telecine-and-inverse-telecine.html.",
+ "documentation": "Ignore this setting unless you set Frame rate (framerateNumerator divided by framerateDenominator) to 29.970. If your input framerate is 23.976, choose Hard. Otherwise, keep the default value None. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/working-with-telecine-and-inverse-telecine.html.",
"enum": [
"NONE",
"HARD"
@@ -12717,7 +12717,7 @@
},
"XavcInterlaceMode": {
"type": "string",
- "documentation": "Choose the scan line type for the output. Keep the default value, Progressive (PROGRESSIVE) to create a progressive output, regardless of the scan type of your input. Use Top field first (TOP_FIELD) or Bottom field first (BOTTOM_FIELD) to create an output that's interlaced with the same field polarity throughout. Use Follow, default top (FOLLOW_TOP_FIELD) or Follow, default bottom (FOLLOW_BOTTOM_FIELD) to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.",
+ "documentation": "Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.",
"enum": [
"PROGRESSIVE",
"TOP_FIELD",
@@ -12743,7 +12743,7 @@
"AdaptiveQuantization": {
"shape": "XavcAdaptiveQuantization",
"locationName": "adaptiveQuantization",
- "documentation": "Keep the default value, Auto (AUTO), for this setting to have MediaConvert automatically apply the best types of quantization for your video content. When you want to apply your quantization settings manually, you must set Adaptive quantization (adaptiveQuantization) to a value other than Auto (AUTO). Use this setting to specify the strength of any adaptive quantization filters that you enable. If you don't want MediaConvert to do any adaptive quantization in this transcode, set Adaptive quantization to Off (OFF). Related settings: The value that you choose here applies to the following settings: Flicker adaptive quantization (flickerAdaptiveQuantization), Spatial adaptive quantization (spatialAdaptiveQuantization), and Temporal adaptive quantization (temporalAdaptiveQuantization)."
+ "documentation": "Keep the default value, Auto, for this setting to have MediaConvert automatically apply the best types of quantization for your video content. When you want to apply your quantization settings manually, you must set Adaptive quantization to a value other than Auto. Use this setting to specify the strength of any adaptive quantization filters that you enable. If you don't want MediaConvert to do any adaptive quantization in this transcode, set Adaptive quantization to Off. Related settings: The value that you choose here applies to the following settings: Flicker adaptive quantization (flickerAdaptiveQuantization), Spatial adaptive quantization, and Temporal adaptive quantization."
},
"EntropyEncoding": {
"shape": "XavcEntropyEncoding",
@@ -12753,7 +12753,7 @@
"FramerateControl": {
"shape": "XavcFramerateControl",
"locationName": "framerateControl",
- "documentation": "If you are using the console, use the Frame rate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list. The framerates shown in the dropdown list are decimal approximations of fractions. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate that you specify in the settings FramerateNumerator and FramerateDenominator."
+ "documentation": "If you are using the console, use the Frame rate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list. The framerates shown in the dropdown list are decimal approximations of fractions."
},
"FramerateConversionAlgorithm": {
"shape": "XavcFramerateConversionAlgorithm",
@@ -12778,54 +12778,54 @@
"SlowPal": {
"shape": "XavcSlowPal",
"locationName": "slowPal",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Frame rate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1."
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Frame rate to 25."
},
"Softness": {
"shape": "__integerMin0Max128",
"locationName": "softness",
- "documentation": "Ignore this setting unless your downstream workflow requires that you specify it explicitly. Otherwise, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness (sharpness) or by enabling a noise reducer filter (noiseReducerFilter). The Softness (softness) setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, for flat quantization. Choose the value 1 or 16 to use the default JVT softening quantization matricies from the H.264 specification. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video."
+ "documentation": "Ignore this setting unless your downstream workflow requires that you specify it explicitly. Otherwise, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness or by enabling a noise reducer filter. The Softness setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, for flat quantization. Choose the value 1 or 16 to use the default JVT softening quantization matricies from the H.264 specification. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video."
},
"SpatialAdaptiveQuantization": {
"shape": "XavcSpatialAdaptiveQuantization",
"locationName": "spatialAdaptiveQuantization",
- "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto (AUTO), for the setting Adaptive quantization (adaptiveQuantization). When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. For this setting, keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization (adaptiveQuantization) depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher."
+ "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. For this setting, keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher."
},
"TemporalAdaptiveQuantization": {
"shape": "XavcTemporalAdaptiveQuantization",
"locationName": "temporalAdaptiveQuantization",
- "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto (AUTO), for the setting Adaptive quantization (adaptiveQuantization). When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. For this setting, keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal adaptive quantization, adjust the strength of the filter with the setting Adaptive quantization (adaptiveQuantization)."
+ "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. For this setting, keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal adaptive quantization, adjust the strength of the filter with the setting Adaptive quantization."
},
"Xavc4kIntraCbgProfileSettings": {
"shape": "Xavc4kIntraCbgProfileSettings",
"locationName": "xavc4kIntraCbgProfileSettings",
- "documentation": "Required when you set (Profile) under (VideoDescription)>(CodecSettings)>(XavcSettings) to the value XAVC_4K_INTRA_CBG."
+ "documentation": "Required when you set Profile to the value XAVC_4K_INTRA_CBG."
},
"Xavc4kIntraVbrProfileSettings": {
"shape": "Xavc4kIntraVbrProfileSettings",
"locationName": "xavc4kIntraVbrProfileSettings",
- "documentation": "Required when you set (Profile) under (VideoDescription)>(CodecSettings)>(XavcSettings) to the value XAVC_4K_INTRA_VBR."
+ "documentation": "Required when you set Profile to the value XAVC_4K_INTRA_VBR."
},
"Xavc4kProfileSettings": {
"shape": "Xavc4kProfileSettings",
"locationName": "xavc4kProfileSettings",
- "documentation": "Required when you set (Profile) under (VideoDescription)>(CodecSettings)>(XavcSettings) to the value XAVC_4K."
+ "documentation": "Required when you set Profile to the value XAVC_4K."
},
"XavcHdIntraCbgProfileSettings": {
"shape": "XavcHdIntraCbgProfileSettings",
"locationName": "xavcHdIntraCbgProfileSettings",
- "documentation": "Required when you set (Profile) under (VideoDescription)>(CodecSettings)>(XavcSettings) to the value XAVC_HD_INTRA_CBG."
+ "documentation": "Required when you set Profile to the value XAVC_HD_INTRA_CBG."
},
"XavcHdProfileSettings": {
"shape": "XavcHdProfileSettings",
"locationName": "xavcHdProfileSettings",
- "documentation": "Required when you set (Profile) under (VideoDescription)>(CodecSettings)>(XavcSettings) to the value XAVC_HD."
+ "documentation": "Required when you set Profile to the value XAVC_HD."
}
},
- "documentation": "Required when you set (Codec) under (VideoDescription)>(CodecSettings) to the value XAVC."
+ "documentation": "Required when you set Codec to the value XAVC."
},
"XavcSlowPal": {
"type": "string",
- "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Frame rate to 25. In your JSON job specification, set (framerateControl) to (SPECIFIED), (framerateNumerator) to 25 and (framerateDenominator) to 1.",
+ "documentation": "Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Frame rate to 25.",
"enum": [
"DISABLED",
"ENABLED"
@@ -12833,7 +12833,7 @@
},
"XavcSpatialAdaptiveQuantization": {
"type": "string",
- "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto (AUTO), for the setting Adaptive quantization (adaptiveQuantization). When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. For this setting, keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization (adaptiveQuantization) depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.",
+ "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. For this setting, keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.",
"enum": [
"DISABLED",
"ENABLED"
@@ -12841,7 +12841,7 @@
},
"XavcTemporalAdaptiveQuantization": {
"type": "string",
- "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto (AUTO), for the setting Adaptive quantization (adaptiveQuantization). When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. For this setting, keep the default value, Enabled (ENABLED), to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal adaptive quantization, adjust the strength of the filter with the setting Adaptive quantization (adaptiveQuantization).",
+ "documentation": "The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. For this setting, keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal adaptive quantization, adjust the strength of the filter with the setting Adaptive quantization.",
"enum": [
"DISABLED",
"ENABLED"
@@ -13873,4 +13873,4 @@
}
},
"documentation": "AWS Elemental MediaConvert"
-}
\ No newline at end of file
+}
diff --git a/services/medialive/pom.xml b/services/medialive/pom.xml
index 17c8c05491e..c2b7f194f31 100644
--- a/services/medialive/pom.xml
+++ b/services/medialive/pom.xml
@@ -20,7 +20,7 @@
servicessoftware.amazon.awssdk
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOT4.0.0medialive
diff --git a/services/medialive/src/main/resources/codegen-resources/service-2.json b/services/medialive/src/main/resources/codegen-resources/service-2.json
index a82d641f438..12e6a1f6179 100644
--- a/services/medialive/src/main/resources/codegen-resources/service-2.json
+++ b/services/medialive/src/main/resources/codegen-resources/service-2.json
@@ -6137,10 +6137,15 @@
"locationName": "uhdDeviceSettings",
"documentation": "Settings that describe an input device that is type UHD."
},
- "Tags": {
+ "Tags": {
"shape": "Tags",
"locationName": "tags",
"documentation": "A collection of key-value pairs."
+ },
+ "AvailabilityZone": {
+ "shape": "__string",
+ "locationName": "availabilityZone",
+ "documentation": "The Availability Zone associated with this input device."
}
},
"documentation": "Placeholder documentation for DescribeInputDeviceResponse"
@@ -9788,10 +9793,15 @@
"locationName": "uhdDeviceSettings",
"documentation": "Settings that describe an input device that is type UHD."
},
- "Tags": {
+ "Tags": {
"shape": "Tags",
"locationName": "tags",
"documentation": "A collection of key-value pairs."
+ },
+ "AvailabilityZone": {
+ "shape": "__string",
+ "locationName": "availabilityZone",
+ "documentation": "The Availability Zone associated with this input device."
}
},
"documentation": "An input device."
@@ -10049,10 +10059,15 @@
"locationName": "uhdDeviceSettings",
"documentation": "Settings that describe an input device that is type UHD."
},
- "Tags": {
+ "Tags": {
"shape": "Tags",
"locationName": "tags",
"documentation": "A collection of key-value pairs."
+ },
+ "AvailabilityZone": {
+ "shape": "__string",
+ "locationName": "availabilityZone",
+ "documentation": "The Availability Zone associated with this input device."
}
},
"documentation": "Details of the input device."
@@ -15721,6 +15736,11 @@
"shape": "InputDeviceConfigurableSettings",
"locationName": "uhdDeviceSettings",
"documentation": "The settings that you want to apply to the UHD input device."
+ },
+ "AvailabilityZone": {
+ "shape": "__string",
+ "locationName": "availabilityZone",
+ "documentation": "The Availability Zone you want associated with this input device."
}
},
"documentation": "Updates an input device."
@@ -15748,6 +15768,11 @@
"shape": "InputDeviceConfigurableSettings",
"locationName": "uhdDeviceSettings",
"documentation": "The settings that you want to apply to the UHD input device."
+ },
+ "AvailabilityZone": {
+ "shape": "__string",
+ "locationName": "availabilityZone",
+ "documentation": "The Availability Zone you want associated with this input device."
}
},
"documentation": "A request to update an input device.",
@@ -15818,10 +15843,15 @@
"locationName": "uhdDeviceSettings",
"documentation": "Settings that describe an input device that is type UHD."
},
- "Tags": {
+ "Tags": {
"shape": "Tags",
"locationName": "tags",
"documentation": "A collection of key-value pairs."
+ },
+ "AvailabilityZone": {
+ "shape": "__string",
+ "locationName": "availabilityZone",
+ "documentation": "The Availability Zone associated with this input device."
}
},
"documentation": "Placeholder documentation for UpdateInputDeviceResponse"
diff --git a/services/mediapackage/pom.xml b/services/mediapackage/pom.xml
index 36e77f2df84..140674caa34 100644
--- a/services/mediapackage/pom.xml
+++ b/services/mediapackage/pom.xml
@@ -20,7 +20,7 @@
servicessoftware.amazon.awssdk
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOT4.0.0mediapackage
diff --git a/services/mediapackagev2/pom.xml b/services/mediapackagev2/pom.xml
index 26b62f13676..16cdfacabf5 100644
--- a/services/mediapackagev2/pom.xml
+++ b/services/mediapackagev2/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmediapackagev2AWS Java SDK :: Services :: Media Package V2
diff --git a/services/mediapackagevod/pom.xml b/services/mediapackagevod/pom.xml
index f02c8ecc99f..9647ed42f00 100644
--- a/services/mediapackagevod/pom.xml
+++ b/services/mediapackagevod/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmediapackagevodAWS Java SDK :: Services :: MediaPackage Vod
diff --git a/services/mediastore/pom.xml b/services/mediastore/pom.xml
index 1e1425ce70d..ba5ffed496c 100644
--- a/services/mediastore/pom.xml
+++ b/services/mediastore/pom.xml
@@ -20,7 +20,7 @@
servicessoftware.amazon.awssdk
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOT4.0.0mediastore
diff --git a/services/mediastoredata/pom.xml b/services/mediastoredata/pom.xml
index 3e5e81a0a21..ca5c667ee50 100644
--- a/services/mediastoredata/pom.xml
+++ b/services/mediastoredata/pom.xml
@@ -20,7 +20,7 @@
servicessoftware.amazon.awssdk
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOT4.0.0mediastoredata
diff --git a/services/mediatailor/pom.xml b/services/mediatailor/pom.xml
index 094a47aab31..ba3f32adef0 100644
--- a/services/mediatailor/pom.xml
+++ b/services/mediatailor/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmediatailorAWS Java SDK :: Services :: MediaTailor
diff --git a/services/medicalimaging/pom.xml b/services/medicalimaging/pom.xml
index ff3035e3d7a..5ebe30ed9f6 100644
--- a/services/medicalimaging/pom.xml
+++ b/services/medicalimaging/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmedicalimagingAWS Java SDK :: Services :: Medical Imaging
diff --git a/services/memorydb/pom.xml b/services/memorydb/pom.xml
index 270d2e3c472..509d976f4b5 100644
--- a/services/memorydb/pom.xml
+++ b/services/memorydb/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmemorydbAWS Java SDK :: Services :: Memory DB
diff --git a/services/mgn/pom.xml b/services/mgn/pom.xml
index 74987b9dfa5..829c6e76179 100644
--- a/services/mgn/pom.xml
+++ b/services/mgn/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmgnAWS Java SDK :: Services :: Mgn
diff --git a/services/migrationhub/pom.xml b/services/migrationhub/pom.xml
index 0c944663c87..07917103ac0 100644
--- a/services/migrationhub/pom.xml
+++ b/services/migrationhub/pom.xml
@@ -20,7 +20,7 @@
servicessoftware.amazon.awssdk
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOT4.0.0migrationhub
diff --git a/services/migrationhubconfig/pom.xml b/services/migrationhubconfig/pom.xml
index 88fac38a5f0..bf216186376 100644
--- a/services/migrationhubconfig/pom.xml
+++ b/services/migrationhubconfig/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmigrationhubconfigAWS Java SDK :: Services :: MigrationHub Config
diff --git a/services/migrationhuborchestrator/pom.xml b/services/migrationhuborchestrator/pom.xml
index b10c7e7bf60..472aeedcb6a 100644
--- a/services/migrationhuborchestrator/pom.xml
+++ b/services/migrationhuborchestrator/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmigrationhuborchestratorAWS Java SDK :: Services :: Migration Hub Orchestrator
diff --git a/services/migrationhubrefactorspaces/pom.xml b/services/migrationhubrefactorspaces/pom.xml
index a90a54e16aa..a08bb023952 100644
--- a/services/migrationhubrefactorspaces/pom.xml
+++ b/services/migrationhubrefactorspaces/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmigrationhubrefactorspacesAWS Java SDK :: Services :: Migration Hub Refactor Spaces
diff --git a/services/migrationhubstrategy/pom.xml b/services/migrationhubstrategy/pom.xml
index af47639b3df..bab3418a884 100644
--- a/services/migrationhubstrategy/pom.xml
+++ b/services/migrationhubstrategy/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmigrationhubstrategyAWS Java SDK :: Services :: Migration Hub Strategy
diff --git a/services/mobile/pom.xml b/services/mobile/pom.xml
index 09d84088a5c..b6c4e7f6612 100644
--- a/services/mobile/pom.xml
+++ b/services/mobile/pom.xml
@@ -20,7 +20,7 @@
servicessoftware.amazon.awssdk
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOT4.0.0mobile
diff --git a/services/mq/pom.xml b/services/mq/pom.xml
index e7c6c7a76f4..606a519775d 100644
--- a/services/mq/pom.xml
+++ b/services/mq/pom.xml
@@ -20,7 +20,7 @@
servicessoftware.amazon.awssdk
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOT4.0.0mq
diff --git a/services/mturk/pom.xml b/services/mturk/pom.xml
index 736ba5facd7..2608820fd5e 100644
--- a/services/mturk/pom.xml
+++ b/services/mturk/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmturkAWS Java SDK :: Services :: Amazon Mechanical Turk Requester
diff --git a/services/mwaa/pom.xml b/services/mwaa/pom.xml
index b375d694da4..67b54ab349a 100644
--- a/services/mwaa/pom.xml
+++ b/services/mwaa/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTmwaaAWS Java SDK :: Services :: MWAA
diff --git a/services/neptune/pom.xml b/services/neptune/pom.xml
index a449a093e4f..2f77db6b822 100644
--- a/services/neptune/pom.xml
+++ b/services/neptune/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTneptuneAWS Java SDK :: Services :: Neptune
diff --git a/services/networkfirewall/pom.xml b/services/networkfirewall/pom.xml
index 8d87e920125..e4f0b6981b0 100644
--- a/services/networkfirewall/pom.xml
+++ b/services/networkfirewall/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTnetworkfirewallAWS Java SDK :: Services :: Network Firewall
diff --git a/services/networkmanager/pom.xml b/services/networkmanager/pom.xml
index 8dfdc89bf71..9664a137455 100644
--- a/services/networkmanager/pom.xml
+++ b/services/networkmanager/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTnetworkmanagerAWS Java SDK :: Services :: NetworkManager
diff --git a/services/nimble/pom.xml b/services/nimble/pom.xml
index 37763703551..b94f402c267 100644
--- a/services/nimble/pom.xml
+++ b/services/nimble/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTnimbleAWS Java SDK :: Services :: Nimble
diff --git a/services/oam/pom.xml b/services/oam/pom.xml
index fca1271a882..b596e7b8b4b 100644
--- a/services/oam/pom.xml
+++ b/services/oam/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOToamAWS Java SDK :: Services :: OAM
diff --git a/services/omics/pom.xml b/services/omics/pom.xml
index 3d4ae275202..988db696927 100644
--- a/services/omics/pom.xml
+++ b/services/omics/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTomicsAWS Java SDK :: Services :: Omics
diff --git a/services/omics/src/main/resources/codegen-resources/endpoint-tests.json b/services/omics/src/main/resources/codegen-resources/endpoint-tests.json
index 8ff05e264f8..ce988072093 100644
--- a/services/omics/src/main/resources/codegen-resources/endpoint-tests.json
+++ b/services/omics/src/main/resources/codegen-resources/endpoint-tests.json
@@ -1,55 +1,55 @@
{
"testCases": [
{
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://omics-fips.us-gov-east-1.api.aws"
+ "url": "https://omics-fips.us-east-1.api.aws"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": true,
- "UseDualStack": true,
- "Region": "us-gov-east-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://omics-fips.us-gov-east-1.amazonaws.com"
+ "url": "https://omics-fips.us-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": true,
- "UseDualStack": false,
- "Region": "us-gov-east-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://omics.us-gov-east-1.api.aws"
+ "url": "https://omics.us-east-1.api.aws"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": false,
- "UseDualStack": true,
- "Region": "us-gov-east-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://omics.us-gov-east-1.amazonaws.com"
+ "url": "https://omics.us-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": false,
- "UseDualStack": false,
- "Region": "us-gov-east-1"
+ "UseDualStack": false
}
},
{
@@ -60,9 +60,9 @@
}
},
"params": {
+ "Region": "cn-north-1",
"UseFIPS": true,
- "UseDualStack": true,
- "Region": "cn-north-1"
+ "UseDualStack": true
}
},
{
@@ -73,9 +73,9 @@
}
},
"params": {
+ "Region": "cn-north-1",
"UseFIPS": true,
- "UseDualStack": false,
- "Region": "cn-north-1"
+ "UseDualStack": false
}
},
{
@@ -86,9 +86,9 @@
}
},
"params": {
+ "Region": "cn-north-1",
"UseFIPS": false,
- "UseDualStack": true,
- "Region": "cn-north-1"
+ "UseDualStack": true
}
},
{
@@ -99,109 +99,109 @@
}
},
"params": {
+ "Region": "cn-north-1",
"UseFIPS": false,
- "UseDualStack": false,
- "Region": "cn-north-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
+ "endpoint": {
+ "url": "https://omics-fips.us-gov-east-1.api.aws"
+ }
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": true,
- "UseDualStack": true,
- "Region": "us-iso-east-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://omics-fips.us-iso-east-1.c2s.ic.gov"
+ "url": "https://omics-fips.us-gov-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": true,
- "UseDualStack": false,
- "Region": "us-iso-east-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "error": "DualStack is enabled but this partition does not support DualStack"
+ "endpoint": {
+ "url": "https://omics.us-gov-east-1.api.aws"
+ }
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": false,
- "UseDualStack": true,
- "Region": "us-iso-east-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://omics.us-iso-east-1.c2s.ic.gov"
+ "url": "https://omics.us-gov-east-1.amazonaws.com"
}
},
"params": {
+ "Region": "us-gov-east-1",
"UseFIPS": false,
- "UseDualStack": false,
- "Region": "us-iso-east-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://omics-fips.us-east-1.api.aws"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": true,
- "UseDualStack": true,
- "Region": "us-east-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://omics-fips.us-east-1.amazonaws.com"
+ "url": "https://omics-fips.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": true,
- "UseDualStack": false,
- "Region": "us-east-1"
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://omics.us-east-1.api.aws"
- }
+ "error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": false,
- "UseDualStack": true,
- "Region": "us-east-1"
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://omics.us-east-1.amazonaws.com"
+ "url": "https://omics.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
+ "Region": "us-iso-east-1",
"UseFIPS": false,
- "UseDualStack": false,
- "Region": "us-east-1"
+ "UseDualStack": false
}
},
{
@@ -210,9 +210,9 @@
"error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
+ "Region": "us-isob-east-1",
"UseFIPS": true,
- "UseDualStack": true,
- "Region": "us-isob-east-1"
+ "UseDualStack": true
}
},
{
@@ -223,9 +223,9 @@
}
},
"params": {
+ "Region": "us-isob-east-1",
"UseFIPS": true,
- "UseDualStack": false,
- "Region": "us-isob-east-1"
+ "UseDualStack": false
}
},
{
@@ -234,9 +234,9 @@
"error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
+ "Region": "us-isob-east-1",
"UseFIPS": false,
- "UseDualStack": true,
- "Region": "us-isob-east-1"
+ "UseDualStack": true
}
},
{
@@ -247,13 +247,27 @@
}
},
"params": {
+ "Region": "us-isob-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
+ }
+ },
+ {
+ "documentation": "For custom endpoint with region set and fips disabled and dualstack disabled",
+ "expect": {
+ "endpoint": {
+ "url": "https://example.com"
+ }
+ },
+ "params": {
+ "Region": "us-east-1",
"UseFIPS": false,
"UseDualStack": false,
- "Region": "us-isob-east-1"
+ "Endpoint": "https://example.com"
}
},
{
- "documentation": "For custom endpoint with fips disabled and dualstack disabled",
+ "documentation": "For custom endpoint with region not set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
"url": "https://example.com"
@@ -262,7 +276,6 @@
"params": {
"UseFIPS": false,
"UseDualStack": false,
- "Region": "us-east-1",
"Endpoint": "https://example.com"
}
},
@@ -272,9 +285,9 @@
"error": "Invalid Configuration: FIPS and custom endpoint are not supported"
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": true,
"UseDualStack": false,
- "Region": "us-east-1",
"Endpoint": "https://example.com"
}
},
@@ -284,11 +297,17 @@
"error": "Invalid Configuration: Dualstack and custom endpoint are not supported"
},
"params": {
+ "Region": "us-east-1",
"UseFIPS": false,
"UseDualStack": true,
- "Region": "us-east-1",
"Endpoint": "https://example.com"
}
+ },
+ {
+ "documentation": "Missing region",
+ "expect": {
+ "error": "Invalid Configuration: Missing Region"
+ }
}
],
"version": "1.0"
diff --git a/services/omics/src/main/resources/codegen-resources/service-2.json b/services/omics/src/main/resources/codegen-resources/service-2.json
index 6279f1c3f23..1d25521ae1b 100644
--- a/services/omics/src/main/resources/codegen-resources/service-2.json
+++ b/services/omics/src/main/resources/codegen-resources/service-2.json
@@ -2625,6 +2625,13 @@
"type":"timestamp",
"timestampFormat":"iso8601"
},
+ "CreationType":{
+ "type":"string",
+ "enum":[
+ "IMPORT",
+ "UPLOAD"
+ ]
+ },
"DeleteAnnotationStoreRequest":{
"type":"structure",
"required":["name"],
@@ -3420,6 +3427,10 @@
"statusMessage":{
"shape":"ReadSetStatusMessage",
"documentation":"
The status message for a read set. It provides more detail as to why the read set has a status.
The default maximum number of messages that a single journey can sent to a single endpoint. The maximum value is 100. If set to 0, this limit will not apply.
"
+ }
+ },
+ "documentation": "
The default sending limits for journeys in the application. To override these limits and define custom limits for a specific journey, use the Journey resource.
The default quiet time for campaigns in the application. Quiet time is a specific time range when messages aren't sent to endpoints, if all the following conditions are met:
The EndpointDemographic.Timezone property of the endpoint is set to a valid value.
The current time in the endpoint's time zone is later than or equal to the time specified by the QuietTime.Start property for the application (or a campaign or journey that has custom quiet time settings).
The current time in the endpoint's time zone is earlier than or equal to the time specified by the QuietTime.End property for the application (or a campaign or journey that has custom quiet time settings).
If any of the preceding conditions isn't met, the endpoint will receive messages from a campaign or journey, even if quiet time is enabled.
The default sending limits for journeys in the application. These limits apply to each journey for the application but can be overridden, on a per journey basis, with the JourneyLimits resource.
"
}
},
"documentation": "
Provides information about an application, including the default settings for an application.
The contents of the JSON file provided by Google during registration in order to generate an access token for authentication. For more information see Migrate from legacy FCM APIs to HTTP v1.
"
}
},
- "documentation": "
Specifies the status and settings of the GCM channel for an application. This channel enables Amazon Pinpoint to send push notifications through the Firebase Cloud Messaging (FCM), formerly Google Cloud Messaging (GCM), service.
Specifies the status and settings of the GCM channel for an application. This channel enables Amazon Pinpoint to send push notifications through the Firebase Cloud Messaging (FCM), formerly Google Cloud Messaging (GCM), service.
(Deprecated) An identifier for the GCM channel. This property is retained only for backward compatibility.
"
@@ -9544,7 +9579,6 @@
},
"documentation": "
Provides information about the status and settings of the GCM channel for an application. The GCM channel enables Amazon Pinpoint to send push notifications through the Firebase Cloud Messaging (FCM), formerly Google Cloud Messaging (GCM), service.
para>normal - The notification might be delayed. Delivery is optimized for battery usage on the recipient's device. Use this value unless immediate delivery is required.
/listitem>
high - The notification is sent immediately and might wake a sleeping device.
/para>
Amazon Pinpoint specifies this value in the FCM priority parameter when it sends the notification message to FCM.
The equivalent values for Apple Push Notification service (APNs) are 5, for normal, and 10, for high. If you specify an APNs value for this property, Amazon Pinpoint accepts and converts the value to the corresponding FCM value.
"
+ "documentation": "
para>normal – The notification might be delayed. Delivery is optimized for battery usage on the recipient's device. Use this value unless immediate delivery is required.
/listitem>
high – The notification is sent immediately and might wake a sleeping device.
/para>
Amazon Pinpoint specifies this value in the FCM priority parameter when it sends the notification message to FCM.
The equivalent values for Apple Push Notification service (APNs) are 5, for normal, and 10, for high. If you specify an APNs value for this property, Amazon Pinpoint accepts and converts the value to the corresponding FCM value.
The maximum number of messages that all journeys can send to an endpoint during the specified timeframe. The maximum value is 100. If set to 0, this limit will not apply.
The default quiet time for campaigns in the application. Quiet time is a specific time range when messages aren't sent to endpoints, if all the following conditions are met:
The EndpointDemographic.Timezone property of the endpoint is set to a valid value.
The current time in the endpoint's time zone is later than or equal to the time specified by the QuietTime.Start property for the application (or a campaign or journey that has custom quiet time settings).
The current time in the endpoint's time zone is earlier than or equal to the time specified by the QuietTime.End property for the application (or a campaign or journey that has custom quiet time settings).
If any of the preceding conditions isn't met, the endpoint will receive messages from a campaign or journey, even if quiet time is enabled.
To override the default quiet time settings for a specific campaign or journey, use the Campaign resource or the Journey resource to define a custom quiet time for the campaign or journey.
The default sending limits for journeys in the application. These limits apply to each journey for the application but can be overridden, on a per journey basis, with the JourneyLimits resource.
"
}
},
"documentation": "
Specifies the default settings for an application.
"
@@ -16033,28 +16101,28 @@
"documentation": "
The status of the journey. Valid values are:
DRAFT - Saves the journey and doesn't publish it.
ACTIVE - Saves and publishes the journey. Depending on the journey's schedule, the journey starts running immediately or at the scheduled start time. If a journey's status is ACTIVE, you can't add, change, or remove activities from it.
PAUSED, CANCELLED, COMPLETED, and CLOSED states are not supported in requests to create or update a journey. To cancel, pause, or resume a journey, use the Journey State resource.
Describes the result of an existing snapshot job that has finished running.
A finished snapshot job will return a COMPLETED or FAILED status when you poll the job with a DescribeDashboardSnapshotJob API call.
If the job has not finished running, this operation returns a message that says Dashboard Snapshot Job with id <SnapshotjobId> has not reached a terminal state..
An Asset Bundle import job imports specified Amazon QuickSight assets into an Amazon QuickSight account. You can also choose to import a naming prefix and specified configuration overrides. The assets that are contained in the bundle file that you provide are used to create or update a new or existing asset in your Amazon QuickSight account. Each Amazon QuickSight account can run up to 5 import jobs concurrently.
The API caller must have the necessary \"create\", \"describe\", and \"update\" permissions in their IAM role to access each resource type that is contained in the bundle file before the resources can be imported.
Starts an asynchronous job that generates a dashboard snapshot. You can request one of the following format configurations per API call.
1 paginated PDF
5 CSVs
Poll job descriptions with a DescribeDashboardSnapshotJob API call. Once the job succeeds, use the DescribeDashboardSnapshotJobResult API to obtain the download URIs that the job generates.
A list of SnapshotJobResultFileGroup objects that contain information on the files that are requested during a StartDashboardSnapshotJob API call. If the job succeeds, these objects contain the location where the snapshot artifacts are stored. If the job fails, the objects contain information about the error that caused the job to fail.
"
+ }
+ },
+ "documentation":"
A structure that contains the file groups that are requested for the artifact generation in a StartDashboardSnapshotJob API call.
Used by the UNIQUE_VALUE aggregation function. If there are multiple values for the field used by the aggregation, the value for this property will be returned instead. Defaults to '*'.
The name of your Amazon QuickSight account. This name is unique over all of Amazon Web Services, and it appears only when users sign in. You can't change AccountName value after the Amazon QuickSight account is created.
The result of the snapshot job. Jobs that have successfully completed will return the S3Uri where they are located. Jobs that have failedwill return information on the error that caused the job to fail.
The visibility setting of a pivot table's collapsed row dimension fields. If the value of this structure is HIDDEN, all collapsed columns in a pivot table are automatically hidden. The default value is VISIBLE.
The region that the Amazon S3 bucket is located in. The bucket must be located in the same region that the StartDashboardSnapshotJob API call is made.
"
+ }
+ },
+ "documentation":"
An optional structure that contains the Amazon S3 bucket configuration that the generated snapshots are stored in. If you don't provide this information, generated snapshots are stored in the default Amazon QuickSight bucket.
The tags to be used for row-level security (RLS). Make sure that the relevant datasets have RLS tags configured before you start a snapshot export job. You can configure the RLS tags of a dataset with a DataSet$RowLevelPermissionTagConfiguration API call.
These are not the tags that are used for Amazon Web Services resource tagging. For more information on row level security in Amazon QuickSight, see Using Row-Level Security (RLS) with Tagsin the Amazon QuickSight User Guide.
"
+ }
+ },
+ "documentation":"
A structure that contains information on the anonymous user configuration.
A list of SnapshotJobResultFileGroup objects that contain information about the snapshot that is generated. This list can hold a maximum of 6 FileGroup configurations.
A list of SnapshotS3DestinationConfiguration objects that contain Amazon S3 destination configurations. This structure can hold a maximum of 1 S3DestinationConfiguration.
"
+ }
+ },
+ "documentation":"
A structure that contains information on the Amazon S3 destinations of the generated snapshot.
A list of SnapshotFileSheetSelection objects that contain information on the dashboard sheet that is exported. These objects provide information about the snapshot artifacts that are generated during the job. This structure can hold a maximum of 5 CSV configurations or 1 configuration for PDF.
The format of the snapshot file to be generated. You can choose between CSV or PDF.
"
+ }
+ },
+ "documentation":"
A structure that contains the information for the snapshot that you want to generate. This information is provided by you when you start a new snapshot job.
A list of SnapshotFile objects that contain the information on the snapshot files that need to be generated. This structure can hold 1 configuration at a time.
"
+ }
+ },
+ "documentation":"
A structure that contains the information on the snapshot files.
A list of visual IDs that are located in the selected sheet. This structure supports tables and pivot tables. This structure is required if you are generating a CSV. You can add a maximum of 1 visual ID to this structure.
"
+ }
+ },
+ "documentation":"
A structure that contains information that identifies the snapshot that needs to be generated.
A list of AnonymousUserSnapshotJobResult objects that contain information on anonymous users and their user configurations. This data provided by you when you make a StartDashboardSnapshotJob API call.
"
+ }
+ },
+ "documentation":"
An object that provides information on the result of a snapshot job. This object provides information about the job, the job status, and the location of the generated file.
An array of error records that describe any failures that occur while the dashboard snapshot job runs.
"
+ }
+ },
+ "documentation":"
The Amazon S3 result from the snapshot job. The result includes the DestinationConfiguration and the Amazon S3 Uri. If an error occured during the job, the result returns information on the error.
An ID for the dashboard snapshot job. This ID is unique to the dashboard while the job is running. This ID can be used to poll the status of a job with a DescribeDashboardSnapshotJob while the job runs. You can reuse this ID for another job 24 hours after the current job is completed.
A structure that contains information about the anonymous users that the generated snapshot is for. This API will not return information about registered Amazon QuickSight.
The type of aggregation that is performed on the column data when it's queried. Valid values for this structure are SUM, MAX, MIN, COUNT, DISTINCT_COUNT, and AVERAGE.
"
+ "documentation":"
The type of aggregation that is performed on the column data when it's queried.
Creates an Aurora global database spread across multiple Amazon Web Services Regions. The global database contains a single primary cluster with read-write capability, and a read-only secondary cluster that receives data from the primary cluster through high-speed replication performed by the Aurora storage subsystem.
You can create a global database that is initially empty, and then add a primary cluster and a secondary cluster to it. Or you can specify an existing Aurora cluster during the create operation, and this cluster becomes the primary cluster of the global database.
This operation applies only to Aurora DB clusters.
"
+ "documentation":"
Creates an Aurora global database spread across multiple Amazon Web Services Regions. The global database contains a single primary cluster with read-write capability, and a read-only secondary cluster that receives data from the primary cluster through high-speed replication performed by the Aurora storage subsystem.
You can create a global database that is initially empty, and then create the primary and secondary DB clusters in the global database. Or you can specify an existing Aurora cluster during the create operation, and this cluster becomes the primary cluster of the global database.
This operation applies only to Aurora DB clusters.
The DeleteDBCluster action deletes a previously provisioned DB cluster. When you delete a DB cluster, all automated backups for that DB cluster are deleted and can't be recovered. Manual DB cluster snapshots of the specified DB cluster are not deleted.
If you're deleting a Multi-AZ DB cluster with read replicas, all cluster members are terminated and read replicas are promoted to standalone instances.
For more information on Amazon Aurora, see What is Amazon Aurora? in the Amazon Aurora User Guide.
Displays backups for both current and deleted DB clusters. For example, use this operation to find details about automated backups for previously deleted clusters. Current clusters are returned for both the DescribeDBClusterAutomatedBackups and DescribeDBClusters operations.
Modify a setting for an Amazon Aurora global cluster. You can change one or more database configuration parameters by specifying these parameters and the new values in the request. For more information on Amazon Aurora, see What is Amazon Aurora? in the Amazon Aurora User Guide.
This action only applies to Aurora DB clusters.
"
+ "documentation":"
Modifies a setting for an Amazon Aurora global cluster. You can change one or more database configuration parameters by specifying these parameters and the new values in the request. For more information on Amazon Aurora, see What is Amazon Aurora? in the Amazon Aurora User Guide.
This operation only applies to Aurora global database clusters.
Restores a DB cluster to an arbitrary point in time. Users can restore to any point in time before LatestRestorableTime for up to BackupRetentionPeriod days. The target DB cluster is created from the source DB cluster with the same configuration as the original DB cluster, except that the new DB cluster is created with the default DB security group.
For Aurora, this action only restores the DB cluster, not the DB instances for that DB cluster. You must invoke the CreateDBInstance action to create DB instances for the restored DB cluster, specifying the identifier of the restored DB cluster in DBClusterIdentifier. You can create DB instances only after the RestoreDBClusterToPointInTime action has completed and the DB cluster is available.
For more information on Amazon Aurora DB clusters, see What is Amazon Aurora? in the Amazon Aurora User Guide.
Before you switch over, production traffic is routed to the databases in the blue environment. After you switch over, production traffic is routed to the databases in the green environment.
Before you switch over, production traffic is routed to the databases in the blue environment. After you switch over, production traffic is routed to the databases in the green environment.
The Amazon Web Services KMS key identifier to encrypt a secret that is automatically generated and managed in Amazon Web Services Secrets Manager.
This setting is valid only if the master user password is managed by RDS in Amazon Web Services Secrets Manager for the DB cluster.
The Amazon Web Services KMS key identifier is the key ARN, key ID, alias ARN, or alias name for the KMS key. To use a KMS key in a different Amazon Web Services account, specify the key ARN or alias ARN.
If you don't specify MasterUserSecretKmsKeyId, then the aws/secretsmanager KMS key is used to encrypt the secret. If the secret is in a different Amazon Web Services account, then you can't use the aws/secretsmanager KMS key to encrypt the secret, and you must use a customer managed KMS key.
There is a default KMS key for your Amazon Web Services account. Your Amazon Web Services account has a different default KMS key for each Amazon Web Services Region.
Valid for Cluster Type: Aurora DB clusters and Multi-AZ DB clusters
Specifies whether read replicas can forward write operations to the writer DB instance in the DB cluster. By default, write operations aren't allowed on reader DB instances.
The meaning of this parameter differs depending on the database engine.
Amazon Aurora MySQL
The name of the database to create when the primary DB instance of the Aurora MySQL DB cluster is created. If you don't specify a value, Amazon RDS doesn't create a database in the DB cluster.
Constraints:
Must contain 1 to 64 alphanumeric characters.
Can't be a word reserved by the database engine.
Amazon Aurora PostgreSQL
The name of the database to create when the primary DB instance of the Aurora PostgreSQL DB cluster is created.
Default: postgres
Constraints:
Must contain 1 to 63 alphanumeric characters.
Must begin with a letter. Subsequent characters can be letters, underscores, or digits (0 to 9).
Can't be a word reserved by the database engine.
Amazon RDS Custom for Oracle
The Oracle System ID (SID) of the created RDS Custom DB instance.
Default: ORCL
Constraints:
Must contain 1 to 8 alphanumeric characters.
Must contain a letter.
Can't be a word reserved by the database engine.
Amazon RDS Custom for SQL Server
Not applicable. Must be null.
RDS for MariaDB
The name of the database to create when the DB instance is created. If you don't specify a value, Amazon RDS doesn't create a database in the DB instance.
Constraints:
Must contain 1 to 64 letters or numbers.
Must begin with a letter. Subsequent characters can be letters, underscores, or digits (0-9).
Can't be a word reserved by the database engine.
RDS for MySQL
The name of the database to create when the DB instance is created. If you don't specify a value, Amazon RDS doesn't create a database in the DB instance.
Constraints:
Must contain 1 to 64 letters or numbers.
Must begin with a letter. Subsequent characters can be letters, underscores, or digits (0-9).
Can't be a word reserved by the database engine.
RDS for Oracle
The Oracle System ID (SID) of the created DB instance.
Default: ORCL
Constraints:
Can't be longer than 8 characters.
Can't be a word reserved by the database engine, such as the string NULL.
RDS for PostgreSQL
The name of the database to create when the DB instance is created.
Default: postgres
Constraints:
Must contain 1 to 63 letters, numbers, or underscores.
Must begin with a letter. Subsequent characters can be letters, underscores, or digits (0-9).
Can't be a word reserved by the database engine.
RDS for SQL Server
Not applicable. Must be null.
"
+ "documentation":"
The meaning of this parameter differs according to the database engine you use.
MySQL
The name of the database to create when the DB instance is created. If this parameter isn't specified, no database is created in the DB instance.
Constraints:
Must contain 1 to 64 letters or numbers.
Must begin with a letter. Subsequent characters can be letters, underscores, or digits (0-9).
Can't be a word reserved by the specified database engine
MariaDB
The name of the database to create when the DB instance is created. If this parameter isn't specified, no database is created in the DB instance.
Constraints:
Must contain 1 to 64 letters or numbers.
Must begin with a letter. Subsequent characters can be letters, underscores, or digits (0-9).
Can't be a word reserved by the specified database engine
PostgreSQL
The name of the database to create when the DB instance is created. If this parameter isn't specified, a database named postgres is created in the DB instance.
Constraints:
Must contain 1 to 63 letters, numbers, or underscores.
Must begin with a letter. Subsequent characters can be letters, underscores, or digits (0-9).
Can't be a word reserved by the specified database engine
Oracle
The Oracle System ID (SID) of the created DB instance. If you don't specify a value, the default value is ORCL. You can't specify the string null, or any other reserved word, for DBName.
Default: ORCL
Constraints:
Can't be longer than 8 characters
Amazon RDS Custom for Oracle
The Oracle System ID (SID) of the created RDS Custom DB instance. If you don't specify a value, the default value is ORCL for non-CDBs and RDSCDB for CDBs.
Default: ORCL
Constraints:
It must contain 1 to 8 alphanumeric characters.
It must contain a letter.
It can't be a word reserved by the database engine.
Amazon RDS Custom for SQL Server
Not applicable. Must be null.
SQL Server
Not applicable. Must be null.
Amazon Aurora MySQL
The name of the database to create when the primary DB instance of the Aurora MySQL DB cluster is created. If this parameter isn't specified for an Aurora MySQL DB cluster, no database is created in the DB cluster.
Constraints:
It must contain 1 to 64 alphanumeric characters.
It can't be a word reserved by the database engine.
Amazon Aurora PostgreSQL
The name of the database to create when the primary DB instance of the Aurora PostgreSQL DB cluster is created. If this parameter isn't specified for an Aurora PostgreSQL DB cluster, a database named postgres is created in the DB cluster.
Constraints:
It must contain 1 to 63 alphanumeric characters.
It must begin with a letter. Subsequent characters can be letters, underscores, or digits (0 to 9).
It can't be a word reserved by the database engine.
The Oracle system identifier (SID), which is the name of the Oracle database instance that manages your database files. In this context, the term \"Oracle database instance\" refers exclusively to the system global area (SGA) and Oracle background processes. If you don't specify a SID, the value defaults to RDSCDB. The Oracle SID is also the name of your CDB.
Specifies whether an Aurora DB cluster has in-cluster write forwarding enabled, not enabled, requested, or is in the process of enabling it.
"
}
},
"documentation":"
Contains the details of an Amazon Aurora DB cluster or Multi-AZ DB cluster.
For an Amazon Aurora DB cluster, this data type is used as a response element in the operations CreateDBCluster, DeleteDBCluster, DescribeDBClusters, FailoverDBCluster, ModifyDBCluster, PromoteReadReplicaDBCluster, RestoreDBClusterFromS3, RestoreDBClusterFromSnapshot, RestoreDBClusterToPointInTime, StartDBCluster, and StopDBCluster.
For a Multi-AZ DB cluster, this data type is used as a response element in the operations CreateDBCluster, DeleteDBCluster, DescribeDBClusters, FailoverDBCluster, ModifyDBCluster, RebootDBCluster, RestoreDBClusterFromSnapshot, and RestoreDBClusterToPointInTime.
For more information on Amazon Aurora DB clusters, see What is Amazon Aurora? in the Amazon Aurora User Guide.
For all database engines except Amazon Aurora, AllocatedStorage specifies the allocated storage size in gibibytes (GiB). For Aurora, AllocatedStorage always returns 1, because Aurora DB cluster storage size isn't fixed, but instead automatically adjusts as needed.
The Availability Zones where instances in the DB cluster can be created. For information on Amazon Web Services Regions and Availability Zones, see Regions and Availability Zones.
The IOPS (I/O operations per second) value for the automated backup.
This setting is only for non-Aurora Multi-AZ DB clusters.
"
+ }
+ },
+ "documentation":"
An automated backup of a DB cluster. It consists of system backups, transaction logs, and the database cluster properties that existed at the time you deleted the source cluster.
The pagination token provided in the previous request. If this parameter is specified the response includes only records beyond the marker, up to MaxRecords.
The quota for retained automated backups was exceeded. This prevents you from retaining any additional automated backups. The retained automated backups quota is the same as your DB cluster quota.
A value that indicates whether the DB engine version supports forwarding write operations from reader DB instances to the writer DB instance in the DB cluster. By default, write operations aren't allowed on reader DB instances.
Valid for: Aurora DB clusters only
"
}
},
"documentation":"
This data type is used as a response element in the action DescribeDBEngineVersions.
The meaning of this parameter differs depending on the database engine.
For RDS for MariaDB, Microsoft SQL Server, MySQL, and PostgreSQL - The name of the initial database specified for this DB instance when it was created, if one was provided. This same name is returned for the life of the DB instance.
For RDS for Oracle - The Oracle System ID (SID) of the created DB instance. This value is only returned when the object returned is an Oracle DB instance.
"
+ "documentation":"
Contains the initial database name that you provided (if required) when you created the DB instance. This name is returned for the life of your DB instance. For an RDS for Oracle CDB instance, the name identifies the PDB rather than the CDB.
The progress of the storage optimization operation as a percentage.
"
}
},
"documentation":"
Contains the details of an Amazon RDS DB instance.
This data type is used as a response element in the operations CreateDBInstance, CreateDBInstanceReadReplica, DeleteDBInstance, DescribeDBInstances, ModifyDBInstance, PromoteReadReplica, RebootDBInstance, RestoreDBInstanceFromDBSnapshot, RestoreDBInstanceFromS3, RestoreDBInstanceToPointInTime, StartDBInstance, and StopDBInstance.
The quota for retained automated backups was exceeded. This prevents you from retaining any additional automated backups. The retained automated backups quota is the same as your DB Instance quota.
",
+ "documentation":"
The quota for retained automated backups was exceeded. This prevents you from retaining any additional automated backups. The retained automated backups quota is the same as your DB instance quota.
The Oracle system identifier (SID), which is the name of the Oracle database instance that manages your database files. The Oracle SID is also the name of your CDB.
"
}
},
"documentation":"
Contains the details of an Amazon RDS DB snapshot.
This data type is used as a response element in the DescribeDBSnapshots action.
A value that indicates whether to delete the resources in the green environment. You can't specify this option if the blue/green deployment status is SWITCHOVER_COMPLETED.
"
+ "documentation":"
Specifies whether to delete the resources in the green environment. You can't specify this option if the blue/green deployment status is SWITCHOVER_COMPLETED.
A value that indicates whether to remove automated backups immediately after the DB cluster is deleted. This parameter isn't case-sensitive. The default is to remove automated backups immediately after the DB cluster is deleted.
The blue/green deployment identifier. If this parameter is specified, information from only the specific blue/green deployment is returned. This parameter isn't case-sensitive.
Constraints:
If supplied, must match an existing blue/green deployment identifier.
"
+ "documentation":"
The blue/green deployment identifier. If you specify this parameter, the response only includes information about the specific blue/green deployment. This parameter isn't case-sensitive.
Constraints:
Must match an existing blue/green deployment identifier.
A filter that specifies one or more blue/green deployments to describe.
Supported filters:
blue-green-deployment-identifier - Accepts system-generated identifiers for blue/green deployments. The results list only includes information about the blue/green deployments with the specified identifiers.
blue-green-deployment-name - Accepts user-supplied names for blue/green deployments. The results list only includes information about the blue/green deployments with the specified names.
source - Accepts source databases for a blue/green deployment. The results list only includes information about the blue/green deployments with the specified source databases.
target - Accepts target databases for a blue/green deployment. The results list only includes information about the blue/green deployments with the specified target databases.
"
+ "documentation":"
A filter that specifies one or more blue/green deployments to describe.
Valid Values:
blue-green-deployment-identifier - Accepts system-generated identifiers for blue/green deployments. The results list only includes information about the blue/green deployments with the specified identifiers.
blue-green-deployment-name - Accepts user-supplied names for blue/green deployments. The results list only includes information about the blue/green deployments with the specified names.
source - Accepts source databases for a blue/green deployment. The results list only includes information about the blue/green deployments with the specified source databases.
target - Accepts target databases for a blue/green deployment. The results list only includes information about the blue/green deployments with the specified target databases.
An optional pagination token provided by a previous DescribeBlueGreenDeployments request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.
"
+ "documentation":"
An optional pagination token provided by a previous DescribeBlueGreenDeployments request. If you specify this parameter, the response only includes records beyond the marker, up to the value specified by MaxRecords.
The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so you can retrieve the remaining results.
Default: 100
Constraints: Minimum 20, maximum 100.
"
+ "documentation":"
The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so you can retrieve the remaining results.
(Optional) The user-supplied DB cluster identifier. If this parameter is specified, it must match the identifier of an existing DB cluster. It returns information from the specific DB cluster's automated backup. This parameter isn't case-sensitive.
A filter that specifies which resources to return based on status.
Supported filters are the following:
status
retained - Automated backups for deleted clusters and after backup replication is stopped.
db-cluster-id - Accepts DB cluster identifiers and Amazon Resource Names (ARNs). The results list includes only information about the DB cluster automated backups identified by these ARNs.
db-cluster-resource-id - Accepts DB resource identifiers and Amazon Resource Names (ARNs). The results list includes only information about the DB cluster resources identified by these ARNs.
Returns all resources by default. The status for each resource is specified in the response.
The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that you can retrieve the remaining results.
The pagination token provided in the previous request. If this parameter is specified the response includes only records beyond the marker, up to MaxRecords.
A value that indicates whether to include manual DB cluster snapshots that are public and can be copied or restored by any Amazon Web Services account. By default, the public snapshots are not included.
(Optional) The user-supplied instance identifier. If this parameter is specified, it must match the identifier of an existing DB instance. It returns information from the specific DB instance' automated backup. This parameter isn't case-sensitive.
"
+ "documentation":"
(Optional) The user-supplied instance identifier. If this parameter is specified, it must match the identifier of an existing DB instance. It returns information from the specific DB instance's automated backup. This parameter isn't case-sensitive.
A filter that specifies which resources to return based on status.
Supported filters are the following:
status
active - automated backups for current instances
retained - automated backups for deleted instances and after backup replication is stopped
creating - automated backups that are waiting for the first automated snapshot to be available
db-instance-id - Accepts DB instance identifiers and Amazon Resource Names (ARNs). The results list includes only information about the DB instance automated backups identified by these ARNs.
dbi-resource-id - Accepts DB resource identifiers and Amazon Resource Names (ARNs). The results list includes only information about the DB instance resources identified by these ARNs.
Returns all resources by default. The status for each resource is specified in the response.
"
+ "documentation":"
A filter that specifies which resources to return based on status.
Supported filters are the following:
status
active - Automated backups for current instances.
creating - Automated backups that are waiting for the first automated snapshot to be available.
retained - Automated backups for deleted instances and after backup replication is stopped.
db-instance-id - Accepts DB instance identifiers and Amazon Resource Names (ARNs). The results list includes only information about the DB instance automated backups identified by these ARNs.
dbi-resource-id - Accepts DB resource identifiers and Amazon Resource Names (ARNs). The results list includes only information about the DB instance resources identified by these ARNs.
Returns all resources by default. The status for each resource is specified in the response.
Specifies whether read replicas can forward write operations to the writer DB instance in the DB cluster. By default, write operations aren't allowed on reader DB instances.
Indicates if the global database cluster has deletion protection enabled. The global database cluster can't be deleted when deletion protection is enabled.
"
+ "documentation":"
Specifies whether to enable deletion protection for the global database cluster. The global database cluster can't be deleted when deletion protection is enabled.
The version number of the database engine to which you want to upgrade. Changing this parameter results in an outage. The change is applied during the next maintenance window unless ApplyImmediately is enabled.
To list all of the available engine versions for aurora-mysql (for MySQL-based Aurora global databases), use the following command:
A value that indicates whether major version upgrades are allowed.
Constraints: You must allow major version upgrades when specifying a value for the EngineVersion parameter that is a different major version than the DB cluster's current version.
If you upgrade the major version of a global database, the cluster and DB instance parameter groups are set to the default parameter groups for the new version. Apply any custom parameter groups after completing the upgrade.
"
+ "documentation":"
Specifies whether to allow major version upgrades.
Constraints: Must be enabled if you specify a value for the EngineVersion parameter that's a different major version than the global cluster's current version.
If you upgrade the major version of a global database, the cluster and DB instance parameter groups are set to the default parameter groups for the new version. Apply any custom parameter groups after completing the upgrade.
The network type is determined by the DBSubnetGroup specified for the DB cluster. A DBSubnetGroup can support only the IPv4 protocol or the IPv4 and the IPv6 protocols (DUAL).
A value that indicates whether the target engine version supports forwarding write operations from reader DB instances to the writer DB instance in the DB cluster. By default, write operations aren't allowed on reader DB instances.
Valid for: Aurora DB clusters only
"
}
},
"documentation":"
The version of the database engine that a DB instance can be upgraded to.
Creates a new Amazon Rekognition Custom Labels dataset. You can create a dataset by using an Amazon Sagemaker format manifest file or by copying an existing Amazon Rekognition Custom Labels dataset.
To create a training dataset for a project, specify train for the value of DatasetType. To create the test dataset for a project, specify test for the value of DatasetType.
The response from CreateDataset is the Amazon Resource Name (ARN) for the dataset. Creating a dataset takes a while to complete. Use DescribeDataset to check the current status. The dataset created successfully if the value of Status is CREATE_COMPLETE.
To check if any non-terminal errors occurred, call ListDatasetEntries and check for the presence of errors lists in the JSON Lines.
Dataset creation fails if a terminal error occurs (Status = CREATE_FAILED). Currently, you can't access the terminal error information.
For more information, see Creating dataset in the Amazon Rekognition Custom Labels Developer Guide.
This operation requires permissions to perform the rekognition:CreateDataset action. If you want to copy an existing dataset, you also require permission to perform the rekognition:ListDatasetEntries action.
"
+ "documentation":"
Creates a new Amazon Rekognition Custom Labels dataset. You can create a dataset by using an Amazon Sagemaker format manifest file or by copying an existing Amazon Rekognition Custom Labels dataset.
To create a training dataset for a project, specify TRAIN for the value of DatasetType. To create the test dataset for a project, specify TEST for the value of DatasetType.
The response from CreateDataset is the Amazon Resource Name (ARN) for the dataset. Creating a dataset takes a while to complete. Use DescribeDataset to check the current status. The dataset created successfully if the value of Status is CREATE_COMPLETE.
To check if any non-terminal errors occurred, call ListDatasetEntries and check for the presence of errors lists in the JSON Lines.
Dataset creation fails if a terminal error occurs (Status = CREATE_FAILED). Currently, you can't access the terminal error information.
For more information, see Creating dataset in the Amazon Rekognition Custom Labels Developer Guide.
This operation requires permissions to perform the rekognition:CreateDataset action. If you want to copy an existing dataset, you also require permission to perform the rekognition:ListDatasetEntries action.
This API operation initiates a Face Liveness session. It returns a SessionId, which you can use to start streaming Face Liveness video and get the results for a Face Liveness session. You can use the OutputConfig option in the Settings parameter to provide an Amazon S3 bucket location. The Amazon S3 bucket stores reference images and audit images. You can use AuditImagesLimit to limit the number of audit images returned. This number is between 0 and 4. By default, it is set to 0. The limit is best effort and based on the duration of the selfie-video.
",
+ "documentation":"
This API operation initiates a Face Liveness session. It returns a SessionId, which you can use to start streaming Face Liveness video and get the results for a Face Liveness session.
You can use the OutputConfig option in the Settings parameter to provide an Amazon S3 bucket location. The Amazon S3 bucket stores reference images and audit images. If no Amazon S3 bucket is defined, raw bytes are sent instead.
You can use AuditImagesLimit to limit the number of audit images returned when GetFaceLivenessSessionResults is called. This number is between 0 and 4. By default, it is set to 0. The limit is best effort and based on the duration of the selfie-video.
Detects instances of real-world entities within an image (JPEG or PNG) provided as input. This includes objects like flower, tree, and table; events like wedding, graduation, and birthday party; and concepts like landscape, evening, and nature.
For an example, see Analyzing images stored in an Amazon S3 bucket in the Amazon Rekognition Developer Guide.
You pass the input image as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.
Optional Parameters
You can specify one or both of the GENERAL_LABELS and IMAGE_PROPERTIES feature types when calling the DetectLabels API. Including GENERAL_LABELS will ensure the response includes the labels detected in the input image, while including IMAGE_PROPERTIES will ensure the response includes information about the image quality and color.
When using GENERAL_LABELS and/or IMAGE_PROPERTIES you can provide filtering criteria to the Settings parameter. You can filter with sets of individual labels or with label categories. You can specify inclusive filters, exclusive filters, or a combination of inclusive and exclusive filters. For more information on filtering see Detecting Labels in an Image.
You can specify MinConfidence to control the confidence threshold for the labels returned. The default is 55%. You can also add the MaxLabels parameter to limit the number of labels returned. The default and upper limit is 1000 labels.
Response Elements
For each object, scene, and concept the API returns one or more labels. The API returns the following types of information about labels:
Name - The name of the detected label.
Confidence - The level of confidence in the label assigned to a detected object.
Parents - The ancestor labels for a detected label. DetectLabels returns a hierarchical taxonomy of detected labels. For example, a detected car might be assigned the label car. The label car has two parent labels: Vehicle (its parent) and Transportation (its grandparent). The response includes the all ancestors for a label, where every ancestor is a unique label. In the previous example, Car, Vehicle, and Transportation are returned as unique labels in the response.
Aliases - Possible Aliases for the label.
Categories - The label categories that the detected label belongs to.
BoundingBox — Bounding boxes are described for all instances of detected common object labels, returned in an array of Instance objects. An Instance object contains a BoundingBox object, describing the location of the label on the input image. It also includes the confidence for the accuracy of the detected bounding box.
The API returns the following information regarding the image, as part of the ImageProperties structure:
Quality - Information about the Sharpness, Brightness, and Contrast of the input image, scored between 0 to 100. Image quality is returned for the entire image, as well as the background and the foreground.
Dominant Color - An array of the dominant colors in the image.
Foreground - Information about the sharpness, brightness, and dominant colors of the input image’s foreground.
Background - Information about the sharpness, brightness, and dominant colors of the input image’s background.
The list of returned labels will include at least one label for every detected object, along with information about that label. In the following example, suppose the input image has a lighthouse, the sea, and a rock. The response includes all three labels, one for each object, as well as the confidence in the label:
{Name: lighthouse, Confidence: 98.4629}
{Name: rock,Confidence: 79.2097}
{Name: sea,Confidence: 75.061}
The list of labels can include multiple labels for the same object. For example, if the input image shows a flower (for example, a tulip), the operation might return the following three labels.
{Name: flower,Confidence: 99.0562}
{Name: plant,Confidence: 99.0562}
{Name: tulip,Confidence: 99.0562}
In this example, the detection algorithm more precisely identifies the flower as a tulip.
If the object detected is a person, the operation doesn't provide the same facial details that the DetectFaces operation provides.
This is a stateless API operation that doesn't return any data.
This operation requires permissions to perform the rekognition:DetectLabels action.
"
+ "documentation":"
Detects instances of real-world entities within an image (JPEG or PNG) provided as input. This includes objects like flower, tree, and table; events like wedding, graduation, and birthday party; and concepts like landscape, evening, and nature.
For an example, see Analyzing images stored in an Amazon S3 bucket in the Amazon Rekognition Developer Guide.
You pass the input image as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.
Optional Parameters
You can specify one or both of the GENERAL_LABELS and IMAGE_PROPERTIES feature types when calling the DetectLabels API. Including GENERAL_LABELS will ensure the response includes the labels detected in the input image, while including IMAGE_PROPERTIES will ensure the response includes information about the image quality and color.
When using GENERAL_LABELS and/or IMAGE_PROPERTIES you can provide filtering criteria to the Settings parameter. You can filter with sets of individual labels or with label categories. You can specify inclusive filters, exclusive filters, or a combination of inclusive and exclusive filters. For more information on filtering see Detecting Labels in an Image.
When getting labels, you can specify MinConfidence to control the confidence threshold for the labels returned. The default is 55%. You can also add the MaxLabels parameter to limit the number of labels returned. The default and upper limit is 1000 labels. These arguments are only valid when supplying GENERAL_LABELS as a feature type.
Response Elements
For each object, scene, and concept the API returns one or more labels. The API returns the following types of information about labels:
Name - The name of the detected label.
Confidence - The level of confidence in the label assigned to a detected object.
Parents - The ancestor labels for a detected label. DetectLabels returns a hierarchical taxonomy of detected labels. For example, a detected car might be assigned the label car. The label car has two parent labels: Vehicle (its parent) and Transportation (its grandparent). The response includes the all ancestors for a label, where every ancestor is a unique label. In the previous example, Car, Vehicle, and Transportation are returned as unique labels in the response.
Aliases - Possible Aliases for the label.
Categories - The label categories that the detected label belongs to.
BoundingBox — Bounding boxes are described for all instances of detected common object labels, returned in an array of Instance objects. An Instance object contains a BoundingBox object, describing the location of the label on the input image. It also includes the confidence for the accuracy of the detected bounding box.
The API returns the following information regarding the image, as part of the ImageProperties structure:
Quality - Information about the Sharpness, Brightness, and Contrast of the input image, scored between 0 to 100. Image quality is returned for the entire image, as well as the background and the foreground.
Dominant Color - An array of the dominant colors in the image.
Foreground - Information about the sharpness, brightness, and dominant colors of the input image’s foreground.
Background - Information about the sharpness, brightness, and dominant colors of the input image’s background.
The list of returned labels will include at least one label for every detected object, along with information about that label. In the following example, suppose the input image has a lighthouse, the sea, and a rock. The response includes all three labels, one for each object, as well as the confidence in the label:
{Name: lighthouse, Confidence: 98.4629}
{Name: rock,Confidence: 79.2097}
{Name: sea,Confidence: 75.061}
The list of labels can include multiple labels for the same object. For example, if the input image shows a flower (for example, a tulip), the operation might return the following three labels.
{Name: flower,Confidence: 99.0562}
{Name: plant,Confidence: 99.0562}
{Name: tulip,Confidence: 99.0562}
In this example, the detection algorithm more precisely identifies the flower as a tulip.
If the object detected is a person, the operation doesn't provide the same facial details that the DetectFaces operation provides.
This is a stateless API operation that doesn't return any data.
This operation requires permissions to perform the rekognition:DetectLabels action.
Gets face detection results for a Amazon Rekognition Video analysis started by StartFaceDetection.
Face detection with Amazon Rekognition Video is an asynchronous operation. You start face detection by calling StartFaceDetection which returns a job identifier (JobId). When the face detection operation finishes, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartFaceDetection. To get the results of the face detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetFaceDetection and pass the job identifier (JobId) from the initial call to StartFaceDetection.
GetFaceDetection returns an array of detected faces (Faces) sorted by the time the faces were detected.
Use MaxResults parameter to limit the number of labels returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetFaceDetection and populate the NextToken request parameter with the token value returned from the previous call to GetFaceDetection.
"
+ "documentation":"
Gets face detection results for a Amazon Rekognition Video analysis started by StartFaceDetection.
Face detection with Amazon Rekognition Video is an asynchronous operation. You start face detection by calling StartFaceDetection which returns a job identifier (JobId). When the face detection operation finishes, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartFaceDetection. To get the results of the face detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetFaceDetection and pass the job identifier (JobId) from the initial call to StartFaceDetection.
GetFaceDetection returns an array of detected faces (Faces) sorted by the time the faces were detected.
Use MaxResults parameter to limit the number of labels returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetFaceDetection and populate the NextToken request parameter with the token value returned from the previous call to GetFaceDetection.
Note that for the GetFaceDetection operation, the returned values for FaceOccluded and EyeDirection will always be \"null\".
Retrieves the results of a specific Face Liveness session. It requires the sessionId as input, which was created using CreateFaceLivenessSession. Returns the corresponding Face Liveness confidence score, a reference image that includes a face bounding box, and audit images that also contain face bounding boxes. The Face Liveness confidence score ranges from 0 to 100. The reference image can optionally be returned.
"
+ "documentation":"
Retrieves the results of a specific Face Liveness session. It requires the sessionId as input, which was created using CreateFaceLivenessSession. Returns the corresponding Face Liveness confidence score, a reference image that includes a face bounding box, and audit images that also contain face bounding boxes. The Face Liveness confidence score ranges from 0 to 100.
The number of audit images returned by GetFaceLivenessSessionResults is defined by the AuditImagesLimit paramater when calling CreateFaceLivenessSession. Reference images are always returned when possible.
Gets the text detection results of a Amazon Rekognition Video analysis started by StartTextDetection.
Text detection with Amazon Rekognition Video is an asynchronous operation. You start text detection by calling StartTextDetection which returns a job identifier (JobId) When the text detection operation finishes, Amazon Rekognition publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartTextDetection. To get the results of the text detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. if so, call GetTextDetection and pass the job identifier (JobId) from the initial call of StartLabelDetection.
GetTextDetection returns an array of detected text (TextDetections) sorted by the time the text was detected, up to 50 words per frame of video.
Each element of the array includes the detected text, the precentage confidence in the acuracy of the detected text, the time the text was detected, bounding box information for where the text was located, and unique identifiers for words and their lines.
Use MaxResults parameter to limit the number of text detections returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetTextDetection and populate the NextToken request parameter with the token value returned from the previous call to GetTextDetection.
"
+ "documentation":"
Gets the text detection results of a Amazon Rekognition Video analysis started by StartTextDetection.
Text detection with Amazon Rekognition Video is an asynchronous operation. You start text detection by calling StartTextDetection which returns a job identifier (JobId) When the text detection operation finishes, Amazon Rekognition publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartTextDetection. To get the results of the text detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. if so, call GetTextDetection and pass the job identifier (JobId) from the initial call of StartLabelDetection.
GetTextDetection returns an array of detected text (TextDetections) sorted by the time the text was detected, up to 100 words per frame of video.
Each element of the array includes the detected text, the precentage confidence in the acuracy of the detected text, the time the text was detected, bounding box information for where the text was located, and unique identifiers for words and their lines.
Use MaxResults parameter to limit the number of text detections returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetTextDetection and populate the NextToken request parameter with the token value returned from the previous call to GetTextDetection.
An array of facial attributes you want to be returned. A DEFAULT subset of facial attributes - BoundingBox, Confidence, Pose, Quality, and Landmarks - will always be returned. You can request for specific facial attributes (in addition to the default list) - by using [\"DEFAULT\", \"FACE_OCCLUDED\"] or just [\"FACE_OCCLUDED\"]. You can request for all facial attributes by using [\"ALL\"]. Requesting more attributes may increase response time.
If you provide both, [\"ALL\", \"DEFAULT\"], the service uses a logical \"AND\" operator to determine which attributes to return (in this case, all attributes).
"
+ "documentation":"
An array of facial attributes you want to be returned. A DEFAULT subset of facial attributes - BoundingBox, Confidence, Pose, Quality, and Landmarks - will always be returned. You can request for specific facial attributes (in addition to the default list) - by using [\"DEFAULT\", \"FACE_OCCLUDED\"] or just [\"FACE_OCCLUDED\"]. You can request for all facial attributes by using [\"ALL\"]. Requesting more attributes may increase response time.
If you provide both, [\"ALL\", \"DEFAULT\"], the service uses a logical \"AND\" operator to determine which attributes to return (in this case, all attributes).
Note that while the FaceOccluded and EyeDirection attributes are supported when using DetectFaces, they aren't supported when analyzing videos with StartFaceDetection and GetFaceDetection.
Maximum number of labels you want the service to return in the response. The service returns the specified number of highest confidence labels.
"
+ "documentation":"
Maximum number of labels you want the service to return in the response. The service returns the specified number of highest confidence labels. Only valid when GENERAL_LABELS is specified as a feature type in the Feature input parameter.
Specifies the minimum confidence level for the labels to return. Amazon Rekognition doesn't return any labels with confidence lower than this specified value.
If MinConfidence is not specified, the operation returns labels with a confidence values greater than or equal to 55 percent.
"
+ "documentation":"
Specifies the minimum confidence level for the labels to return. Amazon Rekognition doesn't return any labels with confidence lower than this specified value.
If MinConfidence is not specified, the operation returns labels with a confidence values greater than or equal to 55 percent. Only valid when GENERAL_LABELS is specified as a feature type in the Feature input parameter.
A set of images from the Face Liveness video that can be used for audit purposes. It includes a bounding box of the face and the Base64-encoded bytes that return an image. If the CreateFaceLivenessSession request included an OutputConfig argument, the image will be uploaded to an S3Object specified in the output configuration.
"
+ "documentation":"
A set of images from the Face Liveness video that can be used for audit purposes. It includes a bounding box of the face and the Base64-encoded bytes that return an image. If the CreateFaceLivenessSession request included an OutputConfig argument, the image will be uploaded to an S3Object specified in the output configuration. If no Amazon S3 bucket is defined, raw bytes are sent instead.
Creates an Resilience Hub application. An Resilience Hub application is a collection of Amazon Web Services resources structured to prevent and recover Amazon Web Services application disruptions. To describe an Resilience Hub application, you provide an application name, resources from one or more CloudFormation stacks, Resource Groups, Terraform state files, AppRegistry applications, and an appropriate resiliency policy. For more information about the number of resources supported per application, see Service Quotas.
After you create an Resilience Hub application, you publish it so that you can run a resiliency assessment on it. You can then use recommendations from the assessment to improve resiliency by running another assessment, comparing results, and then iterating the process until you achieve your goals for recovery time objective (RTO) and recovery point objective (RPO).
"
+ "documentation":"
Creates an Resilience Hub application. An Resilience Hub application is a collection of Amazon Web Services resources structured to prevent and recover Amazon Web Services application disruptions. To describe a Resilience Hub application, you provide an application name, resources from one or more CloudFormation stacks, Resource Groups, Terraform state files, AppRegistry applications, and an appropriate resiliency policy. In addition, you can also add resources that are located on Amazon Elastic Kubernetes Service (Amazon EKS) clusters as optional resources. For more information about the number of resources supported per application, see Service quotas.
After you create an Resilience Hub application, you publish it so that you can run a resiliency assessment on it. You can then use recommendations from the assessment to improve resiliency by running another assessment, comparing results, and then iterating the process until you achieve your goals for recovery time objective (RTO) and recovery point objective (RPO).
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
Mappings used to map logical resources from the template to physical resources. You can use the mapping type CFN_STACK if the application template uses a logical stack name. Or you can map individual resources by using the mapping type RESOURCE. We recommend using the mapping type CFN_STACK if the application is backed by a CloudFormation stack.
"
+ "documentation":"
List of sources that are used to map a logical resource from the template to a physical resource. You can use sources such as CloudFormation, Terraform state files, AppRegistry applications, or Amazon EKS.
The Application Component for the CloudWatch alarm recommendation.
"
+ "documentation":"
Application Component name for the CloudWatch alarm recommendation. This name is saved as the first item in the appComponentNames list.
",
+ "deprecated":true,
+ "deprecatedMessage":"An alarm recommendation can be attached to multiple Application Components, hence this property will be replaced by the new property 'appComponentNames'."
+ },
+ "appComponentNames":{
+ "shape":"AppComponentNameList",
+ "documentation":"
List of Application Component names for the CloudWatch alarm recommendation.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The list of events you would like to subscribe and get notification for. Currently, Resilience Hub supports notifications only for Drift detected and Scheduled assessment failure events.
The Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
Difference type between actual and expected recovery point objective (RPO) and recovery time objective (RTO) values. Currently, Resilience Hub supports only NotEqual difference type.
The type of entity in which drift was detected. For compliance drifts, Resilience Hub supports AWS::ResilienceHub::AppComponent and AWS::ResilienceHub::Application.
Assessment identifier of a previous assessment of the same application version. Resilience Hub uses the previous assessment (associated with the reference identifier) to compare the compliance with the current assessment to identify drifts.
Defines recommendations for an Resilience Hub Application Component, returned as an object. This object contains component names, configuration recommendations, and recommendation statuses.
The list of events you would like to subscribe and get notification for. Currently, Resilience Hub supports only Drift detected and Scheduled assessment failure events notification.
The Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The list of Application Components that this resource belongs to. If an Application Component is not part of the Resilience Hub application, it will be added.
"
+ "documentation":"
List of Application Components that this resource belongs to. If an Application Component is not part of the Resilience Hub application, it will be added.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
A JSON string that provides information about your application structure. To learn more about the appTemplateBody template, see the sample template provided in the Examples section.
The appTemplateBody JSON string has the following structure:
resources
The list of logical resources that must be included in the Resilience Hub application.
Type: Array
Don't add the resources that you want to exclude.
Each resources array item includes the following fields:
logicalResourceId
The logical identifier of the resource.
Type: Object
Each logicalResourceId object includes the following fields:
identifier
The identifier of the resource.
Type: String
logicalStackName
The name of the CloudFormation stack this resource belongs to.
Type: String
resourceGroupName
The name of the resource group this resource belongs to.
Type: String
terraformSourceName
The name of the Terraform S3 state file this resource belongs to.
Type: String
eksSourceName
The name of the Amazon Elastic Kubernetes Service cluster and namespace this resource belongs to.
This parameter accepts values in \"eks-cluster/namespace\" format.
Type: String
type
The type of resource.
Type: string
name
The name of the resource.
Type: String
additionalInfo
Additional configuration parameters for an Resilience Hub application. If you want to implement additionalInfo through the Resilience Hub console rather than using an API call, see Configure the application configuration parameters.
Currently, this parameter accepts a key-value mapping (in a string format) of only one failover region and one associated account.
The list of Application Components that this resource belongs to. If an Application Component is not part of the Resilience Hub application, it will be added.
Type: Array
Each appComponents array item includes the following fields:
The list of included resources that are assigned to the Application Component.
Type: Array of strings
additionalInfo
Additional configuration parameters for an Resilience Hub application. If you want to implement additionalInfo through the Resilience Hub console rather than using an API call, see Configure the application configuration parameters.
Currently, this parameter accepts a key-value mapping (in a string format) of only one failover region and one associated account.
The list of logical resource identifiers to be excluded from the application.
Type: Array
Don't add the resources that you want to include.
Each excludedResources array item includes the following fields:
logicalResourceIds
The logical identifier of the resource.
Type: Object
You can configure only one of the following fields:
logicalStackName
resourceGroupName
terraformSourceName
eksSourceName
Each logicalResourceIds object includes the following fields:
identifier
The identifier of the resource.
Type: String
logicalStackName
The name of the CloudFormation stack this resource belongs to.
Type: String
resourceGroupName
The name of the resource group this resource belongs to.
Type: String
terraformSourceName
The name of the Terraform S3 state file this resource belongs to.
Type: String
eksSourceName
The name of the Amazon Elastic Kubernetes Service cluster and namespace this resource belongs to.
This parameter accepts values in \"eks-cluster/namespace\" format.
Type: String
version
The Resilience Hub application version.
additionalInfo
Additional configuration parameters for an Resilience Hub application. If you want to implement additionalInfo through the Resilience Hub console rather than using an API call, see Configure the application configuration parameters.
Currently, this parameter accepts a key-value mapping (in a string format) of only one failover region and one associated account.
A JSON string that provides information about your application structure. To learn more about the appTemplateBody template, see the sample template provided in the Examples section.
The appTemplateBody JSON string has the following structure:
resources
The list of logical resources that must be included in the Resilience Hub application.
Type: Array
Don't add the resources that you want to exclude.
Each resources array item includes the following fields:
logicalResourceId
Logical identifier of the resource.
Type: Object
Each logicalResourceId object includes the following fields:
identifier
Identifier of the resource.
Type: String
logicalStackName
The name of the CloudFormation stack this resource belongs to.
Type: String
resourceGroupName
The name of the resource group this resource belongs to.
Type: String
terraformSourceName
The name of the Terraform S3 state file this resource belongs to.
Type: String
eksSourceName
Name of the Amazon Elastic Kubernetes Service cluster and namespace this resource belongs to.
This parameter accepts values in \"eks-cluster/namespace\" format.
Type: String
type
The type of resource.
Type: string
name
The name of the resource.
Type: String
additionalInfo
Additional configuration parameters for an Resilience Hub application. If you want to implement additionalInfo through the Resilience Hub console rather than using an API call, see Configure the application configuration parameters.
Currently, this parameter accepts a key-value mapping (in a string format) of only one failover region and one associated account.
List of Application Components that this resource belongs to. If an Application Component is not part of the Resilience Hub application, it will be added.
Type: Array
Each appComponents array item includes the following fields:
The list of included resources that are assigned to the Application Component.
Type: Array of strings
additionalInfo
Additional configuration parameters for an Resilience Hub application. If you want to implement additionalInfo through the Resilience Hub console rather than using an API call, see Configure the application configuration parameters.
Currently, this parameter accepts a key-value mapping (in a string format) of only one failover region and one associated account.
The list of logical resource identifiers to be excluded from the application.
Type: Array
Don't add the resources that you want to include.
Each excludedResources array item includes the following fields:
logicalResourceIds
Logical identifier of the resource.
Type: Object
You can configure only one of the following fields:
logicalStackName
resourceGroupName
terraformSourceName
eksSourceName
Each logicalResourceIds object includes the following fields:
identifier
Identifier of the resource.
Type: String
logicalStackName
The name of the CloudFormation stack this resource belongs to.
Type: String
resourceGroupName
The name of the resource group this resource belongs to.
Type: String
terraformSourceName
The name of the Terraform S3 state file this resource belongs to.
Type: String
eksSourceName
Name of the Amazon Elastic Kubernetes Service cluster and namespace this resource belongs to.
This parameter accepts values in \"eks-cluster/namespace\" format.
Type: String
version
Resilience Hub application version.
additionalInfo
Additional configuration parameters for an Resilience Hub application. If you want to implement additionalInfo through the Resilience Hub console rather than using an API call, see Configure the application configuration parameters.
Currently, this parameter accepts a key-value mapping (in a string format) of only one failover region and one associated account.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Amazon Elastic Kubernetes Service cluster. The format for this ARN is: arn:aws:eks:region:account-id:cluster/cluster-name. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Amazon Elastic Kubernetes Service cluster. The format for this ARN is: arn:aws:eks:region:account-id:cluster/cluster-name. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Amazon Elastic Kubernetes Service cluster. The format for this ARN is: arn:aws:eks:region:account-id:cluster/cluster-name. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Amazon Elastic Kubernetes Service cluster. The format for this ARN is: arn:aws:eks:region:account-id:cluster/cluster-name. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The type of event you would like to subscribe and get notification for. Currently, Resilience Hub supports notifications only for Drift detected (DriftDetected) and Scheduled assessment failure (ScheduledAssessmentFailure) events.
Amazon Resource Name (ARN) of the Amazon Simple Notification Service topic. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ }
+ },
+ "documentation":"
Indicates an event you would like to subscribe and get notification for. Currently, Resilience Hub supports notifications only for Drift detected and Scheduled assessment failure events.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
"
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
",
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
",
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
"
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
"
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
"
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
"
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
"
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
",
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
",
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
",
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
",
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
",
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
"
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
",
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
"
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
"
+ "documentation":"
Maximum number of results to include in the response. If more results exist than the specified MaxResults value, a token is included in the response so that the remaining results can be retrieved.
Defines a list of role Amazon Resource Names (ARNs) to be used in other accounts. These ARNs are used for querying purposes while importing resources and assessing your application.
These ARNs are required only when your resources are in other accounts and you have different role name in these accounts. Else, the invoker role name will be used in the other accounts.
These roles must have a trust policy with iam:AssumeRole permission to the invoker role in the primary account.
Existing Amazon Web Services IAM role name in the primary Amazon Web Services account that will be assumed by Resilience Hub Service Principle to obtain a read-only access to your application resources while running an assessment.
You must have iam:passRole permission for this role while creating or updating the application.
Defines how Resilience Hub scans your resources. It can scan for the resources by using a pre-existing role in your Amazon Web Services account, or by using the credentials of the current IAM user.
"
+ }
+ },
+ "documentation":"
Defines the roles and credentials that Resilience Hub would use while creating the application, importing its resources, and running an assessment.
Defines a physical resource. A physical resource is a resource that exists in your account. It can be identified using an Amazon Resource Name (ARN) or an Resilience Hub-native identifier.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
A JSON string that provides information about your application structure. To learn more about the appTemplateBody template, see the sample template provided in the Examples section.
The appTemplateBody JSON string has the following structure:
resources
The list of logical resources that must be included in the Resilience Hub application.
Type: Array
Don't add the resources that you want to exclude.
Each resources array item includes the following fields:
logicalResourceId
The logical identifier of the resource.
Type: Object
Each logicalResourceId object includes the following fields:
identifier
The identifier of the resource.
Type: String
logicalStackName
The name of the CloudFormation stack this resource belongs to.
Type: String
resourceGroupName
The name of the resource group this resource belongs to.
Type: String
terraformSourceName
The name of the Terraform S3 state file this resource belongs to.
Type: String
eksSourceName
The name of the Amazon Elastic Kubernetes Service cluster and namespace this resource belongs to.
This parameter accepts values in \"eks-cluster/namespace\" format.
Type: String
type
The type of resource.
Type: string
name
The name of the resource.
Type: String
additionalInfo
Additional configuration parameters for an Resilience Hub application. If you want to implement additionalInfo through the Resilience Hub console rather than using an API call, see Configure the application configuration parameters.
Currently, this parameter accepts a key-value mapping (in a string format) of only one failover region and one associated account.
The list of Application Components that this resource belongs to. If an Application Component is not part of the Resilience Hub application, it will be added.
Type: Array
Each appComponents array item includes the following fields:
The list of included resources that are assigned to the Application Component.
Type: Array of strings
additionalInfo
Additional configuration parameters for an Resilience Hub application. If you want to implement additionalInfo through the Resilience Hub console rather than using an API call, see Configure the application configuration parameters.
Currently, this parameter accepts a key-value mapping (in a string format) of only one failover region and one associated account.
The list of logical resource identifiers to be excluded from the application.
Type: Array
Don't add the resources that you want to include.
Each excludedResources array item includes the following fields:
logicalResourceIds
The logical identifier of the resource.
Type: Object
You can configure only one of the following fields:
logicalStackName
resourceGroupName
terraformSourceName
eksSourceName
Each logicalResourceIds object includes the following fields:
identifier
The identifier of the resource.
Type: String
logicalStackName
The name of the CloudFormation stack this resource belongs to.
Type: String
resourceGroupName
The name of the resource group this resource belongs to.
Type: String
terraformSourceName
The name of the Terraform S3 state file this resource belongs to.
Type: String
eksSourceName
The name of the Amazon Elastic Kubernetes Service cluster and namespace this resource belongs to.
This parameter accepts values in \"eks-cluster/namespace\" format.
Type: String
version
The Resilience Hub application version.
additionalInfo
Additional configuration parameters for an Resilience Hub application. If you want to implement additionalInfo through the Resilience Hub console rather than using an API call, see Configure the application configuration parameters.
Currently, this parameter accepts a key-value mapping (in a string format) of only one failover region and one associated account.
A JSON string that provides information about your application structure. To learn more about the appTemplateBody template, see the sample template provided in the Examples section.
The appTemplateBody JSON string has the following structure:
resources
The list of logical resources that must be included in the Resilience Hub application.
Type: Array
Don't add the resources that you want to exclude.
Each resources array item includes the following fields:
logicalResourceId
Logical identifier of the resource.
Type: Object
Each logicalResourceId object includes the following fields:
identifier
Identifier of the resource.
Type: String
logicalStackName
The name of the CloudFormation stack this resource belongs to.
Type: String
resourceGroupName
The name of the resource group this resource belongs to.
Type: String
terraformSourceName
The name of the Terraform S3 state file this resource belongs to.
Type: String
eksSourceName
Name of the Amazon Elastic Kubernetes Service cluster and namespace this resource belongs to.
This parameter accepts values in \"eks-cluster/namespace\" format.
Type: String
type
The type of resource.
Type: string
name
The name of the resource.
Type: String
additionalInfo
Additional configuration parameters for an Resilience Hub application. If you want to implement additionalInfo through the Resilience Hub console rather than using an API call, see Configure the application configuration parameters.
Currently, this parameter accepts a key-value mapping (in a string format) of only one failover region and one associated account.
List of Application Components that this resource belongs to. If an Application Component is not part of the Resilience Hub application, it will be added.
Type: Array
Each appComponents array item includes the following fields:
The list of included resources that are assigned to the Application Component.
Type: Array of strings
additionalInfo
Additional configuration parameters for an Resilience Hub application. If you want to implement additionalInfo through the Resilience Hub console rather than using an API call, see Configure the application configuration parameters.
Currently, this parameter accepts a key-value mapping (in a string format) of only one failover region and one associated account.
The list of logical resource identifiers to be excluded from the application.
Type: Array
Don't add the resources that you want to include.
Each excludedResources array item includes the following fields:
logicalResourceIds
Logical identifier of the resource.
Type: Object
You can configure only one of the following fields:
logicalStackName
resourceGroupName
terraformSourceName
eksSourceName
Each logicalResourceIds object includes the following fields:
identifier
Identifier of the resource.
Type: String
logicalStackName
The name of the CloudFormation stack this resource belongs to.
Type: String
resourceGroupName
The name of the resource group this resource belongs to.
Type: String
terraformSourceName
The name of the Terraform S3 state file this resource belongs to.
Type: String
eksSourceName
Name of the Amazon Elastic Kubernetes Service cluster and namespace this resource belongs to.
This parameter accepts values in \"eks-cluster/namespace\" format.
Type: String
version
Resilience Hub application version.
additionalInfo
Additional configuration parameters for an Resilience Hub application. If you want to implement additionalInfo through the Resilience Hub console rather than using an API call, see Configure the application configuration parameters.
Currently, this parameter accepts a key-value mapping (in a string format) of only one failover region and one associated account.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the assessment. The format for this ARN is: arn:partition:resiliencehub:region:account:app-assessment/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The list of events you would like to subscribe and get notification for. Currently, Resilience Hub supports notifications only for Drift detected and Scheduled assessment failure events.
The Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The list of Application Components that this resource belongs to. If an Application Component is not part of the Resilience Hub application, it will be added.
"
+ "documentation":"
List of Application Components that this resource belongs to. If an Application Component is not part of the Resilience Hub application, it will be added.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the Resilience Hub application. The format for this ARN is: arn:partition:resiliencehub:region:account:app/app-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
The Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
"
+ "documentation":"
Amazon Resource Name (ARN) of the resiliency policy. The format for this ARN is: arn:partition:resiliencehub:region:account:resiliency-policy/policy-id. For more information about ARNs, see Amazon Resource Names (ARNs) in the AWS General Reference guide.
Creates, changes, or deletes a resource record set, which contains authoritative DNS information for a specified domain name or subdomain name. For example, you can use ChangeResourceRecordSets to create a resource record set that routes traffic for test.example.com to a web server that has an IP address of 192.0.2.44.
Deleting Resource Record Sets
To delete a resource record set, you must specify all the same values that you specified when you created it.
Change Batches and Transactional Changes
The request body must include a document with a ChangeResourceRecordSetsRequest element. The request body contains a list of change items, known as a change batch. Change batches are considered transactional changes. Route 53 validates the changes in the request and then either makes all or none of the changes in the change batch request. This ensures that DNS routing isn't adversely affected by partial changes to the resource record sets in a hosted zone.
For example, suppose a change batch request contains two changes: it deletes the CNAME resource record set for www.example.com and creates an alias resource record set for www.example.com. If validation for both records succeeds, Route 53 deletes the first resource record set and creates the second resource record set in a single operation. If validation for either the DELETE or the CREATE action fails, then the request is canceled, and the original CNAME record continues to exist.
If you try to delete the same resource record set more than once in a single change batch, Route 53 returns an InvalidChangeBatch error.
Traffic Flow
To create resource record sets for complex routing configurations, use either the traffic flow visual editor in the Route 53 console or the API actions for traffic policies and traffic policy instances. Save the configuration as a traffic policy, then associate the traffic policy with one or more domain names (such as example.com) or subdomain names (such as www.example.com), in the same hosted zone or in multiple hosted zones. You can roll back the updates if the new configuration isn't performing as expected. For more information, see Using Traffic Flow to Route DNS Traffic in the Amazon Route 53 Developer Guide.
Create, Delete, and Upsert
Use ChangeResourceRecordsSetsRequest to perform the following actions:
CREATE: Creates a resource record set that has the specified values.
DELETE: Deletes an existing resource record set that has the specified values.
UPSERT: If a resource set exists Route 53 updates it with the values in the request.
Syntaxes for Creating, Updating, and Deleting Resource Record Sets
The syntax for a request depends on the type of resource record set that you want to create, delete, or update, such as weighted, alias, or failover. The XML elements in your request must appear in the order listed in the syntax.
For an example for each type of resource record set, see \"Examples.\"
Don't refer to the syntax in the \"Parameter Syntax\" section, which includes all of the elements for every kind of resource record set that you can create, delete, or update by using ChangeResourceRecordSets.
Change Propagation to Route 53 DNS Servers
When you submit a ChangeResourceRecordSets request, Route 53 propagates your changes to all of the Route 53 authoritative DNS servers. While your changes are propagating, GetChange returns a status of PENDING. When propagation is complete, GetChange returns a status of INSYNC. Changes generally propagate to all Route 53 name servers within 60 seconds. For more information, see GetChange.
Limits on ChangeResourceRecordSets Requests
For information about the limits on a ChangeResourceRecordSets request, see Limits in the Amazon Route 53 Developer Guide.
"
+ "documentation":"
Creates, changes, or deletes a resource record set, which contains authoritative DNS information for a specified domain name or subdomain name. For example, you can use ChangeResourceRecordSets to create a resource record set that routes traffic for test.example.com to a web server that has an IP address of 192.0.2.44.
Deleting Resource Record Sets
To delete a resource record set, you must specify all the same values that you specified when you created it.
Change Batches and Transactional Changes
The request body must include a document with a ChangeResourceRecordSetsRequest element. The request body contains a list of change items, known as a change batch. Change batches are considered transactional changes. Route 53 validates the changes in the request and then either makes all or none of the changes in the change batch request. This ensures that DNS routing isn't adversely affected by partial changes to the resource record sets in a hosted zone.
For example, suppose a change batch request contains two changes: it deletes the CNAME resource record set for www.example.com and creates an alias resource record set for www.example.com. If validation for both records succeeds, Route 53 deletes the first resource record set and creates the second resource record set in a single operation. If validation for either the DELETE or the CREATE action fails, then the request is canceled, and the original CNAME record continues to exist.
If you try to delete the same resource record set more than once in a single change batch, Route 53 returns an InvalidChangeBatch error.
Traffic Flow
To create resource record sets for complex routing configurations, use either the traffic flow visual editor in the Route 53 console or the API actions for traffic policies and traffic policy instances. Save the configuration as a traffic policy, then associate the traffic policy with one or more domain names (such as example.com) or subdomain names (such as www.example.com), in the same hosted zone or in multiple hosted zones. You can roll back the updates if the new configuration isn't performing as expected. For more information, see Using Traffic Flow to Route DNS Traffic in the Amazon Route 53 Developer Guide.
Create, Delete, and Upsert
Use ChangeResourceRecordsSetsRequest to perform the following actions:
CREATE: Creates a resource record set that has the specified values.
DELETE: Deletes an existing resource record set that has the specified values.
UPSERT: If a resource set exists Route 53 updates it with the values in the request.
Syntaxes for Creating, Updating, and Deleting Resource Record Sets
The syntax for a request depends on the type of resource record set that you want to create, delete, or update, such as weighted, alias, or failover. The XML elements in your request must appear in the order listed in the syntax.
For an example for each type of resource record set, see \"Examples.\"
Don't refer to the syntax in the \"Parameter Syntax\" section, which includes all of the elements for every kind of resource record set that you can create, delete, or update by using ChangeResourceRecordSets.
Change Propagation to Route 53 DNS Servers
When you submit a ChangeResourceRecordSets request, Route 53 propagates your changes to all of the Route 53 authoritative DNS servers managing the hosted zone. While your changes are propagating, GetChange returns a status of PENDING. When propagation is complete, GetChange returns a status of INSYNC. Changes generally propagate to all Route 53 name servers managing the hosted zone within 60 seconds. For more information, see GetChange.
Limits on ChangeResourceRecordSets Requests
For information about the limits on a ChangeResourceRecordSets request, see Limits in the Amazon Route 53 Developer Guide.
Returns the current status of a change batch request. The status is one of the following values:
PENDING indicates that the changes in this request have not propagated to all Amazon Route 53 DNS servers. This is the initial status of all change batch requests.
INSYNC indicates that the changes have propagated to all Route 53 DNS servers.
"
+ "documentation":"
Returns the current status of a change batch request. The status is one of the following values:
PENDING indicates that the changes in this request have not propagated to all Amazon Route 53 DNS servers managing the hosted zone. This is the initial status of all change batch requests.
INSYNC indicates that the changes have propagated to all Route 53 DNS servers managing the hosted zone.
Gets the value that Amazon Route 53 returns in response to a DNS request for a specified record name and type. You can optionally specify the IP address of a DNS resolver, an EDNS0 client subnet IP address, and a subnet mask.
This call only supports querying public hosted zones.
"
+ "documentation":"
Gets the value that Amazon Route 53 returns in response to a DNS request for a specified record name and type. You can optionally specify the IP address of a DNS resolver, an EDNS0 client subnet IP address, and a subnet mask.
This call only supports querying public hosted zones.
The TestDnsAnswer returns information similar to what you would expect from the answer section of the dig command. Therefore, if you query for the name servers of a subdomain that point to the parent name servers, those will not be returned.
If you want to associate a reusable delegation set with this hosted zone, the ID that Amazon Route 53 assigned to the reusable delegation set when you created it. For more information about reusable delegation sets, see CreateReusableDelegationSet.
"
+ "documentation":"
If you want to associate a reusable delegation set with this hosted zone, the ID that Amazon Route 53 assigned to the reusable delegation set when you created it. For more information about reusable delegation sets, see CreateReusableDelegationSet.
If you are using a reusable delegation set to create a public hosted zone for a subdomain, make sure that the parent hosted zone doesn't use one or more of the same name servers. If you have overlapping nameservers, the operation will cause a ConflictingDomainsExist error.
"
}
},
"documentation":"
A complex type that contains information about the request to create a public or private hosted zone.
"
@@ -5469,7 +5470,8 @@
"af-south-1",
"eu-south-1",
"eu-south-2",
- "ap-southeast-4"
+ "ap-southeast-4",
+ "il-central-1"
],
"max":64,
"min":1
@@ -6344,7 +6346,8 @@
"af-south-1",
"eu-south-1",
"eu-south-2",
- "ap-southeast-4"
+ "ap-southeast-4",
+ "il-central-1"
],
"max":64,
"min":1
diff --git a/services/route53domains/pom.xml b/services/route53domains/pom.xml
index 2c24af10edb..81cf9e46861 100644
--- a/services/route53domains/pom.xml
+++ b/services/route53domains/pom.xml
@@ -22,7 +22,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTroute53domainsAWS Java SDK :: Services :: Amazon Route53 Domains
diff --git a/services/route53recoverycluster/pom.xml b/services/route53recoverycluster/pom.xml
index cd06448e077..0dd3739e76f 100644
--- a/services/route53recoverycluster/pom.xml
+++ b/services/route53recoverycluster/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTroute53recoveryclusterAWS Java SDK :: Services :: Route53 Recovery Cluster
diff --git a/services/route53recoverycontrolconfig/pom.xml b/services/route53recoverycontrolconfig/pom.xml
index eb93cbac853..9e7d93e6b95 100644
--- a/services/route53recoverycontrolconfig/pom.xml
+++ b/services/route53recoverycontrolconfig/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTroute53recoverycontrolconfigAWS Java SDK :: Services :: Route53 Recovery Control Config
diff --git a/services/route53recoveryreadiness/pom.xml b/services/route53recoveryreadiness/pom.xml
index b8167512a7d..f9ec52f370c 100644
--- a/services/route53recoveryreadiness/pom.xml
+++ b/services/route53recoveryreadiness/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTroute53recoveryreadinessAWS Java SDK :: Services :: Route53 Recovery Readiness
diff --git a/services/route53resolver/pom.xml b/services/route53resolver/pom.xml
index e4bd57229fa..a628b91cb9b 100644
--- a/services/route53resolver/pom.xml
+++ b/services/route53resolver/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTroute53resolverAWS Java SDK :: Services :: Route53Resolver
diff --git a/services/rum/pom.xml b/services/rum/pom.xml
index 65009c2b341..1df461cb60a 100644
--- a/services/rum/pom.xml
+++ b/services/rum/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTrumAWS Java SDK :: Services :: RUM
diff --git a/services/s3/pom.xml b/services/s3/pom.xml
index e03dd358d58..f708f8c4238 100644
--- a/services/s3/pom.xml
+++ b/services/s3/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTs3AWS Java SDK :: Services :: Amazon S3
diff --git a/services/s3/src/it/java/software/amazon/awssdk/services/s3/S3IntegrationTestBase.java b/services/s3/src/it/java/software/amazon/awssdk/services/s3/S3IntegrationTestBase.java
index 63dcf2ddc88..03cf42afe5d 100644
--- a/services/s3/src/it/java/software/amazon/awssdk/services/s3/S3IntegrationTestBase.java
+++ b/services/s3/src/it/java/software/amazon/awssdk/services/s3/S3IntegrationTestBase.java
@@ -117,7 +117,7 @@ protected static void deleteBucketAndAllContents(String bucketName) {
S3TestUtils.deleteBucketAndAllContents(s3, bucketName);
}
- private static class UserAgentVerifyingExecutionInterceptor implements ExecutionInterceptor {
+ protected static class UserAgentVerifyingExecutionInterceptor implements ExecutionInterceptor {
private final String clientName;
private final ClientType clientType;
diff --git a/services/s3/src/it/java/software/amazon/awssdk/services/s3/crt/S3CrossRegionCrtIntegrationTest.java b/services/s3/src/it/java/software/amazon/awssdk/services/s3/crt/S3CrossRegionCrtIntegrationTest.java
index 953c6e4b4f4..72c6fce095c 100644
--- a/services/s3/src/it/java/software/amazon/awssdk/services/s3/crt/S3CrossRegionCrtIntegrationTest.java
+++ b/services/s3/src/it/java/software/amazon/awssdk/services/s3/crt/S3CrossRegionCrtIntegrationTest.java
@@ -17,7 +17,7 @@
import static org.assertj.core.api.Assertions.assertThat;
import static org.assertj.core.api.Assertions.assertThatThrownBy;
-import static software.amazon.awssdk.services.s3.crt.S3CrtClientCopyIntegrationTest.randomBytes;
+import static software.amazon.awssdk.services.s3.multipart.S3ClientMultiPartCopyIntegrationTest.randomBytes;
import static software.amazon.awssdk.services.s3.utils.ChecksumUtils.computeCheckSum;
import static software.amazon.awssdk.testutils.service.S3BucketUtils.temporaryBucketName;
diff --git a/services/s3/src/it/java/software/amazon/awssdk/services/s3/crt/S3CrtClientPutObjectIntegrationTest.java b/services/s3/src/it/java/software/amazon/awssdk/services/s3/crt/S3CrtClientPutObjectIntegrationTest.java
index c6761f66e17..f81e700395e 100644
--- a/services/s3/src/it/java/software/amazon/awssdk/services/s3/crt/S3CrtClientPutObjectIntegrationTest.java
+++ b/services/s3/src/it/java/software/amazon/awssdk/services/s3/crt/S3CrtClientPutObjectIntegrationTest.java
@@ -47,7 +47,7 @@
public class S3CrtClientPutObjectIntegrationTest extends S3IntegrationTestBase {
private static final String TEST_BUCKET = temporaryBucketName(S3CrtClientPutObjectIntegrationTest.class);
private static final String TEST_KEY = "8mib_file.dat";
- private static final int OBJ_SIZE = 8 * 1024 * 1024;
+ private static final int OBJ_SIZE = 10 * 1024 * 1024;
private static RandomTempFile testFile;
private static S3AsyncClient s3Crt;
diff --git a/services/s3/src/it/java/software/amazon/awssdk/services/s3/crt/S3CrtClientCopyIntegrationTest.java b/services/s3/src/it/java/software/amazon/awssdk/services/s3/multipart/S3ClientMultiPartCopyIntegrationTest.java
similarity index 70%
rename from services/s3/src/it/java/software/amazon/awssdk/services/s3/crt/S3CrtClientCopyIntegrationTest.java
rename to services/s3/src/it/java/software/amazon/awssdk/services/s3/multipart/S3ClientMultiPartCopyIntegrationTest.java
index f4d2b34c1cd..fc4f31b76b1 100644
--- a/services/s3/src/it/java/software/amazon/awssdk/services/s3/crt/S3CrtClientCopyIntegrationTest.java
+++ b/services/s3/src/it/java/software/amazon/awssdk/services/s3/multipart/S3ClientMultiPartCopyIntegrationTest.java
@@ -13,7 +13,7 @@
* permissions and limitations under the License.
*/
-package software.amazon.awssdk.services.s3.crt;
+package software.amazon.awssdk.services.s3.multipart;
import static org.assertj.core.api.Assertions.assertThat;
import static org.assertj.core.api.Fail.fail;
@@ -24,13 +24,17 @@
import java.nio.ByteBuffer;
import java.security.SecureRandom;
import java.util.Base64;
-import java.util.UUID;
import java.util.concurrent.CompletableFuture;
import java.util.concurrent.ThreadLocalRandom;
+import java.util.concurrent.TimeUnit;
+import java.util.stream.Stream;
import javax.crypto.KeyGenerator;
import org.junit.jupiter.api.AfterAll;
import org.junit.jupiter.api.BeforeAll;
-import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.Timeout;
+import org.junit.jupiter.params.ParameterizedTest;
+import org.junit.jupiter.params.provider.MethodSource;
+import software.amazon.awssdk.core.ClientType;
import software.amazon.awssdk.core.ResponseBytes;
import software.amazon.awssdk.core.async.AsyncRequestBody;
import software.amazon.awssdk.core.sync.ResponseTransformer;
@@ -42,8 +46,9 @@
import software.amazon.awssdk.services.s3.model.MetadataDirective;
import software.amazon.awssdk.utils.Md5Utils;
-public class S3CrtClientCopyIntegrationTest extends S3IntegrationTestBase {
- private static final String BUCKET = temporaryBucketName(S3CrtClientCopyIntegrationTest.class);
+@Timeout(value = 3, unit = TimeUnit.MINUTES)
+public class S3ClientMultiPartCopyIntegrationTest extends S3IntegrationTestBase {
+ private static final String BUCKET = temporaryBucketName(S3ClientMultiPartCopyIntegrationTest.class);
private static final String ORIGINAL_OBJ = "test_file.dat";
private static final String COPIED_OBJ = "test_file_copy.dat";
private static final String ORIGINAL_OBJ_SPECIAL_CHARACTER = "original-special-chars-@$%";
@@ -51,6 +56,8 @@ public class S3CrtClientCopyIntegrationTest extends S3IntegrationTestBase {
private static final long OBJ_SIZE = ThreadLocalRandom.current().nextLong(8 * 1024 * 1024, 16 * 1024 * 1024 + 1);
private static final long SMALL_OBJ_SIZE = 1024 * 1024;
private static S3AsyncClient s3CrtAsyncClient;
+ private static S3AsyncClient s3MpuClient;
+
@BeforeAll
public static void setUp() throws Exception {
S3IntegrationTestBase.setUp();
@@ -59,40 +66,56 @@ public static void setUp() throws Exception {
.credentialsProvider(CREDENTIALS_PROVIDER_CHAIN)
.region(DEFAULT_REGION)
.build();
+ s3MpuClient = S3AsyncClient.builder()
+ .region(DEFAULT_REGION)
+ .credentialsProvider(CREDENTIALS_PROVIDER_CHAIN)
+ .overrideConfiguration(o -> o.addExecutionInterceptor(
+ new UserAgentVerifyingExecutionInterceptor("NettyNio", ClientType.ASYNC)))
+ .multipartEnabled(true)
+ .build();
}
@AfterAll
public static void teardown() throws Exception {
s3CrtAsyncClient.close();
+ s3MpuClient.close();
deleteBucketAndAllContents(BUCKET);
}
- @Test
- void copy_singlePart_hasSameContent() {
+ public static Stream s3AsyncClient() {
+ return Stream.of(s3MpuClient, s3CrtAsyncClient);
+ }
+
+ @ParameterizedTest(autoCloseArguments = false)
+ @MethodSource("s3AsyncClient")
+ void copy_singlePart_hasSameContent(S3AsyncClient s3AsyncClient) {
byte[] originalContent = randomBytes(SMALL_OBJ_SIZE);
createOriginalObject(originalContent, ORIGINAL_OBJ);
- copyObject(ORIGINAL_OBJ, COPIED_OBJ);
+ copyObject(ORIGINAL_OBJ, COPIED_OBJ, s3AsyncClient);
validateCopiedObject(originalContent, ORIGINAL_OBJ);
}
- @Test
- void copy_copiedObject_hasSameContent() {
+ @ParameterizedTest(autoCloseArguments = false)
+ @MethodSource("s3AsyncClient")
+ void copy_copiedObject_hasSameContent(S3AsyncClient s3AsyncClient) {
byte[] originalContent = randomBytes(OBJ_SIZE);
createOriginalObject(originalContent, ORIGINAL_OBJ);
- copyObject(ORIGINAL_OBJ, COPIED_OBJ);
+ copyObject(ORIGINAL_OBJ, COPIED_OBJ, s3AsyncClient);
validateCopiedObject(originalContent, ORIGINAL_OBJ);
}
- @Test
- void copy_specialCharacters_hasSameContent() {
+ @ParameterizedTest(autoCloseArguments = false)
+ @MethodSource("s3AsyncClient")
+ void copy_specialCharacters_hasSameContent(S3AsyncClient s3AsyncClient) {
byte[] originalContent = randomBytes(OBJ_SIZE);
createOriginalObject(originalContent, ORIGINAL_OBJ_SPECIAL_CHARACTER);
- copyObject(ORIGINAL_OBJ_SPECIAL_CHARACTER, COPIED_OBJ_SPECIAL_CHARACTER);
+ copyObject(ORIGINAL_OBJ_SPECIAL_CHARACTER, COPIED_OBJ_SPECIAL_CHARACTER, s3AsyncClient);
validateCopiedObject(originalContent, COPIED_OBJ_SPECIAL_CHARACTER);
}
- @Test
- void copy_ssecServerSideEncryption_shouldSucceed() {
+ @ParameterizedTest(autoCloseArguments = false)
+ @MethodSource("s3AsyncClient")
+ void copy_ssecServerSideEncryption_shouldSucceed(S3AsyncClient s3AsyncClient) {
byte[] originalContent = randomBytes(OBJ_SIZE);
byte[] secretKey = generateSecretKey();
String b64Key = Base64.getEncoder().encodeToString(secretKey);
@@ -102,16 +125,14 @@ void copy_ssecServerSideEncryption_shouldSucceed() {
String newB64Key = Base64.getEncoder().encodeToString(newSecretKey);
String newB64KeyMd5 = Md5Utils.md5AsBase64(newSecretKey);
- // Java S3 client is used because CRT S3 client putObject fails with SSE-C
- // TODO: change back to S3CrtClient once the issue is fixed in CRT
- s3Async.putObject(r -> r.bucket(BUCKET)
- .key(ORIGINAL_OBJ)
- .sseCustomerKey(b64Key)
- .sseCustomerAlgorithm(AES256.name())
- .sseCustomerKeyMD5(b64KeyMd5),
- AsyncRequestBody.fromBytes(originalContent)).join();
+ s3AsyncClient.putObject(r -> r.bucket(BUCKET)
+ .key(ORIGINAL_OBJ)
+ .sseCustomerKey(b64Key)
+ .sseCustomerAlgorithm(AES256.name())
+ .sseCustomerKeyMD5(b64KeyMd5),
+ AsyncRequestBody.fromBytes(originalContent)).join();
- CompletableFuture future = s3CrtAsyncClient.copyObject(c -> c
+ CompletableFuture future = s3AsyncClient.copyObject(c -> c
.sourceBucket(BUCKET)
.sourceKey(ORIGINAL_OBJ)
.metadataDirective(MetadataDirective.REPLACE)
@@ -143,12 +164,12 @@ private static byte[] generateSecretKey() {
private void createOriginalObject(byte[] originalContent, String originalKey) {
s3CrtAsyncClient.putObject(r -> r.bucket(BUCKET)
- .key(originalKey),
+ .key(originalKey),
AsyncRequestBody.fromBytes(originalContent)).join();
}
- private void copyObject(String original, String destination) {
- CompletableFuture future = s3CrtAsyncClient.copyObject(c -> c
+ private void copyObject(String original, String destination, S3AsyncClient s3AsyncClient) {
+ CompletableFuture future = s3AsyncClient.copyObject(c -> c
.sourceBucket(BUCKET)
.sourceKey(original)
.destinationBucket(BUCKET)
diff --git a/services/s3/src/it/java/software/amazon/awssdk/services/s3/multipart/S3MultipartClientPutObjectIntegrationTest.java b/services/s3/src/it/java/software/amazon/awssdk/services/s3/multipart/S3MultipartClientPutObjectIntegrationTest.java
new file mode 100644
index 00000000000..fa31b5453e5
--- /dev/null
+++ b/services/s3/src/it/java/software/amazon/awssdk/services/s3/multipart/S3MultipartClientPutObjectIntegrationTest.java
@@ -0,0 +1,136 @@
+/*
+ * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License").
+ * You may not use this file except in compliance with the License.
+ * A copy of the License is located at
+ *
+ * http://aws.amazon.com/apache2.0
+ *
+ * or in the "license" file accompanying this file. This file is distributed
+ * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
+ * express or implied. See the License for the specific language governing
+ * permissions and limitations under the License.
+ */
+
+package software.amazon.awssdk.services.s3.multipart;
+
+import static java.util.concurrent.TimeUnit.SECONDS;
+import static org.assertj.core.api.Assertions.assertThat;
+import static software.amazon.awssdk.testutils.service.S3BucketUtils.temporaryBucketName;
+
+import java.io.ByteArrayInputStream;
+import java.io.File;
+import java.nio.ByteBuffer;
+import java.nio.charset.Charset;
+import java.nio.file.Files;
+import java.util.Optional;
+import java.util.UUID;
+import org.apache.commons.lang3.RandomStringUtils;
+import org.junit.jupiter.api.AfterAll;
+import org.junit.jupiter.api.BeforeAll;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.Timeout;
+import org.reactivestreams.Subscriber;
+import software.amazon.awssdk.core.ClientType;
+import software.amazon.awssdk.core.ResponseInputStream;
+import software.amazon.awssdk.core.async.AsyncRequestBody;
+import software.amazon.awssdk.core.internal.async.FileAsyncRequestBody;
+import software.amazon.awssdk.core.sync.ResponseTransformer;
+import software.amazon.awssdk.services.s3.S3AsyncClient;
+import software.amazon.awssdk.services.s3.S3IntegrationTestBase;
+import software.amazon.awssdk.services.s3.model.GetObjectResponse;
+import software.amazon.awssdk.services.s3.utils.ChecksumUtils;
+
+@Timeout(value = 30, unit = SECONDS)
+public class S3MultipartClientPutObjectIntegrationTest extends S3IntegrationTestBase {
+
+ private static final String TEST_BUCKET = temporaryBucketName(S3MultipartClientPutObjectIntegrationTest.class);
+ private static final String TEST_KEY = "testfile.dat";
+ private static final int OBJ_SIZE = 19 * 1024 * 1024;
+
+ private static File testFile;
+ private static S3AsyncClient mpuS3Client;
+
+ @BeforeAll
+ public static void setup() throws Exception {
+ S3IntegrationTestBase.setUp();
+ S3IntegrationTestBase.createBucket(TEST_BUCKET);
+ byte[] CONTENT =
+ RandomStringUtils.randomAscii(OBJ_SIZE).getBytes(Charset.defaultCharset());
+
+ testFile = File.createTempFile("SplittingPublisherTest", UUID.randomUUID().toString());
+ Files.write(testFile.toPath(), CONTENT);
+ mpuS3Client = S3AsyncClient
+ .builder()
+ .region(DEFAULT_REGION)
+ .credentialsProvider(CREDENTIALS_PROVIDER_CHAIN)
+ .overrideConfiguration(o -> o.addExecutionInterceptor(
+ new UserAgentVerifyingExecutionInterceptor("NettyNio", ClientType.ASYNC)))
+ .multipartEnabled(true)
+ .build();
+ }
+
+ @AfterAll
+ public static void teardown() throws Exception {
+ mpuS3Client.close();
+ testFile.delete();
+ deleteBucketAndAllContents(TEST_BUCKET);
+ }
+
+ @Test
+ void putObject_fileRequestBody_objectSentCorrectly() throws Exception {
+ AsyncRequestBody body = AsyncRequestBody.fromFile(testFile.toPath());
+ mpuS3Client.putObject(r -> r.bucket(TEST_BUCKET).key(TEST_KEY), body).join();
+
+ ResponseInputStream objContent =
+ S3IntegrationTestBase.s3.getObject(r -> r.bucket(TEST_BUCKET).key(TEST_KEY),
+ ResponseTransformer.toInputStream());
+
+ assertThat(objContent.response().contentLength()).isEqualTo(testFile.length());
+ byte[] expectedSum = ChecksumUtils.computeCheckSum(Files.newInputStream(testFile.toPath()));
+ assertThat(ChecksumUtils.computeCheckSum(objContent)).isEqualTo(expectedSum);
+ }
+
+ @Test
+ void putObject_byteAsyncRequestBody_objectSentCorrectly() throws Exception {
+ byte[] bytes = RandomStringUtils.randomAscii(OBJ_SIZE).getBytes(Charset.defaultCharset());
+ AsyncRequestBody body = AsyncRequestBody.fromBytes(bytes);
+ mpuS3Client.putObject(r -> r.bucket(TEST_BUCKET).key(TEST_KEY), body).join();
+
+ ResponseInputStream objContent =
+ S3IntegrationTestBase.s3.getObject(r -> r.bucket(TEST_BUCKET).key(TEST_KEY),
+ ResponseTransformer.toInputStream());
+
+ assertThat(objContent.response().contentLength()).isEqualTo(OBJ_SIZE);
+ byte[] expectedSum = ChecksumUtils.computeCheckSum(new ByteArrayInputStream(bytes));
+ assertThat(ChecksumUtils.computeCheckSum(objContent)).isEqualTo(expectedSum);
+ }
+
+ @Test
+ void putObject_unknownContentLength_objectSentCorrectly() throws Exception {
+ AsyncRequestBody body = FileAsyncRequestBody.builder()
+ .path(testFile.toPath())
+ .build();
+ mpuS3Client.putObject(r -> r.bucket(TEST_BUCKET).key(TEST_KEY), new AsyncRequestBody() {
+ @Override
+ public Optional contentLength() {
+ return Optional.empty();
+ }
+
+ @Override
+ public void subscribe(Subscriber super ByteBuffer> s) {
+ body.subscribe(s);
+ }
+ }).get(30, SECONDS);
+
+ ResponseInputStream objContent =
+ S3IntegrationTestBase.s3.getObject(r -> r.bucket(TEST_BUCKET).key(TEST_KEY),
+ ResponseTransformer.toInputStream());
+
+ assertThat(objContent.response().contentLength()).isEqualTo(testFile.length());
+ byte[] expectedSum = ChecksumUtils.computeCheckSum(Files.newInputStream(testFile.toPath()));
+ assertThat(ChecksumUtils.computeCheckSum(objContent)).isEqualTo(expectedSum);
+ }
+
+}
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/S3CrtAsyncClientBuilder.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/S3CrtAsyncClientBuilder.java
index 7d119c2f45b..78dc144f661 100644
--- a/services/s3/src/main/java/software/amazon/awssdk/services/s3/S3CrtAsyncClientBuilder.java
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/S3CrtAsyncClientBuilder.java
@@ -234,6 +234,26 @@ default S3CrtAsyncClientBuilder retryConfiguration(Consumer
+ * Multipart uploads are easier to recover from and also potentially faster than single part uploads, especially when the
+ * upload parts can be uploaded in parallel. Because there are additional network API calls, small objects are still
+ * recommended to use a single connection for the upload. See
+ * Uploading and copying objects using
+ * multipart upload.
+ *
+ *
+ * By default, it is the same as {@link #minimumPartSizeInBytes(Long)}.
+ *
+ * @param thresholdInBytes the value of the threshold to set.
+ * @return an instance of this builder.
+ */
+ S3CrtAsyncClientBuilder thresholdInBytes(Long thresholdInBytes);
+
@Override
S3AsyncClient build();
}
\ No newline at end of file
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/client/S3AsyncClientDecorator.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/client/S3AsyncClientDecorator.java
index 2dbb61091da..b751cb29c1b 100644
--- a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/client/S3AsyncClientDecorator.java
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/client/S3AsyncClientDecorator.java
@@ -23,11 +23,17 @@
import software.amazon.awssdk.services.s3.S3AsyncClient;
import software.amazon.awssdk.services.s3.endpoints.S3ClientContextParams;
import software.amazon.awssdk.services.s3.internal.crossregion.S3CrossRegionAsyncClient;
+import software.amazon.awssdk.services.s3.internal.multipart.MultipartS3AsyncClient;
+import software.amazon.awssdk.services.s3.multipart.MultipartConfiguration;
import software.amazon.awssdk.utils.AttributeMap;
import software.amazon.awssdk.utils.ConditionalDecorator;
@SdkInternalApi
public class S3AsyncClientDecorator {
+ public static final AttributeMap.Key MULTIPART_CONFIGURATION_KEY =
+ new AttributeMap.Key(MultipartConfiguration.class){};
+ public static final AttributeMap.Key MULTIPART_ENABLED_KEY =
+ new AttributeMap.Key(Boolean.class){};
public S3AsyncClientDecorator() {
}
@@ -36,14 +42,26 @@ public S3AsyncClient decorate(S3AsyncClient base,
SdkClientConfiguration clientConfiguration,
AttributeMap clientContextParams) {
List> decorators = new ArrayList<>();
- decorators.add(ConditionalDecorator.create(isCrossRegionEnabledAsync(clientContextParams),
- S3CrossRegionAsyncClient::new));
+ decorators.add(ConditionalDecorator.create(
+ isCrossRegionEnabledAsync(clientContextParams),
+ S3CrossRegionAsyncClient::new));
+
+ decorators.add(ConditionalDecorator.create(
+ isMultipartEnable(clientContextParams),
+ client -> {
+ MultipartConfiguration multipartConfiguration = clientContextParams.get(MULTIPART_CONFIGURATION_KEY);
+ return MultipartS3AsyncClient.create(client, multipartConfiguration);
+ }));
return ConditionalDecorator.decorate(base, decorators);
}
private Predicate isCrossRegionEnabledAsync(AttributeMap clientContextParams) {
Boolean crossRegionEnabled = clientContextParams.get(S3ClientContextParams.CROSS_REGION_ACCESS_ENABLED);
- return client -> crossRegionEnabled != null && crossRegionEnabled.booleanValue();
+ return client -> crossRegionEnabled != null && crossRegionEnabled.booleanValue();
}
+ private Predicate isMultipartEnable(AttributeMap clientContextParams) {
+ Boolean multipartEnabled = clientContextParams.get(MULTIPART_ENABLED_KEY);
+ return client -> multipartEnabled != null && multipartEnabled.booleanValue();
+ }
}
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/CopyRequestConversionUtils.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/CopyRequestConversionUtils.java
deleted file mode 100644
index 2a464b10f49..00000000000
--- a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/CopyRequestConversionUtils.java
+++ /dev/null
@@ -1,168 +0,0 @@
-/*
- * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
- *
- * Licensed under the Apache License, Version 2.0 (the "License").
- * You may not use this file except in compliance with the License.
- * A copy of the License is located at
- *
- * http://aws.amazon.com/apache2.0
- *
- * or in the "license" file accompanying this file. This file is distributed
- * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
- * express or implied. See the License for the specific language governing
- * permissions and limitations under the License.
- */
-
-package software.amazon.awssdk.services.s3.internal.crt;
-
-import software.amazon.awssdk.annotations.SdkInternalApi;
-import software.amazon.awssdk.services.s3.model.AbortMultipartUploadRequest;
-import software.amazon.awssdk.services.s3.model.CompleteMultipartUploadResponse;
-import software.amazon.awssdk.services.s3.model.CompletedPart;
-import software.amazon.awssdk.services.s3.model.CopyObjectRequest;
-import software.amazon.awssdk.services.s3.model.CopyObjectResponse;
-import software.amazon.awssdk.services.s3.model.CopyPartResult;
-import software.amazon.awssdk.services.s3.model.CreateMultipartUploadRequest;
-import software.amazon.awssdk.services.s3.model.HeadObjectRequest;
-import software.amazon.awssdk.services.s3.model.UploadPartCopyRequest;
-
-/**
- * Request conversion utility method for POJO classes associated with {@link S3CrtAsyncClient#copyObject(CopyObjectRequest)}
- */
-@SdkInternalApi
-public final class CopyRequestConversionUtils {
-
- private CopyRequestConversionUtils() {
- }
-
- public static HeadObjectRequest toHeadObjectRequest(CopyObjectRequest copyObjectRequest) {
- return HeadObjectRequest.builder()
- .bucket(copyObjectRequest.sourceBucket())
- .key(copyObjectRequest.sourceKey())
- .versionId(copyObjectRequest.sourceVersionId())
- .ifMatch(copyObjectRequest.copySourceIfMatch())
- .ifModifiedSince(copyObjectRequest.copySourceIfModifiedSince())
- .ifNoneMatch(copyObjectRequest.copySourceIfNoneMatch())
- .ifUnmodifiedSince(copyObjectRequest.copySourceIfUnmodifiedSince())
- .expectedBucketOwner(copyObjectRequest.expectedSourceBucketOwner())
- .sseCustomerAlgorithm(copyObjectRequest.copySourceSSECustomerAlgorithm())
- .sseCustomerKey(copyObjectRequest.copySourceSSECustomerKey())
- .sseCustomerKeyMD5(copyObjectRequest.copySourceSSECustomerKeyMD5())
- .build();
- }
-
- public static CompletedPart toCompletedPart(CopyPartResult copyPartResult, int partNumber) {
- return CompletedPart.builder()
- .partNumber(partNumber)
- .eTag(copyPartResult.eTag())
- .checksumCRC32C(copyPartResult.checksumCRC32C())
- .checksumCRC32(copyPartResult.checksumCRC32())
- .checksumSHA1(copyPartResult.checksumSHA1())
- .checksumSHA256(copyPartResult.checksumSHA256())
- .eTag(copyPartResult.eTag())
- .build();
- }
-
- public static CreateMultipartUploadRequest toCreateMultipartUploadRequest(CopyObjectRequest copyObjectRequest) {
- return CreateMultipartUploadRequest.builder()
- .bucket(copyObjectRequest.destinationBucket())
- .contentEncoding(copyObjectRequest.contentEncoding())
- .checksumAlgorithm(copyObjectRequest.checksumAlgorithmAsString())
- .tagging(copyObjectRequest.tagging())
- .contentType(copyObjectRequest.contentType())
- .contentLanguage(copyObjectRequest.contentLanguage())
- .contentDisposition(copyObjectRequest.contentDisposition())
- .cacheControl(copyObjectRequest.cacheControl())
- .expires(copyObjectRequest.expires())
- .key(copyObjectRequest.destinationKey())
- .websiteRedirectLocation(copyObjectRequest.websiteRedirectLocation())
- .expectedBucketOwner(copyObjectRequest.expectedBucketOwner())
- .requestPayer(copyObjectRequest.requestPayerAsString())
- .acl(copyObjectRequest.aclAsString())
- .grantRead(copyObjectRequest.grantRead())
- .grantReadACP(copyObjectRequest.grantReadACP())
- .grantWriteACP(copyObjectRequest.grantWriteACP())
- .grantFullControl(copyObjectRequest.grantFullControl())
- .storageClass(copyObjectRequest.storageClassAsString())
- .ssekmsKeyId(copyObjectRequest.ssekmsKeyId())
- .sseCustomerKey(copyObjectRequest.sseCustomerKey())
- .sseCustomerAlgorithm(copyObjectRequest.sseCustomerAlgorithm())
- .sseCustomerKeyMD5(copyObjectRequest.sseCustomerKeyMD5())
- .ssekmsEncryptionContext(copyObjectRequest.ssekmsEncryptionContext())
- .serverSideEncryption(copyObjectRequest.serverSideEncryptionAsString())
- .bucketKeyEnabled(copyObjectRequest.bucketKeyEnabled())
- .objectLockMode(copyObjectRequest.objectLockModeAsString())
- .objectLockLegalHoldStatus(copyObjectRequest.objectLockLegalHoldStatusAsString())
- .objectLockRetainUntilDate(copyObjectRequest.objectLockRetainUntilDate())
- .metadata(copyObjectRequest.metadata())
- .build();
- }
-
- public static CopyObjectResponse toCopyObjectResponse(CompleteMultipartUploadResponse response) {
- CopyObjectResponse.Builder builder = CopyObjectResponse.builder()
- .versionId(response.versionId())
- .copyObjectResult(b -> b.checksumCRC32(response.checksumCRC32())
- .checksumSHA1(response.checksumSHA1())
- .checksumSHA256(response.checksumSHA256())
- .checksumCRC32C(response.checksumCRC32C())
- .eTag(response.eTag())
- .build())
- .expiration(response.expiration())
- .bucketKeyEnabled(response.bucketKeyEnabled())
- .serverSideEncryption(response.serverSideEncryption())
- .ssekmsKeyId(response.ssekmsKeyId())
- .serverSideEncryption(response.serverSideEncryptionAsString())
- .requestCharged(response.requestChargedAsString());
- if (response.responseMetadata() != null) {
- builder.responseMetadata(response.responseMetadata());
- }
-
- if (response.sdkHttpResponse() != null) {
- builder.sdkHttpResponse(response.sdkHttpResponse());
- }
-
- return builder.build();
- }
-
- public static AbortMultipartUploadRequest toAbortMultipartUploadRequest(CopyObjectRequest copyObjectRequest,
- String uploadId) {
- return AbortMultipartUploadRequest.builder()
- .uploadId(uploadId)
- .bucket(copyObjectRequest.destinationBucket())
- .key(copyObjectRequest.destinationKey())
- .requestPayer(copyObjectRequest.requestPayerAsString())
- .expectedBucketOwner(copyObjectRequest.expectedBucketOwner())
- .build();
- }
-
- public static UploadPartCopyRequest toUploadPartCopyRequest(CopyObjectRequest copyObjectRequest,
- int partNumber,
- String uploadId,
- String range) {
-
- return UploadPartCopyRequest.builder()
- .sourceBucket(copyObjectRequest.sourceBucket())
- .sourceKey(copyObjectRequest.sourceKey())
- .sourceVersionId(copyObjectRequest.sourceVersionId())
- .uploadId(uploadId)
- .partNumber(partNumber)
- .destinationBucket(copyObjectRequest.destinationBucket())
- .destinationKey(copyObjectRequest.destinationKey())
- .copySourceIfMatch(copyObjectRequest.copySourceIfMatch())
- .copySourceIfNoneMatch(copyObjectRequest.copySourceIfNoneMatch())
- .copySourceIfUnmodifiedSince(copyObjectRequest.copySourceIfUnmodifiedSince())
- .copySourceRange(range)
- .copySourceSSECustomerAlgorithm(copyObjectRequest.copySourceSSECustomerAlgorithm())
- .copySourceSSECustomerKeyMD5(copyObjectRequest.copySourceSSECustomerKeyMD5())
- .copySourceSSECustomerKey(copyObjectRequest.copySourceSSECustomerKey())
- .copySourceIfModifiedSince(copyObjectRequest.copySourceIfModifiedSince())
- .expectedBucketOwner(copyObjectRequest.expectedBucketOwner())
- .expectedSourceBucketOwner(copyObjectRequest.expectedSourceBucketOwner())
- .requestPayer(copyObjectRequest.requestPayerAsString())
- .sseCustomerKey(copyObjectRequest.sseCustomerKey())
- .sseCustomerAlgorithm(copyObjectRequest.sseCustomerAlgorithm())
- .sseCustomerKeyMD5(copyObjectRequest.sseCustomerKeyMD5())
- .build();
- }
-
-}
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/DefaultS3CrtAsyncClient.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/DefaultS3CrtAsyncClient.java
index 5ab689637c4..284748f163b 100644
--- a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/DefaultS3CrtAsyncClient.java
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/DefaultS3CrtAsyncClient.java
@@ -53,6 +53,7 @@
import software.amazon.awssdk.services.s3.S3CrtAsyncClientBuilder;
import software.amazon.awssdk.services.s3.crt.S3CrtHttpConfiguration;
import software.amazon.awssdk.services.s3.crt.S3CrtRetryConfiguration;
+import software.amazon.awssdk.services.s3.internal.multipart.CopyObjectHelper;
import software.amazon.awssdk.services.s3.model.CopyObjectRequest;
import software.amazon.awssdk.services.s3.model.CopyObjectResponse;
import software.amazon.awssdk.services.s3.model.GetObjectRequest;
@@ -69,7 +70,10 @@ private DefaultS3CrtAsyncClient(DefaultS3CrtClientBuilder builder) {
super(initializeS3AsyncClient(builder));
long partSizeInBytes = builder.minimalPartSizeInBytes == null ? DEFAULT_PART_SIZE_IN_BYTES :
builder.minimalPartSizeInBytes;
- this.copyObjectHelper = new CopyObjectHelper((S3AsyncClient) delegate(), partSizeInBytes);
+ long thresholdInBytes = builder.thresholdInBytes == null ? partSizeInBytes : builder.thresholdInBytes;
+ this.copyObjectHelper = new CopyObjectHelper((S3AsyncClient) delegate(),
+ partSizeInBytes,
+ thresholdInBytes);
}
@Override
@@ -114,6 +118,7 @@ private static S3CrtAsyncHttpClient.Builder initializeS3CrtAsyncHttpClient(Defau
Validate.isPositiveOrNull(builder.maxConcurrency, "maxConcurrency");
Validate.isPositiveOrNull(builder.targetThroughputInGbps, "targetThroughputInGbps");
Validate.isPositiveOrNull(builder.minimalPartSizeInBytes, "minimalPartSizeInBytes");
+ Validate.isPositiveOrNull(builder.thresholdInBytes, "thresholdInBytes");
S3NativeClientConfiguration.Builder nativeClientBuilder =
S3NativeClientConfiguration.builder()
@@ -125,7 +130,8 @@ private static S3CrtAsyncHttpClient.Builder initializeS3CrtAsyncHttpClient(Defau
.endpointOverride(builder.endpointOverride)
.credentialsProvider(builder.credentialsProvider)
.readBufferSizeInBytes(builder.readBufferSizeInBytes)
- .httpConfiguration(builder.httpConfiguration);
+ .httpConfiguration(builder.httpConfiguration)
+ .thresholdInBytes(builder.thresholdInBytes);
if (builder.retryConfiguration != null) {
nativeClientBuilder.standardRetryOptions(
@@ -153,6 +159,7 @@ public static final class DefaultS3CrtClientBuilder implements S3CrtAsyncClientB
private List executionInterceptors;
private S3CrtRetryConfiguration retryConfiguration;
private boolean crossRegionAccessEnabled;
+ private Long thresholdInBytes;
public AwsCredentialsProvider credentialsProvider() {
return credentialsProvider;
@@ -273,6 +280,12 @@ public S3CrtAsyncClientBuilder crossRegionAccessEnabled(Boolean crossRegionAcces
return this;
}
+ @Override
+ public S3CrtAsyncClientBuilder thresholdInBytes(Long thresholdInBytes) {
+ this.thresholdInBytes = thresholdInBytes;
+ return this;
+ }
+
@Override
public S3CrtAsyncClient build() {
return new DefaultS3CrtAsyncClient(this);
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/S3CrtAsyncHttpClient.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/S3CrtAsyncHttpClient.java
index 149471f3017..f8bf0d809ff 100644
--- a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/S3CrtAsyncHttpClient.java
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/S3CrtAsyncHttpClient.java
@@ -75,6 +75,7 @@ private S3CrtAsyncHttpClient(Builder builder) {
.withCredentialsProvider(s3NativeClientConfiguration.credentialsProvider())
.withClientBootstrap(s3NativeClientConfiguration.clientBootstrap())
.withPartSize(s3NativeClientConfiguration.partSizeBytes())
+ .withMultipartUploadThreshold(s3NativeClientConfiguration.thresholdInBytes())
.withComputeContentMd5(false)
.withMaxConnections(s3NativeClientConfiguration.maxConcurrency())
.withThroughputTargetGbps(s3NativeClientConfiguration.targetThroughputInGbps())
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/S3NativeClientConfiguration.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/S3NativeClientConfiguration.java
index b39cf1ea8e4..fe5bb9a5dba 100644
--- a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/S3NativeClientConfiguration.java
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/S3NativeClientConfiguration.java
@@ -51,6 +51,7 @@ public class S3NativeClientConfiguration implements SdkAutoCloseable {
private final CrtCredentialsProviderAdapter credentialProviderAdapter;
private final CredentialsProvider credentialsProvider;
private final long partSizeInBytes;
+ private final long thresholdInBytes;
private final double targetThroughputInGbps;
private final int maxConcurrency;
private final URI endpointOverride;
@@ -86,6 +87,8 @@ public S3NativeClientConfiguration(Builder builder) {
this.partSizeInBytes = builder.partSizeInBytes == null ? DEFAULT_PART_SIZE_IN_BYTES :
builder.partSizeInBytes;
+ this.thresholdInBytes = builder.thresholdInBytes == null ? this.partSizeInBytes :
+ builder.thresholdInBytes;
this.targetThroughputInGbps = builder.targetThroughputInGbps == null ?
DEFAULT_TARGET_THROUGHPUT_IN_GBPS : builder.targetThroughputInGbps;
@@ -144,6 +147,10 @@ public long partSizeBytes() {
return partSizeInBytes;
}
+ public long thresholdInBytes() {
+ return thresholdInBytes;
+ }
+
public double targetThroughputInGbps() {
return targetThroughputInGbps;
}
@@ -187,6 +194,7 @@ public static final class Builder {
private S3CrtHttpConfiguration httpConfiguration;
private StandardRetryOptions standardRetryOptions;
+ private Long thresholdInBytes;
private Builder() {
}
@@ -247,5 +255,10 @@ public Builder standardRetryOptions(StandardRetryOptions standardRetryOptions) {
this.standardRetryOptions = standardRetryOptions;
return this;
}
+
+ public Builder thresholdInBytes(Long thresholdInBytes) {
+ this.thresholdInBytes = thresholdInBytes;
+ return this;
+ }
}
}
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/UploadPartCopyRequestIterable.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/UploadPartCopyRequestIterable.java
index 84d3c6ac530..da8eea8fc64 100644
--- a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/UploadPartCopyRequestIterable.java
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/UploadPartCopyRequestIterable.java
@@ -19,6 +19,7 @@
import java.util.NoSuchElementException;
import software.amazon.awssdk.annotations.SdkInternalApi;
import software.amazon.awssdk.core.pagination.sync.SdkIterable;
+import software.amazon.awssdk.services.s3.internal.multipart.SdkPojoConversionUtils;
import software.amazon.awssdk.services.s3.model.CopyObjectRequest;
import software.amazon.awssdk.services.s3.model.UploadPartCopyRequest;
@@ -65,10 +66,10 @@ public UploadPartCopyRequest next() {
long partSize = Math.min(optimalPartSize, remainingBytes);
String range = range(partSize);
UploadPartCopyRequest uploadPartCopyRequest =
- CopyRequestConversionUtils.toUploadPartCopyRequest(copyObjectRequest,
- partNumber,
- uploadId,
- range);
+ SdkPojoConversionUtils.toUploadPartCopyRequest(copyObjectRequest,
+ partNumber,
+ uploadId,
+ range);
partNumber++;
offset += partSize;
remainingBytes -= partSize;
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/handlers/PutObjectInterceptor.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/handlers/StreamingRequestInterceptor.java
similarity index 85%
rename from services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/handlers/PutObjectInterceptor.java
rename to services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/handlers/StreamingRequestInterceptor.java
index 92859b96dea..82a47cee6f3 100644
--- a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/handlers/PutObjectInterceptor.java
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/handlers/StreamingRequestInterceptor.java
@@ -21,17 +21,18 @@
import software.amazon.awssdk.core.interceptor.ExecutionInterceptor;
import software.amazon.awssdk.http.SdkHttpRequest;
import software.amazon.awssdk.services.s3.model.PutObjectRequest;
+import software.amazon.awssdk.services.s3.model.UploadPartRequest;
/**
* Interceptor to add an 'Expect: 100-continue' header to the HTTP Request if it represents a PUT Object request.
*/
@SdkInternalApi
//TODO: This should be generalized for all streaming requests
-public final class PutObjectInterceptor implements ExecutionInterceptor {
+public final class StreamingRequestInterceptor implements ExecutionInterceptor {
@Override
public SdkHttpRequest modifyHttpRequest(Context.ModifyHttpRequest context,
ExecutionAttributes executionAttributes) {
- if (context.request() instanceof PutObjectRequest) {
+ if (context.request() instanceof PutObjectRequest || context.request() instanceof UploadPartRequest) {
return context.httpRequest().toBuilder().putHeader("Expect", "100-continue").build();
}
return context.httpRequest();
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/CopyObjectHelper.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/CopyObjectHelper.java
similarity index 71%
rename from services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/CopyObjectHelper.java
rename to services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/CopyObjectHelper.java
index a26c5ab62c5..afb1ca0e4e8 100644
--- a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/crt/CopyObjectHelper.java
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/CopyObjectHelper.java
@@ -13,21 +13,17 @@
* permissions and limitations under the License.
*/
-package software.amazon.awssdk.services.s3.internal.crt;
+package software.amazon.awssdk.services.s3.internal.multipart;
import java.util.ArrayList;
import java.util.List;
import java.util.concurrent.CompletableFuture;
-import java.util.concurrent.CompletionException;
import java.util.concurrent.atomic.AtomicReferenceArray;
-import java.util.function.BiFunction;
-import java.util.function.Supplier;
import java.util.stream.IntStream;
import software.amazon.awssdk.annotations.SdkInternalApi;
-import software.amazon.awssdk.core.exception.SdkClientException;
import software.amazon.awssdk.services.s3.S3AsyncClient;
-import software.amazon.awssdk.services.s3.model.AbortMultipartUploadRequest;
+import software.amazon.awssdk.services.s3.internal.crt.UploadPartCopyRequestIterable;
import software.amazon.awssdk.services.s3.model.CompleteMultipartUploadRequest;
import software.amazon.awssdk.services.s3.model.CompleteMultipartUploadResponse;
import software.amazon.awssdk.services.s3.model.CompletedMultipartUpload;
@@ -50,17 +46,18 @@
public final class CopyObjectHelper {
private static final Logger log = Logger.loggerFor(S3AsyncClient.class);
- /**
- * The max number of parts on S3 side is 10,000
- */
- private static final long MAX_UPLOAD_PARTS = 10_000;
-
private final S3AsyncClient s3AsyncClient;
private final long partSizeInBytes;
+ private final GenericMultipartHelper genericMultipartHelper;
+ private final long uploadThreshold;
- public CopyObjectHelper(S3AsyncClient s3AsyncClient, long partSizeInBytes) {
+ public CopyObjectHelper(S3AsyncClient s3AsyncClient, long partSizeInBytes, long uploadThreshold) {
this.s3AsyncClient = s3AsyncClient;
this.partSizeInBytes = partSizeInBytes;
+ this.genericMultipartHelper = new GenericMultipartHelper<>(s3AsyncClient,
+ SdkPojoConversionUtils::toAbortMultipartUploadRequest,
+ SdkPojoConversionUtils::toCopyObjectResponse);
+ this.uploadThreshold = uploadThreshold;
}
public CompletableFuture copyObject(CopyObjectRequest copyObjectRequest) {
@@ -69,14 +66,15 @@ public CompletableFuture copyObject(CopyObjectRequest copyOb
try {
CompletableFuture headFuture =
- s3AsyncClient.headObject(CopyRequestConversionUtils.toHeadObjectRequest(copyObjectRequest));
+ s3AsyncClient.headObject(SdkPojoConversionUtils.toHeadObjectRequest(copyObjectRequest));
// Ensure cancellations are forwarded to the head future
CompletableFutureUtils.forwardExceptionTo(returnFuture, headFuture);
headFuture.whenComplete((headObjectResponse, throwable) -> {
if (throwable != null) {
- handleException(returnFuture, () -> "Failed to retrieve metadata from the source object", throwable);
+ genericMultipartHelper.handleException(returnFuture, () -> "Failed to retrieve metadata from the source "
+ + "object", throwable);
} else {
doCopyObject(copyObjectRequest, returnFuture, headObjectResponse);
}
@@ -92,7 +90,7 @@ private void doCopyObject(CopyObjectRequest copyObjectRequest, CompletableFuture
HeadObjectResponse headObjectResponse) {
Long contentLength = headObjectResponse.contentLength();
- if (contentLength <= partSizeInBytes) {
+ if (contentLength <= partSizeInBytes || contentLength <= uploadThreshold) {
log.debug(() -> "Starting the copy as a single copy part request");
copyInOneChunk(copyObjectRequest, returnFuture);
} else {
@@ -105,7 +103,7 @@ private void copyInParts(CopyObjectRequest copyObjectRequest,
Long contentLength,
CompletableFuture returnFuture) {
- CreateMultipartUploadRequest request = CopyRequestConversionUtils.toCreateMultipartUploadRequest(copyObjectRequest);
+ CreateMultipartUploadRequest request = SdkPojoConversionUtils.toCreateMultipartUploadRequest(copyObjectRequest);
CompletableFuture createMultipartUploadFuture =
s3AsyncClient.createMultipartUpload(request);
@@ -114,7 +112,7 @@ private void copyInParts(CopyObjectRequest copyObjectRequest,
createMultipartUploadFuture.whenComplete((createMultipartUploadResponse, throwable) -> {
if (throwable != null) {
- handleException(returnFuture, () -> "Failed to initiate multipart upload", throwable);
+ genericMultipartHelper.handleException(returnFuture, () -> "Failed to initiate multipart upload", throwable);
} else {
log.debug(() -> "Initiated new multipart upload, uploadId: " + createMultipartUploadResponse.uploadId());
doCopyInParts(copyObjectRequest, contentLength, returnFuture, createMultipartUploadResponse.uploadId());
@@ -122,17 +120,18 @@ private void copyInParts(CopyObjectRequest copyObjectRequest,
});
}
- private int determinePartCount(long contentLength, long partSize) {
- return (int) Math.ceil(contentLength / (double) partSize);
- }
-
private void doCopyInParts(CopyObjectRequest copyObjectRequest,
Long contentLength,
CompletableFuture returnFuture,
String uploadId) {
- long optimalPartSize = calculateOptimalPartSizeForCopy(contentLength);
- int partCount = determinePartCount(contentLength, optimalPartSize);
+ long optimalPartSize = genericMultipartHelper.calculateOptimalPartSizeFor(contentLength, partSizeInBytes);
+
+ int partCount = genericMultipartHelper.determinePartCount(contentLength, optimalPartSize);
+ if (optimalPartSize > partSizeInBytes) {
+ log.debug(() -> String.format("Configured partSize is %d, but using %d to prevent reaching maximum number of parts "
+ + "allowed", partSizeInBytes, optimalPartSize));
+ }
log.debug(() -> String.format("Starting multipart copy with partCount: %s, optimalPartSize: %s",
partCount, optimalPartSize));
@@ -147,32 +146,15 @@ private void doCopyInParts(CopyObjectRequest copyObjectRequest,
optimalPartSize);
CompletableFutureUtils.allOfExceptionForwarded(futures.toArray(new CompletableFuture[0]))
.thenCompose(ignore -> completeMultipartUpload(copyObjectRequest, uploadId, completedParts))
- .handle(handleExceptionOrResponse(copyObjectRequest, returnFuture, uploadId))
+ .handle(genericMultipartHelper.handleExceptionOrResponse(copyObjectRequest, returnFuture,
+ uploadId))
.exceptionally(throwable -> {
- handleException(returnFuture, () -> "Unexpected exception occurred", throwable);
+ genericMultipartHelper.handleException(returnFuture, () -> "Unexpected exception occurred",
+ throwable);
return null;
});
}
- private BiFunction handleExceptionOrResponse(
- CopyObjectRequest copyObjectRequest,
- CompletableFuture returnFuture,
- String uploadId) {
-
- return (completeMultipartUploadResponse, throwable) -> {
- if (throwable != null) {
- cleanUpParts(copyObjectRequest, uploadId);
- handleException(returnFuture, () -> "Failed to send multipart copy requests.",
- throwable);
- } else {
- returnFuture.complete(CopyRequestConversionUtils.toCopyObjectResponse(
- completeMultipartUploadResponse));
- }
-
- return null;
- };
- }
-
private CompletableFuture completeMultipartUpload(
CopyObjectRequest copyObjectRequest, String uploadId, AtomicReferenceArray completedParts) {
log.debug(() -> String.format("Sending completeMultipartUploadRequest, uploadId: %s",
@@ -193,39 +175,9 @@ private CompletableFuture completeMultipartUplo
.sseCustomerKey(copyObjectRequest.sseCustomerKey())
.sseCustomerKeyMD5(copyObjectRequest.sseCustomerKeyMD5())
.build();
-
return s3AsyncClient.completeMultipartUpload(completeMultipartUploadRequest);
}
- private void cleanUpParts(CopyObjectRequest copyObjectRequest, String uploadId) {
- AbortMultipartUploadRequest abortMultipartUploadRequest =
- CopyRequestConversionUtils.toAbortMultipartUploadRequest(copyObjectRequest, uploadId);
- s3AsyncClient.abortMultipartUpload(abortMultipartUploadRequest)
- .exceptionally(throwable -> {
- log.warn(() -> String.format("Failed to abort previous multipart upload "
- + "(id: %s)"
- + ". You may need to call "
- + "S3AsyncClient#abortMultiPartUpload to "
- + "free all storage consumed by"
- + " all parts. ",
- uploadId), throwable);
- return null;
- });
- }
-
- private static void handleException(CompletableFuture returnFuture,
- Supplier message,
- Throwable throwable) {
- Throwable cause = throwable instanceof CompletionException ? throwable.getCause() : throwable;
-
- if (cause instanceof Error) {
- returnFuture.completeExceptionally(cause);
- } else {
- SdkClientException exception = SdkClientException.create(message.get(), cause);
- returnFuture.completeExceptionally(exception);
- }
- }
-
private List> sendUploadPartCopyRequests(CopyObjectRequest copyObjectRequest,
long contentLength,
String uploadId,
@@ -253,7 +205,8 @@ private void sendIndividualUploadPartCopy(String uploadId,
log.debug(() -> "Sending uploadPartCopyRequest with range: " + uploadPartCopyRequest.copySourceRange() + " uploadId: "
+ uploadId);
- CompletableFuture uploadPartCopyFuture = s3AsyncClient.uploadPartCopy(uploadPartCopyRequest);
+ CompletableFuture uploadPartCopyFuture =
+ s3AsyncClient.uploadPartCopy(uploadPartCopyRequest);
CompletableFuture convertFuture =
uploadPartCopyFuture.thenApply(uploadPartCopyResponse ->
@@ -268,23 +221,13 @@ private static CompletedPart convertUploadPartCopyResponse(AtomicReferenceArray<
UploadPartCopyResponse uploadPartCopyResponse) {
CopyPartResult copyPartResult = uploadPartCopyResponse.copyPartResult();
CompletedPart completedPart =
- CopyRequestConversionUtils.toCompletedPart(copyPartResult,
- partNumber);
+ SdkPojoConversionUtils.toCompletedPart(copyPartResult,
+ partNumber);
completedParts.set(partNumber - 1, completedPart);
return completedPart;
}
- /**
- * Calculates the optimal part size of each part request if the copy operation is carried out as multipart copy.
- */
- private long calculateOptimalPartSizeForCopy(long contentLengthOfSource) {
- double optimalPartSize = contentLengthOfSource / (double) MAX_UPLOAD_PARTS;
-
- optimalPartSize = Math.ceil(optimalPartSize);
- return (long) Math.max(optimalPartSize, partSizeInBytes);
- }
-
private void copyInOneChunk(CopyObjectRequest copyObjectRequest,
CompletableFuture returnFuture) {
CompletableFuture copyObjectFuture =
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/GenericMultipartHelper.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/GenericMultipartHelper.java
new file mode 100644
index 00000000000..38e76394958
--- /dev/null
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/GenericMultipartHelper.java
@@ -0,0 +1,140 @@
+/*
+ * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License").
+ * You may not use this file except in compliance with the License.
+ * A copy of the License is located at
+ *
+ * http://aws.amazon.com/apache2.0
+ *
+ * or in the "license" file accompanying this file. This file is distributed
+ * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
+ * express or implied. See the License for the specific language governing
+ * permissions and limitations under the License.
+ */
+
+package software.amazon.awssdk.services.s3.internal.multipart;
+
+import java.util.concurrent.CompletableFuture;
+import java.util.concurrent.CompletionException;
+import java.util.concurrent.atomic.AtomicReferenceArray;
+import java.util.function.BiFunction;
+import java.util.function.Function;
+import java.util.function.Supplier;
+import java.util.stream.IntStream;
+import software.amazon.awssdk.annotations.SdkInternalApi;
+import software.amazon.awssdk.core.exception.SdkClientException;
+import software.amazon.awssdk.services.s3.S3AsyncClient;
+import software.amazon.awssdk.services.s3.model.AbortMultipartUploadRequest;
+import software.amazon.awssdk.services.s3.model.CompleteMultipartUploadRequest;
+import software.amazon.awssdk.services.s3.model.CompleteMultipartUploadResponse;
+import software.amazon.awssdk.services.s3.model.CompletedMultipartUpload;
+import software.amazon.awssdk.services.s3.model.CompletedPart;
+import software.amazon.awssdk.services.s3.model.S3Request;
+import software.amazon.awssdk.services.s3.model.S3Response;
+import software.amazon.awssdk.utils.Logger;
+
+@SdkInternalApi
+public final class GenericMultipartHelper {
+ private static final Logger log = Logger.loggerFor(GenericMultipartHelper.class);
+ /**
+ * The max number of parts on S3 side is 10,000
+ */
+ private static final long MAX_UPLOAD_PARTS = 10_000;
+
+ private final S3AsyncClient s3AsyncClient;
+ private final Function abortMultipartUploadRequestConverter;
+ private final Function responseConverter;
+
+ public GenericMultipartHelper(S3AsyncClient s3AsyncClient,
+ Function abortMultipartUploadRequestConverter,
+ Function responseConverter) {
+ this.s3AsyncClient = s3AsyncClient;
+ this.abortMultipartUploadRequestConverter = abortMultipartUploadRequestConverter;
+ this.responseConverter = responseConverter;
+ }
+
+ public void handleException(CompletableFuture returnFuture,
+ Supplier message,
+ Throwable throwable) {
+ Throwable cause = throwable instanceof CompletionException ? throwable.getCause() : throwable;
+
+ if (cause instanceof Error) {
+ returnFuture.completeExceptionally(cause);
+ } else {
+ SdkClientException exception = SdkClientException.create(message.get(), cause);
+ returnFuture.completeExceptionally(exception);
+ }
+ }
+
+ public long calculateOptimalPartSizeFor(long contentLengthOfSource, long partSizeInBytes) {
+ double optimalPartSize = contentLengthOfSource / (double) MAX_UPLOAD_PARTS;
+
+ optimalPartSize = Math.ceil(optimalPartSize);
+ return (long) Math.max(optimalPartSize, partSizeInBytes);
+ }
+
+ public int determinePartCount(long contentLength, long partSize) {
+ return (int) Math.ceil(contentLength / (double) partSize);
+ }
+
+ public CompletableFuture completeMultipartUpload(
+ RequestT request, String uploadId, CompletedPart[] parts) {
+ log.debug(() -> String.format("Sending completeMultipartUploadRequest, uploadId: %s",
+ uploadId));
+ CompleteMultipartUploadRequest completeMultipartUploadRequest =
+ CompleteMultipartUploadRequest.builder()
+ .bucket(request.getValueForField("Bucket", String.class).get())
+ .key(request.getValueForField("Key", String.class).get())
+ .uploadId(uploadId)
+ .multipartUpload(CompletedMultipartUpload.builder()
+ .parts(parts)
+ .build())
+ .build();
+ return s3AsyncClient.completeMultipartUpload(completeMultipartUploadRequest);
+ }
+
+ public CompletableFuture completeMultipartUpload(
+ RequestT request, String uploadId, AtomicReferenceArray completedParts) {
+ CompletedPart[] parts =
+ IntStream.range(0, completedParts.length())
+ .mapToObj(completedParts::get)
+ .toArray(CompletedPart[]::new);
+ return completeMultipartUpload(request, uploadId, parts);
+ }
+
+ public BiFunction handleExceptionOrResponse(
+ RequestT request,
+ CompletableFuture returnFuture,
+ String uploadId) {
+
+ return (completeMultipartUploadResponse, throwable) -> {
+ if (throwable != null) {
+ cleanUpParts(uploadId, abortMultipartUploadRequestConverter.apply(request));
+ handleException(returnFuture, () -> "Failed to send multipart requests",
+ throwable);
+ } else {
+ returnFuture.complete(responseConverter.apply(
+ completeMultipartUploadResponse));
+ }
+
+ return null;
+ };
+ }
+
+ public void cleanUpParts(String uploadId, AbortMultipartUploadRequest.Builder abortMultipartUploadRequest) {
+ log.debug(() -> "Aborting multipart upload: " + uploadId);
+ AbortMultipartUploadRequest request = abortMultipartUploadRequest.uploadId(uploadId).build();
+ s3AsyncClient.abortMultipartUpload(request)
+ .exceptionally(throwable -> {
+ log.warn(() -> String.format("Failed to abort previous multipart upload "
+ + "(id: %s)"
+ + ". You may need to call "
+ + "S3AsyncClient#abortMultiPartUpload to "
+ + "free all storage consumed by"
+ + " all parts. ",
+ uploadId), throwable);
+ return null;
+ });
+ }
+}
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/MultipartS3AsyncClient.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/MultipartS3AsyncClient.java
new file mode 100644
index 00000000000..65b26ddec97
--- /dev/null
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/MultipartS3AsyncClient.java
@@ -0,0 +1,108 @@
+/*
+ * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License").
+ * You may not use this file except in compliance with the License.
+ * A copy of the License is located at
+ *
+ * http://aws.amazon.com/apache2.0
+ *
+ * or in the "license" file accompanying this file. This file is distributed
+ * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
+ * express or implied. See the License for the specific language governing
+ * permissions and limitations under the License.
+ */
+
+package software.amazon.awssdk.services.s3.internal.multipart;
+
+
+import java.util.concurrent.CompletableFuture;
+import java.util.function.Function;
+import software.amazon.awssdk.annotations.SdkInternalApi;
+import software.amazon.awssdk.core.ApiName;
+import software.amazon.awssdk.core.async.AsyncRequestBody;
+import software.amazon.awssdk.core.async.AsyncResponseTransformer;
+import software.amazon.awssdk.services.s3.DelegatingS3AsyncClient;
+import software.amazon.awssdk.services.s3.S3AsyncClient;
+import software.amazon.awssdk.services.s3.internal.UserAgentUtils;
+import software.amazon.awssdk.services.s3.model.CopyObjectRequest;
+import software.amazon.awssdk.services.s3.model.CopyObjectResponse;
+import software.amazon.awssdk.services.s3.model.GetObjectRequest;
+import software.amazon.awssdk.services.s3.model.GetObjectResponse;
+import software.amazon.awssdk.services.s3.model.PutObjectRequest;
+import software.amazon.awssdk.services.s3.model.PutObjectResponse;
+import software.amazon.awssdk.services.s3.model.S3Request;
+import software.amazon.awssdk.services.s3.multipart.MultipartConfiguration;
+import software.amazon.awssdk.utils.Validate;
+
+/**
+ * An {@link S3AsyncClient} that automatically converts put, copy requests to their respective multipart call. Note: get is not
+ * yet supported.
+ *
+ * @see MultipartConfiguration
+ */
+@SdkInternalApi
+public final class MultipartS3AsyncClient extends DelegatingS3AsyncClient {
+
+ private static final ApiName USER_AGENT_API_NAME = ApiName.builder().name("hll").version("s3Multipart").build();
+
+ private static final long DEFAULT_MIN_PART_SIZE = 8L * 1024 * 1024;
+ private static final long DEFAULT_THRESHOLD = 8L * 1024 * 1024;
+ private static final long DEFAULT_API_CALL_BUFFER_SIZE = DEFAULT_MIN_PART_SIZE * 4;
+
+ private final UploadObjectHelper mpuHelper;
+ private final CopyObjectHelper copyObjectHelper;
+
+ private MultipartS3AsyncClient(S3AsyncClient delegate, MultipartConfiguration multipartConfiguration) {
+ super(delegate);
+ MultipartConfiguration validConfiguration = Validate.getOrDefault(multipartConfiguration,
+ MultipartConfiguration.builder()::build);
+ long minPartSizeInBytes = Validate.getOrDefault(validConfiguration.minimumPartSizeInBytes(),
+ () -> DEFAULT_MIN_PART_SIZE);
+ long threshold = Validate.getOrDefault(validConfiguration.thresholdInBytes(),
+ () -> DEFAULT_THRESHOLD);
+ long apiCallBufferSizeInBytes = Validate.getOrDefault(validConfiguration.apiCallBufferSizeInBytes(),
+ () -> computeApiCallBufferSize(validConfiguration));
+ mpuHelper = new UploadObjectHelper(delegate, minPartSizeInBytes, threshold, apiCallBufferSizeInBytes);
+ copyObjectHelper = new CopyObjectHelper(delegate, minPartSizeInBytes, threshold);
+ }
+
+ private long computeApiCallBufferSize(MultipartConfiguration multipartConfiguration) {
+ return multipartConfiguration.minimumPartSizeInBytes() != null ? multipartConfiguration.minimumPartSizeInBytes() * 4
+ : DEFAULT_API_CALL_BUFFER_SIZE;
+ }
+
+ @Override
+ public CompletableFuture putObject(PutObjectRequest putObjectRequest, AsyncRequestBody requestBody) {
+ return mpuHelper.uploadObject(putObjectRequest, requestBody);
+ }
+
+ @Override
+ public CompletableFuture copyObject(CopyObjectRequest copyObjectRequest) {
+ return copyObjectHelper.copyObject(copyObjectRequest);
+ }
+
+ @Override
+ public CompletableFuture getObject(
+ GetObjectRequest getObjectRequest, AsyncResponseTransformer asyncResponseTransformer) {
+ throw new UnsupportedOperationException(
+ "Multipart download is not yet supported. Instead use the CRT based S3 client for multipart download.");
+ }
+
+ @Override
+ public void close() {
+ delegate().close();
+ }
+
+ public static MultipartS3AsyncClient create(S3AsyncClient client, MultipartConfiguration multipartConfiguration) {
+ S3AsyncClient clientWithUserAgent = new DelegatingS3AsyncClient(client) {
+ @Override
+ protected CompletableFuture invokeOperation(T request, Function> operation) {
+ T requestWithUserAgent = UserAgentUtils.applyUserAgentInfo(request, c -> c.addApiName(USER_AGENT_API_NAME));
+ return operation.apply(requestWithUserAgent);
+ }
+ };
+ return new MultipartS3AsyncClient(clientWithUserAgent, multipartConfiguration);
+ }
+}
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/MultipartUploadHelper.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/MultipartUploadHelper.java
new file mode 100644
index 00000000000..9754d284f5b
--- /dev/null
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/MultipartUploadHelper.java
@@ -0,0 +1,147 @@
+/*
+ * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License").
+ * You may not use this file except in compliance with the License.
+ * A copy of the License is located at
+ *
+ * http://aws.amazon.com/apache2.0
+ *
+ * or in the "license" file accompanying this file. This file is distributed
+ * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
+ * express or implied. See the License for the specific language governing
+ * permissions and limitations under the License.
+ */
+
+package software.amazon.awssdk.services.s3.internal.multipart;
+
+
+import static software.amazon.awssdk.services.s3.internal.multipart.SdkPojoConversionUtils.toAbortMultipartUploadRequest;
+
+import java.util.Collection;
+import java.util.concurrent.CompletableFuture;
+import java.util.function.Consumer;
+import software.amazon.awssdk.annotations.SdkInternalApi;
+import software.amazon.awssdk.core.async.AsyncRequestBody;
+import software.amazon.awssdk.services.s3.S3AsyncClient;
+import software.amazon.awssdk.services.s3.model.CompletedPart;
+import software.amazon.awssdk.services.s3.model.CreateMultipartUploadRequest;
+import software.amazon.awssdk.services.s3.model.CreateMultipartUploadResponse;
+import software.amazon.awssdk.services.s3.model.PutObjectRequest;
+import software.amazon.awssdk.services.s3.model.PutObjectResponse;
+import software.amazon.awssdk.services.s3.model.UploadPartRequest;
+import software.amazon.awssdk.services.s3.model.UploadPartResponse;
+import software.amazon.awssdk.utils.CompletableFutureUtils;
+import software.amazon.awssdk.utils.Logger;
+import software.amazon.awssdk.utils.Pair;
+
+/**
+ * A base class contains common logic used by {@link UploadWithUnknownContentLengthHelper} and
+ * {@link UploadWithKnownContentLengthHelper}.
+ */
+@SdkInternalApi
+public final class MultipartUploadHelper {
+ private static final Logger log = Logger.loggerFor(MultipartUploadHelper.class);
+
+ private final S3AsyncClient s3AsyncClient;
+ private final long partSizeInBytes;
+ private final GenericMultipartHelper genericMultipartHelper;
+
+ private final long maxMemoryUsageInBytes;
+ private final long multipartUploadThresholdInBytes;
+
+ public MultipartUploadHelper(S3AsyncClient s3AsyncClient,
+ long partSizeInBytes,
+ long multipartUploadThresholdInBytes,
+ long maxMemoryUsageInBytes) {
+ this.s3AsyncClient = s3AsyncClient;
+ this.partSizeInBytes = partSizeInBytes;
+ this.genericMultipartHelper = new GenericMultipartHelper<>(s3AsyncClient,
+ SdkPojoConversionUtils::toAbortMultipartUploadRequest,
+ SdkPojoConversionUtils::toPutObjectResponse);
+ this.maxMemoryUsageInBytes = maxMemoryUsageInBytes;
+ this.multipartUploadThresholdInBytes = multipartUploadThresholdInBytes;
+ }
+
+ CompletableFuture createMultipartUpload(PutObjectRequest putObjectRequest,
+ CompletableFuture returnFuture) {
+ CreateMultipartUploadRequest request = SdkPojoConversionUtils.toCreateMultipartUploadRequest(putObjectRequest);
+ CompletableFuture createMultipartUploadFuture =
+ s3AsyncClient.createMultipartUpload(request);
+
+ // Ensure cancellations are forwarded to the createMultipartUploadFuture future
+ CompletableFutureUtils.forwardExceptionTo(returnFuture, createMultipartUploadFuture);
+ return createMultipartUploadFuture;
+ }
+
+ void completeMultipartUpload(CompletableFuture returnFuture,
+ String uploadId,
+ CompletedPart[] completedParts,
+ PutObjectRequest putObjectRequest) {
+ genericMultipartHelper.completeMultipartUpload(putObjectRequest,
+ uploadId,
+ completedParts)
+ .handle(genericMultipartHelper.handleExceptionOrResponse(putObjectRequest, returnFuture,
+ uploadId))
+ .exceptionally(throwable -> {
+ genericMultipartHelper.handleException(returnFuture, () -> "Unexpected exception occurred",
+ throwable);
+ return null;
+ });
+ }
+
+ CompletableFuture sendIndividualUploadPartRequest(String uploadId,
+ Consumer completedPartsConsumer,
+ Collection> futures,
+ Pair requestPair) {
+ UploadPartRequest uploadPartRequest = requestPair.left();
+ Integer partNumber = uploadPartRequest.partNumber();
+ log.debug(() -> "Sending uploadPartRequest: " + uploadPartRequest.partNumber() + " uploadId: " + uploadId + " "
+ + "contentLength " + requestPair.right().contentLength());
+
+ CompletableFuture uploadPartFuture = s3AsyncClient.uploadPart(uploadPartRequest,
+ requestPair.right());
+
+ CompletableFuture convertFuture =
+ uploadPartFuture.thenApply(uploadPartResponse -> convertUploadPartResponse(completedPartsConsumer, partNumber,
+ uploadPartResponse));
+ futures.add(convertFuture);
+ CompletableFutureUtils.forwardExceptionTo(convertFuture, uploadPartFuture);
+ return convertFuture;
+ }
+
+ void failRequestsElegantly(Collection> futures,
+ Throwable t,
+ String uploadId,
+ CompletableFuture returnFuture,
+ PutObjectRequest putObjectRequest) {
+ genericMultipartHelper.handleException(returnFuture, () -> "Failed to send multipart upload requests", t);
+ if (uploadId != null) {
+ genericMultipartHelper.cleanUpParts(uploadId, toAbortMultipartUploadRequest(putObjectRequest));
+ }
+ cancelingOtherOngoingRequests(futures, t);
+ }
+
+ static void cancelingOtherOngoingRequests(Collection> futures, Throwable t) {
+ log.trace(() -> "cancelling other ongoing requests " + futures.size());
+ futures.forEach(f -> f.completeExceptionally(t));
+ }
+
+ static CompletedPart convertUploadPartResponse(Consumer consumer,
+ Integer partNumber,
+ UploadPartResponse uploadPartResponse) {
+ CompletedPart completedPart = SdkPojoConversionUtils.toCompletedPart(uploadPartResponse, partNumber);
+
+ consumer.accept(completedPart);
+ return completedPart;
+ }
+
+ void uploadInOneChunk(PutObjectRequest putObjectRequest,
+ AsyncRequestBody asyncRequestBody,
+ CompletableFuture returnFuture) {
+ CompletableFuture putObjectResponseCompletableFuture = s3AsyncClient.putObject(putObjectRequest,
+ asyncRequestBody);
+ CompletableFutureUtils.forwardExceptionTo(returnFuture, putObjectResponseCompletableFuture);
+ CompletableFutureUtils.forwardResultTo(putObjectResponseCompletableFuture, returnFuture);
+ }
+}
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/SdkPojoConversionUtils.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/SdkPojoConversionUtils.java
new file mode 100644
index 00000000000..25fde18cada
--- /dev/null
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/SdkPojoConversionUtils.java
@@ -0,0 +1,195 @@
+/*
+ * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License").
+ * You may not use this file except in compliance with the License.
+ * A copy of the License is located at
+ *
+ * http://aws.amazon.com/apache2.0
+ *
+ * or in the "license" file accompanying this file. This file is distributed
+ * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
+ * express or implied. See the License for the specific language governing
+ * permissions and limitations under the License.
+ */
+
+package software.amazon.awssdk.services.s3.internal.multipart;
+
+import java.util.Arrays;
+import java.util.HashMap;
+import java.util.HashSet;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+import software.amazon.awssdk.annotations.SdkInternalApi;
+import software.amazon.awssdk.core.SdkField;
+import software.amazon.awssdk.core.SdkPojo;
+import software.amazon.awssdk.services.s3.model.AbortMultipartUploadRequest;
+import software.amazon.awssdk.services.s3.model.CompleteMultipartUploadResponse;
+import software.amazon.awssdk.services.s3.model.CompletedPart;
+import software.amazon.awssdk.services.s3.model.CopyObjectRequest;
+import software.amazon.awssdk.services.s3.model.CopyObjectResponse;
+import software.amazon.awssdk.services.s3.model.CopyObjectResult;
+import software.amazon.awssdk.services.s3.model.CopyPartResult;
+import software.amazon.awssdk.services.s3.model.CreateMultipartUploadRequest;
+import software.amazon.awssdk.services.s3.model.HeadObjectRequest;
+import software.amazon.awssdk.services.s3.model.PutObjectRequest;
+import software.amazon.awssdk.services.s3.model.PutObjectResponse;
+import software.amazon.awssdk.services.s3.model.UploadPartCopyRequest;
+import software.amazon.awssdk.services.s3.model.UploadPartRequest;
+import software.amazon.awssdk.services.s3.model.UploadPartResponse;
+import software.amazon.awssdk.utils.Logger;
+
+/**
+ * Request conversion utility method for POJO classes associated with multipart feature.
+ */
+@SdkInternalApi
+public final class SdkPojoConversionUtils {
+ private static final Logger log = Logger.loggerFor(SdkPojoConversionUtils.class);
+
+ private static final HashSet PUT_OBJECT_REQUEST_TO_UPLOAD_PART_FIELDS_TO_IGNORE =
+ new HashSet<>(Arrays.asList("ChecksumSHA1", "ChecksumSHA256", "ContentMD5", "ChecksumCRC32C", "ChecksumCRC32"));
+
+ private SdkPojoConversionUtils() {
+ }
+
+ public static UploadPartRequest toUploadPartRequest(PutObjectRequest putObjectRequest, int partNumber, String uploadId) {
+
+ UploadPartRequest.Builder builder = UploadPartRequest.builder();
+
+ setSdkFields(builder, putObjectRequest, PUT_OBJECT_REQUEST_TO_UPLOAD_PART_FIELDS_TO_IGNORE);
+
+ return builder.uploadId(uploadId).partNumber(partNumber).build();
+ }
+
+ public static CreateMultipartUploadRequest toCreateMultipartUploadRequest(PutObjectRequest putObjectRequest) {
+
+ CreateMultipartUploadRequest.Builder builder = CreateMultipartUploadRequest.builder();
+ setSdkFields(builder, putObjectRequest);
+ return builder.build();
+ }
+
+ public static HeadObjectRequest toHeadObjectRequest(CopyObjectRequest copyObjectRequest) {
+
+ // We can't set SdkFields directly because the fields in CopyObjectRequest do not match 100% with the ones in
+ // HeadObjectRequest
+ return HeadObjectRequest.builder()
+ .bucket(copyObjectRequest.sourceBucket())
+ .key(copyObjectRequest.sourceKey())
+ .versionId(copyObjectRequest.sourceVersionId())
+ .ifMatch(copyObjectRequest.copySourceIfMatch())
+ .ifModifiedSince(copyObjectRequest.copySourceIfModifiedSince())
+ .ifNoneMatch(copyObjectRequest.copySourceIfNoneMatch())
+ .ifUnmodifiedSince(copyObjectRequest.copySourceIfUnmodifiedSince())
+ .expectedBucketOwner(copyObjectRequest.expectedSourceBucketOwner())
+ .sseCustomerAlgorithm(copyObjectRequest.copySourceSSECustomerAlgorithm())
+ .sseCustomerKey(copyObjectRequest.copySourceSSECustomerKey())
+ .sseCustomerKeyMD5(copyObjectRequest.copySourceSSECustomerKeyMD5())
+ .build();
+ }
+
+ public static CompletedPart toCompletedPart(CopyPartResult copyPartResult, int partNumber) {
+ CompletedPart.Builder builder = CompletedPart.builder();
+
+ setSdkFields(builder, copyPartResult);
+ return builder.partNumber(partNumber).build();
+ }
+
+ public static CompletedPart toCompletedPart(UploadPartResponse partResponse, int partNumber) {
+ CompletedPart.Builder builder = CompletedPart.builder();
+ setSdkFields(builder, partResponse);
+ return builder.partNumber(partNumber).build();
+ }
+
+ private static void setSdkFields(SdkPojo targetBuilder, SdkPojo sourceObject) {
+ setSdkFields(targetBuilder, sourceObject, new HashSet<>());
+ }
+
+ private static void setSdkFields(SdkPojo targetBuilder, SdkPojo sourceObject, Set fieldsToIgnore) {
+ Map sourceFields = retrieveSdkFields(sourceObject, sourceObject.sdkFields());
+ List> targetSdkFields = targetBuilder.sdkFields();
+
+ for (SdkField> field : targetSdkFields) {
+ if (fieldsToIgnore.contains(field.memberName())) {
+ continue;
+ }
+ field.set(targetBuilder, sourceFields.getOrDefault(field.memberName(), null));
+ }
+ }
+
+ public static CreateMultipartUploadRequest toCreateMultipartUploadRequest(CopyObjectRequest copyObjectRequest) {
+ CreateMultipartUploadRequest.Builder builder = CreateMultipartUploadRequest.builder();
+
+ setSdkFields(builder, copyObjectRequest);
+ builder.bucket(copyObjectRequest.destinationBucket());
+ builder.key(copyObjectRequest.destinationKey());
+ return builder.build();
+ }
+
+ public static CopyObjectResponse toCopyObjectResponse(CompleteMultipartUploadResponse response) {
+ CopyObjectResponse.Builder builder = CopyObjectResponse.builder();
+
+ setSdkFields(builder, response);
+
+ builder.responseMetadata(response.responseMetadata());
+ builder.sdkHttpResponse(response.sdkHttpResponse());
+
+ return builder.copyObjectResult(toCopyObjectResult(response))
+ .build();
+ }
+
+ private static CopyObjectResult toCopyObjectResult(CompleteMultipartUploadResponse response) {
+ CopyObjectResult.Builder builder = CopyObjectResult.builder();
+
+ setSdkFields(builder, response);
+ return builder.build();
+ }
+
+ public static AbortMultipartUploadRequest.Builder toAbortMultipartUploadRequest(CopyObjectRequest copyObjectRequest) {
+ AbortMultipartUploadRequest.Builder builder = AbortMultipartUploadRequest.builder();
+ setSdkFields(builder, copyObjectRequest);
+ builder.bucket(copyObjectRequest.destinationBucket());
+ builder.key(copyObjectRequest.destinationKey());
+ return builder;
+ }
+
+ public static AbortMultipartUploadRequest.Builder toAbortMultipartUploadRequest(PutObjectRequest putObjectRequest) {
+ AbortMultipartUploadRequest.Builder builder = AbortMultipartUploadRequest.builder();
+ setSdkFields(builder, putObjectRequest);
+ return builder;
+ }
+
+ public static UploadPartCopyRequest toUploadPartCopyRequest(CopyObjectRequest copyObjectRequest,
+ int partNumber,
+ String uploadId,
+ String range) {
+ UploadPartCopyRequest.Builder builder = UploadPartCopyRequest.builder();
+ setSdkFields(builder, copyObjectRequest);
+ return builder.copySourceRange(range)
+ .partNumber(partNumber)
+ .uploadId(uploadId)
+ .bucket(copyObjectRequest.destinationBucket())
+ .key(copyObjectRequest.destinationKey())
+ .build();
+ }
+
+ public static PutObjectResponse toPutObjectResponse(CompleteMultipartUploadResponse response) {
+
+ PutObjectResponse.Builder builder = PutObjectResponse.builder();
+
+ setSdkFields(builder, response);
+
+ builder.responseMetadata(response.responseMetadata());
+ builder.sdkHttpResponse(response.sdkHttpResponse());
+
+ return builder.build();
+ }
+
+ private static Map retrieveSdkFields(SdkPojo sourceObject, List> sdkFields) {
+ return sdkFields.stream().collect(
+ HashMap::new,
+ (map, field) -> map.put(field.memberName(),
+ field.getValueOrDefault(sourceObject)),
+ Map::putAll);
+ }
+}
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/UploadObjectHelper.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/UploadObjectHelper.java
new file mode 100644
index 00000000000..0700e8ade5f
--- /dev/null
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/UploadObjectHelper.java
@@ -0,0 +1,73 @@
+/*
+ * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License").
+ * You may not use this file except in compliance with the License.
+ * A copy of the License is located at
+ *
+ * http://aws.amazon.com/apache2.0
+ *
+ * or in the "license" file accompanying this file. This file is distributed
+ * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
+ * express or implied. See the License for the specific language governing
+ * permissions and limitations under the License.
+ */
+
+package software.amazon.awssdk.services.s3.internal.multipart;
+
+import java.util.concurrent.CompletableFuture;
+import software.amazon.awssdk.annotations.SdkInternalApi;
+import software.amazon.awssdk.core.async.AsyncRequestBody;
+import software.amazon.awssdk.services.s3.S3AsyncClient;
+import software.amazon.awssdk.services.s3.model.PutObjectRequest;
+import software.amazon.awssdk.services.s3.model.PutObjectResponse;
+import software.amazon.awssdk.utils.Logger;
+
+/**
+ * An internal helper class that automatically uses multipart upload based on the size of the object.
+ */
+@SdkInternalApi
+public final class UploadObjectHelper {
+ private static final Logger log = Logger.loggerFor(UploadObjectHelper.class);
+
+ private final S3AsyncClient s3AsyncClient;
+ private final long partSizeInBytes;
+ private final GenericMultipartHelper genericMultipartHelper;
+
+ private final long maxMemoryUsageInBytes;
+ private final long multipartUploadThresholdInBytes;
+ private final UploadWithKnownContentLengthHelper uploadWithKnownContentLength;
+ private final UploadWithUnknownContentLengthHelper uploadWithUnknownContentLength;
+
+ public UploadObjectHelper(S3AsyncClient s3AsyncClient,
+ long partSizeInBytes,
+ long multipartUploadThresholdInBytes,
+ long maxMemoryUsageInBytes) {
+ this.s3AsyncClient = s3AsyncClient;
+ this.partSizeInBytes = partSizeInBytes;
+ this.genericMultipartHelper = new GenericMultipartHelper<>(s3AsyncClient,
+ SdkPojoConversionUtils::toAbortMultipartUploadRequest,
+ SdkPojoConversionUtils::toPutObjectResponse);
+ this.maxMemoryUsageInBytes = maxMemoryUsageInBytes;
+ this.multipartUploadThresholdInBytes = multipartUploadThresholdInBytes;
+ this.uploadWithKnownContentLength = new UploadWithKnownContentLengthHelper(s3AsyncClient,
+ partSizeInBytes,
+ multipartUploadThresholdInBytes,
+ maxMemoryUsageInBytes);
+ this.uploadWithUnknownContentLength = new UploadWithUnknownContentLengthHelper(s3AsyncClient,
+ partSizeInBytes,
+ multipartUploadThresholdInBytes,
+ maxMemoryUsageInBytes);
+ }
+
+ public CompletableFuture uploadObject(PutObjectRequest putObjectRequest,
+ AsyncRequestBody asyncRequestBody) {
+ Long contentLength = asyncRequestBody.contentLength().orElseGet(putObjectRequest::contentLength);
+
+ if (contentLength == null) {
+ return uploadWithUnknownContentLength.uploadObject(putObjectRequest, asyncRequestBody);
+ } else {
+ return uploadWithKnownContentLength.uploadObject(putObjectRequest, asyncRequestBody, contentLength.longValue());
+ }
+ }
+}
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/UploadWithKnownContentLengthHelper.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/UploadWithKnownContentLengthHelper.java
new file mode 100644
index 00000000000..f7d199ac3aa
--- /dev/null
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/UploadWithKnownContentLengthHelper.java
@@ -0,0 +1,256 @@
+/*
+ * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License").
+ * You may not use this file except in compliance with the License.
+ * A copy of the License is located at
+ *
+ * http://aws.amazon.com/apache2.0
+ *
+ * or in the "license" file accompanying this file. This file is distributed
+ * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
+ * express or implied. See the License for the specific language governing
+ * permissions and limitations under the License.
+ */
+
+package software.amazon.awssdk.services.s3.internal.multipart;
+
+
+import java.util.Collection;
+import java.util.concurrent.CompletableFuture;
+import java.util.concurrent.ConcurrentLinkedQueue;
+import java.util.concurrent.atomic.AtomicBoolean;
+import java.util.concurrent.atomic.AtomicInteger;
+import java.util.concurrent.atomic.AtomicReferenceArray;
+import java.util.function.Consumer;
+import java.util.stream.IntStream;
+import org.reactivestreams.Subscriber;
+import org.reactivestreams.Subscription;
+import software.amazon.awssdk.annotations.SdkInternalApi;
+import software.amazon.awssdk.core.async.AsyncRequestBody;
+import software.amazon.awssdk.services.s3.S3AsyncClient;
+import software.amazon.awssdk.services.s3.model.CompletedPart;
+import software.amazon.awssdk.services.s3.model.CreateMultipartUploadResponse;
+import software.amazon.awssdk.services.s3.model.PutObjectRequest;
+import software.amazon.awssdk.services.s3.model.PutObjectResponse;
+import software.amazon.awssdk.services.s3.model.UploadPartRequest;
+import software.amazon.awssdk.utils.Logger;
+import software.amazon.awssdk.utils.Pair;
+
+/**
+ * An internal helper class that automatically uses multipart upload based on the size of the object.
+ */
+@SdkInternalApi
+public final class UploadWithKnownContentLengthHelper {
+ private static final Logger log = Logger.loggerFor(UploadWithKnownContentLengthHelper.class);
+
+ private final S3AsyncClient s3AsyncClient;
+ private final long partSizeInBytes;
+ private final GenericMultipartHelper genericMultipartHelper;
+
+ private final long maxMemoryUsageInBytes;
+ private final long multipartUploadThresholdInBytes;
+ private final MultipartUploadHelper multipartUploadHelper;
+
+ public UploadWithKnownContentLengthHelper(S3AsyncClient s3AsyncClient,
+ long partSizeInBytes,
+ long multipartUploadThresholdInBytes,
+ long maxMemoryUsageInBytes) {
+ this.s3AsyncClient = s3AsyncClient;
+ this.partSizeInBytes = partSizeInBytes;
+ this.genericMultipartHelper = new GenericMultipartHelper<>(s3AsyncClient,
+ SdkPojoConversionUtils::toAbortMultipartUploadRequest,
+ SdkPojoConversionUtils::toPutObjectResponse);
+ this.maxMemoryUsageInBytes = maxMemoryUsageInBytes;
+ this.multipartUploadThresholdInBytes = multipartUploadThresholdInBytes;
+ this.multipartUploadHelper = new MultipartUploadHelper(s3AsyncClient, partSizeInBytes, multipartUploadThresholdInBytes,
+ maxMemoryUsageInBytes);
+ }
+
+ public CompletableFuture uploadObject(PutObjectRequest putObjectRequest,
+ AsyncRequestBody asyncRequestBody,
+ long contentLength) {
+ CompletableFuture returnFuture = new CompletableFuture<>();
+
+ try {
+ if (contentLength > multipartUploadThresholdInBytes && contentLength > partSizeInBytes) {
+ log.debug(() -> "Starting the upload as multipart upload request");
+ uploadInParts(putObjectRequest, contentLength, asyncRequestBody, returnFuture);
+ } else {
+ log.debug(() -> "Starting the upload as a single upload part request");
+ multipartUploadHelper.uploadInOneChunk(putObjectRequest, asyncRequestBody, returnFuture);
+ }
+
+ } catch (Throwable throwable) {
+ returnFuture.completeExceptionally(throwable);
+ }
+
+ return returnFuture;
+ }
+
+ private void uploadInParts(PutObjectRequest putObjectRequest, long contentLength, AsyncRequestBody asyncRequestBody,
+ CompletableFuture returnFuture) {
+
+ CompletableFuture createMultipartUploadFuture =
+ multipartUploadHelper.createMultipartUpload(putObjectRequest, returnFuture);
+
+ createMultipartUploadFuture.whenComplete((createMultipartUploadResponse, throwable) -> {
+ if (throwable != null) {
+ genericMultipartHelper.handleException(returnFuture, () -> "Failed to initiate multipart upload", throwable);
+ } else {
+ log.debug(() -> "Initiated a new multipart upload, uploadId: " + createMultipartUploadResponse.uploadId());
+ doUploadInParts(Pair.of(putObjectRequest, asyncRequestBody), contentLength, returnFuture,
+ createMultipartUploadResponse.uploadId());
+ }
+ });
+ }
+
+ private void doUploadInParts(Pair request,
+ long contentLength,
+ CompletableFuture returnFuture,
+ String uploadId) {
+
+ long optimalPartSize = genericMultipartHelper.calculateOptimalPartSizeFor(contentLength, partSizeInBytes);
+ int partCount = genericMultipartHelper.determinePartCount(contentLength, optimalPartSize);
+ if (optimalPartSize > partSizeInBytes) {
+ log.debug(() -> String.format("Configured partSize is %d, but using %d to prevent reaching maximum number of parts "
+ + "allowed", partSizeInBytes, optimalPartSize));
+ }
+
+ log.debug(() -> String.format("Starting multipart upload with partCount: %d, optimalPartSize: %d", partCount,
+ optimalPartSize));
+
+ MpuRequestContext mpuRequestContext = new MpuRequestContext(request, contentLength, optimalPartSize, uploadId);
+
+ request.right()
+ .split(b -> b.chunkSizeInBytes(mpuRequestContext.partSize)
+ .bufferSizeInBytes(maxMemoryUsageInBytes))
+ .subscribe(new KnownContentLengthAsyncRequestBodySubscriber(mpuRequestContext,
+ returnFuture));
+ }
+
+ private static final class MpuRequestContext {
+ private final Pair request;
+ private final long contentLength;
+ private final long partSize;
+
+ private final String uploadId;
+
+ private MpuRequestContext(Pair request,
+ long contentLength,
+ long partSize,
+ String uploadId) {
+ this.request = request;
+ this.contentLength = contentLength;
+ this.partSize = partSize;
+ this.uploadId = uploadId;
+ }
+ }
+
+ private class KnownContentLengthAsyncRequestBodySubscriber implements Subscriber {
+
+ /**
+ * The number of AsyncRequestBody has been received but yet to be processed
+ */
+ private final AtomicInteger asyncRequestBodyInFlight = new AtomicInteger(0);
+
+ /**
+ * Indicates whether CompleteMultipart has been initiated or not.
+ */
+ private final AtomicBoolean completedMultipartInitiated = new AtomicBoolean(false);
+
+ private final AtomicBoolean failureActionInitiated = new AtomicBoolean(false);
+
+ private final AtomicInteger partNumber = new AtomicInteger(1);
+
+ private final AtomicReferenceArray completedParts;
+ private final String uploadId;
+ private final Collection> futures = new ConcurrentLinkedQueue<>();
+
+ private final PutObjectRequest putObjectRequest;
+ private final CompletableFuture returnFuture;
+ private Subscription subscription;
+
+ private volatile boolean isDone;
+
+ KnownContentLengthAsyncRequestBodySubscriber(MpuRequestContext mpuRequestContext,
+ CompletableFuture returnFuture) {
+ long optimalPartSize = genericMultipartHelper.calculateOptimalPartSizeFor(mpuRequestContext.contentLength,
+ partSizeInBytes);
+ int partCount = genericMultipartHelper.determinePartCount(mpuRequestContext.contentLength, optimalPartSize);
+ this.putObjectRequest = mpuRequestContext.request.left();
+ this.returnFuture = returnFuture;
+ this.completedParts = new AtomicReferenceArray<>(partCount);
+ this.uploadId = mpuRequestContext.uploadId;
+ }
+
+ @Override
+ public void onSubscribe(Subscription s) {
+ if (this.subscription != null) {
+ log.warn(() -> "The subscriber has already been subscribed. Cancelling the incoming subscription");
+ subscription.cancel();
+ return;
+ }
+ this.subscription = s;
+ s.request(1);
+ returnFuture.whenComplete((r, t) -> {
+ if (t != null) {
+ s.cancel();
+ multipartUploadHelper.cancelingOtherOngoingRequests(futures, t);
+ }
+ });
+ }
+
+ @Override
+ public void onNext(AsyncRequestBody asyncRequestBody) {
+ log.trace(() -> "Received asyncRequestBody " + asyncRequestBody.contentLength());
+ asyncRequestBodyInFlight.incrementAndGet();
+ UploadPartRequest uploadRequest =
+ SdkPojoConversionUtils.toUploadPartRequest(putObjectRequest,
+ partNumber.getAndIncrement(),
+ uploadId);
+
+ Consumer completedPartConsumer = completedPart -> completedParts.set(completedPart.partNumber() - 1,
+ completedPart);
+ multipartUploadHelper.sendIndividualUploadPartRequest(uploadId, completedPartConsumer, futures,
+ Pair.of(uploadRequest, asyncRequestBody))
+ .whenComplete((r, t) -> {
+ if (t != null) {
+ if (failureActionInitiated.compareAndSet(false, true)) {
+ multipartUploadHelper.failRequestsElegantly(futures, t, uploadId, returnFuture,
+ putObjectRequest);
+ }
+ } else {
+ completeMultipartUploadIfFinish(asyncRequestBodyInFlight.decrementAndGet());
+ }
+ });
+ subscription.request(1);
+ }
+
+ @Override
+ public void onError(Throwable t) {
+ log.debug(() -> "Received onError ", t);
+ if (failureActionInitiated.compareAndSet(false, true)) {
+ multipartUploadHelper.failRequestsElegantly(futures, t, uploadId, returnFuture, putObjectRequest);
+ }
+ }
+
+ @Override
+ public void onComplete() {
+ log.debug(() -> "Received onComplete()");
+ isDone = true;
+ completeMultipartUploadIfFinish(asyncRequestBodyInFlight.get());
+ }
+
+ private void completeMultipartUploadIfFinish(int requestsInFlight) {
+ if (isDone && requestsInFlight == 0 && completedMultipartInitiated.compareAndSet(false, true)) {
+ CompletedPart[] parts =
+ IntStream.range(0, completedParts.length())
+ .mapToObj(completedParts::get)
+ .toArray(CompletedPart[]::new);
+ multipartUploadHelper.completeMultipartUpload(returnFuture, uploadId, parts, putObjectRequest);
+ }
+ }
+
+ }
+}
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/UploadWithUnknownContentLengthHelper.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/UploadWithUnknownContentLengthHelper.java
new file mode 100644
index 00000000000..0c8c3c70b51
--- /dev/null
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/internal/multipart/UploadWithUnknownContentLengthHelper.java
@@ -0,0 +1,249 @@
+/*
+ * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License").
+ * You may not use this file except in compliance with the License.
+ * A copy of the License is located at
+ *
+ * http://aws.amazon.com/apache2.0
+ *
+ * or in the "license" file accompanying this file. This file is distributed
+ * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
+ * express or implied. See the License for the specific language governing
+ * permissions and limitations under the License.
+ */
+
+package software.amazon.awssdk.services.s3.internal.multipart;
+
+
+import java.util.Collection;
+import java.util.Comparator;
+import java.util.Queue;
+import java.util.concurrent.CompletableFuture;
+import java.util.concurrent.ConcurrentLinkedQueue;
+import java.util.concurrent.atomic.AtomicBoolean;
+import java.util.concurrent.atomic.AtomicInteger;
+import org.reactivestreams.Subscriber;
+import org.reactivestreams.Subscription;
+import software.amazon.awssdk.annotations.SdkInternalApi;
+import software.amazon.awssdk.core.async.AsyncRequestBody;
+import software.amazon.awssdk.core.async.SdkPublisher;
+import software.amazon.awssdk.services.s3.S3AsyncClient;
+import software.amazon.awssdk.services.s3.model.CompletedPart;
+import software.amazon.awssdk.services.s3.model.CreateMultipartUploadResponse;
+import software.amazon.awssdk.services.s3.model.PutObjectRequest;
+import software.amazon.awssdk.services.s3.model.PutObjectResponse;
+import software.amazon.awssdk.services.s3.model.UploadPartRequest;
+import software.amazon.awssdk.utils.CompletableFutureUtils;
+import software.amazon.awssdk.utils.Logger;
+import software.amazon.awssdk.utils.Pair;
+
+/**
+ * An internal helper class that uploads streams with unknown content length.
+ */
+@SdkInternalApi
+public final class UploadWithUnknownContentLengthHelper {
+ private static final Logger log = Logger.loggerFor(UploadWithUnknownContentLengthHelper.class);
+
+ private final S3AsyncClient s3AsyncClient;
+ private final long partSizeInBytes;
+ private final GenericMultipartHelper genericMultipartHelper;
+
+ private final long maxMemoryUsageInBytes;
+ private final long multipartUploadThresholdInBytes;
+
+ private final MultipartUploadHelper multipartUploadHelper;
+
+ public UploadWithUnknownContentLengthHelper(S3AsyncClient s3AsyncClient,
+ long partSizeInBytes,
+ long multipartUploadThresholdInBytes,
+ long maxMemoryUsageInBytes) {
+ this.s3AsyncClient = s3AsyncClient;
+ this.partSizeInBytes = partSizeInBytes;
+ this.genericMultipartHelper = new GenericMultipartHelper<>(s3AsyncClient,
+ SdkPojoConversionUtils::toAbortMultipartUploadRequest,
+ SdkPojoConversionUtils::toPutObjectResponse);
+ this.maxMemoryUsageInBytes = maxMemoryUsageInBytes;
+ this.multipartUploadThresholdInBytes = multipartUploadThresholdInBytes;
+ this.multipartUploadHelper = new MultipartUploadHelper(s3AsyncClient, partSizeInBytes, multipartUploadThresholdInBytes,
+ maxMemoryUsageInBytes);
+ }
+
+ public CompletableFuture uploadObject(PutObjectRequest putObjectRequest,
+ AsyncRequestBody asyncRequestBody) {
+ CompletableFuture returnFuture = new CompletableFuture<>();
+
+ SdkPublisher splitAsyncRequestBodyResponse =
+ asyncRequestBody.split(b -> b.chunkSizeInBytes(partSizeInBytes)
+ .bufferSizeInBytes(maxMemoryUsageInBytes));
+
+ splitAsyncRequestBodyResponse.subscribe(new UnknownContentLengthAsyncRequestBodySubscriber(partSizeInBytes,
+ putObjectRequest,
+ returnFuture));
+ return returnFuture;
+ }
+
+ private class UnknownContentLengthAsyncRequestBodySubscriber implements Subscriber {
+ /**
+ * Indicates whether this is the first async request body or not.
+ */
+ private final AtomicBoolean isFirstAsyncRequestBody = new AtomicBoolean(true);
+
+ /**
+ * Indicates whether CreateMultipartUpload has been initiated or not
+ */
+ private final AtomicBoolean createMultipartUploadInitiated = new AtomicBoolean(false);
+
+ /**
+ * Indicates whether CompleteMultipart has been initiated or not.
+ */
+ private final AtomicBoolean completedMultipartInitiated = new AtomicBoolean(false);
+
+ /**
+ * The number of AsyncRequestBody has been received but yet to be processed
+ */
+ private final AtomicInteger asyncRequestBodyInFlight = new AtomicInteger(0);
+
+ private final AtomicBoolean failureActionInitiated = new AtomicBoolean(false);
+
+ private AtomicInteger partNumber = new AtomicInteger(1);
+
+ private final Queue completedParts = new ConcurrentLinkedQueue<>();
+ private final Collection> futures = new ConcurrentLinkedQueue<>();
+
+ private final CompletableFuture uploadIdFuture = new CompletableFuture<>();
+
+ private final long maximumChunkSizeInByte;
+ private final PutObjectRequest putObjectRequest;
+ private final CompletableFuture returnFuture;
+ private Subscription subscription;
+ private AsyncRequestBody firstRequestBody;
+
+ private String uploadId;
+ private volatile boolean isDone;
+
+ UnknownContentLengthAsyncRequestBodySubscriber(long maximumChunkSizeInByte,
+ PutObjectRequest putObjectRequest,
+ CompletableFuture returnFuture) {
+ this.maximumChunkSizeInByte = maximumChunkSizeInByte;
+ this.putObjectRequest = putObjectRequest;
+ this.returnFuture = returnFuture;
+ }
+
+ @Override
+ public void onSubscribe(Subscription s) {
+ if (this.subscription != null) {
+ log.warn(() -> "The subscriber has already been subscribed. Cancelling the incoming subscription");
+ subscription.cancel();
+ return;
+ }
+ this.subscription = s;
+ s.request(1);
+ returnFuture.whenComplete((r, t) -> {
+ if (t != null) {
+ s.cancel();
+ multipartUploadHelper.cancelingOtherOngoingRequests(futures, t);
+ }
+ });
+ }
+
+ @Override
+ public void onNext(AsyncRequestBody asyncRequestBody) {
+ log.trace(() -> "Received asyncRequestBody " + asyncRequestBody.contentLength());
+ asyncRequestBodyInFlight.incrementAndGet();
+
+ if (isFirstAsyncRequestBody.compareAndSet(true, false)) {
+ log.trace(() -> "Received first async request body");
+ // If this is the first AsyncRequestBody received, request another one because we don't know if there is more
+ firstRequestBody = asyncRequestBody;
+ subscription.request(1);
+ return;
+ }
+
+ // If there are more than 1 AsyncRequestBodies, then we know we need to upload this
+ // object using MPU
+ if (createMultipartUploadInitiated.compareAndSet(false, true)) {
+ log.debug(() -> "Starting the upload as multipart upload request");
+ CompletableFuture createMultipartUploadFuture =
+ multipartUploadHelper.createMultipartUpload(putObjectRequest, returnFuture);
+
+ createMultipartUploadFuture.whenComplete((createMultipartUploadResponse, throwable) -> {
+ if (throwable != null) {
+ genericMultipartHelper.handleException(returnFuture, () -> "Failed to initiate multipart upload",
+ throwable);
+ subscription.cancel();
+ } else {
+ uploadId = createMultipartUploadResponse.uploadId();
+ log.debug(() -> "Initiated a new multipart upload, uploadId: " + uploadId);
+
+ sendUploadPartRequest(uploadId, firstRequestBody);
+ sendUploadPartRequest(uploadId, asyncRequestBody);
+
+ // We need to complete the uploadIdFuture *after* the first two requests have been sent
+ uploadIdFuture.complete(uploadId);
+ }
+ });
+ CompletableFutureUtils.forwardExceptionTo(returnFuture, createMultipartUploadFuture);
+ } else {
+ uploadIdFuture.whenComplete((r, t) -> {
+ sendUploadPartRequest(uploadId, asyncRequestBody);
+ });
+ }
+ }
+
+ private void sendUploadPartRequest(String uploadId, AsyncRequestBody asyncRequestBody) {
+ multipartUploadHelper.sendIndividualUploadPartRequest(uploadId, completedParts::add, futures,
+ uploadPart(asyncRequestBody))
+ .whenComplete((r, t) -> {
+ if (t != null) {
+ if (failureActionInitiated.compareAndSet(false, true)) {
+ multipartUploadHelper.failRequestsElegantly(futures, t, uploadId, returnFuture, putObjectRequest);
+ }
+ } else {
+ completeMultipartUploadIfFinish(asyncRequestBodyInFlight.decrementAndGet());
+ }
+ });
+ synchronized (this) {
+ subscription.request(1);
+ };
+ }
+
+ private Pair uploadPart(AsyncRequestBody asyncRequestBody) {
+ UploadPartRequest uploadRequest =
+ SdkPojoConversionUtils.toUploadPartRequest(putObjectRequest,
+ partNumber.getAndIncrement(),
+ uploadId);
+ return Pair.of(uploadRequest, asyncRequestBody);
+ }
+
+ @Override
+ public void onError(Throwable t) {
+ log.debug(() -> "Received onError() ", t);
+ if (failureActionInitiated.compareAndSet(false, true)) {
+ multipartUploadHelper.failRequestsElegantly(futures, t, uploadId, returnFuture, putObjectRequest);
+ }
+ }
+
+ @Override
+ public void onComplete() {
+ log.debug(() -> "Received onComplete()");
+ // If CreateMultipartUpload has not been initiated at this point, we know this is a single object upload
+ if (createMultipartUploadInitiated.get() == false) {
+ log.debug(() -> "Starting the upload as a single object upload request");
+ multipartUploadHelper.uploadInOneChunk(putObjectRequest, firstRequestBody, returnFuture);
+ } else {
+ isDone = true;
+ completeMultipartUploadIfFinish(asyncRequestBodyInFlight.get());
+ }
+ }
+
+ private void completeMultipartUploadIfFinish(int requestsInFlight) {
+ if (isDone && requestsInFlight == 0 && completedMultipartInitiated.compareAndSet(false, true)) {
+ CompletedPart[] parts = completedParts.stream()
+ .sorted(Comparator.comparingInt(CompletedPart::partNumber))
+ .toArray(CompletedPart[]::new);
+ multipartUploadHelper.completeMultipartUpload(returnFuture, uploadId, parts, putObjectRequest);
+ }
+ }
+ }
+}
diff --git a/services/s3/src/main/java/software/amazon/awssdk/services/s3/multipart/MultipartConfiguration.java b/services/s3/src/main/java/software/amazon/awssdk/services/s3/multipart/MultipartConfiguration.java
new file mode 100644
index 00000000000..28e418974db
--- /dev/null
+++ b/services/s3/src/main/java/software/amazon/awssdk/services/s3/multipart/MultipartConfiguration.java
@@ -0,0 +1,199 @@
+/*
+ * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License").
+ * You may not use this file except in compliance with the License.
+ * A copy of the License is located at
+ *
+ * http://aws.amazon.com/apache2.0
+ *
+ * or in the "license" file accompanying this file. This file is distributed
+ * on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
+ * express or implied. See the License for the specific language governing
+ * permissions and limitations under the License.
+ */
+
+package software.amazon.awssdk.services.s3.multipart;
+
+import java.util.function.Consumer;
+import software.amazon.awssdk.annotations.SdkPublicApi;
+import software.amazon.awssdk.core.async.AsyncRequestBody;
+import software.amazon.awssdk.core.async.AsyncResponseTransformer;
+import software.amazon.awssdk.services.s3.S3AsyncClient;
+import software.amazon.awssdk.services.s3.S3AsyncClientBuilder;
+import software.amazon.awssdk.services.s3.model.CopyObjectRequest;
+import software.amazon.awssdk.services.s3.model.GetObjectRequest;
+import software.amazon.awssdk.utils.builder.CopyableBuilder;
+import software.amazon.awssdk.utils.builder.ToCopyableBuilder;
+
+/**
+ * Class that hold configuration properties related to multipart operation for a {@link S3AsyncClient}. Passing this class to the
+ * {@link S3AsyncClientBuilder#multipartConfiguration(MultipartConfiguration)} will enable automatic conversion of
+ * {@link S3AsyncClient#putObject(Consumer, AsyncRequestBody)}, {@link S3AsyncClient#copyObject(CopyObjectRequest)} to their
+ * respective multipart operation.
+ *
+ * Note: The multipart operation for {@link S3AsyncClient#getObject(GetObjectRequest, AsyncResponseTransformer)} is
+ * temporarily disabled and will result in throwing a {@link UnsupportedOperationException} if called when configured for
+ * multipart operation.
+ */
+@SdkPublicApi
+public final class MultipartConfiguration implements ToCopyableBuilder {
+
+ private final Long thresholdInBytes;
+ private final Long minimumPartSizeInBytes;
+ private final Long apiCallBufferSizeInBytes;
+
+ private MultipartConfiguration(DefaultMultipartConfigBuilder builder) {
+ this.thresholdInBytes = builder.thresholdInBytes;
+ this.minimumPartSizeInBytes = builder.minimumPartSizeInBytes;
+ this.apiCallBufferSizeInBytes = builder.apiCallBufferSizeInBytes;
+ }
+
+ public static Builder builder() {
+ return new DefaultMultipartConfigBuilder();
+ }
+
+ @Override
+ public Builder toBuilder() {
+ return builder()
+ .apiCallBufferSizeInBytes(apiCallBufferSizeInBytes)
+ .minimumPartSizeInBytes(minimumPartSizeInBytes)
+ .thresholdInBytes(thresholdInBytes);
+ }
+
+ /**
+ * Indicates the value of the configured threshold, in bytes. Any request whose size is less than the configured value will
+ * not use multipart operation
+ * @return the value of the configured threshold.
+ */
+ public Long thresholdInBytes() {
+ return this.thresholdInBytes;
+ }
+
+ /**
+ * Indicated the size, in bytes, of each individual part of the part requests. The actual part size used might be bigger to
+ * conforms to the maximum number of parts allowed per multipart requests.
+ * @return the value of the configured part size.
+ */
+ public Long minimumPartSizeInBytes() {
+ return this.minimumPartSizeInBytes;
+ }
+
+ /**
+ * The maximum memory, in bytes, that the SDK will use to buffer requests content into memory.
+ * @return the value of the configured maximum memory usage.
+ */
+ public Long apiCallBufferSizeInBytes() {
+ return this.apiCallBufferSizeInBytes;
+ }
+
+ /**
+ * Builder for a {@link MultipartConfiguration}.
+ */
+ public interface Builder extends CopyableBuilder {
+
+ /**
+ * Configures the minimum number of bytes of the body of the request required for requests to be converted to their
+ * multipart equivalent. Only taken into account when converting {@code putObject} and {@code copyObject} requests.
+ * Any request whose size is less than the configured value will not use multipart operation,
+ * even if multipart is enabled via {@link S3AsyncClientBuilder#multipartEnabled(Boolean)}.
+ *
+ *
+ * Default value: 8 Mib
+ *
+ * @param thresholdInBytes the value of the threshold to set.
+ * @return an instance of this builder.
+ */
+ Builder thresholdInBytes(Long thresholdInBytes);
+
+ /**
+ * Indicates the value of the configured threshold.
+ * @return the value of the threshold.
+ */
+ Long thresholdInBytes();
+
+ /**
+ * Configures the part size, in bytes, to be used in each individual part requests.
+ * Only used for putObject and copyObject operations.
+ *
+ * When uploading large payload, the size of the payload of each individual part requests might actually be
+ * bigger than
+ * the configured value since there is a limit to the maximum number of parts possible per multipart request. If the
+ * configured part size would lead to a number of parts higher than the maximum allowed, a larger part size will be
+ * calculated instead to allow fewer part to be uploaded, to avoid the limit imposed on the maximum number of parts.
+ *
+ * In the case where the {@code minimumPartSizeInBytes} is set to a value higher than the {@code thresholdInBytes}, when
+ * the client receive a request with a size smaller than a single part multipart operation will NOT be performed
+ * even if the size of the request is larger than the threshold.
+ *
+ * Default value: 8 Mib
+ *
+ * @param minimumPartSizeInBytes the value of the part size to set
+ * @return an instance of this builder.
+ */
+ Builder minimumPartSizeInBytes(Long minimumPartSizeInBytes);
+
+ /**
+ * Indicated the value of the part configured size.
+ * @return the value of the part size
+ */
+ Long minimumPartSizeInBytes();
+
+ /**
+ * Configures the maximum amount of memory, in bytes, the SDK will use to buffer content of requests in memory.
+ * Increasing this value may lead to better performance at the cost of using more memory.
+ *
Starts an Amazon SageMaker Inference Recommender autoscaling recommendation job. Returns recommendations for autoscaling policies that you can apply to your SageMaker endpoint.
If specified, monitoring jobs substract this time from the end time. For information about using offsets for scheduling monitoring jobs, see Schedule Model Quality Monitoring Jobs.
"
+ "documentation":"
If specified, monitoring jobs subtract this time from the end time. For information about using offsets for scheduling monitoring jobs, see Schedule Model Quality Monitoring Jobs.
SystemUpdating: Endpoint is undergoing maintenance and cannot be updated or deleted or re-scaled until it has completed. This maintenance operation does not change any customer-specified values such as VPC config, KMS encryption, model, instance type, or instance count.
RollingBack: Endpoint fails to scale up or down or change its variant weight and is in the process of rolling back to its previous configuration. Once the rollback completes, endpoint returns to an InService status. This transitional status only applies to an endpoint that has autoscaling enabled and is undergoing variant weight or capacity changes as part of an UpdateEndpointWeightsAndCapacities call or when the UpdateEndpointWeightsAndCapacities operation is called explicitly.
InService: Endpoint is available to process incoming requests.
Failed: Endpoint could not be created, updated, or re-scaled. Use the FailureReason value returned by DescribeEndpoint for information about the failure. DeleteEndpoint is the only operation that can be performed on a failed endpoint.
"
+ "documentation":"
The status of the endpoint.
OutOfService: Endpoint is not available to take incoming requests.
SystemUpdating: Endpoint is undergoing maintenance and cannot be updated or deleted or re-scaled until it has completed. This maintenance operation does not change any customer-specified values such as VPC config, KMS encryption, model, instance type, or instance count.
RollingBack: Endpoint fails to scale up or down or change its variant weight and is in the process of rolling back to its previous configuration. Once the rollback completes, endpoint returns to an InService status. This transitional status only applies to an endpoint that has autoscaling enabled and is undergoing variant weight or capacity changes as part of an UpdateEndpointWeightsAndCapacities call or when the UpdateEndpointWeightsAndCapacities operation is called explicitly.
InService: Endpoint is available to process incoming requests.
Failed: Endpoint could not be created, updated, or re-scaled. Use the FailureReason value returned by DescribeEndpoint for information about the failure. DeleteEndpoint is the only operation that can be performed on a failed endpoint.
UpdateRollbackFailed: Both the rolling deployment and auto-rollback failed. Your endpoint is in service with a mix of the old and new endpoint configurations. For information about how to remedy this issue and restore the endpoint's status to InService, see Rolling Deployments.
The recommendation ID of a previously completed inference recommendation. This ID should come from one of the recommendations returned by the job specified in the InferenceRecommendationsJobName field.
Specify either this field or the EndpointName field.
The name of an endpoint benchmarked during a previously completed inference recommendation job. This name should come from one of the recommendations returned by the job specified in the InferenceRecommendationsJobName field.
Specify either this field or the RecommendationId field.
Specifies additional parameters for compiler options in JSON format. The compiler options are TargetPlatform specific. It is required for NVIDIA accelerators and highly recommended for CPU compilations. For any other cases, it is optional to specify CompilerOptions.
DTYPE: Specifies the data type for the input. When compiling for ml_* (except for ml_inf) instances using PyTorch framework, provide the data type (dtype) of the model's input. \"float32\" is used if \"DTYPE\" is not specified. Options for data type are:
float32: Use either \"float\" or \"float32\".
int64: Use either \"int64\" or \"long\".
For example, {\"dtype\" : \"float32\"}.
CPU: Compilation for CPU supports the following compiler options.
mcpu: CPU micro-architecture. For example, {'mcpu': 'skylake-avx512'}
mattr: CPU flags. For example, {'mattr': ['+neon', '+vfpv4']}
ARM: Details of ARM CPU compilations.
NEON: NEON is an implementation of the Advanced SIMD extension used in ARMv7 processors.
For example, add {'mattr': ['+neon']} to the compiler options if compiling for ARM 32-bit platform with the NEON support.
NVIDIA: Compilation for NVIDIA GPU supports the following compiler options.
gpu_code: Specifies the targeted architecture.
trt-ver: Specifies the TensorRT versions in x.y.z. format.
cuda-ver: Specifies the CUDA version in x.y format.
For example, {'gpu-code': 'sm_72', 'trt-ver': '6.0.1', 'cuda-ver': '10.1'}
ANDROID: Compilation for the Android OS supports the following compiler options:
ANDROID_PLATFORM: Specifies the Android API levels. Available levels range from 21 to 29. For example, {'ANDROID_PLATFORM': 28}.
mattr: Add {'mattr': ['+neon']} to compiler options if compiling for ARM 32-bit platform with NEON support.
INFERENTIA: Compilation for target ml_inf1 uses compiler options passed in as a JSON string. For example, \"CompilerOptions\": \"\\\"--verbose 1 --num-neuroncores 2 -O2\\\"\".
CoreML: Compilation for the CoreML OutputConfigTargetDevice supports the following compiler options:
class_labels: Specifies the classification labels file name inside input tar.gz file. For example, {\"class_labels\": \"imagenet_labels_1000.txt\"}. Labels inside the txt file should be separated by newlines.
EIA: Compilation for the Elastic Inference Accelerator supports the following compiler options:
precision_mode: Specifies the precision of compiled artifacts. Supported values are \"FP16\" and \"FP32\". Default is \"FP32\".
signature_def_key: Specifies the signature to use for models in SavedModel format. Defaults is TensorFlow's default signature def key.
output_names: Specifies a list of output tensor names for models in FrozenGraph format. Set at most one API field, either: signature_def_key or output_names.
For example: {\"precision_mode\": \"FP32\", \"output_names\": [\"output:0\"]}
"
+ "documentation":"
Specifies additional parameters for compiler options in JSON format. The compiler options are TargetPlatform specific. It is required for NVIDIA accelerators and highly recommended for CPU compilations. For any other cases, it is optional to specify CompilerOptions.
DTYPE: Specifies the data type for the input. When compiling for ml_* (except for ml_inf) instances using PyTorch framework, provide the data type (dtype) of the model's input. \"float32\" is used if \"DTYPE\" is not specified. Options for data type are:
float32: Use either \"float\" or \"float32\".
int64: Use either \"int64\" or \"long\".
For example, {\"dtype\" : \"float32\"}.
CPU: Compilation for CPU supports the following compiler options.
mcpu: CPU micro-architecture. For example, {'mcpu': 'skylake-avx512'}
mattr: CPU flags. For example, {'mattr': ['+neon', '+vfpv4']}
ARM: Details of ARM CPU compilations.
NEON: NEON is an implementation of the Advanced SIMD extension used in ARMv7 processors.
For example, add {'mattr': ['+neon']} to the compiler options if compiling for ARM 32-bit platform with the NEON support.
NVIDIA: Compilation for NVIDIA GPU supports the following compiler options.
gpu_code: Specifies the targeted architecture.
trt-ver: Specifies the TensorRT versions in x.y.z. format.
cuda-ver: Specifies the CUDA version in x.y format.
For example, {'gpu-code': 'sm_72', 'trt-ver': '6.0.1', 'cuda-ver': '10.1'}
ANDROID: Compilation for the Android OS supports the following compiler options:
ANDROID_PLATFORM: Specifies the Android API levels. Available levels range from 21 to 29. For example, {'ANDROID_PLATFORM': 28}.
mattr: Add {'mattr': ['+neon']} to compiler options if compiling for ARM 32-bit platform with NEON support.
INFERENTIA: Compilation for target ml_inf1 uses compiler options passed in as a JSON string. For example, \"CompilerOptions\": \"\\\"--verbose 1 --num-neuroncores 2 -O2\\\"\".
CoreML: Compilation for the CoreML OutputConfigTargetDevice supports the following compiler options:
class_labels: Specifies the classification labels file name inside input tar.gz file. For example, {\"class_labels\": \"imagenet_labels_1000.txt\"}. Labels inside the txt file should be separated by newlines.
EIA: Compilation for the Elastic Inference Accelerator supports the following compiler options:
precision_mode: Specifies the precision of compiled artifacts. Supported values are \"FP16\" and \"FP32\". Default is \"FP32\".
signature_def_key: Specifies the signature to use for models in SavedModel format. Defaults is TensorFlow's default signature def key.
output_names: Specifies a list of output tensor names for models in FrozenGraph format. Set at most one API field, either: signature_def_key or output_names.
For example: {\"precision_mode\": \"FP32\", \"output_names\": [\"output:0\"]}
Specifies how long a traffic phase should be. For custom load tests, the value should be between 120 and 3600. This value should not exceed JobDurationInSeconds.
The interval of time taken by a model to respond as viewed from SageMaker. The interval includes the local communication time taken to send the request and to fetch the response from the container of a model and the time taken to complete the inference in the container.
Stops a load test when the number of invocations (TPS) peaks and flattens, which means that the instance has reached capacity. The default value is Stop. If you want the load test to continue after invocations have flattened, set the value to Continue.
"
}
},
"documentation":"
Specifies conditions for stopping a job. When a job reaches a stopping condition limit, SageMaker ends the job.
The number of invocations sent to a model, normalized by InstanceCount in each ProductionVariant. 1/numberOfInstances is sent as the value on each request, where numberOfInstances is the number of active instances for the ProductionVariant behind the endpoint at the time of the request.
The interval of time taken by a model to respond as viewed from SageMaker. This interval includes the local communication times taken to send the request and to fetch the response from the container of a model and the time taken to complete the inference in the container.
Specifies how many new users to spawn in each step.
"
+ }
+ },
+ "documentation":"
Defines the stairs traffic pattern for an Inference Recommender load test. This pattern type consists of multiple steps where the number of users increases at each step.
Specify either the stairs or phases traffic pattern.
A collection of settings that apply to users of Amazon SageMaker Studio. These settings are specified when the CreateUserProfile API is called, and as DefaultUserSettings when the CreateDomain API is called.
SecurityGroups is aggregated when specified in both calls. For all other settings in UserSettings, the values specified in CreateUserProfile take precedence over those specified in CreateDomain.
Unique, case-sensitive identifier you provide to ensure the idempotency of the request. If you do not specify a client token, EventBridge Scheduler uses a randomly generated token for the request to ensure idempotency.
The expression that defines when the schedule runs. The following formats are supported.
at expression - at(yyyy-mm-ddThh:mm:ss)
rate expression - rate(unit value)
cron expression - cron(fields)
You can use at expressions to create one-time schedules that invoke a target once, at the time and in the time zone, that you specify. You can use rate and cron expressions to create recurring schedules. Rate-based schedules are useful when you want to invoke a target at regular intervals, such as every 15 minutes or every five days. Cron-based schedules are useful when you want to invoke a target periodically at a specific time, such as at 8:00 am (UTC+0) every 1st day of the month.
A cron expression consists of six fields separated by white spaces: (minutes hours day_of_month month day_of_week year).
A rate expression consists of a value as a positive integer, and a unit with the following options: minute | minutes | hour | hours | day | days
The expression that defines when the schedule runs. The following formats are supported.
at expression - at(yyyy-mm-ddThh:mm:ss)
rate expression - rate(value unit)
cron expression - cron(fields)
You can use at expressions to create one-time schedules that invoke a target once, at the time and in the time zone, that you specify. You can use rate and cron expressions to create recurring schedules. Rate-based schedules are useful when you want to invoke a target at regular intervals, such as every 15 minutes or every five days. Cron-based schedules are useful when you want to invoke a target periodically at a specific time, such as at 8:00 am (UTC+0) every 1st day of the month.
A cron expression consists of six fields separated by white spaces: (minutes hours day_of_month month day_of_week year).
A rate expression consists of a value as a positive integer, and a unit with the following options: minute | minutes | hour | hours | day | days
The expression that defines when the schedule runs. The following formats are supported.
at expression - at(yyyy-mm-ddThh:mm:ss)
rate expression - rate(unit value)
cron expression - cron(fields)
You can use at expressions to create one-time schedules that invoke a target once, at the time and in the time zone, that you specify. You can use rate and cron expressions to create recurring schedules. Rate-based schedules are useful when you want to invoke a target at regular intervals, such as every 15 minutes or every five days. Cron-based schedules are useful when you want to invoke a target periodically at a specific time, such as at 8:00 am (UTC+0) every 1st day of the month.
A cron expression consists of six fields separated by white spaces: (minutes hours day_of_month month day_of_week year).
A rate expression consists of a value as a positive integer, and a unit with the following options: minute | minutes | hour | hours | day | days
The expression that defines when the schedule runs. The following formats are supported.
at expression - at(yyyy-mm-ddThh:mm:ss)
rate expression - rate(value unit)
cron expression - cron(fields)
You can use at expressions to create one-time schedules that invoke a target once, at the time and in the time zone, that you specify. You can use rate and cron expressions to create recurring schedules. Rate-based schedules are useful when you want to invoke a target at regular intervals, such as every 15 minutes or every five days. Cron-based schedules are useful when you want to invoke a target periodically at a specific time, such as at 8:00 am (UTC+0) every 1st day of the month.
A cron expression consists of six fields separated by white spaces: (minutes hours day_of_month month day_of_week year).
A rate expression consists of a value as a positive integer, and a unit with the following options: minute | minutes | hour | hours | day | days
Unique, case-sensitive identifier you provide to ensure the idempotency of the request. If you do not specify a client token, EventBridge Scheduler uses a randomly generated token for the request to ensure idempotency.
The expression that defines when the schedule runs. The following formats are supported.
at expression - at(yyyy-mm-ddThh:mm:ss)
rate expression - rate(unit value)
cron expression - cron(fields)
You can use at expressions to create one-time schedules that invoke a target once, at the time and in the time zone, that you specify. You can use rate and cron expressions to create recurring schedules. Rate-based schedules are useful when you want to invoke a target at regular intervals, such as every 15 minutes or every five days. Cron-based schedules are useful when you want to invoke a target periodically at a specific time, such as at 8:00 am (UTC+0) every 1st day of the month.
A cron expression consists of six fields separated by white spaces: (minutes hours day_of_month month day_of_week year).
A rate expression consists of a value as a positive integer, and a unit with the following options: minute | minutes | hour | hours | day | days
The expression that defines when the schedule runs. The following formats are supported.
at expression - at(yyyy-mm-ddThh:mm:ss)
rate expression - rate(value unit)
cron expression - cron(fields)
You can use at expressions to create one-time schedules that invoke a target once, at the time and in the time zone, that you specify. You can use rate and cron expressions to create recurring schedules. Rate-based schedules are useful when you want to invoke a target at regular intervals, such as every 15 minutes or every five days. Cron-based schedules are useful when you want to invoke a target periodically at a specific time, such as at 8:00 am (UTC+0) every 1st day of the month.
A cron expression consists of six fields separated by white spaces: (minutes hours day_of_month month day_of_week year).
A rate expression consists of a value as a positive integer, and a unit with the following options: minute | minutes | hour | hours | day | days
Specifies whether a rule is the last to be applied with respect to a finding that matches the rule criteria. This is useful when a finding matches the criteria for multiple rules, and each rule has different actions. If the value of this field is set to true for a rule, Security Hub applies the rule action to a finding that matches the rule criteria and doesn't evaluate other rules for the finding. The default value of this field is false.
"
+ "documentation":"
Specifies whether a rule is the last to be applied with respect to a finding that matches the rule criteria. This is useful when a finding matches the criteria for multiple rules, and each rule has different actions. If a rule is terminal, Security Hub applies the rule action to a finding that matches the rule criteria and doesn't evaluate other rules for the finding. By default, a rule isn't terminal.
Specifies whether a rule is the last to be applied with respect to a finding that matches the rule criteria. This is useful when a finding matches the criteria for multiple rules, and each rule has different actions. If the value of this field is set to true for a rule, Security Hub applies the rule action to a finding that matches the rule criteria and doesn't evaluate other rules for the finding. The default value of this field is false.
"
+ "documentation":"
Specifies whether a rule is the last to be applied with respect to a finding that matches the rule criteria. This is useful when a finding matches the criteria for multiple rules, and each rule has different actions. If a rule is terminal, Security Hub applies the rule action to a finding that matches the rule criteria and doesn't evaluate other rules for the finding. By default, a rule isn't terminal.
The location in Amazon S3 where query and calculation results are stored and the encryption option, if any, used for query and calculation results. These are known as client-side settings. If workgroup settings override client-side settings, then the query uses the workgroup settings.
"
+ }
+ },
+ "documentation":"
The configuration of the workgroup, which includes the location in Amazon Simple Storage Service (Amazon S3) where query results are stored, the encryption option, if any, used for query results, whether Amazon CloudWatch metrics are enabled for the workgroup, and the limit for the amount of bytes scanned (cutoff) per query, if it is specified.
Specifies the method used to encrypt the user’s data stores in the Athena workgroup.
"
+ }
+ },
+ "documentation":"
The location in Amazon Simple Storage Service (Amazon S3) where query and calculation results are stored and the encryption option, if any, used for query and calculation results. These are known as client-side settings. If workgroup settings override client-side settings, then the query uses the workgroup settings.
The configuration of the workgroup, which includes the location in Amazon Simple Storage Service (Amazon S3) where query results are stored, the encryption option, if any, used for query results, whether Amazon CloudWatch metrics are enabled for the workgroup, and the limit for the amount of bytes scanned (cutoff) per query, if it is specified.
"
+ }
+ },
+ "documentation":"
Provides information about an Amazon Athena workgroup.
The name of the manual DB cluster snapshot attribute. The attribute named restore refers to the list of Amazon Web Services accounts that have permission to copy or restore the manual DB cluster snapshot.
The value(s) for the manual DB cluster snapshot attribute. If the AttributeName field is set to restore, then this element returns a list of IDs of the Amazon Web Services accounts that are authorized to copy or restore the manual DB cluster snapshot. If a value of all is in the list, then the manual DB cluster snapshot is public and available for any Amazon Web Services account to copy or restore.
"
+ }
+ },
+ "documentation":"
Contains the name and values of a manual Amazon Relational Database Service (RDS) DB cluster snapshot attribute.
Specifies whether a rule is the last to be applied with respect to a finding that matches the rule criteria. This is useful when a finding matches the criteria for multiple rules, and each rule has different actions. If the value of this field is set to true for a rule, Security Hub applies the rule action to a finding that matches the rule criteria and doesn't evaluate other rules for the finding. The default value of this field is false.
"
+ "documentation":"
Specifies whether a rule is the last to be applied with respect to a finding that matches the rule criteria. This is useful when a finding matches the criteria for multiple rules, and each rule has different actions. If a rule is terminal, Security Hub applies the rule action to a finding that matches the rule criteria and doesn't evaluate other rules for the finding. By default, a rule isn't terminal.
The value for the key in the map filter. Filter values are case sensitive. For example, one of the values for a tag called Department might be Security. If you provide security as the filter value, then there is no match.
"
+ "documentation":"
The value for the key in the map filter. Filter values are case sensitive. For example, one of the values for a tag called Department might be Security. If you provide security as the filter value, then there's no match.
The condition to apply to the key value when querying for findings with a map filter.
To search for values that exactly match the filter value, use EQUALS. For example, for the ResourceTags field, the filter Department EQUALS Security matches findings that have the value Security for the tag Department.
To search for values other than the filter value, use NOT_EQUALS. For example, for the ResourceTags field, the filter Department NOT_EQUALS Finance matches findings that do not have the value Finance for the tag Department.
EQUALS filters on the same field are joined by OR. A finding matches if it matches any one of those filters.
NOT_EQUALS filters on the same field are joined by AND. A finding matches only if it matches all of those filters.
You cannot have both an EQUALS filter and a NOT_EQUALS filter on the same field.
"
+ "documentation":"
The condition to apply to the key value when filtering Security Hub findings with a map filter.
To search for values that have the filter value, use one of the following comparison operators:
To search for values that include the filter value, use CONTAINS. For example, for the ResourceTags field, the filter Department CONTAINS Security matches findings that include the value Security for the Department tag. In the same example, a finding with a value of Security team for the Department tag is a match.
To search for values that exactly match the filter value, use EQUALS. For example, for the ResourceTags field, the filter Department EQUALS Security matches findings that have the value Security for the Department tag.
CONTAINS and EQUALS filters on the same field are joined by OR. A finding matches if it matches any one of those filters. For example, the filters Department CONTAINS Security OR Department CONTAINS Finance match a finding that includes either Security, Finance, or both values.
To search for values that don't have the filter value, use one of the following comparison operators:
To search for values that exclude the filter value, use NOT_CONTAINS. For example, for the ResourceTags field, the filter Department NOT_CONTAINS Finance matches findings that exclude the value Finance for the Department tag.
To search for values other than the filter value, use NOT_EQUALS. For example, for the ResourceTags field, the filter Department NOT_EQUALS Finance matches findings that don’t have the value Finance for the Department tag.
NOT_CONTAINS and NOT_EQUALS filters on the same field are joined by AND. A finding matches only if it matches all of those filters. For example, the filters Department NOT_CONTAINS Security AND Department NOT_CONTAINS Finance match a finding that excludes both the Security and Finance values.
CONTAINS filters can only be used with other CONTAINS filters. NOT_CONTAINS filters can only be used with other NOT_CONTAINS filters.
You can’t have both a CONTAINS filter and a NOT_CONTAINS filter on the same field. Similarly, you can’t have both an EQUALS filter and a NOT_EQUALS filter on the same field. Combining filters in this way returns an error.
CONTAINS and NOT_CONTAINS operators can be used only with automation rules. For more information, see Automation rules in the Security Hub User Guide.
"
}
},
- "documentation":"
A map filter for querying findings. Each map filter provides the field to check, the value to look for, and the comparison operator.
"
+ "documentation":"
A map filter for filtering Security Hub findings. Each map filter provides the field to check for, the value to check for, and the comparison operator.
Provides information about an Amazon Athena workgroup. A workgroup helps you separate users, teams, applications, or workloads. It also helps you set limits on data processing and track costs.
"
}
},
"documentation":"
Additional details about a resource related to a finding.
To provide the details, use the object that corresponds to the resource type. For example, if the resource type is AwsEc2Instance, then you use the AwsEc2Instance object to provide the details.
If the type-specific object does not contain all of the fields you want to populate, then you use the Other object to populate those additional fields.
You also use the Other object to populate the details when the selected type does not have a corresponding object.
The string filter value. Filter values are case sensitive. For example, the product name for control-based findings is Security Hub. If you provide security hub as the filter text, then there is no match.
"
+ "documentation":"
The string filter value. Filter values are case sensitive. For example, the product name for control-based findings is Security Hub. If you provide security hub as the filter value, there's no match.
The condition to apply to a string value when querying for findings. To search for values that contain the filter criteria value, use one of the following comparison operators:
To search for values that exactly match the filter value, use EQUALS.
For example, the filter ResourceType EQUALS AwsEc2SecurityGroup only matches findings that have a resource type of AwsEc2SecurityGroup.
To search for values that start with the filter value, use PREFIX.
For example, the filter ResourceType PREFIX AwsIam matches findings that have a resource type that starts with AwsIam. Findings with a resource type of AwsIamPolicy, AwsIamRole, or AwsIamUser would all match.
EQUALS and PREFIX filters on the same field are joined by OR. A finding matches if it matches any one of those filters.
To search for values that do not contain the filter criteria value, use one of the following comparison operators:
To search for values that do not exactly match the filter value, use NOT_EQUALS.
For example, the filter ResourceType NOT_EQUALS AwsIamPolicy matches findings that have a resource type other than AwsIamPolicy.
To search for values that do not start with the filter value, use PREFIX_NOT_EQUALS.
For example, the filter ResourceType PREFIX_NOT_EQUALS AwsIam matches findings that have a resource type that does not start with AwsIam. Findings with a resource type of AwsIamPolicy, AwsIamRole, or AwsIamUser would all be excluded from the results.
NOT_EQUALS and PREFIX_NOT_EQUALS filters on the same field are joined by AND. A finding matches only if it matches all of those filters.
For filters on the same field, you cannot provide both an EQUALS filter and a NOT_EQUALS or PREFIX_NOT_EQUALS filter. Combining filters in this way always returns an error, even if the provided filter values would return valid results.
You can combine PREFIX filters with NOT_EQUALS or PREFIX_NOT_EQUALS filters for the same field. Security Hub first processes the PREFIX filters, then the NOT_EQUALS or PREFIX_NOT_EQUALS filters.
For example, for the following filter, Security Hub first identifies findings that have resource types that start with either AwsIAM or AwsEc2. It then excludes findings that have a resource type of AwsIamPolicy and findings that have a resource type of AwsEc2NetworkInterface.
ResourceType PREFIX AwsIam
ResourceType PREFIX AwsEc2
ResourceType NOT_EQUALS AwsIamPolicy
ResourceType NOT_EQUALS AwsEc2NetworkInterface
"
+ "documentation":"
The condition to apply to a string value when filtering Security Hub findings.
To search for values that have the filter value, use one of the following comparison operators:
To search for values that include the filter value, use CONTAINS. For example, the filter Title CONTAINS CloudFront matches findings that have a Title that includes the string CloudFront.
To search for values that exactly match the filter value, use EQUALS. For example, the filter AwsAccountId EQUALS 123456789012 only matches findings that have an account ID of 123456789012.
To search for values that start with the filter value, use PREFIX. For example, the filter ResourceRegion PREFIX us matches findings that have a ResourceRegion that starts with us. A ResourceRegion that starts with a different value, such as af, ap, or ca, doesn't match.
CONTAINS, EQUALS, and PREFIX filters on the same field are joined by OR. A finding matches if it matches any one of those filters. For example, the filters Title CONTAINS CloudFront OR Title CONTAINS CloudWatch match a finding that includes either CloudFront, CloudWatch, or both strings in the title.
To search for values that don’t have the filter value, use one of the following comparison operators:
To search for values that exclude the filter value, use NOT_CONTAINS. For example, the filter Title NOT_CONTAINS CloudFront matches findings that have a Title that excludes the string CloudFront.
To search for values other than the filter value, use NOT_EQUALS. For example, the filter AwsAccountId NOT_EQUALS 123456789012 only matches findings that have an account ID other than 123456789012.
To search for values that don't start with the filter value, use PREFIX_NOT_EQUALS. For example, the filter ResourceRegion PREFIX_NOT_EQUALS us matches findings with a ResourceRegion that starts with a value other than us.
NOT_CONTAINS, NOT_EQUALS, and PREFIX_NOT_EQUALS filters on the same field are joined by AND. A finding matches only if it matches all of those filters. For example, the filters Title NOT_CONTAINS CloudFront AND Title NOT_CONTAINS CloudWatch match a finding that excludes both CloudFront and CloudWatch in the title.
You can’t have both a CONTAINS filter and a NOT_CONTAINS filter on the same field. Similarly, you can't provide both an EQUALS filter and a NOT_EQUALS or PREFIX_NOT_EQUALS filter on the same field. Combining filters in this way returns an error. CONTAINS filters can only be used with other CONTAINS filters. NOT_CONTAINS filters can only be used with other NOT_CONTAINS filters.
You can combine PREFIX filters with NOT_EQUALS or PREFIX_NOT_EQUALS filters for the same field. Security Hub first processes the PREFIX filters, and then the NOT_EQUALS or PREFIX_NOT_EQUALS filters.
For example, for the following filters, Security Hub first identifies findings that have resource types that start with either AwsIam or AwsEc2. It then excludes findings that have a resource type of AwsIamPolicy and findings that have a resource type of AwsEc2NetworkInterface.
ResourceType PREFIX AwsIam
ResourceType PREFIX AwsEc2
ResourceType NOT_EQUALS AwsIamPolicy
ResourceType NOT_EQUALS AwsEc2NetworkInterface
CONTAINS and NOT_CONTAINS operators can be used only with automation rules. For more information, see Automation rules in the Security Hub User Guide.
"
}
},
- "documentation":"
A string filter for querying findings.
"
+ "documentation":"
A string filter for filtering Security Hub findings.
Specifies whether a rule is the last to be applied with respect to a finding that matches the rule criteria. This is useful when a finding matches the criteria for multiple rules, and each rule has different actions. If the value of this field is set to true for a rule, Security Hub applies the rule action to a finding that matches the rule criteria and doesn't evaluate other rules for the finding. The default value of this field is false.
"
+ "documentation":"
Specifies whether a rule is the last to be applied with respect to a finding that matches the rule criteria. This is useful when a finding matches the criteria for multiple rules, and each rule has different actions. If a rule is terminal, Security Hub applies the rule action to a finding that matches the rule criteria and doesn't evaluate other rules for the finding. By default, a rule isn't terminal.
Requests the import of a resource as an Service Catalog provisioned product that is associated to an Service Catalog product and provisioning artifact. Once imported, all supported governance actions are supported on the provisioned product.
Resource import only supports CloudFormation stack ARNs. CloudFormation StackSets, and non-root nested stacks are not supported.
The CloudFormation stack must have one of the following statuses to be imported: CREATE_COMPLETE, UPDATE_COMPLETE, UPDATE_ROLLBACK_COMPLETE, IMPORT_COMPLETE, and IMPORT_ROLLBACK_COMPLETE.
Import of the resource requires that the CloudFormation stack template matches the associated Service Catalog product provisioning artifact.
When you import an existing CloudFormation stack into a portfolio, constraints that are associated with the product aren't applied during the import process. The constraints are applied after you call UpdateProvisionedProduct for the provisioned product.
The user or role that performs this operation must have the cloudformation:GetTemplate and cloudformation:DescribeStacks IAM policy permissions.
"
+ "documentation":"
Requests the import of a resource as an Service Catalog provisioned product that is associated to an Service Catalog product and provisioning artifact. Once imported, all supported governance actions are supported on the provisioned product.
Resource import only supports CloudFormation stack ARNs. CloudFormation StackSets, and non-root nested stacks, are not supported.
The CloudFormation stack must have one of the following statuses to be imported: CREATE_COMPLETE, UPDATE_COMPLETE, UPDATE_ROLLBACK_COMPLETE, IMPORT_COMPLETE, and IMPORT_ROLLBACK_COMPLETE.
Import of the resource requires that the CloudFormation stack template matches the associated Service Catalog product provisioning artifact.
When you import an existing CloudFormation stack into a portfolio, Service Catalog does not apply the product's associated constraints during the import process. Service Catalog applies the constraints after you call UpdateProvisionedProduct for the provisioned product.
The user or role that performs this operation must have the cloudformation:GetTemplate and cloudformation:DescribeStacks IAM policy permissions.
You can only import one provisioned product at a time. The product's CloudFormation stack must have the IMPORT_COMPLETE status before you import another.
Updates the specified portfolio share. You can use this API to enable or disable TagOptions sharing or Principal sharing for an existing portfolio share.
The portfolio share cannot be updated if the CreatePortfolioShare operation is IN_PROGRESS, as the share is not available to recipient entities. In this case, you must wait for the portfolio share to be COMPLETED.
You must provide the accountId or organization node in the input, but not both.
If the portfolio is shared to both an external account and an organization node, and both shares need to be updated, you must invoke UpdatePortfolioShare separately for each share type.
This API cannot be used for removing the portfolio share. You must use DeletePortfolioShare API for that action.
When you associate a principal with portfolio, a potential privilege escalation path may occur when that portfolio is then shared with other accounts. For a user in a recipient account who is not an Service Catalog Admin, but still has the ability to create Principals (Users/Groups/Roles), that user could create a role that matches a principal name association for the portfolio. Although this user may not know which principal names are associated through Service Catalog, they may be able to guess the user. If this potential escalation path is a concern, then Service Catalog recommends using PrincipalType as IAM. With this configuration, the PrincipalARN must already exist in the recipient account before it can be associated.
"
+ "documentation":"
Updates the specified portfolio share. You can use this API to enable or disable TagOptions sharing or Principal sharing for an existing portfolio share.
The portfolio share cannot be updated if the CreatePortfolioShare operation is IN_PROGRESS, as the share is not available to recipient entities. In this case, you must wait for the portfolio share to be completed.
You must provide the accountId or organization node in the input, but not both.
If the portfolio is shared to both an external account and an organization node, and both shares need to be updated, you must invoke UpdatePortfolioShare separately for each share type.
This API cannot be used for removing the portfolio share. You must use DeletePortfolioShare API for that action.
When you associate a principal with portfolio, a potential privilege escalation path may occur when that portfolio is then shared with other accounts. For a user in a recipient account who is not an Service Catalog Admin, but still has the ability to create Principals (Users/Groups/Roles), that user could create a role that matches a principal name association for the portfolio. Although this user may not know which principal names are associated through Service Catalog, they may be able to guess the user. If this potential escalation path is a concern, then Service Catalog recommends using PrincipalType as IAM. With this configuration, the PrincipalARN must already exist in the recipient account before it can be associated.
When the key is SearchQuery, the searchable fields are arn, createdTime, id, lastRecordId, idempotencyToken, name, physicalId, productId, provisioningArtifact, type, status, tags, userArn, userArnSession, lastProvisioningRecordId, lastSuccessfulProvisioningRecordId, productName, and provisioningArtifactName.
Example: \"SearchQuery\":[\"status:AVAILABLE\"]
"
+ "documentation":"
The search filters.
When the key is SearchQuery, the searchable fields are arn, createdTime, id, lastRecordId, idempotencyToken, name, physicalId, productId, provisioningArtifactId, type, status, tags, userArn, userArnSession, lastProvisioningRecordId, lastSuccessfulProvisioningRecordId, productName, and provisioningArtifactName.
Example: \"SearchQuery\":[\"status:AVAILABLE\"]
"
},
"SortBy":{
"shape":"SortField",
diff --git a/services/servicecatalogappregistry/pom.xml b/services/servicecatalogappregistry/pom.xml
index 58c02124c4f..5a4fa7ee125 100644
--- a/services/servicecatalogappregistry/pom.xml
+++ b/services/servicecatalogappregistry/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTservicecatalogappregistryAWS Java SDK :: Services :: Service Catalog App Registry
diff --git a/services/servicediscovery/pom.xml b/services/servicediscovery/pom.xml
index 8ff747a3d86..34e9cd51d09 100644
--- a/services/servicediscovery/pom.xml
+++ b/services/servicediscovery/pom.xml
@@ -20,7 +20,7 @@
servicessoftware.amazon.awssdk
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOT4.0.0servicediscovery
diff --git a/services/servicequotas/pom.xml b/services/servicequotas/pom.xml
index 480c984c98b..b2d90fbba0c 100644
--- a/services/servicequotas/pom.xml
+++ b/services/servicequotas/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTservicequotasAWS Java SDK :: Services :: Service Quotas
diff --git a/services/ses/pom.xml b/services/ses/pom.xml
index c6a5c4c02d4..ea8beadd35e 100644
--- a/services/ses/pom.xml
+++ b/services/ses/pom.xml
@@ -21,7 +21,7 @@
software.amazon.awssdkservices
- 2.20.109-SNAPSHOT
+ 2.20.126-SNAPSHOTsesAWS Java SDK :: Services :: Amazon SES
diff --git a/services/ses/src/main/resources/codegen-resources/endpoint-rule-set.json b/services/ses/src/main/resources/codegen-resources/endpoint-rule-set.json
index 103753b59b9..1d567c5f974 100644
--- a/services/ses/src/main/resources/codegen-resources/endpoint-rule-set.json
+++ b/services/ses/src/main/resources/codegen-resources/endpoint-rule-set.json
@@ -32,13 +32,12 @@
{
"conditions": [
{
- "fn": "aws.partition",
+ "fn": "isSet",
"argv": [
{
- "ref": "Region"
+ "ref": "Endpoint"
}
- ],
- "assign": "PartitionResult"
+ ]
}
],
"type": "tree",
@@ -46,23 +45,20 @@
{
"conditions": [
{
- "fn": "isSet",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "Endpoint"
- }
+ "ref": "UseFIPS"
+ },
+ true
]
- },
- {
- "fn": "parseURL",
- "argv": [
- {
- "ref": "Endpoint"
- }
- ],
- "assign": "url"
}
],
+ "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "type": "error"
+ },
+ {
+ "conditions": [],
"type": "tree",
"rules": [
{
@@ -71,67 +67,42 @@
"fn": "booleanEquals",
"argv": [
{
- "ref": "UseFIPS"
+ "ref": "UseDualStack"
},
true
]
}
],
- "error": "Invalid Configuration: FIPS and custom endpoint are not supported",
+ "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
"type": "error"
},
{
"conditions": [],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
- },
- true
- ]
- }
- ],
- "error": "Invalid Configuration: Dualstack and custom endpoint are not supported",
- "type": "error"
+ "endpoint": {
+ "url": {
+ "ref": "Endpoint"
},
- {
- "conditions": [],
- "endpoint": {
- "url": {
- "ref": "Endpoint"
- },
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
}
]
- },
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
{
"conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
- },
- true
- ]
- },
- {
- "fn": "booleanEquals",
+ "fn": "isSet",
"argv": [
{
- "ref": "UseDualStack"
- },
- true
+ "ref": "Region"
+ }
]
}
],
@@ -140,90 +111,215 @@
{
"conditions": [
{
- "fn": "booleanEquals",
+ "fn": "aws.partition",
"argv": [
- true,
{
- "fn": "getAttr",
+ "ref": "Region"
+ }
+ ],
+ "assign": "PartitionResult"
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseFIPS"
},
- "supportsFIPS"
+ true
+ ]
+ },
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ {
+ "ref": "UseDualStack"
+ },
+ true
]
}
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsFIPS"
+ ]
+ }
+ ]
+ },
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://email-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
+ "type": "error"
+ }
]
},
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseFIPS"
},
- "supportsDualStack"
+ true
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsFIPS"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://email-fips.{Region}.{PartitionResult#dnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
]
+ },
+ {
+ "conditions": [],
+ "error": "FIPS is enabled but this partition does not support FIPS",
+ "type": "error"
}
]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
- "endpoint": {
- "url": "https://email-fips.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
- },
- {
- "conditions": [],
- "error": "FIPS and DualStack are enabled, but this partition does not support one or both",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseFIPS"
},
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
{
- "fn": "booleanEquals",
- "argv": [
- true,
+ "conditions": [
{
- "fn": "getAttr",
+ "fn": "booleanEquals",
"argv": [
{
- "ref": "PartitionResult"
+ "ref": "UseDualStack"
},
- "supportsFIPS"
+ true
]
}
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [
+ {
+ "fn": "booleanEquals",
+ "argv": [
+ true,
+ {
+ "fn": "getAttr",
+ "argv": [
+ {
+ "ref": "PartitionResult"
+ },
+ "supportsDualStack"
+ ]
+ }
+ ]
+ }
+ ],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "type": "tree",
+ "rules": [
+ {
+ "conditions": [],
+ "endpoint": {
+ "url": "https://email.{Region}.{PartitionResult#dualStackDnsSuffix}",
+ "properties": {},
+ "headers": {}
+ },
+ "type": "endpoint"
+ }
+ ]
+ }
+ ]
+ },
+ {
+ "conditions": [],
+ "error": "DualStack is enabled but this partition does not support DualStack",
+ "type": "error"
+ }
]
- }
- ],
- "type": "tree",
- "rules": [
+ },
{
"conditions": [],
"type": "tree",
@@ -231,7 +327,7 @@
{
"conditions": [],
"endpoint": {
- "url": "https://email-fips.{Region}.{PartitionResult#dnsSuffix}",
+ "url": "https://email.{Region}.{PartitionResult#dnsSuffix}",
"properties": {},
"headers": {}
},
@@ -240,74 +336,13 @@
]
}
]
- },
- {
- "conditions": [],
- "error": "FIPS is enabled but this partition does not support FIPS",
- "type": "error"
- }
- ]
- },
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- {
- "ref": "UseDualStack"
- },
- true
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [
- {
- "fn": "booleanEquals",
- "argv": [
- true,
- {
- "fn": "getAttr",
- "argv": [
- {
- "ref": "PartitionResult"
- },
- "supportsDualStack"
- ]
- }
- ]
- }
- ],
- "type": "tree",
- "rules": [
- {
- "conditions": [],
- "endpoint": {
- "url": "https://email.{Region}.{PartitionResult#dualStackDnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
- }
- ]
- },
- {
- "conditions": [],
- "error": "DualStack is enabled but this partition does not support DualStack",
- "type": "error"
}
]
},
{
"conditions": [],
- "endpoint": {
- "url": "https://email.{Region}.{PartitionResult#dnsSuffix}",
- "properties": {},
- "headers": {}
- },
- "type": "endpoint"
+ "error": "Invalid Configuration: Missing Region",
+ "type": "error"
}
]
}
diff --git a/services/ses/src/main/resources/codegen-resources/endpoint-tests.json b/services/ses/src/main/resources/codegen-resources/endpoint-tests.json
index a2a048c63dc..fa4feb98bdd 100644
--- a/services/ses/src/main/resources/codegen-resources/endpoint-tests.json
+++ b/services/ses/src/main/resources/codegen-resources/endpoint-tests.json
@@ -1,822 +1,198 @@
{
"testCases": [
{
- "documentation": "For region ap-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.ap-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.ap-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email.ap-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email.ap-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.eu-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.eu-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email.eu-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email.eu-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.ca-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ca-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.ca-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ca-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email.ca-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ca-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email.ca-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ca-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.eu-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.eu-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-central-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email.eu-central-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email.eu-central-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-central-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.us-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.us-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email.us-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email.us-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.us-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.us-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email.us-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email.us-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "us-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.af-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "af-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.af-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "af-south-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email.af-south-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "af-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email.af-south-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "af-south-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.eu-north-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-north-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.eu-north-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-north-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email.eu-north-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-north-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email.eu-north-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-north-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.eu-west-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.eu-west-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email.eu-west-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email.eu-west-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.eu-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.eu-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-2",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email.eu-west-2.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email.eu-west-2.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-2",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.eu-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.eu-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-1",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email.eu-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "eu-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email.eu-west-1.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "eu-west-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.ap-northeast-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS enabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.ap-northeast-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-3",
- "UseFIPS": true
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email.ap-northeast-3.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "ap-northeast-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack disabled",
- "expect": {
- "endpoint": {
- "url": "https://email.ap-northeast-3.amazonaws.com"
- }
- },
- "params": {
- "UseDualStack": false,
- "Region": "ap-northeast-3",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.ap-northeast-2.api.aws"
+ "url": "https://email.af-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-northeast-2",
- "UseFIPS": true
+ "Region": "af-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.ap-northeast-2.amazonaws.com"
+ "url": "https://email.ap-northeast-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-northeast-2",
- "UseFIPS": true
+ "Region": "ap-northeast-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.ap-northeast-2.api.aws"
+ "url": "https://email.ap-northeast-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
"Region": "ap-northeast-2",
- "UseFIPS": false
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ap-northeast-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.ap-northeast-2.amazonaws.com"
+ "url": "https://email.ap-northeast-3.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-northeast-2",
- "UseFIPS": false
+ "Region": "ap-northeast-3",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.ap-northeast-1.api.aws"
+ "url": "https://email.ap-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-northeast-1",
- "UseFIPS": true
+ "Region": "ap-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.ap-northeast-1.amazonaws.com"
+ "url": "https://email.ap-southeast-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-northeast-1",
- "UseFIPS": true
+ "Region": "ap-southeast-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.ap-northeast-1.api.aws"
+ "url": "https://email.ap-southeast-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-northeast-1",
- "UseFIPS": false
+ "Region": "ap-southeast-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.ap-northeast-1.amazonaws.com"
+ "url": "https://email.ca-central-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-northeast-1",
- "UseFIPS": false
+ "Region": "ca-central-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.me-south-1.api.aws"
+ "url": "https://email.eu-central-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "me-south-1",
- "UseFIPS": true
+ "Region": "eu-central-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.me-south-1.amazonaws.com"
+ "url": "https://email.eu-north-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "me-south-1",
- "UseFIPS": true
+ "Region": "eu-north-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region eu-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.me-south-1.api.aws"
+ "url": "https://email.eu-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "me-south-1",
- "UseFIPS": false
+ "Region": "eu-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.me-south-1.amazonaws.com"
+ "url": "https://email.eu-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "me-south-1",
- "UseFIPS": false
+ "Region": "eu-west-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region sa-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.sa-east-1.api.aws"
+ "url": "https://email.eu-west-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "sa-east-1",
- "UseFIPS": true
+ "Region": "eu-west-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region sa-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.sa-east-1.amazonaws.com"
+ "url": "https://email.eu-west-3.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "sa-east-1",
- "UseFIPS": true
+ "Region": "eu-west-3",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region sa-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.sa-east-1.api.aws"
+ "url": "https://email.me-south-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "sa-east-1",
- "UseFIPS": false
+ "Region": "me-south-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
@@ -827,386 +203,365 @@
}
},
"params": {
- "UseDualStack": false,
"Region": "sa-east-1",
- "UseFIPS": false
- }
- },
- {
- "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack enabled",
- "expect": {
- "endpoint": {
- "url": "https://email-fips.us-gov-west-1.api.aws"
- }
- },
- "params": {
- "UseDualStack": true,
- "Region": "us-gov-west-1",
- "UseFIPS": true
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.us-gov-west-1.amazonaws.com"
+ "url": "https://email.us-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-gov-west-1",
- "UseFIPS": true
+ "Region": "us-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.us-gov-west-1.api.aws"
+ "url": "https://email-fips.us-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-gov-west-1",
- "UseFIPS": false
+ "Region": "us-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.us-gov-west-1.amazonaws.com"
+ "url": "https://email.us-east-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-gov-west-1",
- "UseFIPS": false
+ "Region": "us-east-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.ap-southeast-1.api.aws"
+ "url": "https://email.us-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-1",
- "UseFIPS": true
+ "Region": "us-west-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.ap-southeast-1.amazonaws.com"
+ "url": "https://email.us-west-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-1",
- "UseFIPS": true
+ "Region": "us-west-2",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-west-2 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.ap-southeast-1.api.aws"
+ "url": "https://email-fips.us-west-2.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-1",
- "UseFIPS": false
+ "Region": "us-west-2",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://email.ap-southeast-1.amazonaws.com"
+ "url": "https://email-fips.us-east-1.api.aws"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-1",
- "UseFIPS": false
+ "Region": "us-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.ap-southeast-2.api.aws"
+ "url": "https://email.us-east-1.api.aws"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-2",
- "UseFIPS": true
+ "Region": "us-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.ap-southeast-2.amazonaws.com"
+ "url": "https://email-fips.cn-north-1.api.amazonwebservices.com.cn"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-2",
- "UseFIPS": true
+ "Region": "cn-north-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.ap-southeast-2.api.aws"
+ "url": "https://email-fips.cn-north-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-2",
- "UseFIPS": false
+ "Region": "cn-north-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://email.ap-southeast-2.amazonaws.com"
+ "url": "https://email.cn-north-1.api.amazonwebservices.com.cn"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-2",
- "UseFIPS": false
+ "Region": "cn-north-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS enabled and DualStack enabled",
+ "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.ap-southeast-3.api.aws"
+ "url": "https://email.cn-north-1.amazonaws.com.cn"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-3",
- "UseFIPS": true
+ "Region": "cn-north-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.ap-southeast-3.amazonaws.com"
+ "url": "https://email.us-gov-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-3",
- "UseFIPS": true
+ "Region": "us-gov-west-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.ap-southeast-3.api.aws"
+ "url": "https://email-fips.us-gov-west-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "ap-southeast-3",
- "UseFIPS": false
+ "Region": "us-gov-west-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region ap-southeast-3 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://email.ap-southeast-3.amazonaws.com"
+ "url": "https://email-fips.us-gov-east-1.api.aws"
}
},
"params": {
- "UseDualStack": false,
- "Region": "ap-southeast-3",
- "UseFIPS": false
+ "Region": "us-gov-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.us-east-1.api.aws"
+ "url": "https://email-fips.us-gov-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-1",
- "UseFIPS": true
+ "Region": "us-gov-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.us-east-1.amazonaws.com"
+ "url": "https://email.us-gov-east-1.api.aws"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-1",
- "UseFIPS": true
+ "Region": "us-gov-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.us-east-1.api.aws"
+ "url": "https://email.us-gov-east-1.amazonaws.com"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-1",
- "UseFIPS": false
+ "Region": "us-gov-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://email.us-east-1.amazonaws.com"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-1",
- "UseFIPS": false
+ "Region": "us-iso-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-2 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.us-east-2.api.aws"
+ "url": "https://email-fips.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-2",
- "UseFIPS": true
+ "Region": "us-iso-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://email-fips.us-east-2.amazonaws.com"
- }
+ "error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-2",
- "UseFIPS": true
+ "Region": "us-iso-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region us-east-2 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.us-east-2.api.aws"
+ "url": "https://email.us-iso-east-1.c2s.ic.gov"
}
},
"params": {
- "UseDualStack": true,
- "Region": "us-east-2",
- "UseFIPS": false
+ "Region": "us-iso-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled",
+ "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://email.us-east-2.amazonaws.com"
- }
+ "error": "FIPS and DualStack are enabled, but this partition does not support one or both"
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-2",
- "UseFIPS": false
+ "Region": "us-isob-east-1",
+ "UseFIPS": true,
+ "UseDualStack": true
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS enabled and DualStack enabled",
+ "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email-fips.cn-northwest-1.api.amazonwebservices.com.cn"
+ "url": "https://email-fips.us-isob-east-1.sc2s.sgov.gov"
}
},
"params": {
- "UseDualStack": true,
- "Region": "cn-northwest-1",
- "UseFIPS": true
+ "Region": "us-isob-east-1",
+ "UseFIPS": true,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS enabled and DualStack disabled",
+ "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack enabled",
"expect": {
- "endpoint": {
- "url": "https://email-fips.cn-northwest-1.amazonaws.com.cn"
- }
+ "error": "DualStack is enabled but this partition does not support DualStack"
},
"params": {
- "UseDualStack": false,
- "Region": "cn-northwest-1",
- "UseFIPS": true
+ "Region": "us-isob-east-1",
+ "UseFIPS": false,
+ "UseDualStack": true
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack enabled",
+ "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack disabled",
"expect": {
"endpoint": {
- "url": "https://email.cn-northwest-1.api.amazonwebservices.com.cn"
+ "url": "https://email.us-isob-east-1.sc2s.sgov.gov"
}
},
"params": {
- "UseDualStack": true,
- "Region": "cn-northwest-1",
- "UseFIPS": false
+ "Region": "us-isob-east-1",
+ "UseFIPS": false,
+ "UseDualStack": false
}
},
{
- "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack disabled",
+ "documentation": "For custom endpoint with region set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
- "url": "https://email.cn-northwest-1.amazonaws.com.cn"
+ "url": "https://example.com"
}
},
"params": {
+ "Region": "us-east-1",
+ "UseFIPS": false,
"UseDualStack": false,
- "Region": "cn-northwest-1",
- "UseFIPS": false
+ "Endpoint": "https://example.com"
}
},
{
- "documentation": "For custom endpoint with fips disabled and dualstack disabled",
+ "documentation": "For custom endpoint with region not set and fips disabled and dualstack disabled",
"expect": {
"endpoint": {
"url": "https://example.com"
}
},
"params": {
- "UseDualStack": false,
- "Region": "us-east-1",
"UseFIPS": false,
+ "UseDualStack": false,
"Endpoint": "https://example.com"
}
},
@@ -1216,9 +571,9 @@
"error": "Invalid Configuration: FIPS and custom endpoint are not supported"
},
"params": {
- "UseDualStack": false,
"Region": "us-east-1",
"UseFIPS": true,
+ "UseDualStack": false,
"Endpoint": "https://example.com"
}
},
@@ -1228,11 +583,17 @@
"error": "Invalid Configuration: Dualstack and custom endpoint are not supported"
},
"params": {
- "UseDualStack": true,
"Region": "us-east-1",
"UseFIPS": false,
+ "UseDualStack": true,
"Endpoint": "https://example.com"
}
+ },
+ {
+ "documentation": "Missing region",
+ "expect": {
+ "error": "Invalid Configuration: Missing Region"
+ }
}
],
"version": "1.0"
diff --git a/services/ses/src/main/resources/codegen-resources/service-2.json b/services/ses/src/main/resources/codegen-resources/service-2.json
index 5c376aa4938..0e89c30bb7a 100644
--- a/services/ses/src/main/resources/codegen-resources/service-2.json
+++ b/services/ses/src/main/resources/codegen-resources/service-2.json
@@ -29,7 +29,7 @@
{"shape":"AlreadyExistsException"},
{"shape":"LimitExceededException"}
],
- "documentation":"
Creates a receipt rule set by cloning an existing one. All receipt rules and configurations are copied to the new receipt rule set and are completely independent of the source rule set.
You can execute this operation no more than once per second.
"
+ "documentation":"
Creates a receipt rule set by cloning an existing one. All receipt rules and configurations are copied to the new receipt rule set and are completely independent of the source rule set.
When you create or update an event destination, you must provide one, and only one, destination. The destination can be CloudWatch, Amazon Kinesis Firehose, or Amazon Simple Notification Service (Amazon SNS).
An event destination is the AWS service to which Amazon SES publishes the email sending events associated with a configuration set. For information about using configuration sets, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
"
+ "documentation":"
Creates a configuration set event destination.
When you create or update an event destination, you must provide one, and only one, destination. The destination can be CloudWatch, Amazon Kinesis Firehose, or Amazon Simple Notification Service (Amazon SNS).
An event destination is the Amazon Web Services service to which Amazon SES publishes the email sending events associated with a configuration set. For information about using configuration sets, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
Creates an association between a configuration set and a custom domain for open and click event tracking.
By default, images and links used for tracking open and click events are hosted on domains operated by Amazon SES. You can configure a subdomain of your own to handle these events. For information about using custom domains, see the Amazon SES Developer Guide.
"
+ "documentation":"
Creates an association between a configuration set and a custom domain for open and click event tracking.
By default, images and links used for tracking open and click events are hosted on domains operated by Amazon SES. You can configure a subdomain of your own to handle these events. For information about using custom domains, see the Amazon SES Developer Guide.
Creates an email template. Email templates enable you to send personalized email to one or more destinations in a single API operation. For more information, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
"
+ "documentation":"
Creates an email template. Email templates enable you to send personalized email to one or more destinations in a single operation. For more information, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
Deletes a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
"
+ "documentation":"
Deletes a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
Deletes a configuration set event destination. Configuration set event destinations are associated with configuration sets, which enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
"
+ "documentation":"
Deletes a configuration set event destination. Configuration set event destinations are associated with configuration sets, which enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
Deletes an association between a configuration set and a custom domain for open and click event tracking.
By default, images and links used for tracking open and click events are hosted on domains operated by Amazon SES. You can configure a subdomain of your own to handle these events. For information about using custom domains, see the Amazon SES Developer Guide.
Deleting this kind of association will result in emails sent using the specified configuration set to capture open and click events using the standard, Amazon SES-operated domains.
"
+ "documentation":"
Deletes an association between a configuration set and a custom domain for open and click event tracking.
By default, images and links used for tracking open and click events are hosted on domains operated by Amazon SES. You can configure a subdomain of your own to handle these events. For information about using custom domains, see the Amazon SES Developer Guide.
Deleting this kind of association results in emails sent using the specified configuration set to capture open and click events using the standard, Amazon SES-operated domains.
Deletes the specified sending authorization policy for the given identity (an email address or a domain). This API returns successfully even if a policy with the specified name does not exist.
This API is for the identity owner only. If you have not verified the identity, this API will return an error.
Sending authorization is a feature that enables an identity owner to authorize other senders to use its identities. For information about using sending authorization, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
"
+ "documentation":"
Deletes the specified sending authorization policy for the given identity (an email address or a domain). This operation returns successfully even if a policy with the specified name does not exist.
This operation is for the identity owner only. If you have not verified the identity, it returns an error.
Sending authorization is a feature that enables an identity owner to authorize other senders to use its identities. For information about using sending authorization, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
Returns the current status of Easy DKIM signing for an entity. For domain name identities, this operation also returns the DKIM tokens that are required for Easy DKIM signing, and whether Amazon SES has successfully verified that these tokens have been published.
This operation takes a list of identities as input and returns the following information for each:
Whether Easy DKIM signing is enabled or disabled.
A set of DKIM tokens that represent the identity. If the identity is an email address, the tokens represent the domain of that address.
Whether Amazon SES has successfully verified the DKIM tokens published in the domain's DNS. This information is only returned for domain name identities, not for email addresses.
This operation is throttled at one request per second and can only get DKIM attributes for up to 100 identities at a time.
For more information about creating DNS records using DKIM tokens, go to the Amazon SES Developer Guide.
"
+ "documentation":"
Returns the current status of Easy DKIM signing for an entity. For domain name identities, this operation also returns the DKIM tokens that are required for Easy DKIM signing, and whether Amazon SES has successfully verified that these tokens have been published.
This operation takes a list of identities as input and returns the following information for each:
Whether Easy DKIM signing is enabled or disabled.
A set of DKIM tokens that represent the identity. If the identity is an email address, the tokens represent the domain of that address.
Whether Amazon SES has successfully verified the DKIM tokens published in the domain's DNS. This information is only returned for domain name identities, not for email addresses.
This operation is throttled at one request per second and can only get DKIM attributes for up to 100 identities at a time.
For more information about creating DNS records using DKIM tokens, go to the Amazon SES Developer Guide.
Returns the requested sending authorization policies for the given identity (an email address or a domain). The policies are returned as a map of policy names to policy contents. You can retrieve a maximum of 20 policies at a time.
This API is for the identity owner only. If you have not verified the identity, this API will return an error.
Sending authorization is a feature that enables an identity owner to authorize other senders to use its identities. For information about using sending authorization, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
"
+ "documentation":"
Returns the requested sending authorization policies for the given identity (an email address or a domain). The policies are returned as a map of policy names to policy contents. You can retrieve a maximum of 20 policies at a time.
This operation is for the identity owner only. If you have not verified the identity, it returns an error.
Sending authorization is a feature that enables an identity owner to authorize other senders to use its identities. For information about using sending authorization, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
Given a list of identities (email addresses and/or domains), returns the verification status and (for domain identities) the verification token for each identity.
The verification status of an email address is \"Pending\" until the email address owner clicks the link within the verification email that Amazon SES sent to that address. If the email address owner clicks the link within 24 hours, the verification status of the email address changes to \"Success\". If the link is not clicked within 24 hours, the verification status changes to \"Failed.\" In that case, if you still want to verify the email address, you must restart the verification process from the beginning.
For domain identities, the domain's verification status is \"Pending\" as Amazon SES searches for the required TXT record in the DNS settings of the domain. When Amazon SES detects the record, the domain's verification status changes to \"Success\". If Amazon SES is unable to detect the record within 72 hours, the domain's verification status changes to \"Failed.\" In that case, if you still want to verify the domain, you must restart the verification process from the beginning.
This operation is throttled at one request per second and can only get verification attributes for up to 100 identities at a time.
"
+ "documentation":"
Given a list of identities (email addresses and/or domains), returns the verification status and (for domain identities) the verification token for each identity.
The verification status of an email address is \"Pending\" until the email address owner clicks the link within the verification email that Amazon SES sent to that address. If the email address owner clicks the link within 24 hours, the verification status of the email address changes to \"Success\". If the link is not clicked within 24 hours, the verification status changes to \"Failed.\" In that case, to verify the email address, you must restart the verification process from the beginning.
For domain identities, the domain's verification status is \"Pending\" as Amazon SES searches for the required TXT record in the DNS settings of the domain. When Amazon SES detects the record, the domain's verification status changes to \"Success\". If Amazon SES is unable to detect the record within 72 hours, the domain's verification status changes to \"Failed.\" In that case, to verify the domain, you must restart the verification process from the beginning.
This operation is throttled at one request per second and can only get verification attributes for up to 100 identities at a time.
Provides sending statistics for the current AWS Region. The result is a list of data points, representing the last two weeks of sending activity. Each data point in the list contains statistics for a 15-minute period of time.
You can execute this operation no more than once per second.
"
+ "documentation":"
Provides sending statistics for the current Amazon Web Services Region. The result is a list of data points, representing the last two weeks of sending activity. Each data point in the list contains statistics for a 15-minute period of time.
You can execute this operation no more than once per second.
Provides a list of the configuration sets associated with your Amazon SES account in the current AWS Region. For information about using configuration sets, see Monitoring Your Amazon SES Sending Activity in the Amazon SES Developer Guide.
You can execute this operation no more than once per second. This operation will return up to 1,000 configuration sets each time it is run. If your Amazon SES account has more than 1,000 configuration sets, this operation will also return a NextToken element. You can then execute the ListConfigurationSets operation again, passing the NextToken parameter and the value of the NextToken element to retrieve additional results.
"
+ "documentation":"
Provides a list of the configuration sets associated with your Amazon SES account in the current Amazon Web Services Region. For information about using configuration sets, see Monitoring Your Amazon SES Sending Activity in the Amazon SES Developer Guide.
You can execute this operation no more than once per second. This operation returns up to 1,000 configuration sets each time it is run. If your Amazon SES account has more than 1,000 configuration sets, this operation also returns NextToken. You can then execute the ListConfigurationSets operation again, passing the NextToken parameter and the value of the NextToken element to retrieve additional results.
Returns a list containing all of the identities (email addresses and domains) for your AWS account in the current AWS Region, regardless of verification status.
You can execute this operation no more than once per second.
"
+ "documentation":"
Returns a list containing all of the identities (email addresses and domains) for your Amazon Web Services account in the current Amazon Web Services Region, regardless of verification status.
You can execute this operation no more than once per second.
It's recommended that for successive pagination calls of this API, you continue to the use the same parameter/value pairs as used in the original call, e.g., if you used IdentityType=Domain in the the original call and received a NextToken in the response, you should continue providing the IdentityType=Domain parameter for further NextToken calls; however, if you didn't provide the IdentityType parameter in the original call, then continue to not provide it for successive pagination calls. Using this protocol will ensure consistent results.
Returns a list of sending authorization policies that are attached to the given identity (an email address or a domain). This API returns only a list. If you want the actual policy content, you can use GetIdentityPolicies.
This API is for the identity owner only. If you have not verified the identity, this API will return an error.
Sending authorization is a feature that enables an identity owner to authorize other senders to use its identities. For information about using sending authorization, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
"
+ "documentation":"
Returns a list of sending authorization policies that are attached to the given identity (an email address or a domain). This operation returns only a list. To get the actual policy content, use GetIdentityPolicies.
This operation is for the identity owner only. If you have not verified the identity, it returns an error.
Sending authorization is a feature that enables an identity owner to authorize other senders to use its identities. For information about using sending authorization, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
Lists the receipt rule sets that exist under your AWS account in the current AWS Region. If there are additional receipt rule sets to be retrieved, you will receive a NextToken that you can provide to the next call to ListReceiptRuleSets to retrieve the additional entries.
You can execute this operation no more than once per second.
"
+ "documentation":"
Lists the receipt rule sets that exist under your Amazon Web Services account in the current Amazon Web Services Region. If there are additional receipt rule sets to be retrieved, you receive a NextToken that you can provide to the next call to ListReceiptRuleSets to retrieve the additional entries.
Adds or updates a sending authorization policy for the specified identity (an email address or a domain).
This API is for the identity owner only. If you have not verified the identity, this API will return an error.
Sending authorization is a feature that enables an identity owner to authorize other senders to use its identities. For information about using sending authorization, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
"
+ "documentation":"
Adds or updates a sending authorization policy for the specified identity (an email address or a domain).
This operation is for the identity owner only. If you have not verified the identity, it returns an error.
Sending authorization is a feature that enables an identity owner to authorize other senders to use its identities. For information about using sending authorization, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
Reorders the receipt rules within a receipt rule set.
All of the rules in the rule set must be represented in this request. That is, this API will return an error if the reorder request doesn't explicitly position all of the rules.
You can execute this operation no more than once per second.
"
+ "documentation":"
Reorders the receipt rules within a receipt rule set.
All of the rules in the rule set must be represented in this request. That is, it is error if the reorder request doesn't explicitly position all of the rules.
Generates and sends a bounce message to the sender of an email you received through Amazon SES. You can only use this API on an email up to 24 hours after you receive it.
You cannot use this API to send generic bounces for mail that was not received by Amazon SES.
You can execute this operation no more than once per second.
"
+ "documentation":"
Generates and sends a bounce message to the sender of an email you received through Amazon SES. You can only use this operation on an email up to 24 hours after you receive it.
You cannot use this operation to send generic bounces for mail that was not received by Amazon SES.
Composes an email message to multiple destinations. The message body is created using an email template.
In order to send email using the SendBulkTemplatedEmail operation, your call to the API must meet the following requirements:
The call must refer to an existing email template. You can create email templates using the CreateTemplate operation.
The message must be sent from a verified email address or domain.
If your account is still in the Amazon SES sandbox, you may only send to verified addresses or domains, or to email addresses associated with the Amazon SES Mailbox Simulator. For more information, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.
The maximum message size is 10 MB.
Each Destination parameter must include at least one recipient email address. The recipient address can be a To: address, a CC: address, or a BCC: address. If a recipient email address is invalid (that is, it is not in the format UserName@[SubDomain.]Domain.TopLevelDomain), the entire message will be rejected, even if the message contains other recipients that are valid.
The message may not include more than 50 recipients, across the To:, CC: and BCC: fields. If you need to send an email message to a larger audience, you can divide your recipient list into groups of 50 or fewer, and then call the SendBulkTemplatedEmail operation several times to send the message to each group.
The number of destinations you can contact in a single call to the API may be limited by your account's maximum sending rate.
"
+ "documentation":"
Composes an email message to multiple destinations. The message body is created using an email template.
To send email using this operation, your call must meet the following requirements:
The call must refer to an existing email template. You can create email templates using CreateTemplate.
The message must be sent from a verified email address or domain.
If your account is still in the Amazon SES sandbox, you may send only to verified addresses or domains, or to email addresses associated with the Amazon SES Mailbox Simulator. For more information, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.
The maximum message size is 10 MB.
Each Destination parameter must include at least one recipient email address. The recipient address can be a To: address, a CC: address, or a BCC: address. If a recipient email address is invalid (that is, it is not in the format UserName@[SubDomain.]Domain.TopLevelDomain), the entire message is rejected, even if the message contains other recipients that are valid.
The message may not include more than 50 recipients, across the To:, CC: and BCC: fields. If you need to send an email message to a larger audience, you can divide your recipient list into groups of 50 or fewer, and then call the SendBulkTemplatedEmail operation several times to send the message to each group.
The number of destinations you can contact in a single call can be limited by your account's maximum sending rate.
Adds an email address to the list of identities for your Amazon SES account in the current AWS Region and attempts to verify it. As a result of executing this operation, a customized verification email is sent to the specified address.
To use this operation, you must first create a custom verification email template. For more information about creating and using custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
"
+ "documentation":"
Adds an email address to the list of identities for your Amazon SES account in the current Amazon Web Services Region and attempts to verify it. As a result of executing this operation, a customized verification email is sent to the specified address.
To use this operation, you must first create a custom verification email template. For more information about creating and using custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
Composes an email message and immediately queues it for sending. In order to send email using the SendEmail operation, your message must meet the following requirements:
The message must be sent from a verified email address or domain. If you attempt to send email using a non-verified address or domain, the operation will result in an \"Email address not verified\" error.
If your account is still in the Amazon SES sandbox, you may only send to verified addresses or domains, or to email addresses associated with the Amazon SES Mailbox Simulator. For more information, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.
The maximum message size is 10 MB.
The message must include at least one recipient email address. The recipient address can be a To: address, a CC: address, or a BCC: address. If a recipient email address is invalid (that is, it is not in the format UserName@[SubDomain.]Domain.TopLevelDomain), the entire message will be rejected, even if the message contains other recipients that are valid.
The message may not include more than 50 recipients, across the To:, CC: and BCC: fields. If you need to send an email message to a larger audience, you can divide your recipient list into groups of 50 or fewer, and then call the SendEmail operation several times to send the message to each group.
For every message that you send, the total number of recipients (including each recipient in the To:, CC: and BCC: fields) is counted against the maximum number of emails you can send in a 24-hour period (your sending quota). For more information about sending quotas in Amazon SES, see Managing Your Amazon SES Sending Limits in the Amazon SES Developer Guide.
"
+ "documentation":"
Composes an email message and immediately queues it for sending. To send email using this operation, your message must meet the following requirements:
The message must be sent from a verified email address or domain. If you attempt to send email using a non-verified address or domain, the operation results in an \"Email address not verified\" error.
If your account is still in the Amazon SES sandbox, you may only send to verified addresses or domains, or to email addresses associated with the Amazon SES Mailbox Simulator. For more information, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.
The maximum message size is 10 MB.
The message must include at least one recipient email address. The recipient address can be a To: address, a CC: address, or a BCC: address. If a recipient email address is invalid (that is, it is not in the format UserName@[SubDomain.]Domain.TopLevelDomain), the entire message is rejected, even if the message contains other recipients that are valid.
The message may not include more than 50 recipients, across the To:, CC: and BCC: fields. If you need to send an email message to a larger audience, you can divide your recipient list into groups of 50 or fewer, and then call the SendEmail operation several times to send the message to each group.
For every message that you send, the total number of recipients (including each recipient in the To:, CC: and BCC: fields) is counted against the maximum number of emails you can send in a 24-hour period (your sending quota). For more information about sending quotas in Amazon SES, see Managing Your Amazon SES Sending Limits in the Amazon SES Developer Guide.
Composes an email message and immediately queues it for sending.
This operation is more flexible than the SendEmail API operation. When you use the SendRawEmail operation, you can specify the headers of the message as well as its content. This flexibility is useful, for example, when you want to send a multipart MIME email (such a message that contains both a text and an HTML version). You can also use this operation to send messages that include attachments.
The SendRawEmail operation has the following requirements:
You can only send email from verified email addresses or domains. If you try to send email from an address that isn't verified, the operation results in an \"Email address not verified\" error.
If your account is still in the Amazon SES sandbox, you can only send email to other verified addresses in your account, or to addresses that are associated with the Amazon SES mailbox simulator.
The maximum message size, including attachments, is 10 MB.
Each message has to include at least one recipient address. A recipient address includes any address on the To:, CC:, or BCC: lines.
If you send a single message to more than one recipient address, and one of the recipient addresses isn't in a valid format (that is, it's not in the format UserName@[SubDomain.]Domain.TopLevelDomain), Amazon SES rejects the entire message, even if the other addresses are valid.
Each message can include up to 50 recipient addresses across the To:, CC:, or BCC: lines. If you need to send a single message to more than 50 recipients, you have to split the list of recipient addresses into groups of less than 50 recipients, and send separate messages to each group.
Amazon SES allows you to specify 8-bit Content-Transfer-Encoding for MIME message parts. However, if Amazon SES has to modify the contents of your message (for example, if you use open and click tracking), 8-bit content isn't preserved. For this reason, we highly recommend that you encode all content that isn't 7-bit ASCII. For more information, see MIME Encoding in the Amazon SES Developer Guide.
Additionally, keep the following considerations in mind when using the SendRawEmail operation:
Although you can customize the message headers when using the SendRawEmail operation, Amazon SES will automatically apply its own Message-ID and Date headers; if you passed these headers when creating the message, they will be overwritten by the values that Amazon SES provides.
If you are using sending authorization to send on behalf of another user, SendRawEmail enables you to specify the cross-account identity for the email's Source, From, and Return-Path parameters in one of two ways: you can pass optional parameters SourceArn, FromArn, and/or ReturnPathArn to the API, or you can include the following X-headers in the header of your raw email:
X-SES-SOURCE-ARN
X-SES-FROM-ARN
X-SES-RETURN-PATH-ARN
Don't include these X-headers in the DKIM signature. Amazon SES removes these before it sends the email.
If you only specify the SourceIdentityArn parameter, Amazon SES sets the From and Return-Path addresses to the same identity that you specified.
For every message that you send, the total number of recipients (including each recipient in the To:, CC: and BCC: fields) is counted against the maximum number of emails you can send in a 24-hour period (your sending quota). For more information about sending quotas in Amazon SES, see Managing Your Amazon SES Sending Limits in the Amazon SES Developer Guide.
"
+ "documentation":"
Composes an email message and immediately queues it for sending.
This operation is more flexible than the SendEmail operation. When you use the SendRawEmail operation, you can specify the headers of the message as well as its content. This flexibility is useful, for example, when you need to send a multipart MIME email (such a message that contains both a text and an HTML version). You can also use this operation to send messages that include attachments.
The SendRawEmail operation has the following requirements:
You can only send email from verified email addresses or domains. If you try to send email from an address that isn't verified, the operation results in an \"Email address not verified\" error.
If your account is still in the Amazon SES sandbox, you can only send email to other verified addresses in your account, or to addresses that are associated with the Amazon SES mailbox simulator.
The maximum message size, including attachments, is 10 MB.
Each message has to include at least one recipient address. A recipient address includes any address on the To:, CC:, or BCC: lines.
If you send a single message to more than one recipient address, and one of the recipient addresses isn't in a valid format (that is, it's not in the format UserName@[SubDomain.]Domain.TopLevelDomain), Amazon SES rejects the entire message, even if the other addresses are valid.
Each message can include up to 50 recipient addresses across the To:, CC:, or BCC: lines. If you need to send a single message to more than 50 recipients, you have to split the list of recipient addresses into groups of less than 50 recipients, and send separate messages to each group.
Amazon SES allows you to specify 8-bit Content-Transfer-Encoding for MIME message parts. However, if Amazon SES has to modify the contents of your message (for example, if you use open and click tracking), 8-bit content isn't preserved. For this reason, we highly recommend that you encode all content that isn't 7-bit ASCII. For more information, see MIME Encoding in the Amazon SES Developer Guide.
Additionally, keep the following considerations in mind when using the SendRawEmail operation:
Although you can customize the message headers when using the SendRawEmail operation, Amazon SES automatically applies its own Message-ID and Date headers; if you passed these headers when creating the message, they are overwritten by the values that Amazon SES provides.
If you are using sending authorization to send on behalf of another user, SendRawEmail enables you to specify the cross-account identity for the email's Source, From, and Return-Path parameters in one of two ways: you can pass optional parameters SourceArn, FromArn, and/or ReturnPathArn, or you can include the following X-headers in the header of your raw email:
X-SES-SOURCE-ARN
X-SES-FROM-ARN
X-SES-RETURN-PATH-ARN
Don't include these X-headers in the DKIM signature. Amazon SES removes these before it sends the email.
If you only specify the SourceIdentityArn parameter, Amazon SES sets the From and Return-Path addresses to the same identity that you specified.
For every message that you send, the total number of recipients (including each recipient in the To:, CC: and BCC: fields) is counted against the maximum number of emails you can send in a 24-hour period (your sending quota). For more information about sending quotas in Amazon SES, see Managing Your Amazon SES Sending Limits in the Amazon SES Developer Guide.
Composes an email message using an email template and immediately queues it for sending.
In order to send email using the SendTemplatedEmail operation, your call to the API must meet the following requirements:
The call must refer to an existing email template. You can create email templates using the CreateTemplate operation.
The message must be sent from a verified email address or domain.
If your account is still in the Amazon SES sandbox, you may only send to verified addresses or domains, or to email addresses associated with the Amazon SES Mailbox Simulator. For more information, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.
The maximum message size is 10 MB.
Calls to the SendTemplatedEmail operation may only include one Destination parameter. A destination is a set of recipients who will receive the same version of the email. The Destination parameter can include up to 50 recipients, across the To:, CC: and BCC: fields.
The Destination parameter must include at least one recipient email address. The recipient address can be a To: address, a CC: address, or a BCC: address. If a recipient email address is invalid (that is, it is not in the format UserName@[SubDomain.]Domain.TopLevelDomain), the entire message will be rejected, even if the message contains other recipients that are valid.
If your call to the SendTemplatedEmail operation includes all of the required parameters, Amazon SES accepts it and returns a Message ID. However, if Amazon SES can't render the email because the template contains errors, it doesn't send the email. Additionally, because it already accepted the message, Amazon SES doesn't return a message stating that it was unable to send the email.
For these reasons, we highly recommend that you set up Amazon SES to send you notifications when Rendering Failure events occur. For more information, see Sending Personalized Email Using the Amazon SES API in the Amazon Simple Email Service Developer Guide.
"
+ "documentation":"
Composes an email message using an email template and immediately queues it for sending.
To send email using this operation, your call must meet the following requirements:
The call must refer to an existing email template. You can create email templates using the CreateTemplate operation.
The message must be sent from a verified email address or domain.
If your account is still in the Amazon SES sandbox, you may only send to verified addresses or domains, or to email addresses associated with the Amazon SES Mailbox Simulator. For more information, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.
The maximum message size is 10 MB.
Calls to the SendTemplatedEmail operation may only include one Destination parameter. A destination is a set of recipients that receives the same version of the email. The Destination parameter can include up to 50 recipients, across the To:, CC: and BCC: fields.
The Destination parameter must include at least one recipient email address. The recipient address can be a To: address, a CC: address, or a BCC: address. If a recipient email address is invalid (that is, it is not in the format UserName@[SubDomain.]Domain.TopLevelDomain), the entire message is rejected, even if the message contains other recipients that are valid.
If your call to the SendTemplatedEmail operation includes all of the required parameters, Amazon SES accepts it and returns a Message ID. However, if Amazon SES can't render the email because the template contains errors, it doesn't send the email. Additionally, because it already accepted the message, Amazon SES doesn't return a message stating that it was unable to send the email.
For these reasons, we highly recommend that you set up Amazon SES to send you notifications when Rendering Failure events occur. For more information, see Sending Personalized Email Using the Amazon SES API in the Amazon Simple Email Service Developer Guide.
Enables or disables Easy DKIM signing of email sent from an identity. If Easy DKIM signing is enabled for a domain, then Amazon SES uses DKIM to sign all email that it sends from addresses on that domain. If Easy DKIM signing is enabled for an email address, then Amazon SES uses DKIM to sign all email it sends from that address.
For email addresses (for example, user@example.com), you can only enable DKIM signing if the corresponding domain (in this case, example.com) has been set up to use Easy DKIM.
You can enable DKIM signing for an identity at any time after you start the verification process for the identity, even if the verification process isn't complete.
You can execute this operation no more than once per second.
Enables or disables Easy DKIM signing of email sent from an identity. If Easy DKIM signing is enabled for a domain, then Amazon SES uses DKIM to sign all email that it sends from addresses on that domain. If Easy DKIM signing is enabled for an email address, then Amazon SES uses DKIM to sign all email it sends from that address.
For email addresses (for example, user@example.com), you can only enable DKIM signing if the corresponding domain (in this case, example.com) has been set up to use Easy DKIM.
You can enable DKIM signing for an identity at any time after you start the verification process for the identity, even if the verification process isn't complete.
You can execute this operation no more than once per second.
Given an identity (an email address or a domain), enables or disables whether Amazon SES forwards bounce and complaint notifications as email. Feedback forwarding can only be disabled when Amazon Simple Notification Service (Amazon SNS) topics are specified for both bounces and complaints.
Feedback forwarding does not apply to delivery notifications. Delivery notifications are only available through Amazon SNS.
You can execute this operation no more than once per second.
Given an identity (an email address or a domain), enables or disables whether Amazon SES forwards bounce and complaint notifications as email. Feedback forwarding can only be disabled when Amazon Simple Notification Service (Amazon SNS) topics are specified for both bounces and complaints.
Feedback forwarding does not apply to delivery notifications. Delivery notifications are only available through Amazon SNS.
You can execute this operation no more than once per second.
Given an identity (an email address or a domain), sets whether Amazon SES includes the original email headers in the Amazon Simple Notification Service (Amazon SNS) notifications of a specified type.
You can execute this operation no more than once per second.
Given an identity (an email address or a domain), sets whether Amazon SES includes the original email headers in the Amazon Simple Notification Service (Amazon SNS) notifications of a specified type.
You can execute this operation no more than once per second.
Enables or disables the custom MAIL FROM domain setup for a verified identity (an email address or a domain).
To send emails using the specified MAIL FROM domain, you must add an MX record to your MAIL FROM domain's DNS settings. If you want your emails to pass Sender Policy Framework (SPF) checks, you must also add or update an SPF record. For more information, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
"
+ "documentation":"
Enables or disables the custom MAIL FROM domain setup for a verified identity (an email address or a domain).
To send emails using the specified MAIL FROM domain, you must add an MX record to your MAIL FROM domain's DNS settings. To ensure that your emails pass Sender Policy Framework (SPF) checks, you must also add or update an SPF record. For more information, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
Sets an Amazon Simple Notification Service (Amazon SNS) topic to use when delivering notifications. When you use this operation, you specify a verified identity, such as an email address or domain. When you send an email that uses the chosen identity in the Source field, Amazon SES sends notifications to the topic you specified. You can send bounce, complaint, or delivery notifications (or any combination of the three) to the Amazon SNS topic that you specify.
You can execute this operation no more than once per second.
Sets an Amazon Simple Notification Service (Amazon SNS) topic to use when delivering notifications. When you use this operation, you specify a verified identity, such as an email address or domain. When you send an email that uses the chosen identity in the Source field, Amazon SES sends notifications to the topic you specified. You can send bounce, complaint, or delivery notifications (or any combination of the three) to the Amazon SNS topic that you specify.
You can execute this operation no more than once per second.
Enables or disables email sending across your entire Amazon SES account in the current AWS Region. You can use this operation in conjunction with Amazon CloudWatch alarms to temporarily pause email sending across your Amazon SES account in a given AWS Region when reputation metrics (such as your bounce or complaint rates) reach certain thresholds.
You can execute this operation no more than once per second.
"
+ "documentation":"
Enables or disables email sending across your entire Amazon SES account in the current Amazon Web Services Region. You can use this operation in conjunction with Amazon CloudWatch alarms to temporarily pause email sending across your Amazon SES account in a given Amazon Web Services Region when reputation metrics (such as your bounce or complaint rates) reach certain thresholds.
You can execute this operation no more than once per second.
Updates the event destination of a configuration set. Event destinations are associated with configuration sets, which enable you to publish email sending events to Amazon CloudWatch, Amazon Kinesis Firehose, or Amazon Simple Notification Service (Amazon SNS). For information about using configuration sets, see Monitoring Your Amazon SES Sending Activity in the Amazon SES Developer Guide.
When you create or update an event destination, you must provide one, and only one, destination. The destination can be Amazon CloudWatch, Amazon Kinesis Firehose, or Amazon Simple Notification Service (Amazon SNS).
You can execute this operation no more than once per second.
"
+ "documentation":"
Updates the event destination of a configuration set. Event destinations are associated with configuration sets, which enable you to publish email sending events to Amazon CloudWatch, Amazon Kinesis Firehose, or Amazon Simple Notification Service (Amazon SNS). For information about using configuration sets, see Monitoring Your Amazon SES Sending Activity in the Amazon SES Developer Guide.
When you create or update an event destination, you must provide one, and only one, destination. The destination can be Amazon CloudWatch, Amazon Kinesis Firehose, or Amazon Simple Notification Service (Amazon SNS).
You can execute this operation no more than once per second.
Enables or disables the publishing of reputation metrics for emails sent using a specific configuration set in a given AWS Region. Reputation metrics include bounce and complaint rates. These metrics are published to Amazon CloudWatch. By using CloudWatch, you can create alarms when bounce or complaint rates exceed certain thresholds.
You can execute this operation no more than once per second.
"
+ "documentation":"
Enables or disables the publishing of reputation metrics for emails sent using a specific configuration set in a given Amazon Web Services Region. Reputation metrics include bounce and complaint rates. These metrics are published to Amazon CloudWatch. By using CloudWatch, you can create alarms when bounce or complaint rates exceed certain thresholds.
You can execute this operation no more than once per second.
Enables or disables email sending for messages sent using a specific configuration set in a given AWS Region. You can use this operation in conjunction with Amazon CloudWatch alarms to temporarily pause email sending for a configuration set when the reputation metrics for that configuration set (such as your bounce on complaint rate) exceed certain thresholds.
You can execute this operation no more than once per second.
"
+ "documentation":"
Enables or disables email sending for messages sent using a specific configuration set in a given Amazon Web Services Region. You can use this operation in conjunction with Amazon CloudWatch alarms to temporarily pause email sending for a configuration set when the reputation metrics for that configuration set (such as your bounce on complaint rate) exceed certain thresholds.
You can execute this operation no more than once per second.
Modifies an association between a configuration set and a custom domain for open and click event tracking.
By default, images and links used for tracking open and click events are hosted on domains operated by Amazon SES. You can configure a subdomain of your own to handle these events. For information about using custom domains, see the Amazon SES Developer Guide.
"
+ "documentation":"
Modifies an association between a configuration set and a custom domain for open and click event tracking.
By default, images and links used for tracking open and click events are hosted on domains operated by Amazon SES. You can configure a subdomain of your own to handle these events. For information about using custom domains, see the Amazon SES Developer Guide.
Updates an email template. Email templates enable you to send personalized email to one or more destinations in a single API operation. For more information, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
"
+ "documentation":"
Updates an email template. Email templates enable you to send personalized email to one or more destinations in a single operation. For more information, see the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
Returns a set of DKIM tokens for a domain identity.
When you execute the VerifyDomainDkim operation, the domain that you specify is added to the list of identities that are associated with your account. This is true even if you haven't already associated the domain with your account by using the VerifyDomainIdentity operation. However, you can't send email from the domain until you either successfully verify it or you successfully set up DKIM for it.
You use the tokens that are generated by this operation to create CNAME records. When Amazon SES detects that you've added these records to the DNS configuration for a domain, you can start sending email from that domain. You can start sending email even if you haven't added the TXT record provided by the VerifyDomainIdentity operation to the DNS configuration for your domain. All email that you send from the domain is authenticated using DKIM.
To create the CNAME records for DKIM authentication, use the following values:
Name: token._domainkey.example.com
Type: CNAME
Value: token.dkim.amazonses.com
In the preceding example, replace token with one of the tokens that are generated when you execute this operation. Replace example.com with your domain. Repeat this process for each token that's generated by this operation.
You can execute this operation no more than once per second.
"
+ "documentation":"
Returns a set of DKIM tokens for a domain identity.
When you execute the VerifyDomainDkim operation, the domain that you specify is added to the list of identities that are associated with your account. This is true even if you haven't already associated the domain with your account by using the VerifyDomainIdentity operation. However, you can't send email from the domain until you either successfully verify it or you successfully set up DKIM for it.
You use the tokens that are generated by this operation to create CNAME records. When Amazon SES detects that you've added these records to the DNS configuration for a domain, you can start sending email from that domain. You can start sending email even if you haven't added the TXT record provided by the VerifyDomainIdentity operation to the DNS configuration for your domain. All email that you send from the domain is authenticated using DKIM.
To create the CNAME records for DKIM authentication, use the following values:
Name: token._domainkey.example.com
Type: CNAME
Value: token.dkim.amazonses.com
In the preceding example, replace token with one of the tokens that are generated when you execute this operation. Replace example.com with your domain. Repeat this process for each token that's generated by this operation.
You can execute this operation no more than once per second.
Adds a domain to the list of identities for your Amazon SES account in the current AWS Region and attempts to verify it. For more information about verifying domains, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
"
+ "documentation":"
Adds a domain to the list of identities for your Amazon SES account in the current Amazon Web Services Region and attempts to verify it. For more information about verifying domains, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.
You can execute this operation no more than once per second.
Adds an email address to the list of identities for your Amazon SES account in the current AWS region and attempts to verify it. As a result of executing this operation, a verification email is sent to the specified address.
You can execute this operation no more than once per second.
"
+ "documentation":"
Adds an email address to the list of identities for your Amazon SES account in the current Amazon Web Services Region and attempts to verify it. As a result of executing this operation, a verification email is sent to the specified address.
You can execute this operation no more than once per second.
The name of the header to add. Must be between 1 and 50 characters, inclusive, and consist of alphanumeric (a-z, A-Z, 0-9) characters and dashes only.
"
+ "documentation":"
The name of the header to add to the incoming message. The name must contain at least one character, and can contain up to 50 characters. It consists of alphanumeric (a–z, A–Z, 0–9) characters and dashes.
Must be less than 2048 characters, and must not contain newline characters (\"\\r\" or \"\\n\").
"
+ "documentation":"
The content to include in the header. This value can contain up to 2048 characters. It can't contain newline (\\n) or carriage return (\\r) characters.
"
}
},
- "documentation":"
When included in a receipt rule, this action adds a header to the received email.
The Amazon Resource Name (ARN) of the Amazon SNS topic to notify when the bounce action is taken. An example of an Amazon SNS topic ARN is arn:aws:sns:us-west-2:123456789012:MyTopic. For more information about Amazon SNS topics, see the Amazon SNS Developer Guide.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the Amazon SNS topic to notify when the bounce action is taken. You can find the ARN of a topic by using the ListTopics operation in Amazon SNS.
The email address of the sender of the bounced email. This is the address from which the bounce message will be sent.
"
+ "documentation":"
The email address of the sender of the bounced email. This is the address from which the bounce message is sent.
"
}
},
- "documentation":"
When included in a receipt rule, this action rejects the received email by returning a bounce response to the sender and, optionally, publishes a notification to Amazon Simple Notification Service (Amazon SNS).
For information about sending a bounce message in response to a received email, see the Amazon SES Developer Guide.
"
+ "documentation":"
When included in a receipt rule, this action rejects the received email by returning a bounce response to the sender and, optionally, publishes a notification to Amazon Simple Notification Service (Amazon SNS).
For information about sending a bounce message in response to a received email, see the Amazon SES Developer Guide.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to receive email for the recipient of the bounced email. For more information about sending authorization, see the Amazon SES Developer Guide.
"
+ "documentation":"
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to receive email for the recipient of the bounced email. For more information about sending authorization, see the Amazon SES Developer Guide.
Recipient-related DSN fields, most of which would normally be filled in automatically when provided with a BounceType. You must provide either this parameter or BounceType.
"
}
},
- "documentation":"
Recipient-related information to include in the Delivery Status Notification (DSN) when an email that Amazon SES receives on your behalf bounces.
InvalidParameterValue: One or more of the parameters you specified when calling this operation was invalid. See the error message for additional information.
TransientFailure: Amazon SES was unable to process your request because of a temporary issue.
Failed: Amazon SES was unable to process your request. See the error message for additional information.
"
+ "documentation":"
The status of a message sent using the SendBulkTemplatedEmail operation.
Possible values for this parameter include:
Success: Amazon SES accepted the message, and attempts to deliver it to the recipients.
MessageRejected: The message was rejected because it contained a virus.
MailFromDomainNotVerified: The sender's email address or domain was not verified.
ConfigurationSetDoesNotExist: The configuration set you specified does not exist.
TemplateDoesNotExist: The template you specified does not exist.
AccountSuspended: Your account has been shut down because of issues related to your email sending practices.
AccountThrottled: The number of emails you can send has been reduced because your account has exceeded its allocated sending limit.
AccountDailyQuotaExceeded: You have reached or exceeded the maximum number of emails you can send from your account in a 24-hour period.
InvalidSendingPoolName: The configuration set you specified refers to an IP pool that does not exist.
AccountSendingPaused: Email sending for the Amazon SES account was disabled using the UpdateAccountSendingEnabled operation.
InvalidParameterValue: One or more of the parameters you specified when calling this operation was invalid. See the error message for additional information.
TransientFailure: Amazon SES was unable to process your request because of a temporary issue.
Failed: Amazon SES was unable to process your request. See the error message for additional information.
Represents a request to create a receipt rule set by cloning an existing one. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to create a receipt rule set by cloning an existing one. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
A list of dimensions upon which to categorize your emails when you publish email sending events to Amazon CloudWatch.
"
}
},
- "documentation":"
Contains information associated with an Amazon CloudWatch event destination to which email sending events are published.
Event destinations, such as Amazon CloudWatch, are associated with configuration sets, which enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Contains information associated with an Amazon CloudWatch event destination to which email sending events are published.
Event destinations, such as Amazon CloudWatch, are associated with configuration sets, which enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
The place where Amazon SES finds the value of a dimension to publish to Amazon CloudWatch. If you want Amazon SES to use the message tags that you specify using an X-SES-MESSAGE-TAGS header or a parameter to the SendEmail/SendRawEmail API, choose messageTag. If you want Amazon SES to use your own email headers, choose emailHeader.
"
+ "documentation":"
The place where Amazon SES finds the value of a dimension to publish to Amazon CloudWatch. To use the message tags that you specify using an X-SES-MESSAGE-TAGS header or a parameter to the SendEmail/SendRawEmail API, specify messageTag. To use your own email headers, specify emailHeader. To put a custom tag on any link included in your email, specify linkTag.
The default value of the dimension that is published to Amazon CloudWatch if you do not provide the value of the dimension when you send an email. The default value must:
This value can only contain ASCII letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-).
Contain less than 256 characters.
"
+ "documentation":"
The default value of the dimension that is published to Amazon CloudWatch if you do not provide the value of the dimension when you send an email. The default value must meet the following requirements:
Contain only ASCII letters (a-z, A-Z), numbers (0-9), underscores (_), dashes (-), at signs (@), or periods (.).
Contain 256 characters or fewer.
"
}
},
- "documentation":"
Contains the dimension configuration to use when you publish email sending events to Amazon CloudWatch.
For information about publishing email sending events to Amazon CloudWatch, see the Amazon SES Developer Guide.
"
+ "documentation":"
Contains the dimension configuration to use when you publish email sending events to Amazon CloudWatch.
For information about publishing email sending events to Amazon CloudWatch, see the Amazon SES Developer Guide.
An object that describes the AWS service that email sending event information will be published to.
"
+ "documentation":"
An object that describes the Amazon Web Services service that email sending event where information is published.
"
}
},
- "documentation":"
Represents a request to create a configuration set event destination. A configuration set event destination, which can be either Amazon CloudWatch or Amazon Kinesis Firehose, describes an AWS service in which Amazon SES publishes the email sending events associated with a configuration set. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to create a configuration set event destination. A configuration set event destination, which can be either Amazon CloudWatch or Amazon Kinesis Firehose, describes an Amazon Web Services service in which Amazon SES publishes the email sending events associated with a configuration set. For information about using configuration sets, see the Amazon SES Developer Guide.
A data structure that contains the name of the configuration set.
"
}
},
- "documentation":"
Represents a request to create a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to create a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
The content of the custom verification email. The total size of the email must be less than 10 MB. The message body may contain HTML, with some limitations. For more information, see Custom Verification Email Frequently Asked Questions in the Amazon SES Developer Guide.
"
+ "documentation":"
The content of the custom verification email. The total size of the email must be less than 10 MB. The message body may contain HTML, with some limitations. For more information, see Custom Verification Email Frequently Asked Questions in the Amazon SES Developer Guide.
A data structure that describes the IP address filter to create, which consists of a name, an IP address range, and whether to allow or block mail from it.
"
}
},
- "documentation":"
Represents a request to create a new IP address filter. You use IP address filters when you receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to create a new IP address filter. You use IP address filters when you receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
The name of an existing rule after which the new rule will be placed. If this parameter is null, the new rule will be inserted at the beginning of the rule list.
"
+ "documentation":"
The name of an existing rule after which the new rule is placed. If this parameter is null, the new rule is inserted at the beginning of the rule list.
A data structure that contains the specified rule's name, actions, recipients, domains, enabled status, scan status, and TLS policy.
"
}
},
- "documentation":"
Represents a request to create a receipt rule. You use receipt rules to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to create a receipt rule. You use receipt rules to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
The name of the rule set to create. The name must:
This value can only contain ASCII letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-).
Start and end with a letter or number.
Contain less than 64 characters.
"
+ "documentation":"
The name of the rule set to create. The name must meet the following requirements:
Contain only ASCII letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-).
Start and end with a letter or number.
Contain 64 characters or fewer.
"
}
},
- "documentation":"
Represents a request to create an empty receipt rule set. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to create an empty receipt rule set. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
Represents a request to delete a configuration set event destination. Configuration set event destinations are associated with configuration sets, which enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to delete a configuration set event destination. Configuration set event destinations are associated with configuration sets, which enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
Represents a request to delete a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to delete a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
The identity that is associated with the policy that you want to delete. You can specify the identity by using its name or by using its Amazon Resource Name (ARN). Examples: user@example.com, example.com, arn:aws:ses:us-east-1:123456789012:identity/example.com.
To successfully call this API, you must own the identity.
"
+ "documentation":"
The identity that is associated with the policy to delete. You can specify the identity by using its name or by using its Amazon Resource Name (ARN). Examples: user@example.com, example.com, arn:aws:ses:us-east-1:123456789012:identity/example.com.
To successfully call this operation, you must own the identity.
Represents a request to delete a sending authorization policy for an identity. Sending authorization is an Amazon SES feature that enables you to authorize other senders to use your identities. For information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to delete a sending authorization policy for an identity. Sending authorization is an Amazon SES feature that enables you to authorize other senders to use your identities. For information, see the Amazon SES Developer Guide.
The identity to be removed from the list of identities for the AWS Account.
"
+ "documentation":"
The identity to be removed from the list of identities for the Amazon Web Services account.
"
}
},
"documentation":"
Represents a request to delete one of your Amazon SES identities (an email address or domain).
"
@@ -1846,7 +1846,7 @@
"documentation":"
The name of the IP address filter to delete.
"
}
},
- "documentation":"
Represents a request to delete an IP address filter. You use IP address filters when you receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to delete an IP address filter. You use IP address filters when you receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
Represents a request to delete a receipt rule. You use receipt rules to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to delete a receipt rule. You use receipt rules to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
Represents a request to delete a receipt rule set and all of the receipt rules it contains. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to delete a receipt rule set and all of the receipt rules it contains. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
Represents a request to return the metadata and receipt rules for the receipt rule set that is currently active. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to return the metadata and receipt rules for the receipt rule set that is currently active. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
Represents a request to return the details of a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to return the details of a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
An object that represents the reputation settings for the configuration set.
"
}
},
- "documentation":"
Represents the details of a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents the details of a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
Represents a request to return the details of a receipt rule. You use receipt rules to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to return the details of a receipt rule. You use receipt rules to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
Represents a request to return the details of a receipt rule set. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to return the details of a receipt rule set. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
The recipients to place on the BCC: line of the message.
"
}
},
- "documentation":"
Represents the destination of the message, consisting of To:, CC:, and BCC: fields.
Amazon SES does not support the SMTPUTF8 extension, as described in RFC6531. For this reason, the local part of a destination email address (the part of the email address that precedes the @ sign) may only contain 7-bit ASCII characters. If the domain part of an address (the part after the @ sign) contains non-ASCII characters, they must be encoded using Punycode, as described in RFC3492.
"
+ "documentation":"
Represents the destination of the message, consisting of To:, CC:, and BCC: fields.
Amazon SES does not support the SMTPUTF8 extension, as described in RFC6531. For this reason, the email address string must be 7-bit ASCII. If you want to send to or from email addresses that contain Unicode characters in the domain part of an address, you must encode the domain using Punycode. Punycode is not permitted in the local part of the email address (the part before the @ sign) nor in the \"friendly from\" name. If you want to use Unicode characters in the \"friendly from\" name, you must encode the \"friendly from\" name using MIME encoded-word syntax, as described in Sending raw email using the Amazon SES API. For more information about Punycode, see RFC 3492.
An object that contains the topic ARN associated with an Amazon Simple Notification Service (Amazon SNS) event destination.
"
}
},
- "documentation":"
Contains information about the event destination that the specified email sending events will be published to.
When you create or update an event destination, you must provide one, and only one, destination. The destination can be Amazon CloudWatch, Amazon Kinesis Firehose or Amazon Simple Notification Service (Amazon SNS).
Event destinations are associated with configuration sets, which enable you to publish email sending events to Amazon CloudWatch, Amazon Kinesis Firehose, or Amazon Simple Notification Service (Amazon SNS). For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Contains information about an event destination.
When you create or update an event destination, you must provide one, and only one, destination. The destination can be Amazon CloudWatch, Amazon Kinesis Firehose or Amazon Simple Notification Service (Amazon SNS).
Event destinations are associated with configuration sets, which enable you to publish email sending events to Amazon CloudWatch, Amazon Kinesis Firehose, or Amazon Simple Notification Service (Amazon SNS). For information about using configuration sets, see the Amazon SES Developer Guide.
The name of the custom verification email template that you want to retrieve.
"
+ "documentation":"
The name of the custom verification email template to retrieve.
"
}
},
"documentation":"
Represents a request to retrieve an existing custom verification email template.
"
@@ -2289,7 +2289,7 @@
"documentation":"
A list of one or more verified identities - email addresses, domains, or both.
"
}
},
- "documentation":"
Represents a request for the status of Amazon SES Easy DKIM signing for an identity. For domain identities, this request also returns the DKIM tokens that are required for Easy DKIM signing, and whether Amazon SES successfully verified that these tokens were published. For more information about Easy DKIM, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request for the status of Amazon SES Easy DKIM signing for an identity. For domain identities, this request also returns the DKIM tokens that are required for Easy DKIM signing, and whether Amazon SES successfully verified that these tokens were published. For more information about Easy DKIM, see the Amazon SES Developer Guide.
Represents a request to return the Amazon SES custom MAIL FROM attributes for a list of identities. For information about using a custom MAIL FROM domain, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to return the Amazon SES custom MAIL FROM attributes for a list of identities. For information about using a custom MAIL FROM domain, see the Amazon SES Developer Guide.
A list of one or more identities. You can specify an identity by using its name or by using its Amazon Resource Name (ARN). Examples: user@example.com, example.com, arn:aws:ses:us-east-1:123456789012:identity/example.com.
"
}
},
- "documentation":"
Represents a request to return the notification attributes for a list of identities you verified with Amazon SES. For information about Amazon SES notifications, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to return the notification attributes for a list of identities you verified with Amazon SES. For information about Amazon SES notifications, see the Amazon SES Developer Guide.
The identity for which the policies will be retrieved. You can specify an identity by using its name or by using its Amazon Resource Name (ARN). Examples: user@example.com, example.com, arn:aws:ses:us-east-1:123456789012:identity/example.com.
To successfully call this API, you must own the identity.
"
+ "documentation":"
The identity for which the policies are retrieved. You can specify an identity by using its name or by using its Amazon Resource Name (ARN). Examples: user@example.com, example.com, arn:aws:ses:us-east-1:123456789012:identity/example.com.
To successfully call this operation, you must own the identity.
A list of the names of policies to be retrieved. You can retrieve a maximum of 20 policies at a time. If you do not know the names of the policies that are attached to the identity, you can use ListIdentityPolicies.
"
}
},
- "documentation":"
Represents a request to return the requested sending authorization policies for an identity. Sending authorization is an Amazon SES feature that enables you to authorize other senders to use your identities. For information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to return the requested sending authorization policies for an identity. Sending authorization is an Amazon SES feature that enables you to authorize other senders to use your identities. For information, see the Amazon SES Developer Guide.
Represents a request to return the Amazon SES verification status of a list of identities. For domain identities, this request also returns the verification token. For information about verifying identities with Amazon SES, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to return the Amazon SES verification status of a list of identities. For domain identities, this request also returns the verification token. For information about verifying identities with Amazon SES, see the Amazon SES Developer Guide.
A set of character strings that represent the domain's identity. Using these tokens, you need to create DNS CNAME records that point to DKIM public keys that are hosted by Amazon SES. Amazon Web Services eventually detects that you've updated your DNS records. This detection process might take up to 72 hours. After successful detection, Amazon SES is able to DKIM-sign email originating from that domain. (This only applies to domain identities, not email address identities.)
A set of character strings that represent the domain's identity. Using these tokens, you need to create DNS CNAME records that point to DKIM public keys that are hosted by Amazon SES. Amazon Web Services eventually detects that you've updated your DNS records. This detection process might take up to 72 hours. After successful detection, Amazon SES is able to DKIM-sign email originating from that domain. (This only applies to domain identities, not email address identities.)
Describes whether Amazon SES will forward bounce and complaint notifications as email. true indicates that Amazon SES will forward bounce and complaint notifications as email, while false indicates that bounce and complaint notifications will be published only to the specified bounce and complaint Amazon SNS topics.
"
+ "documentation":"
Describes whether Amazon SES forwards bounce and complaint notifications as email. true indicates that Amazon SES forwards bounce and complaint notifications as email, while false indicates that bounce and complaint notifications are published only to the specified bounce and complaint Amazon SNS topics.
Describes whether Amazon SES includes the original email headers in Amazon SNS notifications of type Bounce. A value of true specifies that Amazon SES will include headers in bounce notifications, and a value of false specifies that Amazon SES will not include headers in bounce notifications.
"
+ "documentation":"
Describes whether Amazon SES includes the original email headers in Amazon SNS notifications of type Bounce. A value of true specifies that Amazon SES includes headers in bounce notifications, and a value of false specifies that Amazon SES does not include headers in bounce notifications.
Describes whether Amazon SES includes the original email headers in Amazon SNS notifications of type Complaint. A value of true specifies that Amazon SES will include headers in complaint notifications, and a value of false specifies that Amazon SES will not include headers in complaint notifications.
"
+ "documentation":"
Describes whether Amazon SES includes the original email headers in Amazon SNS notifications of type Complaint. A value of true specifies that Amazon SES includes headers in complaint notifications, and a value of false specifies that Amazon SES does not include headers in complaint notifications.
Describes whether Amazon SES includes the original email headers in Amazon SNS notifications of type Delivery. A value of true specifies that Amazon SES will include headers in delivery notifications, and a value of false specifies that Amazon SES will not include headers in delivery notifications.
"
+ "documentation":"
Describes whether Amazon SES includes the original email headers in Amazon SNS notifications of type Delivery. A value of true specifies that Amazon SES includes headers in delivery notifications, and a value of false specifies that Amazon SES does not include headers in delivery notifications.
"
}
},
"documentation":"
Represents the notification attributes of an identity, including whether an identity has Amazon Simple Notification Service (Amazon SNS) topics set for bounce, complaint, and/or delivery notifications, and whether feedback forwarding is enabled for bounce and complaint notifications.
"
@@ -2628,7 +2628,7 @@
"documentation":"
Indicates that the ARN of the function was not found.
"
}
},
- "documentation":"
Indicates that the provided AWS Lambda function is invalid, or that Amazon SES could not execute the provided function, possibly due to permissions issues. For information about giving permissions, see the Amazon SES Developer Guide.
",
+ "documentation":"
Indicates that the provided Amazon Web Services Lambda function is invalid, or that Amazon SES could not execute the provided function, possibly due to permissions issues. For information about giving permissions, see the Amazon SES Developer Guide.
Indicates that the provided Amazon S3 bucket or AWS KMS encryption key is invalid, or that Amazon SES could not publish to the bucket, possibly due to permissions issues. For information about giving permissions, see the Amazon SES Developer Guide.
",
+ "documentation":"
Indicates that the provided Amazon S3 bucket or Amazon Web Services KMS encryption key is invalid, or that Amazon SES could not publish to the bucket, possibly due to permissions issues. For information about giving permissions, see the Amazon SES Developer Guide.
The ARN of the Amazon Kinesis Firehose stream that email sending events should be published to.
"
}
},
- "documentation":"
Contains the delivery stream ARN and the IAM role ARN associated with an Amazon Kinesis Firehose event destination.
Event destinations, such as Amazon Kinesis Firehose, are associated with configuration sets, which enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Contains the delivery stream ARN and the IAM role ARN associated with an Amazon Kinesis Firehose event destination.
Event destinations, such as Amazon Kinesis Firehose, are associated with configuration sets, which enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
The Amazon Resource Name (ARN) of the Amazon SNS topic to notify when the Lambda action is taken. An example of an Amazon SNS topic ARN is arn:aws:sns:us-west-2:123456789012:MyTopic. For more information about Amazon SNS topics, see the Amazon SNS Developer Guide.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the Amazon SNS topic to notify when the Lambda action is executed. You can find the ARN of a topic by using the ListTopics operation in Amazon SNS.
The Amazon Resource Name (ARN) of the AWS Lambda function. An example of an AWS Lambda function ARN is arn:aws:lambda:us-west-2:account-id:function:MyFunction. For more information about AWS Lambda, see the AWS Lambda Developer Guide.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the Amazon Web Services Lambda function. An example of an Amazon Web Services Lambda function ARN is arn:aws:lambda:us-west-2:account-id:function:MyFunction. For more information about Amazon Web Services Lambda, see the Amazon Web Services Lambda Developer Guide.
The invocation type of the AWS Lambda function. An invocation type of RequestResponse means that the execution of the function will immediately result in a response, and a value of Event means that the function will be invoked asynchronously. The default value is Event. For information about AWS Lambda invocation types, see the AWS Lambda Developer Guide.
There is a 30-second timeout on RequestResponse invocations. You should use Event invocation in most cases. Use RequestResponse only when you want to make a mail flow decision, such as whether to stop the receipt rule or the receipt rule set.
"
+ "documentation":"
The invocation type of the Amazon Web Services Lambda function. An invocation type of RequestResponse means that the execution of the function immediately results in a response, and a value of Event means that the function is invoked asynchronously. The default value is Event. For information about Amazon Web Services Lambda invocation types, see the Amazon Web Services Lambda Developer Guide.
There is a 30-second timeout on RequestResponse invocations. You should use Event invocation in most cases. Use RequestResponse only to make a mail flow decision, such as whether to stop the receipt rule or the receipt rule set.
"
}
},
- "documentation":"
When included in a receipt rule, this action calls an AWS Lambda function and, optionally, publishes a notification to Amazon Simple Notification Service (Amazon SNS).
To enable Amazon SES to call your AWS Lambda function or to publish to an Amazon SNS topic of another account, Amazon SES must have permission to access those resources. For information about giving permissions, see the Amazon SES Developer Guide.
When included in a receipt rule, this action calls an Amazon Web Services Lambda function and, optionally, publishes a notification to Amazon Simple Notification Service (Amazon SNS).
To enable Amazon SES to call your Amazon Web Services Lambda function or to publish to an Amazon SNS topic of another account, Amazon SES must have permission to access those resources. For information about giving permissions, see the Amazon SES Developer Guide.
For information about using Amazon Web Services Lambda actions in receipt rules, see the Amazon SES Developer Guide.
Represents a request to list the configuration sets associated with your AWS account. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to list the configuration sets associated with your Amazon Web Services account. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
A token indicating that there are additional configuration sets available to be listed. Pass this token to successive calls of ListConfigurationSets.
"
}
},
- "documentation":"
A list of configuration sets associated with your AWS account. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
A list of configuration sets associated with your Amazon Web Services account. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
The maximum number of custom verification email templates to return. This value must be at least 1 and less than or equal to 50. If you do not specify a value, or if you specify a value less than 1 or greater than 50, the operation will return up to 50 results.
"
+ "documentation":"
The maximum number of custom verification email templates to return. This value must be at least 1 and less than or equal to 50. If you do not specify a value, or if you specify a value less than 1 or greater than 50, the operation returns up to 50 results.
"
}
},
- "documentation":"
Represents a request to list the existing custom verification email templates for your account.
The type of the identities to list. Possible values are \"EmailAddress\" and \"Domain\". If this parameter is omitted, then all identities will be listed.
"
+ "documentation":"
The type of the identities to list. Possible values are \"EmailAddress\" and \"Domain\". If this parameter is omitted, then all identities are listed.
The maximum number of identities per page. Possible values are 1-1000 inclusive.
"
}
},
- "documentation":"
Represents a request to return a list of all identities (email addresses and domains) that you have attempted to verify under your AWS account, regardless of verification status.
"
+ "documentation":"
Represents a request to return a list of all identities (email addresses and domains) that you have attempted to verify under your Amazon Web Services account, regardless of verification status.
The identity that is associated with the policy for which the policies will be listed. You can specify an identity by using its name or by using its Amazon Resource Name (ARN). Examples: user@example.com, example.com, arn:aws:ses:us-east-1:123456789012:identity/example.com.
To successfully call this API, you must own the identity.
"
+ "documentation":"
The identity that is associated with the policy for which the policies are listed. You can specify an identity by using its name or by using its Amazon Resource Name (ARN). Examples: user@example.com, example.com, arn:aws:ses:us-east-1:123456789012:identity/example.com.
To successfully call this operation, you must own the identity.
"
}
},
- "documentation":"
Represents a request to return a list of sending authorization policies that are attached to an identity. Sending authorization is an Amazon SES feature that enables you to authorize other senders to use your identities. For information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to return a list of sending authorization policies that are attached to an identity. Sending authorization is an Amazon SES feature that enables you to authorize other senders to use your identities. For information, see the Amazon SES Developer Guide.
Represents a request to list the IP address filters that exist under your AWS account. You use IP address filters when you receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to list the IP address filters that exist under your Amazon Web Services account. You use IP address filters when you receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
A token returned from a previous call to ListReceiptRuleSets to indicate the position in the receipt rule set list.
"
}
},
- "documentation":"
Represents a request to list the receipt rule sets that exist under your AWS account. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to list the receipt rule sets that exist under your Amazon Web Services account. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
A token indicating that there are additional receipt rule sets available to be listed. Pass this token to successive calls of ListReceiptRuleSets to retrieve up to 100 receipt rule sets at a time.
"
}
},
- "documentation":"
A list of receipt rule sets that exist under your AWS account.
"
+ "documentation":"
A list of receipt rule sets that exist under your Amazon Web Services account.
The maximum number of templates to return. This value must be at least 1 and less than or equal to 10. If you do not specify a value, or if you specify a value less than 1 or greater than 10, the operation will return up to 10 results.
"
+ "documentation":"
The maximum number of templates to return. This value must be at least 1 and less than or equal to 100. If more than 100 items are requested, the page size will automatically set to 100. If you do not specify a value, 10 is the default page size.
A token indicating that there are additional email templates available to be listed. Pass this token to a subsequent call to ListTemplates to retrieve the next 50 email templates.
"
+ "documentation":"
A token indicating that there are additional email templates available to be listed. Pass this token to a subsequent call to ListTemplates to retrieve the next set of email templates within your page size.
"
}
}
},
@@ -2981,7 +2981,7 @@
"documentation":"
A list of email addresses that have been verified.
"
}
},
- "documentation":"
A list of email addresses that you have verified with Amazon SES under your AWS account.
"
+ "documentation":"
A list of email addresses that you have verified with Amazon SES under your Amazon Web Services account.
This value can only contain ASCII letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-).
Contain less than 256 characters.
"
+ "documentation":"
The value of the tag. The value must meet the following requirements:
Contain only ASCII letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-).
Contain 256 characters or fewer.
"
}
},
- "documentation":"
Contains the name and value of a tag that you can provide to SendEmail or SendRawEmail to apply to an email.
Message tags, which you use with configuration sets, enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Contains the name and value of a tag that you can provide to SendEmail or SendRawEmail to apply to an email.
Message tags, which you use with configuration sets, enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
The identity that the policy will apply to. You can specify an identity by using its name or by using its Amazon Resource Name (ARN). Examples: user@example.com, example.com, arn:aws:ses:us-east-1:123456789012:identity/example.com.
To successfully call this API, you must own the identity.
"
+ "documentation":"
The identity to which that the policy applies. You can specify an identity by using its name or by using its Amazon Resource Name (ARN). Examples: user@example.com, example.com, arn:aws:ses:us-east-1:123456789012:identity/example.com.
To successfully call this operation, you must own the identity.
Represents a request to add or update a sending authorization policy for an identity. Sending authorization is an Amazon SES feature that enables you to authorize other senders to use your identities. For information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to add or update a sending authorization policy for an identity. Sending authorization is an Amazon SES feature that enables you to authorize other senders to use your identities. For information, see the Amazon SES Developer Guide.
The raw data of the message. This data needs to base64-encoded if you are accessing Amazon SES directly through the HTTPS interface. If you are accessing Amazon SES using an AWS SDK, the SDK takes care of the base 64-encoding for you. In all cases, the client must ensure that the message format complies with Internet email standards regarding email header fields, MIME types, and MIME encoding.
The To:, CC:, and BCC: headers in the raw message can contain a group list.
If you are using SendRawEmail with sending authorization, you can include X-headers in the raw message to specify the \"Source,\" \"From,\" and \"Return-Path\" addresses. For more information, see the documentation for SendRawEmail.
Do not include these X-headers in the DKIM signature, because they are removed by Amazon SES before sending the email.
The raw data of the message. This data needs to base64-encoded if you are accessing Amazon SES directly through the HTTPS interface. If you are accessing Amazon SES using an Amazon Web Services SDK, the SDK takes care of the base 64-encoding for you. In all cases, the client must ensure that the message format complies with Internet email standards regarding email header fields, MIME types, and MIME encoding.
The To:, CC:, and BCC: headers in the raw message can contain a group list.
If you are using SendRawEmail with sending authorization, you can include X-headers in the raw message to specify the \"Source,\" \"From,\" and \"Return-Path\" addresses. For more information, see the documentation for SendRawEmail.
Do not include these X-headers in the DKIM signature, because they are removed by Amazon SES before sending the email.
Publishes the email content within a notification to Amazon SNS.
"
}
},
- "documentation":"
An action that Amazon SES can take when it receives an email on behalf of one or more email addresses or domains that you own. An instance of this data type can represent only one action.
An action that Amazon SES can take when it receives an email on behalf of one or more email addresses or domains that you own. An instance of this data type can represent only one action.
A single IP address or a range of IP addresses that you want to block or allow, specified in Classless Inter-Domain Routing (CIDR) notation. An example of a single email address is 10.0.0.1. An example of a range of IP addresses is 10.0.0.1/24. For more information about CIDR notation, see RFC 2317.
"
+ "documentation":"
A single IP address or a range of IP addresses to block or allow, specified in Classless Inter-Domain Routing (CIDR) notation. An example of a single email address is 10.0.0.1. An example of a range of IP addresses is 10.0.0.1/24. For more information about CIDR notation, see RFC 2317.
"
}
},
- "documentation":"
A receipt IP address filter enables you to specify whether to accept or reject mail originating from an IP address or range of IP addresses.
Specifies whether Amazon SES should require that incoming email is delivered over a connection encrypted with Transport Layer Security (TLS). If this parameter is set to Require, Amazon SES will bounce emails that are not received over TLS. The default is Optional.
"
+ "documentation":"
Specifies whether Amazon SES should require that incoming email is delivered over a connection encrypted with Transport Layer Security (TLS). If this parameter is set to Require, Amazon SES bounces emails that are not received over TLS. The default is Optional.
The recipient domains and email addresses that the receipt rule applies to. If this field is not specified, this rule will match all recipients under all verified domains.
"
+ "documentation":"
The recipient domains and email addresses that the receipt rule applies to. If this field is not specified, this rule matches all recipients on all verified domains.
If true, then messages that this receipt rule applies to are scanned for spam and viruses. The default value is false.
"
}
},
- "documentation":"
Receipt rules enable you to specify which actions Amazon SES should take when it receives mail on behalf of one or more email addresses or domains that you own.
Each receipt rule defines a set of email addresses or domains that it applies to. If the email addresses or domains match at least one recipient address of the message, Amazon SES executes all of the receipt rule's actions on the message.
Receipt rules enable you to specify which actions Amazon SES should take when it receives mail on behalf of one or more email addresses or domains that you own.
Each receipt rule defines a set of email addresses or domains that it applies to. If the email addresses or domains match at least one recipient address of the message, Amazon SES executes all of the receipt rule's actions on the message.
The email address that the message was ultimately delivered to. This corresponds to the Final-Recipient in the DSN. If not specified, FinalRecipient will be set to the Recipient specified in the BouncedRecipientInfo structure. Either FinalRecipient or the recipient in BouncedRecipientInfo must be a recipient of the original bounced message.
Do not prepend the FinalRecipient email address with rfc 822;, as described in RFC 3798.
"
+ "documentation":"
The email address that the message was ultimately delivered to. This corresponds to the Final-Recipient in the DSN. If not specified, FinalRecipient is set to the Recipient specified in the BouncedRecipientInfo structure. Either FinalRecipient or the recipient in BouncedRecipientInfo must be a recipient of the original bounced message.
Do not prepend the FinalRecipient email address with rfc 822;, as described in RFC 3798.
A list of the specified receipt rule set's receipt rules in the order that you want to put them.
"
+ "documentation":"
The specified receipt rule set's receipt rules, in order.
"
}
},
- "documentation":"
Represents a request to reorder the receipt rules within a receipt rule set. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to reorder the receipt rules within a receipt rule set. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
Describes whether email sending is enabled or disabled for the configuration set. If the value is true, then Amazon SES will send emails that use the configuration set. If the value is false, Amazon SES will not send emails that use the configuration set. The default value is true. You can change this setting using UpdateConfigurationSetSendingEnabled.
"
+ "documentation":"
Describes whether email sending is enabled or disabled for the configuration set. If the value is true, then Amazon SES sends emails that use the configuration set. If the value is false, Amazon SES does not send emails that use the configuration set. The default value is true. You can change this setting using UpdateConfigurationSetSendingEnabled.
The ARN of the Amazon SNS topic to notify when the message is saved to the Amazon S3 bucket. An example of an Amazon SNS topic ARN is arn:aws:sns:us-west-2:123456789012:MyTopic. For more information about Amazon SNS topics, see the Amazon SNS Developer Guide.
"
+ "documentation":"
The ARN of the Amazon SNS topic to notify when the message is saved to the Amazon S3 bucket. You can find the ARN of a topic by using the ListTopics operation in Amazon SNS.
The customer master key that Amazon SES should use to encrypt your emails before saving them to the Amazon S3 bucket. You can use the default master key or a custom master key you created in AWS KMS as follows:
To use the default master key, provide an ARN in the form of arn:aws:kms:REGION:ACCOUNT-ID-WITHOUT-HYPHENS:alias/aws/ses. For example, if your AWS account ID is 123456789012 and you want to use the default master key in the US West (Oregon) region, the ARN of the default master key would be arn:aws:kms:us-west-2:123456789012:alias/aws/ses. If you use the default master key, you don't need to perform any extra steps to give Amazon SES permission to use the key.
To use a custom master key you created in AWS KMS, provide the ARN of the master key and ensure that you add a statement to your key's policy to give Amazon SES permission to use it. For more information about giving permissions, see the Amazon SES Developer Guide.
For more information about key policies, see the AWS KMS Developer Guide. If you do not specify a master key, Amazon SES will not encrypt your emails.
Your mail is encrypted by Amazon SES using the Amazon S3 encryption client before the mail is submitted to Amazon S3 for storage. It is not encrypted using Amazon S3 server-side encryption. This means that you must use the Amazon S3 encryption client to decrypt the email after retrieving it from Amazon S3, as the service has no access to use your AWS KMS keys for decryption. This encryption client is currently available with the AWS SDK for Java and AWS SDK for Ruby only. For more information about client-side encryption using AWS KMS master keys, see the Amazon S3 Developer Guide.
"
+ "documentation":"
The customer master key that Amazon SES should use to encrypt your emails before saving them to the Amazon S3 bucket. You can use the default master key or a custom master key that you created in Amazon Web Services KMS as follows:
To use the default master key, provide an ARN in the form of arn:aws:kms:REGION:ACCOUNT-ID-WITHOUT-HYPHENS:alias/aws/ses. For example, if your Amazon Web Services account ID is 123456789012 and you want to use the default master key in the US West (Oregon) Region, the ARN of the default master key would be arn:aws:kms:us-west-2:123456789012:alias/aws/ses. If you use the default master key, you don't need to perform any extra steps to give Amazon SES permission to use the key.
To use a custom master key that you created in Amazon Web Services KMS, provide the ARN of the master key and ensure that you add a statement to your key's policy to give Amazon SES permission to use it. For more information about giving permissions, see the Amazon SES Developer Guide.
For more information about key policies, see the Amazon Web Services KMS Developer Guide. If you do not specify a master key, Amazon SES does not encrypt your emails.
Your mail is encrypted by Amazon SES using the Amazon S3 encryption client before the mail is submitted to Amazon S3 for storage. It is not encrypted using Amazon S3 server-side encryption. This means that you must use the Amazon S3 encryption client to decrypt the email after retrieving it from Amazon S3, as the service has no access to use your Amazon Web Services KMS keys for decryption. This encryption client is currently available with the Amazon Web Services SDK for Java and Amazon Web Services SDK for Ruby only. For more information about client-side encryption using Amazon Web Services KMS master keys, see the Amazon S3 Developer Guide.
"
}
},
- "documentation":"
When included in a receipt rule, this action saves the received message to an Amazon Simple Storage Service (Amazon S3) bucket and, optionally, publishes a notification to Amazon Simple Notification Service (Amazon SNS).
To enable Amazon SES to write emails to your Amazon S3 bucket, use an AWS KMS key to encrypt your emails, or publish to an Amazon SNS topic of another account, Amazon SES must have permission to access those resources. For information about giving permissions, see the Amazon SES Developer Guide.
When you save your emails to an Amazon S3 bucket, the maximum email size (including headers) is 30 MB. Emails larger than that will bounce.
For information about specifying Amazon S3 actions in receipt rules, see the Amazon SES Developer Guide.
"
+ "documentation":"
When included in a receipt rule, this action saves the received message to an Amazon Simple Storage Service (Amazon S3) bucket and, optionally, publishes a notification to Amazon Simple Notification Service (Amazon SNS).
To enable Amazon SES to write emails to your Amazon S3 bucket, use an Amazon Web Services KMS key to encrypt your emails, or publish to an Amazon SNS topic of another account, Amazon SES must have permission to access those resources. For information about granting permissions, see the Amazon SES Developer Guide.
When you save your emails to an Amazon S3 bucket, the maximum email size (including headers) is 40 MB. Emails larger than that bounces.
For information about specifying Amazon S3 actions in receipt rules, see the Amazon SES Developer Guide.
The Amazon Resource Name (ARN) of the Amazon SNS topic to notify. An example of an Amazon SNS topic ARN is arn:aws:sns:us-west-2:123456789012:MyTopic. For more information about Amazon SNS topics, see the Amazon SNS Developer Guide.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the Amazon SNS topic to notify. You can find the ARN of a topic by using the ListTopics operation in Amazon SNS.
The encoding to use for the email within the Amazon SNS notification. UTF-8 is easier to use, but may not preserve all special characters when a message was encoded with a different encoding format. Base64 preserves all special characters. The default value is UTF-8.
"
}
},
- "documentation":"
When included in a receipt rule, this action publishes a notification to Amazon Simple Notification Service (Amazon SNS). This action includes a complete copy of the email content in the Amazon SNS notifications. Amazon SNS notifications for all other actions simply provide information about the email. They do not include the email content itself.
If you own the Amazon SNS topic, you don't need to do anything to give Amazon SES permission to publish emails to it. However, if you don't own the Amazon SNS topic, you need to attach a policy to the topic to give Amazon SES permissions to access it. For information about giving permissions, see the Amazon SES Developer Guide.
You can only publish emails that are 150 KB or less (including the header) to Amazon SNS. Larger emails will bounce. If you anticipate emails larger than 150 KB, use the S3 action instead.
For information about using a receipt rule to publish an Amazon SNS notification, see the Amazon SES Developer Guide.
"
+ "documentation":"
When included in a receipt rule, this action publishes a notification to Amazon Simple Notification Service (Amazon SNS). This action includes a complete copy of the email content in the Amazon SNS notifications. Amazon SNS notifications for all other actions simply provide information about the email. They do not include the email content itself.
If you own the Amazon SNS topic, you don't need to do anything to give Amazon SES permission to publish emails to it. However, if you don't own the Amazon SNS topic, you need to attach a policy to the topic to give Amazon SES permissions to access it. For information about giving permissions, see the Amazon SES Developer Guide.
You can only publish emails that are 150 KB or less (including the header) to Amazon SNS. Larger emails bounce. If you anticipate emails larger than 150 KB, use the S3 action instead.
For information about using a receipt rule to publish an Amazon SNS notification, see the Amazon SES Developer Guide.
The ARN of the Amazon SNS topic that email sending events will be published to. An example of an Amazon SNS topic ARN is arn:aws:sns:us-west-2:123456789012:MyTopic. For more information about Amazon SNS topics, see the Amazon SNS Developer Guide.
"
+ "documentation":"
The ARN of the Amazon SNS topic for email sending events. You can find the ARN of a topic by using the ListTopics Amazon SNS operation.
Contains the topic ARN associated with an Amazon Simple Notification Service (Amazon SNS) event destination.
Event destinations, such as Amazon SNS, are associated with configuration sets, which enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Contains the topic ARN associated with an Amazon Simple Notification Service (Amazon SNS) event destination.
Event destinations, such as Amazon SNS, are associated with configuration sets, which enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
Human-readable text for the bounce message to explain the failure. If not specified, the text will be auto-generated based on the bounced recipient information.
"
+ "documentation":"
Human-readable text for the bounce message to explain the failure. If not specified, the text is auto-generated based on the bounced recipient information.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to use the address in the \"From\" header of the bounce. For more information about sending authorization, see the Amazon SES Developer Guide.
"
+ "documentation":"
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to use the address in the \"From\" header of the bounce. For more information about sending authorization, see the Amazon SES Developer Guide.
"
}
},
"documentation":"
Represents a request to send a bounce message to the sender of an email you received through Amazon SES.
The email address that is sending the email. This email address must be either individually verified with Amazon SES, or from a domain that has been verified with Amazon SES. For information about verifying identities, see the Amazon SES Developer Guide.
If you are sending on behalf of another user and have been permitted to do so by a sending authorization policy, then you must also specify the SourceArn parameter. For more information about sending authorization, see the Amazon SES Developer Guide.
Amazon SES does not support the SMTPUTF8 extension, as described in RFC6531. For this reason, the local part of a source email address (the part of the email address that precedes the @ sign) may only contain 7-bit ASCII characters. If the domain part of an address (the part after the @ sign) contains non-ASCII characters, they must be encoded using Punycode, as described in RFC3492. The sender name (also known as the friendly name) may contain non-ASCII characters. These characters must be encoded using MIME encoded-word syntax, as described in RFC 2047. MIME encoded-word syntax uses the following form: =?charset?encoding?encoded-text?=.
"
+ "documentation":"
The email address that is sending the email. This email address must be either individually verified with Amazon SES, or from a domain that has been verified with Amazon SES. For information about verifying identities, see the Amazon SES Developer Guide.
If you are sending on behalf of another user and have been permitted to do so by a sending authorization policy, then you must also specify the SourceArn parameter. For more information about sending authorization, see the Amazon SES Developer Guide.
Amazon SES does not support the SMTPUTF8 extension, as described in RFC6531. For this reason, the email address string must be 7-bit ASCII. If you want to send to or from email addresses that contain Unicode characters in the domain part of an address, you must encode the domain using Punycode. Punycode is not permitted in the local part of the email address (the part before the @ sign) nor in the \"friendly from\" name. If you want to use Unicode characters in the \"friendly from\" name, you must encode the \"friendly from\" name using MIME encoded-word syntax, as described in Sending raw email using the Amazon SES API. For more information about Punycode, see RFC 3492.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to send for the email address specified in the Source parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to send from user@example.com, then you would specify the SourceArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the Source to be user@example.com.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to send for the email address specified in the Source parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to send from user@example.com, then you would specify the SourceArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the Source to be user@example.com.
The email address that bounces and complaints will be forwarded to when feedback forwarding is enabled. If the message cannot be delivered to the recipient, then an error message will be returned from the recipient's ISP; this message will then be forwarded to the email address specified by the ReturnPath parameter. The ReturnPath parameter is never overwritten. This email address must be either individually verified with Amazon SES, or from a domain that has been verified with Amazon SES.
"
+ "documentation":"
The email address that bounces and complaints are forwarded to when feedback forwarding is enabled. If the message cannot be delivered to the recipient, then an error message is returned from the recipient's ISP; this message is forwarded to the email address specified by the ReturnPath parameter. The ReturnPath parameter is never overwritten. This email address must be either individually verified with Amazon SES, or from a domain that has been verified with Amazon SES.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to use the email address specified in the ReturnPath parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to use feedback@example.com, then you would specify the ReturnPathArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the ReturnPath to be feedback@example.com.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to use the email address specified in the ReturnPath parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to use feedback@example.com, then you would specify the ReturnPathArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the ReturnPath to be feedback@example.com.
One or more Destination objects. All of the recipients in a Destination will receive the same version of the email. You can specify up to 50 Destination objects within a Destinations array.
"
+ "documentation":"
One or more Destination objects. All of the recipients in a Destination receive the same version of the email. You can specify up to 50 Destination objects within a Destinations array.
"
}
},
- "documentation":"
Represents a request to send a templated email to multiple destinations using Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to send a templated email to multiple destinations using Amazon SES. For more information, see the Amazon SES Developer Guide.
The unique message identifier returned from the SendBulkTemplatedEmail action.
"
+ "documentation":"
One object per intended recipient. Check each response object and retry any messages with a failure status. (Note that order of responses will be respective to order of destinations in the request.)Receipt rules enable you to specify which actions
The email address that is sending the email. This email address must be either individually verified with Amazon SES, or from a domain that has been verified with Amazon SES. For information about verifying identities, see the Amazon SES Developer Guide.
If you are sending on behalf of another user and have been permitted to do so by a sending authorization policy, then you must also specify the SourceArn parameter. For more information about sending authorization, see the Amazon SES Developer Guide.
Amazon SES does not support the SMTPUTF8 extension, as described in RFC6531. For this reason, the local part of a source email address (the part of the email address that precedes the @ sign) may only contain 7-bit ASCII characters. If the domain part of an address (the part after the @ sign) contains non-ASCII characters, they must be encoded using Punycode, as described in RFC3492. The sender name (also known as the friendly name) may contain non-ASCII characters. These characters must be encoded using MIME encoded-word syntax, as described in RFC 2047. MIME encoded-word syntax uses the following form: =?charset?encoding?encoded-text?=.
"
+ "documentation":"
The email address that is sending the email. This email address must be either individually verified with Amazon SES, or from a domain that has been verified with Amazon SES. For information about verifying identities, see the Amazon SES Developer Guide.
If you are sending on behalf of another user and have been permitted to do so by a sending authorization policy, then you must also specify the SourceArn parameter. For more information about sending authorization, see the Amazon SES Developer Guide.
Amazon SES does not support the SMTPUTF8 extension, as described in RFC6531. For this reason, the email address string must be 7-bit ASCII. If you want to send to or from email addresses that contain Unicode characters in the domain part of an address, you must encode the domain using Punycode. Punycode is not permitted in the local part of the email address (the part before the @ sign) nor in the \"friendly from\" name. If you want to use Unicode characters in the \"friendly from\" name, you must encode the \"friendly from\" name using MIME encoded-word syntax, as described in Sending raw email using the Amazon SES API. For more information about Punycode, see RFC 3492.
The email address that bounces and complaints will be forwarded to when feedback forwarding is enabled. If the message cannot be delivered to the recipient, then an error message will be returned from the recipient's ISP; this message will then be forwarded to the email address specified by the ReturnPath parameter. The ReturnPath parameter is never overwritten. This email address must be either individually verified with Amazon SES, or from a domain that has been verified with Amazon SES.
"
+ "documentation":"
The email address that bounces and complaints are forwarded to when feedback forwarding is enabled. If the message cannot be delivered to the recipient, then an error message is returned from the recipient's ISP; this message is forwarded to the email address specified by the ReturnPath parameter. The ReturnPath parameter is never overwritten. This email address must be either individually verified with Amazon SES, or from a domain that has been verified with Amazon SES.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to send for the email address specified in the Source parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to send from user@example.com, then you would specify the SourceArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the Source to be user@example.com.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to send for the email address specified in the Source parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to send from user@example.com, then you would specify the SourceArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the Source to be user@example.com.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to use the email address specified in the ReturnPath parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to use feedback@example.com, then you would specify the ReturnPathArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the ReturnPath to be feedback@example.com.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to use the email address specified in the ReturnPath parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to use feedback@example.com, then you would specify the ReturnPathArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the ReturnPath to be feedback@example.com.
The identity's email address. If you do not provide a value for this parameter, you must specify a \"From\" address in the raw text of the message. (You can also specify both.)
Amazon SES does not support the SMTPUTF8 extension, as described inRFC6531. For this reason, the local part of a source email address (the part of the email address that precedes the @ sign) may only contain 7-bit ASCII characters. If the domain part of an address (the part after the @ sign) contains non-ASCII characters, they must be encoded using Punycode, as described in RFC3492. The sender name (also known as the friendly name) may contain non-ASCII characters. These characters must be encoded using MIME encoded-word syntax, as described in RFC 2047. MIME encoded-word syntax uses the following form: =?charset?encoding?encoded-text?=.
If you specify the Source parameter and have feedback forwarding enabled, then bounces and complaints will be sent to this email address. This takes precedence over any Return-Path header that you might include in the raw text of the message.
"
+ "documentation":"
The identity's email address. If you do not provide a value for this parameter, you must specify a \"From\" address in the raw text of the message. (You can also specify both.)
Amazon SES does not support the SMTPUTF8 extension, as described inRFC6531. For this reason, the email address string must be 7-bit ASCII. If you want to send to or from email addresses that contain Unicode characters in the domain part of an address, you must encode the domain using Punycode. Punycode is not permitted in the local part of the email address (the part before the @ sign) nor in the \"friendly from\" name. If you want to use Unicode characters in the \"friendly from\" name, you must encode the \"friendly from\" name using MIME encoded-word syntax, as described in Sending raw email using the Amazon SES API. For more information about Punycode, see RFC 3492.
If you specify the Source parameter and have feedback forwarding enabled, then bounces and complaints are sent to this email address. This takes precedence over any Return-Path header that you might include in the raw text of the message.
The raw email message itself. The message has to meet the following criteria:
The message has to contain a header and a body, separated by a blank line.
All of the required header fields must be present in the message.
Each part of a multipart MIME message must be formatted properly.
Attachments must be of a content type that Amazon SES supports. For a list on unsupported content types, see Unsupported Attachment Types in the Amazon SES Developer Guide.
The entire message must be base64-encoded.
If any of the MIME parts in your message contain content that is outside of the 7-bit ASCII character range, we highly recommend that you encode that content. For more information, see Sending Raw Email in the Amazon SES Developer Guide.
Per RFC 5321, the maximum length of each line of text, including the <CRLF>, must not exceed 1,000 characters.
"
+ "documentation":"
The raw email message itself. The message has to meet the following criteria:
The message has to contain a header and a body, separated by a blank line.
All of the required header fields must be present in the message.
Each part of a multipart MIME message must be formatted properly.
Attachments must be of a content type that Amazon SES supports. For a list on unsupported content types, see Unsupported Attachment Types in the Amazon SES Developer Guide.
The entire message must be base64-encoded.
If any of the MIME parts in your message contain content that is outside of the 7-bit ASCII character range, we highly recommend that you encode that content. For more information, see Sending Raw Email in the Amazon SES Developer Guide.
Per RFC 5321, the maximum length of each line of text, including the <CRLF>, must not exceed 1,000 characters.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to specify a particular \"From\" address in the header of the raw email.
Instead of using this parameter, you can use the X-header X-SES-FROM-ARN in the raw message of the email. If you use both the FromArn parameter and the corresponding X-header, Amazon SES uses the value of the FromArn parameter.
For information about when to use this parameter, see the description of SendRawEmail in this guide, or see the Amazon SES Developer Guide.
"
+ "documentation":"
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to specify a particular \"From\" address in the header of the raw email.
Instead of using this parameter, you can use the X-header X-SES-FROM-ARN in the raw message of the email. If you use both the FromArn parameter and the corresponding X-header, Amazon SES uses the value of the FromArn parameter.
For information about when to use this parameter, see the description of SendRawEmail in this guide, or see the Amazon SES Developer Guide.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to send for the email address specified in the Source parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to send from user@example.com, then you would specify the SourceArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the Source to be user@example.com.
Instead of using this parameter, you can use the X-header X-SES-SOURCE-ARN in the raw message of the email. If you use both the SourceArn parameter and the corresponding X-header, Amazon SES uses the value of the SourceArn parameter.
For information about when to use this parameter, see the description of SendRawEmail in this guide, or see the Amazon SES Developer Guide.
"
+ "documentation":"
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to send for the email address specified in the Source parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to send from user@example.com, then you would specify the SourceArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the Source to be user@example.com.
Instead of using this parameter, you can use the X-header X-SES-SOURCE-ARN in the raw message of the email. If you use both the SourceArn parameter and the corresponding X-header, Amazon SES uses the value of the SourceArn parameter.
For information about when to use this parameter, see the description of SendRawEmail in this guide, or see the Amazon SES Developer Guide.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to use the email address specified in the ReturnPath parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to use feedback@example.com, then you would specify the ReturnPathArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the ReturnPath to be feedback@example.com.
Instead of using this parameter, you can use the X-header X-SES-RETURN-PATH-ARN in the raw message of the email. If you use both the ReturnPathArn parameter and the corresponding X-header, Amazon SES uses the value of the ReturnPathArn parameter.
For information about when to use this parameter, see the description of SendRawEmail in this guide, or see the Amazon SES Developer Guide.
"
+ "documentation":"
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to use the email address specified in the ReturnPath parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to use feedback@example.com, then you would specify the ReturnPathArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the ReturnPath to be feedback@example.com.
Instead of using this parameter, you can use the X-header X-SES-RETURN-PATH-ARN in the raw message of the email. If you use both the ReturnPathArn parameter and the corresponding X-header, Amazon SES uses the value of the ReturnPathArn parameter.
For information about when to use this parameter, see the description of SendRawEmail in this guide, or see the Amazon SES Developer Guide.
The email address that is sending the email. This email address must be either individually verified with Amazon SES, or from a domain that has been verified with Amazon SES. For information about verifying identities, see the Amazon SES Developer Guide.
If you are sending on behalf of another user and have been permitted to do so by a sending authorization policy, then you must also specify the SourceArn parameter. For more information about sending authorization, see the Amazon SES Developer Guide.
Amazon SES does not support the SMTPUTF8 extension, as described in RFC6531. For this reason, the local part of a source email address (the part of the email address that precedes the @ sign) may only contain 7-bit ASCII characters. If the domain part of an address (the part after the @ sign) contains non-ASCII characters, they must be encoded using Punycode, as described in RFC3492. The sender name (also known as the friendly name) may contain non-ASCII characters. These characters must be encoded using MIME encoded-word syntax, as described inRFC 2047. MIME encoded-word syntax uses the following form: =?charset?encoding?encoded-text?=.
"
+ "documentation":"
The email address that is sending the email. This email address must be either individually verified with Amazon SES, or from a domain that has been verified with Amazon SES. For information about verifying identities, see the Amazon SES Developer Guide.
If you are sending on behalf of another user and have been permitted to do so by a sending authorization policy, then you must also specify the SourceArn parameter. For more information about sending authorization, see the Amazon SES Developer Guide.
Amazon SES does not support the SMTPUTF8 extension, as described in RFC6531. for this reason, The email address string must be 7-bit ASCII. If you want to send to or from email addresses that contain Unicode characters in the domain part of an address, you must encode the domain using Punycode. Punycode is not permitted in the local part of the email address (the part before the @ sign) nor in the \"friendly from\" name. If you want to use Unicode characters in the \"friendly from\" name, you must encode the \"friendly from\" name using MIME encoded-word syntax, as described in Sending raw email using the Amazon SES API. For more information about Punycode, see RFC 3492.
The email address that bounces and complaints will be forwarded to when feedback forwarding is enabled. If the message cannot be delivered to the recipient, then an error message will be returned from the recipient's ISP; this message will then be forwarded to the email address specified by the ReturnPath parameter. The ReturnPath parameter is never overwritten. This email address must be either individually verified with Amazon SES, or from a domain that has been verified with Amazon SES.
"
+ "documentation":"
The email address that bounces and complaints are forwarded to when feedback forwarding is enabled. If the message cannot be delivered to the recipient, then an error message is returned from the recipient's ISP; this message is forwarded to the email address specified by the ReturnPath parameter. The ReturnPath parameter is never overwritten. This email address must be either individually verified with Amazon SES, or from a domain that has been verified with Amazon SES.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to send for the email address specified in the Source parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to send from user@example.com, then you would specify the SourceArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the Source to be user@example.com.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to send for the email address specified in the Source parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to send from user@example.com, then you would specify the SourceArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the Source to be user@example.com.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to use the email address specified in the ReturnPath parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to use feedback@example.com, then you would specify the ReturnPathArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the ReturnPath to be feedback@example.com.
This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to use the email address specified in the ReturnPath parameter.
For example, if the owner of example.com (which has ARN arn:aws:ses:us-east-1:123456789012:identity/example.com) attaches a policy to it that authorizes you to use feedback@example.com, then you would specify the ReturnPathArn to be arn:aws:ses:us-east-1:123456789012:identity/example.com, and the ReturnPath to be feedback@example.com.
A list of replacement values to apply to the template. This parameter is a JSON object, typically consisting of key-value pairs in which the keys correspond to replacement tags in the email template.
"
}
},
- "documentation":"
Represents a request to send a templated email using Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to send a templated email using Amazon SES. For more information, see the Amazon SES Developer Guide.
The name of the receipt rule set to make active. Setting this value to null disables all email receiving.
"
}
},
- "documentation":"
Represents a request to set a receipt rule set as the active receipt rule set. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to set a receipt rule set as the active receipt rule set. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
Sets whether DKIM signing is enabled for an identity. Set to true to enable DKIM signing for this identity; false to disable it.
"
}
},
- "documentation":"
Represents a request to enable or disable Amazon SES Easy DKIM signing for an identity. For more information about setting up Easy DKIM, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to enable or disable Amazon SES Easy DKIM signing for an identity. For more information about setting up Easy DKIM, see the Amazon SES Developer Guide.
Sets whether Amazon SES will forward bounce and complaint notifications as email. true specifies that Amazon SES will forward bounce and complaint notifications as email, in addition to any Amazon SNS topic publishing otherwise specified. false specifies that Amazon SES will publish bounce and complaint notifications only through Amazon SNS. This value can only be set to false when Amazon SNS topics are set for both Bounce and Complaint notification types.
"
+ "documentation":"
Sets whether Amazon SES forwards bounce and complaint notifications as email. true specifies that Amazon SES forwards bounce and complaint notifications as email, in addition to any Amazon SNS topic publishing otherwise specified. false specifies that Amazon SES publishes bounce and complaint notifications only through Amazon SNS. This value can only be set to false when Amazon SNS topics are set for both Bounce and Complaint notification types.
"
}
},
- "documentation":"
Represents a request to enable or disable whether Amazon SES forwards you bounce and complaint notifications through email. For information about email feedback forwarding, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to enable or disable whether Amazon SES forwards you bounce and complaint notifications through email. For information about email feedback forwarding, see the Amazon SES Developer Guide.
Sets whether Amazon SES includes the original email headers in Amazon SNS notifications of the specified notification type. A value of true specifies that Amazon SES will include headers in notifications, and a value of false specifies that Amazon SES will not include headers in notifications.
This value can only be set when NotificationType is already set to use a particular Amazon SNS topic.
"
+ "documentation":"
Sets whether Amazon SES includes the original email headers in Amazon SNS notifications of the specified notification type. A value of true specifies that Amazon SES includes headers in notifications, and a value of false specifies that Amazon SES does not include headers in notifications.
This value can only be set when NotificationType is already set to use a particular Amazon SNS topic.
"
}
},
- "documentation":"
Represents a request to set whether Amazon SES includes the original email headers in the Amazon SNS notifications of a specified type. For information about notifications, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to set whether Amazon SES includes the original email headers in the Amazon SNS notifications of a specified type. For information about notifications, see the Amazon SES Developer Guide.
The custom MAIL FROM domain that you want the verified identity to use. The MAIL FROM domain must 1) be a subdomain of the verified identity, 2) not be used in a \"From\" address if the MAIL FROM domain is the destination of email feedback forwarding (for more information, see the Amazon SES Developer Guide), and 3) not be used to receive emails. A value of null disables the custom MAIL FROM setting for the identity.
"
+ "documentation":"
The custom MAIL FROM domain for the verified identity to use. The MAIL FROM domain must 1) be a subdomain of the verified identity, 2) not be used in a \"From\" address if the MAIL FROM domain is the destination of email feedback forwarding (for more information, see the Amazon SES Developer Guide), and 3) not be used to receive emails. A value of null disables the custom MAIL FROM setting for the identity.
The action that you want Amazon SES to take if it cannot successfully read the required MX record when you send an email. If you choose UseDefaultValue, Amazon SES will use amazonses.com (or a subdomain of that) as the MAIL FROM domain. If you choose RejectMessage, Amazon SES will return a MailFromDomainNotVerified error and not send the email.
The action specified in BehaviorOnMXFailure is taken when the custom MAIL FROM domain setup is in the Pending, Failed, and TemporaryFailure states.
"
+ "documentation":"
The action for Amazon SES to take if it cannot successfully read the required MX record when you send an email. If you choose UseDefaultValue, Amazon SES uses amazonses.com (or a subdomain of that) as the MAIL FROM domain. If you choose RejectMessage, Amazon SES returns a MailFromDomainNotVerified error and not send the email.
The action specified in BehaviorOnMXFailure is taken when the custom MAIL FROM domain setup is in the Pending, Failed, and TemporaryFailure states.
"
}
},
- "documentation":"
Represents a request to enable or disable the Amazon SES custom MAIL FROM domain setup for a verified identity. For information about using a custom MAIL FROM domain, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to enable or disable the Amazon SES custom MAIL FROM domain setup for a verified identity. For information about using a custom MAIL FROM domain, see the Amazon SES Developer Guide.
The identity (email address or domain) that you want to set the Amazon SNS topic for.
You can only specify a verified identity for this parameter.
You can specify an identity by using its name or by using its Amazon Resource Name (ARN). The following examples are all valid identities: sender@example.com, example.com, arn:aws:ses:us-east-1:123456789012:identity/example.com.
"
+ "documentation":"
The identity (email address or domain) for the Amazon SNS topic.
You can only specify a verified identity for this parameter.
You can specify an identity by using its name or by using its Amazon Resource Name (ARN). The following examples are all valid identities: sender@example.com, example.com, arn:aws:ses:us-east-1:123456789012:identity/example.com.
The Amazon Resource Name (ARN) of the Amazon SNS topic. If the parameter is omitted from the request or a null value is passed, SnsTopic is cleared and publishing is disabled.
"
}
},
- "documentation":"
Represents a request to specify the Amazon SNS topic to which Amazon SES will publish bounce, complaint, or delivery notifications for emails sent with that identity as the Source. For information about Amazon SES notifications, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to specify the Amazon SNS topic to which Amazon SES publishes bounce, complaint, or delivery notifications for emails sent with that identity as the source. For information about Amazon SES notifications, see the Amazon SES Developer Guide.
The name of the receipt rule after which to place the specified receipt rule.
"
}
},
- "documentation":"
Represents a request to set the position of a receipt rule in a receipt rule set. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to set the position of a receipt rule in a receipt rule set. You use receipt rule sets to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
The Amazon Resource Name (ARN) of the Amazon SNS topic to notify when the stop action is taken. An example of an Amazon SNS topic ARN is arn:aws:sns:us-west-2:123456789012:MyTopic. For more information about Amazon SNS topics, see the Amazon SNS Developer Guide.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the Amazon SNS topic to notify when the stop action is taken. You can find the ARN of a topic by using the ListTopics Amazon SNS operation.
When included in a receipt rule, this action terminates the evaluation of the receipt rule set and, optionally, publishes a notification to Amazon Simple Notification Service (Amazon SNS).
When included in a receipt rule, this action terminates the evaluation of the receipt rule set and, optionally, publishes a notification to Amazon Simple Notification Service (Amazon SNS).
The custom subdomain that will be used to redirect email recipients to the Amazon SES event tracking domain.
"
+ "documentation":"
The custom subdomain that is used to redirect email recipients to the Amazon SES event tracking domain.
"
}
},
- "documentation":"
A domain that is used to redirect email recipients to an Amazon SES-operated domain. This domain captures open and click events generated by Amazon SES emails.
A domain that is used to redirect email recipients to an Amazon SES-operated domain. This domain captures open and click events generated by Amazon SES emails.
The event destination object that you want to apply to the specified configuration set.
"
+ "documentation":"
The event destination object.
"
}
},
- "documentation":"
Represents a request to update the event destination of a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to update the event destination of a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.
Describes whether or not Amazon SES will publish reputation metrics for the configuration set, such as bounce and complaint rates, to Amazon CloudWatch.
"
+ "documentation":"
Describes whether or not Amazon SES publishes reputation metrics for the configuration set, such as bounce and complaint rates, to Amazon CloudWatch.
"
}
},
"documentation":"
Represents a request to modify the reputation metric publishing settings for a configuration set.
The content of the custom verification email. The total size of the email must be less than 10 MB. The message body may contain HTML, with some limitations. For more information, see Custom Verification Email Frequently Asked Questions in the Amazon SES Developer Guide.
"
+ "documentation":"
The content of the custom verification email. The total size of the email must be less than 10 MB. The message body may contain HTML, with some limitations. For more information, see Custom Verification Email Frequently Asked Questions in the Amazon SES Developer Guide.
A data structure that contains the updated receipt rule information.
"
}
},
- "documentation":"
Represents a request to update a receipt rule. You use receipt rules to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to update a receipt rule. You use receipt rules to receive email with Amazon SES. For more information, see the Amazon SES Developer Guide.
The name of the domain to be verified for Easy DKIM signing.
"
}
},
- "documentation":"
Represents a request to generate the CNAME records needed to set up Easy DKIM with Amazon SES. For more information about setting up Easy DKIM, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to generate the CNAME records needed to set up Easy DKIM with Amazon SES. For more information about setting up Easy DKIM, see the Amazon SES Developer Guide.
A set of character strings that represent the domain's identity. If the identity is an email address, the tokens represent the domain of that address.
Using these tokens, you need to create DNS CNAME records that point to DKIM public keys that are hosted by Amazon SES. Amazon Web Services eventually detects that you've updated your DNS records. This detection process might take up to 72 hours. After successful detection, Amazon SES is able to DKIM-sign email originating from that domain. (This only applies to domain identities, not email address identities.)
A set of character strings that represent the domain's identity. If the identity is an email address, the tokens represent the domain of that address.
Using these tokens, you need to create DNS CNAME records that point to DKIM public keys that are hosted by Amazon SES. Amazon Web Services eventually detects that you've updated your DNS records. This detection process might take up to 72 hours. After successful detection, Amazon SES is able to DKIM-sign email originating from that domain. (This only applies to domain identities, not email address identities.)
Returns CNAME records that you must publish to the DNS server of your domain to set up Easy DKIM with Amazon SES.
"
@@ -4418,7 +4418,7 @@
"documentation":"
The domain to be verified.
"
}
},
- "documentation":"
Represents a request to begin Amazon SES domain verification and to generate the TXT records that you must publish to the DNS server of your domain to complete the verification. For information about domain verification, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to begin Amazon SES domain verification and to generate the TXT records that you must publish to the DNS server of your domain to complete the verification. For information about domain verification, see the Amazon SES Developer Guide.
A TXT record that you must place in the DNS settings of the domain to complete domain verification with Amazon SES.
As Amazon SES searches for the TXT record, the domain's verification status is \"Pending\". When Amazon SES detects the record, the domain's verification status changes to \"Success\". If Amazon SES is unable to detect the record within 72 hours, the domain's verification status changes to \"Failed.\" In that case, if you still want to verify the domain, you must restart the verification process from the beginning.
"
+ "documentation":"
A TXT record that you must place in the DNS settings of the domain to complete domain verification with Amazon SES.
As Amazon SES searches for the TXT record, the domain's verification status is \"Pending\". When Amazon SES detects the record, the domain's verification status changes to \"Success\". If Amazon SES is unable to detect the record within 72 hours, the domain's verification status changes to \"Failed.\" In that case, to verify the domain, you must restart the verification process from the beginning. The domain's verification status also changes to \"Success\" when it is DKIM verified.
"
}
},
"documentation":"
Returns a TXT record that you must publish to the DNS server of your domain to complete domain verification with Amazon SES.
"
@@ -4440,7 +4440,7 @@
"documentation":"
The email address to be verified.
"
}
},
- "documentation":"
Represents a request to begin email address verification with Amazon SES. For information about email address verification, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to begin email address verification with Amazon SES. For information about email address verification, see the Amazon SES Developer Guide.
Represents a request to begin email address verification with Amazon SES. For information about email address verification, see the Amazon SES Developer Guide.
"
+ "documentation":"
Represents a request to begin email address verification with Amazon SES. For information about email address verification, see the Amazon SES Developer Guide.
The Amazon Resource Name (ARN) of the Amazon SNS topic to notify when the WorkMail action is called. An example of an Amazon SNS topic ARN is arn:aws:sns:us-west-2:123456789012:MyTopic. For more information about Amazon SNS topics, see the Amazon SNS Developer Guide.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the Amazon SNS topic to notify when the WorkMail action is called. You can find the ARN of a topic by using the ListTopics operation in Amazon SNS.
The ARN of the Amazon WorkMail organization. An example of an Amazon WorkMail organization ARN is arn:aws:workmail:us-west-2:123456789012:organization/m-68755160c4cb4e29a2b2f8fb58f359d7. For information about Amazon WorkMail organizations, see the Amazon WorkMail Administrator Guide.
"
+ "documentation":"
The Amazon Resource Name (ARN) of the Amazon WorkMail organization. Amazon WorkMail ARNs use the following format:
You can find the ID of your organization by using the ListOrganizations operation in Amazon WorkMail. Amazon WorkMail organization IDs begin with \"m-\", followed by a string of alphanumeric characters.
When included in a receipt rule, this action calls Amazon WorkMail and, optionally, publishes a notification to Amazon Simple Notification Service (Amazon SNS). You will typically not use this action directly because Amazon WorkMail adds the rule automatically during its setup procedure.
When included in a receipt rule, this action calls Amazon WorkMail and, optionally, publishes a notification to Amazon Simple Notification Service (Amazon SNS). It usually isn't necessary to set this up manually, because Amazon WorkMail adds the rule automatically during its setup procedure.
A message movement can only be cancelled when the current status is RUNNING.
Cancelling a message movement task does not revert the messages that have already been moved. It can only stop the messages that have not been moved yet.
"
+ "documentation":"
Cancels a specified message movement task. A message movement can only be cancelled when the current status is RUNNING. Cancelling a message movement task does not revert the messages that have already been moved. It can only stop the messages that have not been moved yet.
This action is currently limited to supporting message redrive from dead-letter queues (DLQs) only. In this context, the source queue is the dead-letter queue (DLQ), while the destination queue can be the original source queue (from which the messages were driven to the dead-letter-queue), or a custom destination queue.
Currently, only standard queues are supported.
Only one active message movement task is supported per queue at any given time.
Gets the most recent message movement tasks (up to 10) under a specific source queue.
"
+ "documentation":"
Gets the most recent message movement tasks (up to 10) under a specific source queue.
This action is currently limited to supporting message redrive from dead-letter queues (DLQs) only. In this context, the source queue is the dead-letter queue (DLQ), while the destination queue can be the original source queue (from which the messages were driven to the dead-letter-queue), or a custom destination queue.
Currently, only standard queues are supported.
Only one active message movement task is supported per queue at any given time.
Starts an asynchronous task to move messages from a specified source queue to a specified destination queue.
This action is currently limited to supporting message redrive from dead-letter queues (DLQs) only. In this context, the source queue is the dead-letter queue (DLQ), while the destination queue can be the original source queue (from which the messages were driven to the dead-letter-queue), or a custom destination queue.
Currently, only standard queues are supported.
Only one active message movement task is supported per queue at any given time.
"
+ "documentation":"
Starts an asynchronous task to move messages from a specified source queue to a specified destination queue.
This action is currently limited to supporting message redrive from queues that are configured as dead-letter queues (DLQs) of other Amazon SQS queues only. Non-SQS queue sources of dead-letter queues, such as Lambda or Amazon SNS topics, are currently not supported.
In dead-letter queues redrive context, the StartMessageMoveTask the source queue is the DLQ, while the destination queue can be the original source queue (from which the messages were driven to the dead-letter-queue), or a custom destination queue.
Currently, only standard queues support redrive. FIFO queues don't support redrive.
Only one active message movement task is supported per queue at any given time.
The ARN of the queue that contains the messages to be moved to another queue. Currently, only dead-letter queue (DLQ) ARNs are accepted.
"
+ "documentation":"
The ARN of the queue that contains the messages to be moved to another queue. Currently, only ARNs of dead-letter queues (DLQs) whose sources are other Amazon SQS queues are accepted. DLQs whose sources are non-SQS queues, such as Lambda or Amazon SNS topics, are not currently supported.
The source identity specified by the principal that is calling the AssumeRole operation.
You can require users to specify a source identity when they assume a role. You do this by using the sts:SourceIdentity condition key in a role trust policy. You can use source identity information in CloudTrail logs to determine who took actions with a role. You can use the aws:SourceIdentity condition key to further control access to Amazon Web Services resources based on the value of source identity. For more information about using source identity, see Monitor and control actions taken with assumed roles in the IAM User Guide.
The regex used to validate this parameter is a string of characters consisting of upper- and lower-case alphanumeric characters with no spaces. You can also include underscores or any of the following characters: =,.@-. You cannot use a value that begins with the text aws:. This prefix is reserved for Amazon Web Services internal use.
The OAuth 2.0 access token or OpenID Connect ID token that is provided by the identity provider. Your application must get this token by authenticating the user who is using your application with a web identity provider before the application makes an AssumeRoleWithWebIdentity call.
"
+ "documentation":"
The OAuth 2.0 access token or OpenID Connect ID token that is provided by the identity provider. Your application must get this token by authenticating the user who is using your application with a web identity provider before the application makes an AssumeRoleWithWebIdentity call. Only tokens with RSA algorithms (RS256) are supported.
The ID of the DecisionTaskStarted event recorded when this decision task was started. This information can be useful for diagnosing problems by tracing back the chain of events leading up to this event.
The task list to use for the future decision tasks of this workflow execution. This list overrides the original task list you specified while starting the workflow execution.
Specifies a timeout (in seconds) for the task list override. When this parameter is missing, the task list override is permanent. This parameter makes it possible to temporarily override the task list. If a decision task scheduled on the override task list is not started within the timeout, the decision task will time out. Amazon SWF will revert the override and schedule a new decision task to the original task list.
If a decision task scheduled on the override task list is started within the timeout, but not completed within the start-to-close timeout, Amazon SWF will also revert the override and schedule a new decision task to the original task list.
"
}
},
"documentation":"
Input data for a TaskCompleted response to a decision task.
Creates the connector, which captures the parameters for an outbound connection for the AS2 protocol. The connector is required for sending files to an externally hosted AS2 server. For more details about connectors, see Create AS2 connectors.
"
+ "documentation":"
Creates the connector, which captures the parameters for a connection for the AS2 or SFTP protocol. For AS2, the connector is required for sending files to an externally hosted AS2 server. For SFTP, the connector is required when sending files to an SFTP server or receiving files from an SFTP server. For more details about connectors, see Create AS2 connectors and Create SFTP connectors.
You must specify exactly one configuration object: either for AS2 (As2Config) or SFTP (SftpConfig).
Begins an outbound file transfer to a remote AS2 server. You specify the ConnectorId and the file paths for where to send the files.
"
+ "documentation":"
Begins a file transfer between local Amazon Web Services storage and a remote AS2 or SFTP server.
For an AS2 connector, you specify the ConnectorId and one or more SendFilePaths to identify the files you want to transfer.
For an SFTP connector, the file transfer can be either outbound or inbound. In both cases, you specify the ConnectorId. Depending on the direction of the transfer, you also specify the following items:
If you are transferring file from a partner's SFTP server to Amazon Web Services storage, you specify one or more RetreiveFilePaths to identify the files you want to transfer, and a LocalDirectoryPath to specify the destination folder.
If you are transferring file to a partner's SFTP server from Amazon Web Services storage, you specify one or more SendFilePaths to identify the files you want to transfer, and a RemoteDirectoryPath to specify the destination folder.
Tests whether your SFTP connector is set up successfully. We highly recommend that you call this operation to test your ability to transfer files between a Transfer Family server and a trading partner's SFTP server.
Provides Basic authentication support to the AS2 Connectors API. To use Basic authentication, you must provide the name or Amazon Resource Name (ARN) of a secret in Secrets Manager.
The default value for this parameter is null, which indicates that Basic authentication is not enabled for the connector.
If the connector should use Basic authentication, the secret needs to be in the following format:
Replace user-name and user-password with the credentials for the actual user that is being authenticated.
Note the following:
You are storing these credentials in Secrets Manager, not passing them directly into this API.
If you are using the API, SDKs, or CloudFormation to configure your connector, then you must create the secret before you can enable Basic authentication. However, if you are using the Amazon Web Services management console, you can have the system create the secret for you.
If you have previously enabled Basic authentication for a connector, you can disable it by using the UpdateConnector API call. For example, if you are using the CLI, you can run the following command to remove Basic authentication:
Contains the details for a connector object. The connector object is used for AS2 outbound processes, to connect the Transfer Family customer with the trading partner.
"
+ "documentation":"
Contains the details for an AS2 connector object. The connector object is used for AS2 outbound processes, to connect the Transfer Family customer with the trading partner.
With AS2, you can send files by calling StartFileTransfer and specifying the file paths in the request parameter, SendFilePaths. We use the file’s parent directory (for example, for --send-file-paths /bucket/dir/file.txt, parent directory is /bucket/dir/) to temporarily store a processed AS2 message file, store the MDN when we receive them from the partner, and write a final JSON file containing relevant metadata of the transmission. So, the AccessRole needs to provide read and write access to the parent directory of the file location used in the StartFileTransfer request. Additionally, you need to provide read and write access to the parent directory of the files that you intend to send with StartFileTransfer.
If you are using Basic authentication for your AS2 connector, the access role requires the secretsmanager:GetSecretValue permission for the secret. If the secret is encrypted using a customer-managed key instead of the Amazon Web Services managed key in Secrets Manager, then the role also needs the kms:Decrypt permission for that key.
"
+ "documentation":"
Connectors are used to send files using either the AS2 or SFTP protocol. For the access role, provide the Amazon Resource Name (ARN) of the Identity and Access Management role to use.
For AS2 connectors
With AS2, you can send files by calling StartFileTransfer and specifying the file paths in the request parameter, SendFilePaths. We use the file’s parent directory (for example, for --send-file-paths /bucket/dir/file.txt, parent directory is /bucket/dir/) to temporarily store a processed AS2 message file, store the MDN when we receive them from the partner, and write a final JSON file containing relevant metadata of the transmission. So, the AccessRole needs to provide read and write access to the parent directory of the file location used in the StartFileTransfer request. Additionally, you need to provide read and write access to the parent directory of the files that you intend to send with StartFileTransfer.
If you are using Basic authentication for your AS2 connector, the access role requires the secretsmanager:GetSecretValue permission for the secret. If the secret is encrypted using a customer-managed key instead of the Amazon Web Services managed key in Secrets Manager, then the role also needs the kms:Decrypt permission for that key.
For SFTP connectors
Make sure that the access role provides read and write access to the parent directory of the file location that's used in the StartFileTransfer request. Additionally, make sure that the role provides secretsmanager:GetSecretValue permission to Secrets Manager.
With AS2, you can send files by calling StartFileTransfer and specifying the file paths in the request parameter, SendFilePaths. We use the file’s parent directory (for example, for --send-file-paths /bucket/dir/file.txt, parent directory is /bucket/dir/) to temporarily store a processed AS2 message file, store the MDN when we receive them from the partner, and write a final JSON file containing relevant metadata of the transmission. So, the AccessRole needs to provide read and write access to the parent directory of the file location used in the StartFileTransfer request. Additionally, you need to provide read and write access to the parent directory of the files that you intend to send with StartFileTransfer.
If you are using Basic authentication for your AS2 connector, the access role requires the secretsmanager:GetSecretValue permission for the secret. If the secret is encrypted using a customer-managed key instead of the Amazon Web Services managed key in Secrets Manager, then the role also needs the kms:Decrypt permission for that key.
"
+ "documentation":"
Connectors are used to send files using either the AS2 or SFTP protocol. For the access role, provide the Amazon Resource Name (ARN) of the Identity and Access Management role to use.
For AS2 connectors
With AS2, you can send files by calling StartFileTransfer and specifying the file paths in the request parameter, SendFilePaths. We use the file’s parent directory (for example, for --send-file-paths /bucket/dir/file.txt, parent directory is /bucket/dir/) to temporarily store a processed AS2 message file, store the MDN when we receive them from the partner, and write a final JSON file containing relevant metadata of the transmission. So, the AccessRole needs to provide read and write access to the parent directory of the file location used in the StartFileTransfer request. Additionally, you need to provide read and write access to the parent directory of the files that you intend to send with StartFileTransfer.
If you are using Basic authentication for your AS2 connector, the access role requires the secretsmanager:GetSecretValue permission for the secret. If the secret is encrypted using a customer-managed key instead of the Amazon Web Services managed key in Secrets Manager, then the role also needs the kms:Decrypt permission for that key.
For SFTP connectors
Make sure that the access role provides read and write access to the parent directory of the file location that's used in the StartFileTransfer request. Additionally, make sure that the role provides secretsmanager:GetSecretValue permission to Secrets Manager.
Logical directory mappings that specify what Amazon S3 or Amazon EFS paths and keys should be visible to your user and how you want to make them visible. You must specify the Entry and Target pair, where Entry shows how the path is made visible and Target is the actual Amazon S3 or Amazon EFS path. If you only specify a target, it is displayed as is. You also must ensure that your Identity and Access Management (IAM) role provides access to paths in Target. This value can be set only when HomeDirectoryType is set to LOGICAL.
The following is an Entry and Target pair example.
In most cases, you can use this value instead of the session policy to lock your user down to the designated home directory (\"chroot\"). To do this, you can set Entry to / and set Target to the HomeDirectory parameter value.
The following is an Entry and Target pair example for chroot.
Logical directory mappings that specify what Amazon S3 or Amazon EFS paths and keys should be visible to your user and how you want to make them visible. You must specify the Entry and Target pair, where Entry shows how the path is made visible and Target is the actual Amazon S3 or Amazon EFS path. If you only specify a target, it is displayed as is. You also must ensure that your Identity and Access Management (IAM) role provides access to paths in Target. This value can be set only when HomeDirectoryType is set to LOGICAL.
The following is an Entry and Target pair example.
In most cases, you can use this value instead of the session policy to lock your user down to the designated home directory (\"chroot\"). To do this, you can set Entry to / and set Target to the value the user should see for their home directory when they log in.
The following is an Entry and Target pair example for chroot.
With AS2, you can send files by calling StartFileTransfer and specifying the file paths in the request parameter, SendFilePaths. We use the file’s parent directory (for example, for --send-file-paths /bucket/dir/file.txt, parent directory is /bucket/dir/) to temporarily store a processed AS2 message file, store the MDN when we receive them from the partner, and write a final JSON file containing relevant metadata of the transmission. So, the AccessRole needs to provide read and write access to the parent directory of the file location used in the StartFileTransfer request. Additionally, you need to provide read and write access to the parent directory of the files that you intend to send with StartFileTransfer.
If you are using Basic authentication for your AS2 connector, the access role requires the secretsmanager:GetSecretValue permission for the secret. If the secret is encrypted using a customer-managed key instead of the Amazon Web Services managed key in Secrets Manager, then the role also needs the kms:Decrypt permission for that key.
"
+ "documentation":"
Connectors are used to send files using either the AS2 or SFTP protocol. For the access role, provide the Amazon Resource Name (ARN) of the Identity and Access Management role to use.
For AS2 connectors
With AS2, you can send files by calling StartFileTransfer and specifying the file paths in the request parameter, SendFilePaths. We use the file’s parent directory (for example, for --send-file-paths /bucket/dir/file.txt, parent directory is /bucket/dir/) to temporarily store a processed AS2 message file, store the MDN when we receive them from the partner, and write a final JSON file containing relevant metadata of the transmission. So, the AccessRole needs to provide read and write access to the parent directory of the file location used in the StartFileTransfer request. Additionally, you need to provide read and write access to the parent directory of the files that you intend to send with StartFileTransfer.
If you are using Basic authentication for your AS2 connector, the access role requires the secretsmanager:GetSecretValue permission for the secret. If the secret is encrypted using a customer-managed key instead of the Amazon Web Services managed key in Secrets Manager, then the role also needs the kms:Decrypt permission for that key.
For SFTP connectors
Make sure that the access role provides read and write access to the parent directory of the file location that's used in the StartFileTransfer request. Additionally, make sure that the role provides secretsmanager:GetSecretValue permission to Secrets Manager.
With AS2, you can send files by calling StartFileTransfer and specifying the file paths in the request parameter, SendFilePaths. We use the file’s parent directory (for example, for --send-file-paths /bucket/dir/file.txt, parent directory is /bucket/dir/) to temporarily store a processed AS2 message file, store the MDN when we receive them from the partner, and write a final JSON file containing relevant metadata of the transmission. So, the AccessRole needs to provide read and write access to the parent directory of the file location used in the StartFileTransfer request. Additionally, you need to provide read and write access to the parent directory of the files that you intend to send with StartFileTransfer.
If you are using Basic authentication for your AS2 connector, the access role requires the secretsmanager:GetSecretValue permission for the secret. If the secret is encrypted using a customer-managed key instead of the Amazon Web Services managed key in Secrets Manager, then the role also needs the kms:Decrypt permission for that key.
"
+ "documentation":"
Connectors are used to send files using either the AS2 or SFTP protocol. For the access role, provide the Amazon Resource Name (ARN) of the Identity and Access Management role to use.
For AS2 connectors
With AS2, you can send files by calling StartFileTransfer and specifying the file paths in the request parameter, SendFilePaths. We use the file’s parent directory (for example, for --send-file-paths /bucket/dir/file.txt, parent directory is /bucket/dir/) to temporarily store a processed AS2 message file, store the MDN when we receive them from the partner, and write a final JSON file containing relevant metadata of the transmission. So, the AccessRole needs to provide read and write access to the parent directory of the file location used in the StartFileTransfer request. Additionally, you need to provide read and write access to the parent directory of the files that you intend to send with StartFileTransfer.
If you are using Basic authentication for your AS2 connector, the access role requires the secretsmanager:GetSecretValue permission for the secret. If the secret is encrypted using a customer-managed key instead of the Amazon Web Services managed key in Secrets Manager, then the role also needs the kms:Decrypt permission for that key.
For SFTP connectors
Make sure that the access role provides read and write access to the parent directory of the file location that's used in the StartFileTransfer request. Additionally, make sure that the role provides secretsmanager:GetSecretValue permission to Secrets Manager.
The public portion of the host key, or keys, that are used to authenticate the user to the external server to which you are connecting. You can use the ssh-keyscan command against the SFTP server to retrieve the necessary key.
The three standard SSH public key format elements are <key type>, <body base64>, and an optional <comment>, with spaces between each element.
For the trusted host key, Transfer Family accepts RSA and ECDSA keys.
For RSA keys, the key type is ssh-rsa.
For ECDSA keys, the key type is either ecdsa-sha2-nistp256, ecdsa-sha2-nistp384, or ecdsa-sha2-nistp521, depending on the size of the key you generated.
"
+ }
+ },
+ "documentation":"
Contains the details for an SFTP connector object. The connector object is used for transferring files to and from a partner's SFTP server.
An array of strings. Each string represents the absolute path for one outbound file transfer. For example, DOC-EXAMPLE-BUCKET/myfile.txt.
"
+ "documentation":"
One or more source paths for the Transfer Family server. Each string represents a source file path for one outbound file transfer. For example, DOC-EXAMPLE-BUCKET/myfile.txt.
Replace DOC-EXAMPLE-BUCKETwith one of your actual buckets.
For an outbound transfer, the RemoteDirectoryPath specifies the destination for one or more files that are transferred to the partner's SFTP server. If you don't specify a RemoteDirectoryPath, the destination for transferred files is the SFTP user's home directory.
Returns Connection succeeded if the test is successful. Or, returns a descriptive error message if the test fails. The following list provides the details for some error messages and troubleshooting steps for each.
Unable to access secrets manager: Verify that your secret name aligns with the one in Transfer Role permissions.
Unknown Host/Connection failed: Verify the server URL in the connector configuration , and verify that the login credentials work successfully outside of the connector.
Private key not found: Verify that the secret exists and is formatted correctly.
Invalid trusted host keys: Verify that the trusted host key in the connector configuration matches the ssh-keyscan output.
With AS2, you can send files by calling StartFileTransfer and specifying the file paths in the request parameter, SendFilePaths. We use the file’s parent directory (for example, for --send-file-paths /bucket/dir/file.txt, parent directory is /bucket/dir/) to temporarily store a processed AS2 message file, store the MDN when we receive them from the partner, and write a final JSON file containing relevant metadata of the transmission. So, the AccessRole needs to provide read and write access to the parent directory of the file location used in the StartFileTransfer request. Additionally, you need to provide read and write access to the parent directory of the files that you intend to send with StartFileTransfer.
If you are using Basic authentication for your AS2 connector, the access role requires the secretsmanager:GetSecretValue permission for the secret. If the secret is encrypted using a customer-managed key instead of the Amazon Web Services managed key in Secrets Manager, then the role also needs the kms:Decrypt permission for that key.
"
+ "documentation":"
Connectors are used to send files using either the AS2 or SFTP protocol. For the access role, provide the Amazon Resource Name (ARN) of the Identity and Access Management role to use.
For AS2 connectors
With AS2, you can send files by calling StartFileTransfer and specifying the file paths in the request parameter, SendFilePaths. We use the file’s parent directory (for example, for --send-file-paths /bucket/dir/file.txt, parent directory is /bucket/dir/) to temporarily store a processed AS2 message file, store the MDN when we receive them from the partner, and write a final JSON file containing relevant metadata of the transmission. So, the AccessRole needs to provide read and write access to the parent directory of the file location used in the StartFileTransfer request. Additionally, you need to provide read and write access to the parent directory of the files that you intend to send with StartFileTransfer.
If you are using Basic authentication for your AS2 connector, the access role requires the secretsmanager:GetSecretValue permission for the secret. If the secret is encrypted using a customer-managed key instead of the Amazon Web Services managed key in Secrets Manager, then the role also needs the kms:Decrypt permission for that key.
For SFTP connectors
Make sure that the access role provides read and write access to the parent directory of the file location that's used in the StartFileTransfer request. Additionally, make sure that the role provides secretsmanager:GetSecretValue permission to Secrets Manager.
With AS2, you can send files by calling StartFileTransfer and specifying the file paths in the request parameter, SendFilePaths. We use the file’s parent directory (for example, for --send-file-paths /bucket/dir/file.txt, parent directory is /bucket/dir/) to temporarily store a processed AS2 message file, store the MDN when we receive them from the partner, and write a final JSON file containing relevant metadata of the transmission. So, the AccessRole needs to provide read and write access to the parent directory of the file location used in the StartFileTransfer request. Additionally, you need to provide read and write access to the parent directory of the files that you intend to send with StartFileTransfer.
If you are using Basic authentication for your AS2 connector, the access role requires the secretsmanager:GetSecretValue permission for the secret. If the secret is encrypted using a customer-managed key instead of the Amazon Web Services managed key in Secrets Manager, then the role also needs the kms:Decrypt permission for that key.
"
+ "documentation":"
Connectors are used to send files using either the AS2 or SFTP protocol. For the access role, provide the Amazon Resource Name (ARN) of the Identity and Access Management role to use.
For AS2 connectors
With AS2, you can send files by calling StartFileTransfer and specifying the file paths in the request parameter, SendFilePaths. We use the file’s parent directory (for example, for --send-file-paths /bucket/dir/file.txt, parent directory is /bucket/dir/) to temporarily store a processed AS2 message file, store the MDN when we receive them from the partner, and write a final JSON file containing relevant metadata of the transmission. So, the AccessRole needs to provide read and write access to the parent directory of the file location used in the StartFileTransfer request. Additionally, you need to provide read and write access to the parent directory of the files that you intend to send with StartFileTransfer.
If you are using Basic authentication for your AS2 connector, the access role requires the secretsmanager:GetSecretValue permission for the secret. If the secret is encrypted using a customer-managed key instead of the Amazon Web Services managed key in Secrets Manager, then the role also needs the kms:Decrypt permission for that key.
For SFTP connectors
Make sure that the access role provides read and write access to the parent directory of the file location that's used in the StartFileTransfer request. Additionally, make sure that the role provides secretsmanager:GetSecretValue permission to Secrets Manager.
The Amazon Resource Name (ARN) of the Identity and Access Management (IAM) role that allows a connector to turn on CloudWatch logging for Amazon S3 events. When set, you can view connector activity in your CloudWatch logs.