forked from aws/aws-cdk
-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
chore: add integ test for postgres with access to S3 bucket #1
Open
udaypant
wants to merge
20
commits into
main
Choose a base branch
from
chore/udaypant
base: main
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
comcalvi
suggested changes
Oct 18, 2023
s3ImportBuckets: [importExportBucket], | ||
s3ExportBuckets: [importExportBucket], | ||
}); | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
can we delete this extra blank line?
aws#27981) Steps: 1. Get the list of regions from our single source of truth. 2. Convert that list into the TypeScript code that should be the content of `packages/aws-cdk-lib/region-info/build-tools/metadata.ts`. 3. Create a PR if there were any changes. The workflow is triggered every day at 00:00. ---- *By submitting this pull request, I confirm that my contribution is made under the terms of the Apache-2.0 license*
With aws#26763 the default runtime was reverted from NODEJS_18_X to NODEJS_16_X This PR reflects this in the docs of the NodejsFunctionProps ---- *By submitting this pull request, I confirm that my contribution is made under the terms of the Apache-2.0 license*
… changes (aws#28007) Only automated changes to the `packages/aws-cdk-lib/region-info/build-tools/metadata.ts` file are allowed. ---- *By submitting this pull request, I confirm that my contribution is made under the terms of the Apache-2.0 license*
Adds a Java 21 runtime to Lambda. ---- *By submitting this pull request, I confirm that my contribution is made under the terms of the Apache-2.0 license*
…agedKafka events (aws#28010) ## Summary This PR will include following features - add destination(currently sns/sqs) for Kafka event sources (SelfManagedKafkaEventSource andManagedKafkaEventSource) - add a dedicated S3 destination to the kafka event sources (non-kafka event source doesn't support S3 as destination) ## Backgrounds Lambda Event Source Mapping (ESM) processes events from event sources in a sequence manner. However, a potential issue with this approach is that if a record is deemed a “poison pill”, it will be retried indefinitely until it is processed successfully or until the record expires. This can cause delays in processing others records in the queue. Additionally, for events that exceed the Lambda payload limit of 6 MB, they might be dropped as they cannot be processed. Today, Lambda supports configuring an OnSuccess destination and OnFailure destination for asynchronous invocations. For stream-based event sources, such as Kinesis, and DynamoDB streams, Lambda supports configuring an OnFailure destination. Regarding SQS, SQS poller doesn’t support OnFailure destination, but SQS support Dead Letter Queue (DLQ) natively. For CDK, Some event source mappings (events) can have onFailure destination through a DestinationConfig. Right now that supports only DynamoDB and Kinesis event sources and only SQS and SNS destinations. ## Solution - Add a new `s3onFailureDestination` destination in `s3-ofd.ts` file - Add a new field `onFailure` to `KafkaEventSourceProps` that customer can use `s3onFailureDestination` to pass in. - Add a check in `enrichMappingOptions` for every event type against `EventSourceMappingOptions.supportS3OFD` to check if they support S3 as onFailure. ## User Experience ```diff import { Secret } from 'aws-cdk-lib/aws-secretsmanager'; + import { ManagedKafkaEventSource, S3OnFailureDestination } from 'aws-cdk-lib/aws-lambda-event-sources'; import * as lambda from 'aws-cdk-lib/aws-lambda' import { App, StackProps, Stack } from 'aws-cdk-lib'; + import { Bucket } from 'aws-cdk-lib/aws-s3'; export class CdkTestStack extends Stack { constructor(scope: App, id: string, props?: StackProps) { super(scope, id, props); const myFunction = new lambda.Function(this, 'myFunction', { runtime: lambda.Runtime.NODEJS_16_X, handler: 'index.handler', code: lambda.Code.fromInline('//handler_code_here'), }); // Your MSK cluster arn const clusterArn = 'arn:aws:kafka:us-east-1:0123456789019:cluster/SalesCluster/abcd1234-abcd-cafe-abab-9876543210ab-4'; // The Kafka topic you want to subscribe to const topic = 'some-cool-topic'; // The secret that allows access to your MSK cluster // You still have to make sure that it is associated with your cluster as described in the documentation const secret = new Secret(this, 'Secret', { secretName: 'AmazonMSK_KafkaSecret' }); // Your bucket for Kafka's onFailure Destination + const bucket = Bucket.fromBucketName(this, 'BucketByName', 'my-bucket'); + const s3ofd = new S3OnFailureDestination(bucket); myFunction.addEventSource(new ManagedKafkaEventSource({ clusterArn, topic: topic, secret: secret, batchSize: 100, // default startingPosition: lambda.StartingPosition.TRIM_HORIZON, + onFailure: s3ofd, })); } } ``` ## sample synth output ``` myFunctionServiceRoleDefaultPolicyECBA61F7: Type: AWS::IAM::Policy Properties: PolicyDocument: Statement: - Action: - s3:Abort* - s3:DeleteObject* - s3:GetBucket* - s3:GetObject* - s3:List* - s3:PutObject - s3:PutObjectLegalHold - s3:PutObjectRetention - s3:PutObjectTagging - s3:PutObjectVersionTagging Effect: Allow Resource: - Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - :s3:::my-bucket - Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - :s3:::my-bucket/* myFunctionKafkaEventSourceCdkTestStackmyFunction09D44E80somecooltopic1B4580FF: Type: AWS::Lambda::EventSourceMapping Properties: BatchSize: 100 DestinationConfig: OnFailure: Destination: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - :s3:::my-bucket ``` ## To discuss ### Change parent class name The parent class of `s3ofd` is called `IEventSourceDlq`. `DLQ` stands for Dead Letter Queue, Which is very specific for SQS, we might want to work out a new naming for S3 - Option 1: Rename the `IEventSourceDlq` to `IEventSourceOfd` (Ofd: OnFailureDestinaion)And create `IEventSourceDlq` again to extends from it for keeping backwards compatibility. For S3, Create a new class `IEventSourceS3ofd`. also extend from the parent class and will be implemented by our new S3 Destination class. - Option 2: Don’t create new parent classes, the new S3 Destination class just extend the original `IEventSourceDlq` Class. Like what is shown in the current code. ### Where to add checking for s3 on failure desintation support Currently in the commit, I was checking for s3ofd support in `enrichMappingOptions` Which will be called by all event source classes. Does this design make sense, or should we create a new dedicated function for chcecking? ## Questions: - Check out the current UX Closes #<issue number here>. ---- *By submitting this pull request, I confirm that my contribution is made under the terms of the Apache-2.0 license*
This PR adds new versions for Aurora MySQL. - 3.04.1 - https://docs.aws.amazon.com/AmazonRDS/latest/AuroraMySQLReleaseNotes/AuroraMySQL.Updates.3041.html ---- *By submitting this pull request, I confirm that my contribution is made under the terms of the Apache-2.0 license*
AWS Service Spec packages to latest versions.
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
chore: add integ test for postgres with access to S3 bucket for import and export.
By submitting this pull request, I confirm that my contribution is made under the terms of the Apache-2.0 license