Skip to content

Conversation

lastshogun13
Copy link
Contributor

Description of Changes

This PR adds support for Qwen3 model (qwen.qwen3-235b-a22b-2507-v1:0, qwen.qwen3-32b-v1:0, qwen.qwen3-coder-480b-a35b-v1:0, qwen.qwen3-coder-30b-a3b-v1:0, ) to the Generative AI Use Cases application.

Changes include:

  • Added Qwen3 model configuration to the backend model registry (packages/cdk/lambda/utils/models.ts)
  • Added model metadata for Qwen3 in the common package (packages/common/src/application/model.ts)
    • Model supports text and document features (MODEL_FEATURE.TEXT_DOC)
    • Display name: "Qwen3"
  • Updated deployment option documentation in English, Japanese, and Korean (docs/*/DEPLOY_OPTION.md)
    • Added Qwen3 to the list of supported text generation models
    • Added configuration examples with us-west-2 region specification
    • Added to cross-region inference examples and config preset examples

Impact on existing users:

  • N/A

Checklist

  • Modified relevant documentation
  • Verified operation in local environment
  • Executed npm run cdk:test and if there are snapshot differences, execute npm run cdk:test:update-snapshot to update snapshots

Related Issues

N/A

@kazuhitogo kazuhitogo self-requested a review October 6, 2025 02:05
kazuhitogo
kazuhitogo previously approved these changes Oct 6, 2025
Copy link
Collaborator

@kazuhitogo kazuhitogo left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

There is a problem with the operation check

Copy link
Collaborator

@kazuhitogo kazuhitogo left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

image Only Qwen3 32B returns an error (the other three worked fine), but did it work in your environment?

@kazuhitogo kazuhitogo dismissed their stale review October 6, 2025 05:34

There is a problem with the operation check

@lastshogun13
Copy link
Contributor Author

Adjusted their maxTokens parameters based on actual testing to ensure they all work correctly.

The token limits were verified through systematic testing with AWS Bedrock:

  • qwen3-235b-a22b-2507: 196,608 tokens
  • qwen3-32b: 16,384 tokens
  • qwen3-coder-480b-a35b: 65,536 tokens
  • qwen3-coder-30b-a3b: 196,608 tokens

All models have been tested and confirmed working with these configurations.

Note: The lower token limit for qwen3-coder-480b-a35b was determined through direct Bedrock API testing. The reason for this limitation is currently unknown.

Copy link
Collaborator

@kazuhitogo kazuhitogo left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM, Thank you.

@kazuhitogo kazuhitogo merged commit 8d7dddd into aws-samples:main Oct 14, 2025
2 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants