@@ -45,7 +45,6 @@ import type { SingleCompletionHandler, ApiHandlerCreateMessageMetadata } from ".
4545interface BedrockInferenceConfig {
4646 maxTokens : number
4747 temperature ?: number
48- topP ?: number
4948}
5049
5150// Define interface for Bedrock additional model request fields
@@ -374,12 +373,8 @@ export class AwsBedrockHandler extends BaseProvider implements SingleCompletionH
374373 maxTokens : modelConfig . maxTokens || ( modelConfig . info . maxTokens as number ) ,
375374 temperature : modelConfig . temperature ?? ( this . options . modelTemperature as number ) ,
376375 }
377-
378- if ( ! thinkingEnabled ) {
379- inferenceConfig . topP = 0.1
380- }
381-
382- // Check if 1M context is enabled for Claude Sonnet 4 / 4.5
376+
377+ // Check if 1M context is enabled for Claude Sonnet 4
383378 // Use parseBaseModelId to handle cross-region inference prefixes
384379 const baseModelId = this . parseBaseModelId ( modelConfig . id )
385380 const is1MContextEnabled =
@@ -649,7 +644,6 @@ export class AwsBedrockHandler extends BaseProvider implements SingleCompletionH
649644 const inferenceConfig : BedrockInferenceConfig = {
650645 maxTokens : modelConfig . maxTokens || ( modelConfig . info . maxTokens as number ) ,
651646 temperature : modelConfig . temperature ?? ( this . options . modelTemperature as number ) ,
652- ...( thinkingEnabled ? { } : { topP : 0.1 } ) , // Only set topP when thinking is NOT enabled
653647 }
654648
655649 // For completePrompt, use a unique conversation ID based on the prompt
0 commit comments