Skip to content

Commit 7b7bb49

Browse files
daniel-lxsmrubens
andauthored
fix: remove topP parameter from Bedrock inference config (#8388)
Co-authored-by: Matt Rubens <[email protected]>
1 parent 8111da6 commit 7b7bb49

File tree

2 files changed

+5
-11
lines changed

2 files changed

+5
-11
lines changed

src/api/providers/__tests__/bedrock-reasoning.spec.ts

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -182,7 +182,7 @@ describe("AwsBedrockHandler - Extended Thinking", () => {
182182
)
183183
})
184184

185-
it("should include topP when thinking is disabled", async () => {
185+
it("should not include topP when thinking is disabled (global removal)", async () => {
186186
handler = new AwsBedrockHandler({
187187
apiProvider: "bedrock",
188188
apiModelId: "anthropic.claude-3-7-sonnet-20250219-v1:0",
@@ -216,10 +216,10 @@ describe("AwsBedrockHandler - Extended Thinking", () => {
216216
chunks.push(chunk)
217217
}
218218

219-
// Verify that topP IS present when thinking is disabled
219+
// Verify that topP is NOT present for any model (removed globally)
220220
expect(mockSend).toHaveBeenCalledTimes(1)
221221
expect(capturedPayload).toBeDefined()
222-
expect(capturedPayload.inferenceConfig).toHaveProperty("topP", 0.1)
222+
expect(capturedPayload.inferenceConfig).not.toHaveProperty("topP")
223223

224224
// Verify that additionalModelRequestFields is not present or empty
225225
expect(capturedPayload.additionalModelRequestFields).toBeUndefined()

src/api/providers/bedrock.ts

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,6 @@ import type { SingleCompletionHandler, ApiHandlerCreateMessageMetadata } from ".
4545
interface BedrockInferenceConfig {
4646
maxTokens: number
4747
temperature?: number
48-
topP?: number
4948
}
5049

5150
// Define interface for Bedrock additional model request fields
@@ -374,12 +373,8 @@ export class AwsBedrockHandler extends BaseProvider implements SingleCompletionH
374373
maxTokens: modelConfig.maxTokens || (modelConfig.info.maxTokens as number),
375374
temperature: modelConfig.temperature ?? (this.options.modelTemperature as number),
376375
}
377-
378-
if (!thinkingEnabled) {
379-
inferenceConfig.topP = 0.1
380-
}
381-
382-
// Check if 1M context is enabled for Claude Sonnet 4 / 4.5
376+
377+
// Check if 1M context is enabled for Claude Sonnet 4
383378
// Use parseBaseModelId to handle cross-region inference prefixes
384379
const baseModelId = this.parseBaseModelId(modelConfig.id)
385380
const is1MContextEnabled =
@@ -649,7 +644,6 @@ export class AwsBedrockHandler extends BaseProvider implements SingleCompletionH
649644
const inferenceConfig: BedrockInferenceConfig = {
650645
maxTokens: modelConfig.maxTokens || (modelConfig.info.maxTokens as number),
651646
temperature: modelConfig.temperature ?? (this.options.modelTemperature as number),
652-
...(thinkingEnabled ? {} : { topP: 0.1 }), // Only set topP when thinking is NOT enabled
653647
}
654648

655649
// For completePrompt, use a unique conversation ID based on the prompt

0 commit comments

Comments
 (0)