From 5b55f6e2c4b224acf481481638177e1d4d356f59 Mon Sep 17 00:00:00 2001 From: Roo Code Date: Thu, 4 Sep 2025 17:40:51 +0000 Subject: [PATCH] fix: provide fallback context window values for Ollama and LM Studio models - Add fallback ModelInfo when routerModels.ollama or lmStudioModels return undefined - Fixes context window display showing "used/1" instead of actual max tokens - Ensures proper context window management for Ollama and LM Studio providers Fixes #7674 --- .../components/ui/hooks/useSelectedModel.ts | 24 +++++++++++++++++-- 1 file changed, 22 insertions(+), 2 deletions(-) diff --git a/webview-ui/src/components/ui/hooks/useSelectedModel.ts b/webview-ui/src/components/ui/hooks/useSelectedModel.ts index e9470e090262..d087ab92ebf1 100644 --- a/webview-ui/src/components/ui/hooks/useSelectedModel.ts +++ b/webview-ui/src/components/ui/hooks/useSelectedModel.ts @@ -255,17 +255,37 @@ function getSelectedModel({ case "ollama": { const id = apiConfiguration.ollamaModelId ?? "" const info = routerModels.ollama && routerModels.ollama[id] + // Provide fallback values when info is undefined to fix context window display return { id, - info: info || undefined, + info: + info || + (id + ? { + maxTokens: 8192, + contextWindow: 8192, + supportsImages: false, + supportsPromptCache: true, + } + : undefined), } } case "lmstudio": { const id = apiConfiguration.lmStudioModelId ?? "" const info = lmStudioModels && lmStudioModels[apiConfiguration.lmStudioModelId!] + // Provide fallback values when info is undefined to fix context window display return { id, - info: info || undefined, + info: + info || + (id + ? { + maxTokens: 8192, + contextWindow: 8192, + supportsImages: false, + supportsPromptCache: false, + } + : undefined), } } case "vscode-lm": {