Skip to content

server : fix processing of multiple back-to-back mtmd chunks#21107

Merged
ggerganov merged 1 commit into
masterfrom
gg/server-mtmd-fix-multi-chunk
Mar 28, 2026
Merged

server : fix processing of multiple back-to-back mtmd chunks#21107
ggerganov merged 1 commit into
masterfrom
gg/server-mtmd-fix-multi-chunk

Conversation

@ggerganov
Copy link
Copy Markdown
Member

Overview

fix #21104

When there are no text tokens in-between mtmd chunks, it's better to just process all the mtmd chunks together in one pass. This also avoids tripping the kill switch.

Requirements

@ggerganov ggerganov requested a review from a team as a code owner March 28, 2026 08:16
@ggerganov ggerganov merged commit edfb440 into master Mar 28, 2026
42 of 45 checks passed
@ggerganov ggerganov deleted the gg/server-mtmd-fix-multi-chunk branch March 28, 2026 14:27
slartibardfast pushed a commit to slartibardfast/llama.cpp that referenced this pull request Apr 12, 2026
Seunghhon pushed a commit to Seunghhon/llama.cpp that referenced this pull request Apr 26, 2026
rsenthilkumar6 pushed a commit to rsenthilkumar6/llama.cpp that referenced this pull request May 1, 2026
ljubomirj pushed a commit to ljubomirj/llama.cpp that referenced this pull request May 6, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Eval bug: Kill switch tripped while encoding audio slices (server-context.cpp:2686: fatal error - please provide logs)

1 participant