Skip to content

Conversation

ggerganov
Copy link
Member

fix #15215 #15104

When seq_id == -1 we remove the tokens from all KV cache buffers.

@ggerganov ggerganov merged commit 228f724 into master Aug 11, 2025
53 of 55 checks passed
@ggerganov ggerganov deleted the gg/kv-cache-fix-seq-rm branch August 11, 2025 10:58
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Misc. bug: llama-kv-cache-unified.cpp:222: GGML_ASSERT(seq_id >= 0 && (size_t) seq_id < seq_to_stream.size()) failed when loading processed prompt again
1 participant