From 3579aee11426b0ce2d98043c07685d104e55c80a Mon Sep 17 00:00:00 2001 From: Kasper Marinus Date: Thu, 29 Jan 2026 15:35:42 +0100 Subject: [PATCH] fix: allow interrupting on each token stream ref: N25B-452 --- src/control_backend/agents/llm/llm_agent.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/src/control_backend/agents/llm/llm_agent.py b/src/control_backend/agents/llm/llm_agent.py index 08a77e3..5a77b81 100644 --- a/src/control_backend/agents/llm/llm_agent.py +++ b/src/control_backend/agents/llm/llm_agent.py @@ -179,6 +179,9 @@ class LLMAgent(BaseAgent): full_message = "" current_chunk = "" async for token in self._stream_query_llm(messages): + if self._interrupted: + return + full_message += token current_chunk += token