From dfefa3a8bdcfa675de7bec4be39d9830db3c0eae Mon Sep 17 00:00:00 2001 From: Dave Date: Thu, 25 Dec 2025 12:40:38 +0000 Subject: [PATCH] chore: archive story 06 --- .../archive/06_fix_ui_responsiveness.md | 19 +++++++++++++++++++ 1 file changed, 19 insertions(+) create mode 100644 .living_spec/stories/archive/06_fix_ui_responsiveness.md diff --git a/.living_spec/stories/archive/06_fix_ui_responsiveness.md b/.living_spec/stories/archive/06_fix_ui_responsiveness.md new file mode 100644 index 0000000..b6988ba --- /dev/null +++ b/.living_spec/stories/archive/06_fix_ui_responsiveness.md @@ -0,0 +1,19 @@ +# Story: Fix UI Responsiveness (Tech Debt) + +## User Story +**As a** User +**I want** the UI to remain interactive and responsive while the Agent is thinking or executing tools +**So that** I don't feel like the application has crashed. + +## Context +Currently, the UI locks up or becomes unresponsive during long LLM generations or tool executions. Even though the backend commands are async, the frontend experience degrades. + +## Acceptance Criteria +* [ ] Investigate the root cause of the freezing (JS Main Thread blocking vs. Tauri IPC blocking). +* [ ] Implement a "Streaming" architecture for Chat if necessary (getting partial tokens instead of waiting for full response). + * *Note: This might overlap with future streaming stories, but basic responsiveness is the priority here.* +* [ ] Add visual indicators (Spinner/Progress Bar) that animate smoothly during the wait. +* [ ] Ensure the "Stop Generation" button (if added) can actually interrupt the backend task. + +## Out of Scope +* Full streaming text (unless that is the only way to fix the freezing).