diff --git a/.living_spec/stories/archive/06_fix_ui_responsiveness.md b/.living_spec/stories/archive/06_fix_ui_responsiveness.md new file mode 100644 index 0000000..b6988ba --- /dev/null +++ b/.living_spec/stories/archive/06_fix_ui_responsiveness.md @@ -0,0 +1,19 @@ +# Story: Fix UI Responsiveness (Tech Debt) + +## User Story +**As a** User +**I want** the UI to remain interactive and responsive while the Agent is thinking or executing tools +**So that** I don't feel like the application has crashed. + +## Context +Currently, the UI locks up or becomes unresponsive during long LLM generations or tool executions. Even though the backend commands are async, the frontend experience degrades. + +## Acceptance Criteria +* [ ] Investigate the root cause of the freezing (JS Main Thread blocking vs. Tauri IPC blocking). +* [ ] Implement a "Streaming" architecture for Chat if necessary (getting partial tokens instead of waiting for full response). + * *Note: This might overlap with future streaming stories, but basic responsiveness is the priority here.* +* [ ] Add visual indicators (Spinner/Progress Bar) that animate smoothly during the wait. +* [ ] Ensure the "Stop Generation" button (if added) can actually interrupt the backend task. + +## Out of Scope +* Full streaming text (unless that is the only way to fix the freezing).