# Story: Fix UI Responsiveness (Tech Debt) ## User Story **As a** User **I want** the UI to remain interactive and responsive while the Agent is thinking or executing tools **So that** I don't feel like the application has crashed. ## Context Currently, the UI locks up or becomes unresponsive during long LLM generations or tool executions. Even though the backend commands are async, the frontend experience degrades. ## Acceptance Criteria * [ ] Investigate the root cause of the freezing (JS Main Thread blocking vs. Tauri IPC blocking). * [ ] Implement a "Streaming" architecture for Chat if necessary (getting partial tokens instead of waiting for full response). * *Note: This might overlap with future streaming stories, but basic responsiveness is the priority here.* * [ ] Add visual indicators (Spinner/Progress Bar) that animate smoothly during the wait. * [ ] Ensure the "Stop Generation" button (if added) can actually interrupt the backend task. ## Out of Scope * Full streaming text (unless that is the only way to fix the freezing).