1.0 KiB
1.0 KiB
Story: Fix UI Responsiveness (Tech Debt)
User Story
As a User I want the UI to remain interactive and responsive while the Agent is thinking or executing tools So that I don't feel like the application has crashed.
Context
Currently, the UI locks up or becomes unresponsive during long LLM generations or tool executions. Even though the backend commands are async, the frontend experience degrades.
Acceptance Criteria
- Investigate the root cause of the freezing (JS Main Thread blocking vs. Tauri IPC blocking).
- Implement a "Streaming" architecture for Chat if necessary (getting partial tokens instead of waiting for full response).
- Note: This might overlap with future streaming stories, but basic responsiveness is the priority here.
- Add visual indicators (Spinner/Progress Bar) that animate smoothly during the wait.
- Ensure the "Stop Generation" button (if added) can actually interrupt the backend task.
Out of Scope
- Full streaming text (unless that is the only way to fix the freezing).