20 lines
1.0 KiB
Markdown
20 lines
1.0 KiB
Markdown
|
|
# Story: Fix UI Responsiveness (Tech Debt)
|
||
|
|
|
||
|
|
## User Story
|
||
|
|
**As a** User
|
||
|
|
**I want** the UI to remain interactive and responsive while the Agent is thinking or executing tools
|
||
|
|
**So that** I don't feel like the application has crashed.
|
||
|
|
|
||
|
|
## Context
|
||
|
|
Currently, the UI locks up or becomes unresponsive during long LLM generations or tool executions. Even though the backend commands are async, the frontend experience degrades.
|
||
|
|
|
||
|
|
## Acceptance Criteria
|
||
|
|
* [ ] Investigate the root cause of the freezing (JS Main Thread blocking vs. Tauri IPC blocking).
|
||
|
|
* [ ] Implement a "Streaming" architecture for Chat if necessary (getting partial tokens instead of waiting for full response).
|
||
|
|
* *Note: This might overlap with future streaming stories, but basic responsiveness is the priority here.*
|
||
|
|
* [ ] Add visual indicators (Spinner/Progress Bar) that animate smoothly during the wait.
|
||
|
|
* [ ] Ensure the "Stop Generation" button (if added) can actually interrupt the backend task.
|
||
|
|
|
||
|
|
## Out of Scope
|
||
|
|
* Full streaming text (unless that is the only way to fix the freezing).
|