- Backend: Added OllamaProvider::chat_stream() with newline-delimited JSON parsing - Backend: Emit chat:token events for each token received from Ollama - Backend: Added futures dependency and stream feature for reqwest - Frontend: Added streamingContent state and chat:token event listener - Frontend: Real-time token display with auto-scroll - Frontend: Markdown and syntax highlighting support for streaming content - Fixed all TypeScript errors (tsc --noEmit) - Fixed all Biome warnings and errors - Fixed all Clippy warnings - Added comprehensive code quality documentation - Added tsc --noEmit to verification checklist Tested and verified: - Tokens stream in real-time - Auto-scroll works during streaming - Tool calls interrupt streaming correctly - Multi-turn conversations work - Smooth performance with no lag
Project Specs
This folder contains the "Living Specification" for the project. It serves as the source of truth for all AI sessions.
Structure
- 00_CONTEXT.md: The high-level overview, goals, domain definition, and glossary. Start here.
- tech/: Implementation details, including the Tech Stack, Architecture, and Constraints.
- STACK.md: The technical "Constitution" (Languages, Libraries, Patterns).
- functional/: Domain logic and behavior descriptions, platform-agnostic.
- 01_CORE.md: Core functional specifications.
Usage for LLMs
- Always read 00_CONTEXT.md and tech/STACK.md at the beginning of a session.
- Before writing code, ensure the spec in this folder reflects the desired reality.
- If a Story changes behavior, update the spec first, get approval, then write code.