Files
storkit/.living_spec/specs
Dave 64d1b788be Story 18: Token-by-token streaming responses
- Backend: Added OllamaProvider::chat_stream() with newline-delimited JSON parsing
- Backend: Emit chat:token events for each token received from Ollama
- Backend: Added futures dependency and stream feature for reqwest
- Frontend: Added streamingContent state and chat:token event listener
- Frontend: Real-time token display with auto-scroll
- Frontend: Markdown and syntax highlighting support for streaming content
- Fixed all TypeScript errors (tsc --noEmit)
- Fixed all Biome warnings and errors
- Fixed all Clippy warnings
- Added comprehensive code quality documentation
- Added tsc --noEmit to verification checklist

Tested and verified:
- Tokens stream in real-time
- Auto-scroll works during streaming
- Tool calls interrupt streaming correctly
- Multi-turn conversations work
- Smooth performance with no lag
2025-12-27 16:50:18 +00:00
..
2025-12-24 16:29:33 +00:00
2025-12-24 16:29:33 +00:00

Project Specs

This folder contains the "Living Specification" for the project. It serves as the source of truth for all AI sessions.

Structure

  • 00_CONTEXT.md: The high-level overview, goals, domain definition, and glossary. Start here.
  • tech/: Implementation details, including the Tech Stack, Architecture, and Constraints.
    • STACK.md: The technical "Constitution" (Languages, Libraries, Patterns).
  • functional/: Domain logic and behavior descriptions, platform-agnostic.
    • 01_CORE.md: Core functional specifications.

Usage for LLMs

  1. Always read 00_CONTEXT.md and tech/STACK.md at the beginning of a session.
  2. Before writing code, ensure the spec in this folder reflects the desired reality.
  3. If a Story changes behavior, update the spec first, get approval, then write code.