Dave 64d1b788be Story 18: Token-by-token streaming responses
- Backend: Added OllamaProvider::chat_stream() with newline-delimited JSON parsing
- Backend: Emit chat:token events for each token received from Ollama
- Backend: Added futures dependency and stream feature for reqwest
- Frontend: Added streamingContent state and chat:token event listener
- Frontend: Real-time token display with auto-scroll
- Frontend: Markdown and syntax highlighting support for streaming content
- Fixed all TypeScript errors (tsc --noEmit)
- Fixed all Biome warnings and errors
- Fixed all Clippy warnings
- Added comprehensive code quality documentation
- Added tsc --noEmit to verification checklist

Tested and verified:
- Tokens stream in real-time
- Auto-scroll works during streaming
- Tool calls interrupt streaming correctly
- Multi-turn conversations work
- Smooth performance with no lag
2025-12-27 16:50:18 +00:00
2025-12-24 16:29:33 +00:00
2025-12-24 16:29:33 +00:00
2025-12-24 16:29:33 +00:00
2025-12-24 16:29:33 +00:00
2025-12-24 16:29:33 +00:00
2025-12-24 16:29:33 +00:00
2025-12-24 16:29:33 +00:00
2025-12-24 16:29:33 +00:00

Tauri + React + Typescript

This template should help get you started developing with Tauri, React and Typescript in Vite.

S
Description
No description provided
https://huskies.dev
Readme GPL-3.0 23 MiB
v0.10.4 Latest
2026-04-21 12:01:43 +00:00
Languages
Rust 85.9%
TypeScript 11.4%
HTML 1.6%
Shell 0.6%
CSS 0.4%
Other 0.1%