-
-
Notifications
You must be signed in to change notification settings - Fork 35
[FEAT] Wire End-to-End Token Streaming from Model APIs to Frontend #170
Copy link
Copy link
Open
Labels
HARDOSCG26enhancementNew feature or requestNew feature or requestv2.0Planned for v2.0 releasePlanned for v2.0 release
Description
Description
The WebSocket endpoint at /ws processes requests synchronously — it calls ai_council.process_request(), waits for the complete response, then sends the entire result as one JSON payload. No token-by-token streaming from model APIs to the frontend.
In BEST_QUALITY mode, the user sees a blank screen for the entire multi-model execution duration.
Component
web_app/backend/main.py:332-378, ai_council/execution/agent.py
Acceptance Criteria
In BEST_QUALITY mode, the user sees partial responses appearing progressively rather than a blank screen followed by a complete answer. Each model's response streams as tokens arrive.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
HARDOSCG26enhancementNew feature or requestNew feature or requestv2.0Planned for v2.0 releasePlanned for v2.0 release