Skip to content

[FEAT] Wire End-to-End Token Streaming from Model APIs to Frontend #170

@shrixtacy

Description

@shrixtacy

Description

The WebSocket endpoint at /ws processes requests synchronously — it calls ai_council.process_request(), waits for the complete response, then sends the entire result as one JSON payload. No token-by-token streaming from model APIs to the frontend.

In BEST_QUALITY mode, the user sees a blank screen for the entire multi-model execution duration.

Component

web_app/backend/main.py:332-378, ai_council/execution/agent.py

Acceptance Criteria

In BEST_QUALITY mode, the user sees partial responses appearing progressively rather than a blank screen followed by a complete answer. Each model's response streams as tokens arrive.

Metadata

Metadata

Assignees

No one assigned

    Labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions