Real-time streaming output for multi-agent execution. Specifically, enabling users to see LLM responses as they are generated, rather than waiting for a full response.
Raw Developer Origin & Technical Request
GitHub Issue
Apr 1, 2026
## Summary
Support real-time streaming output during agent execution, so users can see LLM responses as they are generated.
## Motivation
Currently `AgentRunner` uses `adapter.chat()` which waits for the full response. For long-running tasks, users have no visibility into what the agent is doing until it finishes. Streaming would enable:
- Real-time progress feedback in CLI or web UI
- Lower perceived latency
- Early termination if the agent goes off track
## Proposed Approach
- Add `stream` mode option to `AgentRunner.run()`
- Use `adapter.stream()` instead of `adapter.chat()` when enabled
- Emit events via callback or AsyncIterable for consumer integration
- Handle tool calls within streaming context
## Open Questions
- Should streaming be opt-in per agent or per team?
- How to handle tool execution interleaved with streaming output?
Developer Debate & Comments
No active discussions extracted for this entry yet.
Adjacent Repository Pain Points
Other highly discussed features and pain points extracted from JackChen-me/open-multi-agent.
Engagement Signals
Cross-Market Term Frequency
Quantifies the cross-market adoption of foundational terms like CLI and real-time by tracking occurrence frequency across active SaaS architectures and enterprise developer debates.
Macro Market Trends
Correlated public search velocity for adjacent technologies.
Market Trends