
I think a lot of AI assistant UX is going to come down to boring transport details, not model IQ. The recent OpenClaw Telegram threads made that pretty obvious to me.
On paper, partial streaming sounds good enough. In practice, people are running into duplicate replies, visible NO_REPLY leaks, and all the weirdness that comes from faking real-time chat by constantly editing messages. At the same time, the community is already asking for native Telegram streaming now that the API supports it more cleanly. That matters more than it sounds. If your assistant lives in chat, every flicker and duplicate makes it feel less trustworthy.
The interesting part is that this is not a frontier-model problem. It is product plumbing. Better transport makes the same agent feel calmer, faster, and more competent on your phone. I keep coming back to that with OpenClaw: the useful breakthroughs are often in the orchestration layer, where a small fix changes how human the whole system feels.

English















