-
Notifications
You must be signed in to change notification settings - Fork 742
Description
Hi Telegram team,
In the recent update you mentioned that bots can now stream responses as they’re generated, which is especially useful for AI chatbots.
However, as far as I can tell, the only practical way to simulate streaming today is to repeatedly call editMessage while receiving partial output from an LLM. That approach quickly runs into API rate limits and requires complex throttling logic.
Could you please clarify whether the new “streaming response” feature introduces any new backend mechanism or protocol to handle continuous output more efficiently (i.e., without hitting rate limits)?
If not, could you elaborate on the recommended implementation pattern on both sides (bot backend and Telegram client) to achieve smooth streaming behavior?
It also states that developers need to enable these features, but then it refers to the API page, which provides no explanation. BotFather also does not give any clues in this regard.
Thanks a lot for your work and for clarifying this!
