dsdtszi0520538 2019-08-06 11:46 采纳率: 100%
浏览 1188


I would like to know, how flow control works in a client-streaming gRPC service in Go.

Specifically, I am interested in knowing when will a call to stream.SendMsg() function in the client-side block? According to the documentation :

SendMsg() blocks until :

  • There is sufficient flow control to schedule m with the transport, or ...

So what is the specification for the flow control mechanism of the stream? For example, if the server-side code responsible for reading the messages from the stream, isn't reading the messages fast enough, at what point will calls to SendMsg() block?

Is there some kind of backpressure mechanism implemented for the server to tell the client that it is not ready to receive more data? In the meantime, where are all the messages that have been successfully sent before the backpressure signal, queued?

  • 写回答

1条回答 默认 最新

  • douyan8027 2019-08-07 20:19

    gRPC flow control is based on http2 flow control: https://http2.github.io/http2-spec/#FlowControl

    There will be backpressure. Messages are only successfully sent when there's enough flow control window for them, otherwise SendMsg() will block.

    The signal from the receiving side is not to add backpressure, it's to release backpressure. It's like saying "now I'm ready to receive another 1MB of messages, send them".

    本回答被题主选为最佳回答 , 对您是否有帮助呢?



    • ¥15 multisim电路设计(相关搜索:设计报告)
    • ¥15 从github上下载的项目到ecplise上咋运行
    • ¥15 PC-lint Plus
    • ¥15 gpl24676注释
    • ¥15 php5.3内存泄露
    • ¥15 DigSilent如何复制复合模型到自己案例?
    • ¥15 求日版华为b610s-77a 官方公版固件,有偿
    • ¥15 关于#java#的问题,请各位专家解答!(相关搜索:java程序)
    • ¥15 linux tsi721的驱动编译后 insmod 提示 报错
    • ¥20 multisim测数据