Blog
1 week ago
Streaming Faster Made Our LLM Hub Slower
At 200 tok/s × N users, per-token streaming floods the hub with pure overhead. Our adaptive batcher caps 100ms latency and POST rate. The trick: measure TPS at the producer, not the round trip — or you build a feedback loop that eats the hub.
Source: HackerNoon →