Cutting LLM Batch Inference Time in Half: Dynamic Prefix Bucketing at Scale | Not Hacker News!