▲ 2 Cutting LLM Batch Inference Time by Half with Dynamic Prefix Bucketing (daft.ai) by DISCURSIVE | Nov 20, 2025 | 0 comments on HN Visit Link