Hogwild! Inference: Parallel LLM Generation via Concurrent Attention Paper • 2504.06261 • Published Apr 8, 2025 • 110
RetroInfer: A Vector-Storage Approach for Scalable Long-Context LLM Inference Paper • 2505.02922 • Published May 5, 2025 • 28
InfiniPot-V: Memory-Constrained KV Cache Compression for Streaming Video Understanding Paper • 2506.15745 • Published Jun 18, 2025 • 13
Sparse-dLLM: Accelerating Diffusion LLMs with Dynamic Cache Eviction Paper • 2508.02558 • Published Aug 4, 2025 • 10