PulseAugur
LIVE 08:33:30
research · [26 sources] ·
0
research

Researchers unveil new methods to boost LLM inference speed and efficiency

Google Research has introduced "speculative cascades," a novel method to enhance Large Language Model (LLM) efficiency by merging speculative decoding with standard cascades. This hybrid approach aims to reduce computational costs and inference latency without compromising output quality. By strategically using smaller models to predict tokens and then verifying them with larger models, speculative cascades offer improved cost-quality trade-offs compared to either technique used in isolation, as demonstrated with Gemma and T5 models. AI

Summary written by gemini-2.5-flash-lite from 26 sources. How we write summaries →

IMPACT New inference techniques like speculative cascades and KV cache compression could significantly reduce operational costs for LLM deployments.

RANK_REASON The cluster contains research papers detailing new methods for improving LLM inference efficiency.

Read on Hugging Face Blog →

COVERAGE [26]

  1. Google AI / Research TIER_1 ·

    Speculative cascades — A hybrid approach for smarter, faster LLM inference

    Generative AI

  2. Hugging Face Blog TIER_1 ·

    🚀 Accelerating LLM Inference with TGI on Intel Gaudi

  3. Hugging Face Blog TIER_1 ·

    Optimum-NVIDIA Unlocking blazingly fast LLM inference in just 1 line of code

  4. arXiv cs.AI TIER_1 · Sanjeev Rao Ganjihal ·

    Predictive Multi-Tier Memory Management for KV Cache in Large-Scale GPU Inference

    arXiv:2604.26968v1 Announce Type: cross Abstract: Key-value (KV) cache memory management is the primary bottleneck limiting throughput and cost-efficiency in large-scale GPU inference serving. Current systems suffer from three compounding inefficiencies: (1) the absence of unifie…

  5. arXiv cs.LG TIER_1 · Aditya Ukarande, Deep Shekhar, Marc Blackstein, Ram Rangan ·

    Efficient, VRAM-Constrained xLM Inference on Clients

    arXiv:2604.26334v1 Announce Type: cross Abstract: To usher in the next round of client AI innovation, there is an urgent need to enable efficient, lossless inference of high-accuracy large language models (LLMs) and vision language models (VLMs), jointly referred to as xLMs, on c…

  6. arXiv cs.AI TIER_1 · Bodon Jeong, Hongsu Byun, Youngjae Kim, Weikuan Yu, Kyungkeun Lee, Jihoon Yang, Sungyong Park ·

    DUAL-BLADE: Dual-Path NVMe-Direct KV-Cache Offloading for Edge LLM Inference

    arXiv:2604.26557v1 Announce Type: cross Abstract: The increasing deployment of Large Language Model (LLM) inference on edge AI systems demands efficient execution under tight memory budgets. A key challenge arises from Key-Value (KV) caches, which often exceed available device me…

  7. arXiv cs.AI TIER_1 · Sungyong Park ·

    DUAL-BLADE: Dual-Path NVMe-Direct KV-Cache Offloading for Edge LLM Inference

    The increasing deployment of Large Language Model (LLM) inference on edge AI systems demands efficient execution under tight memory budgets. A key challenge arises from Key-Value (KV) caches, which often exceed available device memory. Although NVMe-based offloading offers scalab…

  8. arXiv cs.LG TIER_1 · Ram Rangan ·

    Efficient, VRAM-Constrained xLM Inference on Clients

    To usher in the next round of client AI innovation, there is an urgent need to enable efficient, lossless inference of high-accuracy large language models (LLMs) and vision language models (VLMs), jointly referred to as xLMs, on client systems. To address this, we present pipelin…

  9. arXiv cs.LG TIER_1 · Nada Zine, Cl\'ement Quinton, Romain Rouvoy ·

    Pimp My LLM: Leveraging Variability Modeling to Tune Inference Hyperparameters

    arXiv:2602.17697v2 Announce Type: replace Abstract: Large Language Models (LLMs) are being increasingly used across a wide range of tasks. However, their substantial computational demands raise concerns about the energy efficiency and sustainability of both training and inference…

  10. arXiv cs.CL TIER_1 · Ishan Patel, Ishan Joshi ·

    PolyKV: A Shared Asymmetrically-Compressed KV Cache Pool for Multi-Agent LLM Inference

    arXiv:2604.24971v1 Announce Type: cross Abstract: We present PolyKV, a system in which multiple concurrent inference agents share a single, asymmetrically compressed KV cache pool. Rather than allocating a separate KV cache per agent -- the standard paradigm -- PolyKV writes a co…

  11. arXiv cs.CL TIER_1 · Zahra Dehghanighobadi, Asja Fischer ·

    DepthKV: Layer-Dependent KV Cache Pruning for Long-Context LLM Inference

    arXiv:2604.24647v1 Announce Type: new Abstract: Long-context reasoning is a critical capability of large language models (LLMs), enabling applications such as long-document understanding, summarization, and code generation. However, efficient autoregressive inference relies on th…

  12. arXiv cs.CL TIER_1 · Marta Adamska, Daria Smirnova, Hamid Nasiri, Zhengxin Yu, Peter Garraghan ·

    Green Prompting: Characterizing Prompt-driven Energy Costs of LLM Inference

    arXiv:2503.10666v4 Announce Type: replace Abstract: Large Language Models (LLMs) have become widely used across various domains spanning search engines, code generation, and text creation. However, a major concern associated with their adoption is the high cost of inference, impa…

  13. arXiv cs.CL TIER_1 · Yi Su, Zhenxu Tian, Dan Qiao, Yuechi Zhou, Juntao Li, Min Zhang ·

    LongFlow: Efficient KV Cache Compression for Reasoning Models

    arXiv:2603.11504v2 Announce Type: replace-cross Abstract: Recent reasoning models such as OpenAI-o1 and DeepSeek-R1 have shown strong performance on complex tasks including mathematical reasoning and code generation. However, this performance gain comes with substantially longer …

  14. arXiv cs.LG TIER_1 · Yaoqi Chen, Jinkai Zhang, Baotong Lu, Qianxi Zhang, Chengruidong Zhang, Jing Liu, Jingjia Luo, Di Liu, Huiqiang Jiang, Qi Chen, Bailu Ding, Xiao Yan, Jiawei Jiang, Chen Chen, Mingxing Zhang, Cheng Li, Yuqing Yang, Fan Yang, Mao Yang ·

    RetroInfer: A Vector Storage Engine for Scalable Long-Context LLM Inference

    arXiv:2505.02922v3 Announce Type: replace Abstract: Recent large language models (LLMs) are rapidly extending their context windows, yet inference throughput lags due to increasing GPU memory and bandwidth demands. This is because the key-value (KV) cache, an intermediate structu…

  15. arXiv cs.CL TIER_1 · Ishan Joshi ·

    PolyKV: A Shared Asymmetrically-Compressed KV Cache Pool for Multi-Agent LLM Inference

    We present PolyKV, a system in which multiple concurrent inference agents share a single, asymmetrically compressed KV cache pool. Rather than allocating a separate KV cache per agent -- the standard paradigm -- PolyKV writes a compressed cache once and injects it into N independ…

  16. arXiv cs.CL TIER_1 · Asja Fischer ·

    DepthKV: Layer-Dependent KV Cache Pruning for Long-Context LLM Inference

    Long-context reasoning is a critical capability of large language models (LLMs), enabling applications such as long-document understanding, summarization, and code generation. However, efficient autoregressive inference relies on the key-value (KV) cache, whose memory footprint g…

  17. Hugging Face Daily Papers TIER_1 ·

    DepthKV: Layer-Dependent KV Cache Pruning for Long-Context LLM Inference

    Long-context reasoning is a critical capability of large language models (LLMs), enabling applications such as long-document understanding, summarization, and code generation. However, efficient autoregressive inference relies on the key-value (KV) cache, whose memory footprint g…

  18. arXiv cs.CL TIER_1 · Dinghong Song, Jierui Xu, Weichu Yang, Pengfei Su, Dong Li ·

    NeuronMLP: Efficient LLM Inference via Singular Value Decomposition Compression and Tiling on AWS Trainium

    arXiv:2510.25977v4 Announce Type: replace Abstract: Emerging AI accelerators have started to gain attention and offer new opportunities for efficient inference of large language models (LLMs). Trainium, an AI accelerator recently developed by Amazon Web Services (AWS), provides a…

  19. arXiv cs.LG TIER_1 · Anurita Das ·

    MCAP: Deployment-Time Layer Profiling for Memory-Constrained LLM Inference

    arXiv:2604.21026v2 Announce Type: replace Abstract: Deploying large language models to heterogeneous hardware is often constrained by memory, not compute. We introduce MCAP (Monte Carlo Activation Profiling), a load-time per-layer importance estimator that enables dynamic precisi…

  20. X — Cohere TIER_1 · cohere ·

    Enjoyed the read? If you have deep experience in ML frameworks (training or inference) and love working on problems like these, our team is hiring!

    Enjoyed the read? If you have deep experience in ML frameworks (training or inference) and love working on problems like these, our team is hiring! ML Systems Engineer, Frameworks & Tooling: https://t.co/IyMnsfplXv Audio Inference Engineer, Model Efficiency:

  21. X — Cohere TIER_1 · cohere ·

    For real agentic workloads (North), short-context calibration wasn't enough. We calibrated AWQ on long internal agentic traces (up to 64k tokens) and added toke

    For real agentic workloads (North), short-context calibration wasn't enough. We calibrated AWQ on long internal agentic traces (up to 64k tokens) and added token masking in llm-compressor to exclude repetitive chat templates/tool descriptions from calibration stats. Plus QAD http…

  22. X — Cohere TIER_1 · cohere ·

    🔧 The tricky part: naïvely casting BF16 group scales to FP8 dropped the quality. Our fix: quantize scales per-channel (outer vector scaling) + rescale by 1/8 to

    🔧 The tricky part: naïvely casting BF16 group scales to FP8 dropped the quality. Our fix: quantize scales per-channel (outer vector scaling) + rescale by 1/8 to avoid FP8 clipping. Result: >99.5% of W4A16 accuracy recovered on Command A & Cohere MoE. Paired with a CUTLASS …

  23. X — Cohere TIER_1 · cohere ·

    Excited to share our work on production-ready W4A8 inference, now integrated in vLLM! By combining 4-bit weights (low memory) with 8-bit activations (high compu

    Excited to share our work on production-ready W4A8 inference, now integrated in vLLM! By combining 4-bit weights (low memory) with 8-bit activations (high compute), we hit the sweet spot for both decoding and prefill — up to 58% faster TTFT and 45% faster TPOT vs W4A16 on Hopper.…

  24. Hugging Face Daily Papers TIER_1 ·

    DASH-KV: Accelerating Long-Context LLM Inference via Asymmetric KV Cache Hashing

    The quadratic computational complexity of the standard attention mechanism constitutes a fundamental bottleneck for large language models in long-context inference. While existing KV cache compression methods alleviate memory pressure, they often sacrifice generation quality and …

  25. Mastodon — mastodon.social TIER_1 · aihaberleri ·

    📰 5 KV Cache Compression Techniques in 2026 That Slash LLM Memory Overhead by Up to 7.7x Top KV cache compression techniques are transforming LLM inference by r

    📰 5 KV Cache Compression Techniques in 2026 That Slash LLM Memory Overhead by Up to 7.7x Top KV cache compression techniques are transforming LLM inference by reducing memory overhead through entropy coding, quantization, and rematerialization. These methods enable faster, cheape…

  26. Mastodon — mastodon.social TIER_1 Türkçe(TR) · aihaberleri ·

    📰 KV Cache Compression: 10 Proven Methods to Reduce LLM Memory Overflow in 2026 10 innovative methods to solve the KV cache memory overflow, the biggest challenge for LLMs

    📰 KV Cache Sıkıştırma: 2026'da LLM Bellek Aşırısını Azaltan 10 Kanıtlanmış Yöntem LLM'lerin en büyük zorluğu olan KV cache bellek aşırısını çözen 10 yenilikçi yöntem, entropy coding, low-rank decompositions ve rematerialization ile birleşiyor. Bu teknikler, model boyutunu yarıya …