SenCache: Accelerating Diffusion Model Inference via Sensitivity-Aware Caching
Abstract
A sensitivity-aware caching framework improves diffusion model inference efficiency by dynamically selecting cache timesteps based on model output sensitivity to input perturbations.
Diffusion models achieve state-of-the-art video generation quality, but their inference remains expensive due to the large number of sequential denoising steps. This has motivated a growing line of research on accelerating diffusion inference. Among training-free acceleration methods, caching reduces computation by reusing previously computed model outputs across timesteps. Existing caching methods rely on heuristic criteria to choose cache/reuse timesteps and require extensive tuning. We address this limitation with a principled sensitivity-aware caching framework. Specifically, we formalize the caching error through an analysis of the model output sensitivity to perturbations in the denoising inputs, i.e., the noisy latent and the timestep, and show that this sensitivity is a key predictor of caching error. Based on this analysis, we propose Sensitivity-Aware Caching (SenCache), a dynamic caching policy that adaptively selects caching timesteps on a per-sample basis. Our framework provides a theoretical basis for adaptive caching, explains why prior empirical heuristics can be partially effective, and extends them to a dynamic, sample-specific approach. Experiments on Wan 2.1, CogVideoX, and LTX-Video show that SenCache achieves better visual quality than existing caching methods under similar computational budgets.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- AdaCorrection: Adaptive Offset Cache Correction for Accurate Diffusion Transformers (2026)
- SeaCache: Spectral-Evolution-Aware Cache for Accelerating Diffusion Models (2026)
- Denoising as Path Planning: Training-Free Acceleration of Diffusion Models with DPCache (2026)
- Predict to Skip: Linear Multistep Feature Forecasting for Efficient Diffusion Transformers (2026)
- Forecast the Principal, Stabilize the Residual: Subspace-Aware Feature Caching for Efficient Diffusion Transformers (2026)
- Relational Feature Caching for Accelerating Diffusion Transformers (2026)
- CHAI: CacHe Attention Inference for text2video (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper