Spaces:
Running
on
Zero
Running
on
Zero
A newer version of the Gradio SDK is available:
5.34.2
Caching methods
Cache methods speedup diffusion transformers by storing and reusing intermediate outputs of specific layers, such as attention and feedforward layers, instead of recalculating them at each inference step.
CacheMixin
[[autodoc]] CacheMixin
PyramidAttentionBroadcastConfig
[[autodoc]] PyramidAttentionBroadcastConfig
[[autodoc]] apply_pyramid_attention_broadcast
FasterCacheConfig
[[autodoc]] FasterCacheConfig
[[autodoc]] apply_faster_cache