Spaces:
Running
on
Zero
Running
on
Zero
Update qwenimage/transformer_qwenimage.py
Browse files
qwenimage/transformer_qwenimage.py
CHANGED
@@ -24,7 +24,7 @@ from diffusers.configuration_utils import ConfigMixin, register_to_config
|
|
24 |
from diffusers.loaders import FromOriginalModelMixin, PeftAdapterMixin
|
25 |
from diffusers.utils import USE_PEFT_BACKEND, logging, scale_lora_layers, unscale_lora_layers
|
26 |
from diffusers.utils.torch_utils import maybe_allow_in_graph
|
27 |
-
from diffusers.models.attention import FeedForward
|
28 |
from diffusers.models.attention_dispatch import dispatch_attention_fn
|
29 |
from diffusers.models.attention_processor import Attention
|
30 |
from diffusers.models.cache_utils import CacheMixin
|
@@ -469,7 +469,7 @@ class QwenImageTransformerBlock(nn.Module):
|
|
469 |
return encoder_hidden_states, hidden_states
|
470 |
|
471 |
|
472 |
-
class QwenImageTransformer2DModel(ModelMixin, ConfigMixin, PeftAdapterMixin, FromOriginalModelMixin, CacheMixin):
|
473 |
"""
|
474 |
The Transformer model introduced in Qwen.
|
475 |
|
|
|
24 |
from diffusers.loaders import FromOriginalModelMixin, PeftAdapterMixin
|
25 |
from diffusers.utils import USE_PEFT_BACKEND, logging, scale_lora_layers, unscale_lora_layers
|
26 |
from diffusers.utils.torch_utils import maybe_allow_in_graph
|
27 |
+
from diffusers.models.attention import FeedForward, AttentionMixin
|
28 |
from diffusers.models.attention_dispatch import dispatch_attention_fn
|
29 |
from diffusers.models.attention_processor import Attention
|
30 |
from diffusers.models.cache_utils import CacheMixin
|
|
|
469 |
return encoder_hidden_states, hidden_states
|
470 |
|
471 |
|
472 |
+
class QwenImageTransformer2DModel(ModelMixin, ConfigMixin, PeftAdapterMixin, FromOriginalModelMixin, CacheMixin, AttentionMixin):
|
473 |
"""
|
474 |
The Transformer model introduced in Qwen.
|
475 |
|