diff --git a/src/diffusers/models/attention.py b/src/diffusers/models/attention.py index f9e43e4dd585..1d359864ba85 100644 --- a/src/diffusers/models/attention.py +++ b/src/diffusers/models/attention.py @@ -473,7 +473,8 @@ def set_use_memory_efficient_attention_xformers(self, use_memory_efficient_atten except Exception as e: raise e self.attn1._use_memory_efficient_attention_xformers = use_memory_efficient_attention_xformers - self.attn2._use_memory_efficient_attention_xformers = use_memory_efficient_attention_xformers + if self.attn2 is not None: + self.attn2._use_memory_efficient_attention_xformers = use_memory_efficient_attention_xformers def forward(self, hidden_states, encoder_hidden_states=None, timestep=None, attention_mask=None): # 1. Self-Attention