From a44f962b8e5ff81866f4cad688691b952a67b656 Mon Sep 17 00:00:00 2001 From: huangzehuan Date: Thu, 9 Jan 2025 12:13:41 +0800 Subject: [PATCH] remove unused kwargs --- src/diffusers/models/attention_processor.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/src/diffusers/models/attention_processor.py b/src/diffusers/models/attention_processor.py index fef1593b673b..4d7ae6bef26e 100644 --- a/src/diffusers/models/attention_processor.py +++ b/src/diffusers/models/attention_processor.py @@ -2813,7 +2813,6 @@ def __call__( encoder_hidden_states: torch.Tensor, attention_mask: Optional[torch.Tensor] = None, image_rotary_emb: Optional[torch.Tensor] = None, - **kwargs, ) -> torch.Tensor: text_seq_length = encoder_hidden_states.size(1) @@ -2885,7 +2884,6 @@ def __call__( encoder_hidden_states: torch.Tensor, attention_mask: Optional[torch.Tensor] = None, image_rotary_emb: Optional[torch.Tensor] = None, - **kwargs, ) -> torch.Tensor: text_seq_length = encoder_hidden_states.size(1)