Skip to content

Commit 6ac18d0

Browse files
committed
more
1 parent 5f8a9b6 commit 6ac18d0

10 files changed

Lines changed: 15 additions & 36 deletions

src/diffusers/pipelines/audioldm2/pipeline_audioldm2.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -223,7 +223,7 @@ def __init__(
223223
)
224224
self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
225225

226-
# Copied from diffusers.pipelines.pipeline_utils.StableDiffusionMixin.enable_vae_slicing
226+
# Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_utils.StableDiffusionMixin.enable_vae_slicing
227227
def enable_vae_slicing(self):
228228
r"""
229229
Enable sliced VAE decoding. When this option is enabled, the VAE will split the input tensor in slices to
@@ -237,7 +237,7 @@ def enable_vae_slicing(self):
237237
)
238238
self.vae.enable_slicing()
239239

240-
# Copied from diffusers.pipelines.pipeline_utils.StableDiffusionMixin.disable_vae_slicing
240+
# Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_utils.StableDiffusionMixin.disable_vae_slicing
241241
def disable_vae_slicing(self):
242242
r"""
243243
Disable sliced VAE decoding. If `enable_vae_slicing` was previously enabled, this method will go back to

src/diffusers/pipelines/controlnet/pipeline_controlnet_inpaint_sd_xl.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1136,10 +1136,7 @@ def upcast_vae(self):
11361136
self.vae.to(dtype=torch.float32)
11371137
use_torch_2_0_or_xformers = isinstance(
11381138
self.vae.decoder.mid_block.attentions[0].processor,
1139-
(
1140-
AttnProcessor2_0,
1141-
XFormersAttnProcessor,
1142-
),
1139+
(AttnProcessor2_0, XFormersAttnProcessor),
11431140
)
11441141
# if xformers or torch_2_0 is used attention block does not need
11451142
# to be in float32 which can save lots of memory

src/diffusers/pipelines/controlnet/pipeline_controlnet_sd_xl.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -933,10 +933,7 @@ def upcast_vae(self):
933933
self.vae.to(dtype=torch.float32)
934934
use_torch_2_0_or_xformers = isinstance(
935935
self.vae.decoder.mid_block.attentions[0].processor,
936-
(
937-
AttnProcessor2_0,
938-
XFormersAttnProcessor,
939-
),
936+
(AttnProcessor2_0, XFormersAttnProcessor),
940937
)
941938
# if xformers or torch_2_0 is used attention block does not need
942939
# to be in float32 which can save lots of memory

src/diffusers/pipelines/controlnet/pipeline_controlnet_sd_xl_img2img.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1048,10 +1048,7 @@ def upcast_vae(self):
10481048
self.vae.to(dtype=torch.float32)
10491049
use_torch_2_0_or_xformers = isinstance(
10501050
self.vae.decoder.mid_block.attentions[0].processor,
1051-
(
1052-
AttnProcessor2_0,
1053-
XFormersAttnProcessor,
1054-
),
1051+
(AttnProcessor2_0, XFormersAttnProcessor),
10551052
)
10561053
# if xformers or torch_2_0 is used attention block does not need
10571054
# to be in float32 which can save lots of memory

src/diffusers/pipelines/controlnet/pipeline_controlnet_union_inpaint_sd_xl.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1115,10 +1115,7 @@ def upcast_vae(self):
11151115
self.vae.to(dtype=torch.float32)
11161116
use_torch_2_0_or_xformers = isinstance(
11171117
self.vae.decoder.mid_block.attentions[0].processor,
1118-
(
1119-
AttnProcessor2_0,
1120-
XFormersAttnProcessor,
1121-
),
1118+
(AttnProcessor2_0, XFormersAttnProcessor),
11221119
)
11231120
# if xformers or torch_2_0 is used attention block does not need
11241121
# to be in float32 which can save lots of memory

src/diffusers/pipelines/controlnet/pipeline_controlnet_union_sd_xl.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -908,10 +908,7 @@ def upcast_vae(self):
908908
self.vae.to(dtype=torch.float32)
909909
use_torch_2_0_or_xformers = isinstance(
910910
self.vae.decoder.mid_block.attentions[0].processor,
911-
(
912-
AttnProcessor2_0,
913-
XFormersAttnProcessor,
914-
),
911+
(AttnProcessor2_0, XFormersAttnProcessor),
915912
)
916913
# if xformers or torch_2_0 is used attention block does not need
917914
# to be in float32 which can save lots of memory

src/diffusers/pipelines/controlnet/pipeline_controlnet_union_sd_xl_img2img.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1036,10 +1036,7 @@ def upcast_vae(self):
10361036
self.vae.to(dtype=torch.float32)
10371037
use_torch_2_0_or_xformers = isinstance(
10381038
self.vae.decoder.mid_block.attentions[0].processor,
1039-
(
1040-
AttnProcessor2_0,
1041-
XFormersAttnProcessor,
1042-
),
1039+
(AttnProcessor2_0, XFormersAttnProcessor),
10431040
)
10441041
# if xformers or torch_2_0 is used attention block does not need
10451042
# to be in float32 which can save lots of memory

src/diffusers/pipelines/ledits_pp/pipeline_leditspp_stable_diffusion_xl.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -692,10 +692,7 @@ def upcast_vae(self):
692692
self.vae.to(dtype=torch.float32)
693693
use_torch_2_0_or_xformers = isinstance(
694694
self.vae.decoder.mid_block.attentions[0].processor,
695-
(
696-
AttnProcessor2_0,
697-
XFormersAttnProcessor,
698-
),
695+
(AttnProcessor2_0, XFormersAttnProcessor),
699696
)
700697
# if xformers or torch_2_0 is used attention block does not need
701698
# to be in float32 which can save lots of memory

src/diffusers/pipelines/stable_audio/pipeline_stable_audio.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -124,7 +124,7 @@ def __init__(
124124
)
125125
self.rotary_embed_dim = self.transformer.config.attention_head_dim // 2
126126

127-
# Copied from diffusers.pipelines.pipeline_utils.StableDiffusionMixin.enable_vae_slicing
127+
# Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_utils.StableDiffusionMixin.enable_vae_slicing
128128
def enable_vae_slicing(self):
129129
r"""
130130
Enable sliced VAE decoding. When this option is enabled, the VAE will split the input tensor in slices to
@@ -138,7 +138,7 @@ def enable_vae_slicing(self):
138138
)
139139
self.vae.enable_slicing()
140140

141-
# Copied from diffusers.pipelines.pipeline_utils.StableDiffusionMixin.disable_vae_slicing
141+
# Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_utils.StableDiffusionMixin.disable_vae_slicing
142142
def disable_vae_slicing(self):
143143
r"""
144144
Disable sliced VAE decoding. If `enable_vae_slicing` was previously enabled, this method will go back to

src/diffusers/pipelines/unidiffuser/pipeline_unidiffuser.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -226,7 +226,7 @@ def _infer_mode(self, prompt, prompt_embeds, image, latents, prompt_latents, vae
226226

227227
return mode
228228

229-
# Copied from diffusers.pipelines.pipeline_utils.StableDiffusionMixin.enable_vae_slicing
229+
# Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_utils.StableDiffusionMixin.enable_vae_slicing
230230
def enable_vae_slicing(self):
231231
r"""
232232
Enable sliced VAE decoding. When this option is enabled, the VAE will split the input tensor in slices to
@@ -240,7 +240,7 @@ def enable_vae_slicing(self):
240240
)
241241
self.vae.enable_slicing()
242242

243-
# Copied from diffusers.pipelines.pipeline_utils.StableDiffusionMixin.disable_vae_slicing
243+
# Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_utils.StableDiffusionMixin.disable_vae_slicing
244244
def disable_vae_slicing(self):
245245
r"""
246246
Disable sliced VAE decoding. If `enable_vae_slicing` was previously enabled, this method will go back to
@@ -254,7 +254,7 @@ def disable_vae_slicing(self):
254254
)
255255
self.vae.disable_slicing()
256256

257-
# Copied from diffusers.pipelines.pipeline_utils.StableDiffusionMixin.enable_vae_tiling
257+
# Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_utils.StableDiffusionMixin.enable_vae_tiling
258258
def enable_vae_tiling(self):
259259
r"""
260260
Enable tiled VAE decoding. When this option is enabled, the VAE will split the input tensor into tiles to
@@ -269,7 +269,7 @@ def enable_vae_tiling(self):
269269
)
270270
self.vae.enable_tiling()
271271

272-
# Copied from diffusers.pipelines.pipeline_utils.StableDiffusionMixin.disable_vae_tiling
272+
# Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_utils.StableDiffusionMixin.disable_vae_tiling
273273
def disable_vae_tiling(self):
274274
r"""
275275
Disable tiled VAE decoding. If `enable_vae_tiling` was previously enabled, this method will go back to

0 commit comments

Comments
 (0)