1717from typing import Any , Dict , List , Optional , Tuple , Union
1818
1919import torch
20- from transformers import (
20+ from transformers import (
2121 CLIPTextModel ,
2222 CLIPTextModelWithProjection ,
23- CLIPTokenizer ,
23+ CLIPTokenizer ,
2424)
2525
2626from diffusers .image_processor import VaeImageProcessor
2727from diffusers .loaders import (
28- FromSingleFileMixin ,
28+ FromSingleFileMixin ,
2929 StableDiffusionXLLoraLoaderMixin ,
3030 TextualInversionLoaderMixin ,
3131)
@@ -244,7 +244,7 @@ class StableDiffusionXLTilingPipeline(
244244 StableDiffusionMixin ,
245245 FromSingleFileMixin ,
246246 StableDiffusionXLLoraLoaderMixin ,
247- TextualInversionLoaderMixin ,
247+ TextualInversionLoaderMixin ,
248248):
249249 r"""
250250 Pipeline for text-to-image generation using Stable Diffusion XL.
@@ -256,7 +256,7 @@ class StableDiffusionXLTilingPipeline(
256256 - [`~loaders.TextualInversionLoaderMixin.load_textual_inversion`] for loading textual inversion embeddings
257257 - [`~loaders.FromSingleFileMixin.from_single_file`] for loading `.ckpt` files
258258 - [`~loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights`] for loading LoRA weights
259- - [`~loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights`] for saving LoRA weights
259+ - [`~loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights`] for saving LoRA weights
260260
261261 Args:
262262 vae ([`AutoencoderKL`]):
@@ -295,7 +295,7 @@ class StableDiffusionXLTilingPipeline(
295295 "tokenizer" ,
296296 "tokenizer_2" ,
297297 "text_encoder" ,
298- "text_encoder_2" ,
298+ "text_encoder_2" ,
299299 ]
300300
301301 def __init__ (
@@ -306,7 +306,7 @@ def __init__(
306306 tokenizer : CLIPTokenizer ,
307307 tokenizer_2 : CLIPTokenizer ,
308308 unet : UNet2DConditionModel ,
309- scheduler : KarrasDiffusionSchedulers ,
309+ scheduler : KarrasDiffusionSchedulers ,
310310 force_zeros_for_empty_prompt : bool = True ,
311311 add_watermarker : Optional [bool ] = None ,
312312 ):
@@ -319,7 +319,7 @@ def __init__(
319319 tokenizer = tokenizer ,
320320 tokenizer_2 = tokenizer_2 ,
321321 unet = unet ,
322- scheduler = scheduler ,
322+ scheduler = scheduler ,
323323 )
324324 self .register_to_config (force_zeros_for_empty_prompt = force_zeros_for_empty_prompt )
325325 self .vae_scale_factor = 2 ** (len (self .vae .config .block_out_channels ) - 1 ) if getattr (self , "vae" , None ) else 8
0 commit comments