Skip to content
28 changes: 6 additions & 22 deletions src/diffusers/pipelines/pipeline_loading_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -191,37 +191,21 @@ def variant_compatible_siblings(filenames, variant=None) -> Union[List[os.PathLi
# all variant filenames will be used by default
usable_filenames = set(variant_filenames)

def convert_to_variant(filename):
if "index" in filename:
variant_filename = filename.replace("index", f"index.{variant}")
elif re.compile(f"^(.*?){transformers_index_format}").match(filename) is not None:
variant_filename = f"{filename.split('-')[0]}.{variant}-{'-'.join(filename.split('-')[1:])}"
else:
variant_filename = f"{filename.split('.')[0]}.{variant}.{filename.split('.')[1]}"
return variant_filename

def find_component(filename):
if not len(filename.split("/")) == 2:
return
component = filename.split("/")[0]
return component

def has_sharded_variant(component, variant, variant_filenames):
# If component exists check for sharded variant index filename
# If component doesn't exist check main dir for sharded variant index filename
def has_variant(filename, variant_filenames):
component = find_component(filename)
component = component + "/" if component else ""
variant_index_re = re.compile(
rf"{component}({'|'.join(weight_prefixes)})\.({'|'.join(weight_suffixs)})\.index\.{variant}\.json$"
)
return any(f for f in variant_filenames if variant_index_re.match(f) is not None)

for filename in non_variant_filenames:
if convert_to_variant(filename) in variant_filenames:
continue
return any(f.startswith(component) for f in variant_filenames)
Copy link
Collaborator

@yiyixuxu yiyixuxu Jan 28, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

what happends if like we only have a bf16.bin and this is a non-variant safetensors?

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

As in we are trying to fetch something like this?

        variant = "fp16"
        filenames = [
                f"vae/diffusion_pytorch_model.{variant}.bin",
                f"text_encoder/model.{variant}.bin",
                f"unet/diffusion_pytorch_model.{variant}.bin",
        ]
        model_filenames, variant_filenames = variant_compatible_siblings(filenames, variant=None)

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

like this, I think we should fetch the non-variant safetensors in this case, no?

    variant = "fp16"
    filenames = [
            f"vae/diffusion_pytorch_model.{variant}.bin",
            f"text_encoder/model.{variant}.bin",
            f"unet/diffusion_pytorch_model.{variant}.bin",
            f"vae/diffusion_pytorch_model.safetensors",
            f"text_encoder/model.safetensors",
            f"unet/diffusion_pytorch_model.safetensors",
    ]

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hmm currently the behaviour on main is to return all the files in that list (both bin and safetensors) as usable_filenames and I think the ignore patterns would remove the bin files, resulting in just the safetensors being downloaded.

With this change only the fp16.bin files would be downloaded. Which feels technically "correct" to me since they are the "variant" files of each component. IMO non-variants should only be downloaded if no variant exists (regardless of format)

But this case implies that the proposal here is a breaking change, so I'll update to account for it.


component = find_component(filename)
# If a sharded variant exists skip adding to allowed patterns
if has_sharded_variant(component, variant, variant_filenames):
for filename in non_variant_filenames:
# If a variant exists skip adding to allowed patterns
if has_variant(filename, variant_filenames):
continue

usable_filenames.add(filename)
Expand Down
79 changes: 75 additions & 4 deletions tests/pipelines/test_pipeline_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -259,8 +259,6 @@ def test_non_variants_in_main_dir_downloaded(self):
"diffusion_pytorch_model.safetensors",
"model.safetensors",
f"model.{variant}.safetensors",
f"diffusion_pytorch_model.{variant}.safetensors",
"diffusion_pytorch_model.safetensors",
]
model_filenames, variant_filenames = variant_compatible_siblings(filenames, variant=None)
assert all(variant not in f for f in model_filenames)
Expand All @@ -285,12 +283,50 @@ def test_mixed_variants_in_main_dir_downloaded(self):
f"diffusion_pytorch_model.{variant}.safetensors",
"diffusion_pytorch_model.safetensors",
"model.safetensors",
f"diffusion_pytorch_model.{variant}.safetensors",
"diffusion_pytorch_model.safetensors",
]
model_filenames, variant_filenames = variant_compatible_siblings(filenames, variant=variant)
assert all(variant in f if f != non_variant_file else variant not in f for f in model_filenames)

def test_sharded_variants_in_main_dir_downloaded(self):
variant = "fp16"
filenames = [
"diffusion_pytorch_model.safetensors.index.json",
"diffusion_pytorch_model-00001-of-00003.safetensors",
"diffusion_pytorch_model-00002-of-00003.safetensors",
"diffusion_pytorch_model-00003-of-00003.safetensors",
f"diffusion_pytorch_model.{variant}-00001-of-00002.safetensors",
f"diffusion_pytorch_model.{variant}-00002-of-00002.safetensors",
f"diffusion_pytorch_model.safetensors.index.{variant}.json",
]
model_filenames, variant_filenames = variant_compatible_siblings(filenames, variant=variant)
assert all(variant in f for f in model_filenames)

def test_mixed_sharded_and_variant_in_main_dir_downloaded(self):
variant = "fp16"
filenames = [
"diffusion_pytorch_model.safetensors.index.json",
"diffusion_pytorch_model-00001-of-00003.safetensors",
"diffusion_pytorch_model-00002-of-00003.safetensors",
"diffusion_pytorch_model-00003-of-00003.safetensors",
f"diffusion_pytorch_model.{variant}.safetensors",
]
model_filenames, variant_filenames = variant_compatible_siblings(filenames, variant=variant)
assert all(variant in f for f in model_filenames)

def test_mixed_sharded_non_variants_in_main_dir_downloaded(self):
variant = "fp16"
filenames = [
f"diffusion_pytorch_model.safetensors.index.{variant}.json",
"diffusion_pytorch_model.safetensors.index.json",
"diffusion_pytorch_model-00001-of-00003.safetensors",
"diffusion_pytorch_model-00002-of-00003.safetensors",
"diffusion_pytorch_model-00003-of-00003.safetensors",
f"diffusion_pytorch_model.{variant}-00001-of-00002.safetensors",
f"diffusion_pytorch_model.{variant}-00002-of-00002.safetensors",
]
model_filenames, variant_filenames = variant_compatible_siblings(filenames, variant=None)
assert all(variant not in f for f in model_filenames)

def test_sharded_non_variants_downloaded(self):
variant = "fp16"
filenames = [
Expand Down Expand Up @@ -319,6 +355,35 @@ def test_sharded_variants_downloaded(self):
model_filenames, variant_filenames = variant_compatible_siblings(filenames, variant=variant)
assert all(variant in f for f in model_filenames)

def test_single_variant_with_sharded_non_variant_downloaded(self):
variant = "fp16"
filenames = [
"unet/diffusion_pytorch_model.safetensors.index.json",
"unet/diffusion_pytorch_model-00001-of-00003.safetensors",
"unet/diffusion_pytorch_model-00002-of-00003.safetensors",
"unet/diffusion_pytorch_model-00003-of-00003.safetensors",
f"unet/diffusion_pytorch_model.{variant}.safetensors",
]
model_filenames, variant_filenames = variant_compatible_siblings(filenames, variant=variant)
assert all(variant in f for f in model_filenames)

def test_mixed_single_variant_with_sharded_non_variant_downloaded(self):
variant = "fp16"
allowed_non_variant = "unet"
filenames = [
"vae/diffusion_pytorch_model.safetensors.index.json",
"vae/diffusion_pytorch_model-00001-of-00003.safetensors",
"vae/diffusion_pytorch_model-00002-of-00003.safetensors",
"vae/diffusion_pytorch_model-00003-of-00003.safetensors",
f"vae/diffusion_pytorch_model.{variant}.safetensors",
"unet/diffusion_pytorch_model.safetensors.index.json",
"unet/diffusion_pytorch_model-00001-of-00003.safetensors",
"unet/diffusion_pytorch_model-00002-of-00003.safetensors",
"unet/diffusion_pytorch_model-00003-of-00003.safetensors",
]
model_filenames, variant_filenames = variant_compatible_siblings(filenames, variant=variant)
assert all(variant in f if allowed_non_variant not in f else variant not in f for f in model_filenames)

def test_sharded_mixed_variants_downloaded(self):
variant = "fp16"
allowed_non_variant = "unet"
Expand All @@ -338,6 +403,12 @@ def test_sharded_mixed_variants_downloaded(self):
model_filenames, variant_filenames = variant_compatible_siblings(filenames, variant=variant)
assert all(variant in f if allowed_non_variant not in f else variant not in f for f in model_filenames)

def test_downloading_when_no_variant_exists(self):
variant = "fp16"
filenames = ["model.safetensors", "diffusion_pytorch_model.safetensors"]
model_filenames, variant_filenames = variant_compatible_siblings(filenames, variant=variant)
assert len(model_filenames) != 0


class ProgressBarTests(unittest.TestCase):
def get_dummy_components_image_generation(self):
Expand Down