-
Couldn't load subscription status.
- Fork 6.5k
[LoRA] make set_adapters() robust on silent failures.
#9618
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 16 commits
2c250e7
45c15d3
ce4fcbf
a6f7c90
473bbad
2714043
47ef802
e75a283
5338275
1989c71
3105832
91749a1
0c9d532
2464b20
f2d2461
47172fd
33bc47a
1c864a1
d6298b5
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -199,3 +199,7 @@ def test_simple_inference_with_text_lora_fused(self): | |
| @unittest.skip("Text encoder LoRA is not supported in CogVideoX.") | ||
| def test_simple_inference_with_text_lora_save_load(self): | ||
| pass | ||
|
|
||
| @unittest.skip("Not supported in CogVideoX.") | ||
| def test_simple_inference_with_text_denoiser_multi_adapter_block_lora(self): | ||
| pass | ||
|
Comment on lines
+159
to
+161
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Now that we're catching the error appropriately within the code, we should skip this test for unsupported models. |
||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -163,6 +163,10 @@ def test_with_alpha_in_state_dict(self): | |
| ) | ||
| self.assertFalse(np.allclose(images_lora_with_alpha, images_lora, atol=1e-3, rtol=1e-3)) | ||
|
|
||
| @unittest.skip("Not supported in Flux.") | ||
| def test_simple_inference_with_text_denoiser_block_scale(self): | ||
| pass | ||
|
|
||
| @unittest.skip("Not supported in Flux.") | ||
| def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self): | ||
| pass | ||
|
|
@@ -171,6 +175,10 @@ def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(se | |
| def test_modify_padding_mode(self): | ||
| pass | ||
|
|
||
| @unittest.skip("Not supported in Flux.") | ||
| def test_simple_inference_with_text_denoiser_multi_adapter_block_lora(self): | ||
| pass | ||
|
|
||
|
|
||
| class FluxControlLoRATests(unittest.TestCase, PeftLoraLoaderMixinTests): | ||
| pipeline_class = FluxControlPipeline | ||
|
|
@@ -546,6 +554,10 @@ def test_lora_expanding_shape_with_normal_lora_raises_error(self): | |
| "adapter-2", | ||
| ) | ||
|
|
||
| @unittest.skip("Not supported in Flux.") | ||
| def test_simple_inference_with_text_denoiser_block_scale(self): | ||
| pass | ||
|
|
||
| @unittest.skip("Not supported in Flux.") | ||
| def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self): | ||
| pass | ||
|
|
@@ -554,6 +566,10 @@ def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(se | |
| def test_modify_padding_mode(self): | ||
| pass | ||
|
|
||
| @unittest.skip("Not supported in Flux.") | ||
|
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. |
||
| def test_simple_inference_with_text_denoiser_multi_adapter_block_lora(self): | ||
| pass | ||
|
|
||
|
|
||
| @slow | ||
| @nightly | ||
|
|
||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -29,6 +29,7 @@ | |
| from diffusers.utils import load_image | ||
| from diffusers.utils.import_utils import is_accelerate_available | ||
| from diffusers.utils.testing_utils import ( | ||
| is_flaky, | ||
| is_peft_available, | ||
| numpy_cosine_similarity_distance, | ||
| require_peft_backend, | ||
|
|
@@ -129,6 +130,10 @@ def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(se | |
| def test_modify_padding_mode(self): | ||
| pass | ||
|
|
||
| @is_flaky | ||
|
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Same as #9618 (comment). |
||
| def test_multiple_wrong_adapter_name_raises_error(self): | ||
| super().test_multiple_wrong_adapter_name_raises_error() | ||
|
|
||
|
|
||
| @require_torch_gpu | ||
| @require_peft_backend | ||
|
|
||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -37,6 +37,7 @@ | |
| from diffusers.utils.import_utils import is_accelerate_available | ||
| from diffusers.utils.testing_utils import ( | ||
| CaptureLogger, | ||
| is_flaky, | ||
| load_image, | ||
| nightly, | ||
| numpy_cosine_similarity_distance, | ||
|
|
@@ -111,6 +112,10 @@ def tearDown(self): | |
| gc.collect() | ||
| torch.cuda.empty_cache() | ||
|
|
||
| @is_flaky | ||
|
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. There's absolutely no reason for it to be flaky but I think okay for now. |
||
| def test_multiple_wrong_adapter_name_raises_error(self): | ||
| super().test_multiple_wrong_adapter_name_raises_error() | ||
|
|
||
|
|
||
| @slow | ||
| @nightly | ||
|
|
||
Uh oh!
There was an error while loading. Please reload this page.