From f394604bb587e4493115ff756e4ad7303f5c1454 Mon Sep 17 00:00:00 2001 From: "Liu, Kaixuan" Date: Mon, 13 Oct 2025 06:14:11 +0000 Subject: [PATCH 1/2] fix CI bug for kandinsky3_img2img case Signed-off-by: Liu, Kaixuan --- tests/pipelines/test_pipelines_common.py | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/tests/pipelines/test_pipelines_common.py b/tests/pipelines/test_pipelines_common.py index db8209835be4..d5adc374e73c 100644 --- a/tests/pipelines/test_pipelines_common.py +++ b/tests/pipelines/test_pipelines_common.py @@ -2124,6 +2124,10 @@ def test_encode_prompt_works_in_isolation(self, extra_required_param_value_dict= f"encode_prompt has no default in either encode_prompt or __call__." ) + if "_cut_context" in encode_prompt_param_names and "_cut_context" not in encode_prompt_inputs: + # As in full_pipeline, `_cut_context` is set to True. + encode_prompt_inputs["_cut_context"] = True + # Compute `encode_prompt()`. with torch.no_grad(): encoded_prompt_outputs = pipe_with_just_text_encoder.encode_prompt(**encode_prompt_inputs) From 6bb48a81b315be55a6d0f2da954c90c3719e08c1 Mon Sep 17 00:00:00 2001 From: "Liu, Kaixuan" Date: Thu, 23 Oct 2025 11:21:06 +0800 Subject: [PATCH 2/2] update code Signed-off-by: Liu, Kaixuan --- .../pipelines/kandinsky3/pipeline_kandinsky3_img2img.py | 2 +- tests/pipelines/test_pipelines_common.py | 4 ---- 2 files changed, 1 insertion(+), 5 deletions(-) diff --git a/src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3_img2img.py b/src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3_img2img.py index c7b8022c22b1..73c268897502 100644 --- a/src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3_img2img.py +++ b/src/diffusers/pipelines/kandinsky3/pipeline_kandinsky3_img2img.py @@ -113,7 +113,7 @@ def encode_prompt( negative_prompt=None, prompt_embeds: Optional[torch.Tensor] = None, negative_prompt_embeds: Optional[torch.Tensor] = None, - _cut_context=False, + _cut_context=True, attention_mask: Optional[torch.Tensor] = None, negative_attention_mask: Optional[torch.Tensor] = None, ): diff --git a/tests/pipelines/test_pipelines_common.py b/tests/pipelines/test_pipelines_common.py index d5adc374e73c..db8209835be4 100644 --- a/tests/pipelines/test_pipelines_common.py +++ b/tests/pipelines/test_pipelines_common.py @@ -2124,10 +2124,6 @@ def test_encode_prompt_works_in_isolation(self, extra_required_param_value_dict= f"encode_prompt has no default in either encode_prompt or __call__." ) - if "_cut_context" in encode_prompt_param_names and "_cut_context" not in encode_prompt_inputs: - # As in full_pipeline, `_cut_context` is set to True. - encode_prompt_inputs["_cut_context"] = True - # Compute `encode_prompt()`. with torch.no_grad(): encoded_prompt_outputs = pipe_with_just_text_encoder.encode_prompt(**encode_prompt_inputs)