Skip to content

Commit 4ffc993

Browse files
authored
Merge branch 'main' into enable-compilation
2 parents ebd8289 + b793deb commit 4ffc993

File tree

3 files changed

+14
-5
lines changed

3 files changed

+14
-5
lines changed

src/diffusers/pipelines/audioldm2/pipeline_audioldm2.py

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -312,15 +312,14 @@ def generate_language_model(
312312
The sequence of generated hidden-states.
313313
"""
314314
cache_position_kwargs = {}
315-
if is_transformers_version("<", "4.52.0.dev0"):
315+
if is_transformers_version("<", "4.52.1"):
316316
cache_position_kwargs["input_ids"] = inputs_embeds
317-
cache_position_kwargs["model_kwargs"] = model_kwargs
318317
else:
319318
cache_position_kwargs["seq_length"] = inputs_embeds.shape[0]
320319
cache_position_kwargs["device"] = (
321320
self.language_model.device if getattr(self, "language_model", None) is not None else self.device
322321
)
323-
cache_position_kwargs["model_kwargs"] = model_kwargs
322+
cache_position_kwargs["model_kwargs"] = model_kwargs
324323
max_new_tokens = max_new_tokens if max_new_tokens is not None else self.language_model.config.max_new_tokens
325324
model_kwargs = self.language_model._get_initial_cache_position(**cache_position_kwargs)
326325

tests/pipelines/audioldm2/test_audioldm2.py

Lines changed: 11 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -45,6 +45,7 @@
4545
LMSDiscreteScheduler,
4646
PNDMScheduler,
4747
)
48+
from diffusers.utils import is_transformers_version
4849
from diffusers.utils.testing_utils import (
4950
backend_empty_cache,
5051
enable_full_determinism,
@@ -220,6 +221,11 @@ def get_dummy_inputs(self, device, seed=0):
220221
}
221222
return inputs
222223

224+
@pytest.mark.xfail(
225+
condition=is_transformers_version(">=", "4.54.1"),
226+
reason="Test currently fails on Transformers version 4.54.1.",
227+
strict=False,
228+
)
223229
def test_audioldm2_ddim(self):
224230
device = "cpu" # ensure determinism for the device-dependent torch.Generator
225231

@@ -312,7 +318,6 @@ def test_audioldm2_negative_prompt_embeds(self):
312318
components = self.get_dummy_components()
313319
audioldm_pipe = AudioLDM2Pipeline(**components)
314320
audioldm_pipe = audioldm_pipe.to(torch_device)
315-
audioldm_pipe = audioldm_pipe.to(torch_device)
316321
audioldm_pipe.set_progress_bar_config(disable=None)
317322

318323
inputs = self.get_dummy_inputs(torch_device)
@@ -371,6 +376,11 @@ def test_audioldm2_negative_prompt_embeds(self):
371376

372377
assert np.abs(audio_1 - audio_2).max() < 1e-2
373378

379+
@pytest.mark.xfail(
380+
condition=is_transformers_version(">=", "4.54.1"),
381+
reason="Test currently fails on Transformers version 4.54.1.",
382+
strict=False,
383+
)
374384
def test_audioldm2_negative_prompt(self):
375385
device = "cpu" # ensure determinism for the device-dependent torch.Generator
376386
components = self.get_dummy_components()

tests/pipelines/qwenimage/test_qwenimage.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -160,7 +160,7 @@ def test_inference(self):
160160
self.assertEqual(generated_image.shape, (3, 32, 32))
161161

162162
# fmt: off
163-
expected_slice = torch.tensor([0.563, 0.6358, 0.6028, 0.5656, 0.5806, 0.5512, 0.5712, 0.6331, 0.4147, 0.3558, 0.5625, 0.4831, 0.4957, 0.5258, 0.4075, 0.5018])
163+
expected_slice = torch.tensor([0.56331, 0.63677, 0.6015, 0.56369, 0.58166, 0.55277, 0.57176, 0.63261, 0.41466, 0.35561, 0.56229, 0.48334, 0.49714, 0.52622, 0.40872, 0.50208])
164164
# fmt: on
165165

166166
generated_slice = generated_image.flatten()

0 commit comments

Comments
 (0)