Skip to content

Commit 2addc4b

Browse files
authored
Merge branch 'main' into tests/add-test-slices-wan
2 parents 3f6fb29 + 9db9be6 commit 2addc4b

File tree

10 files changed

+86
-41
lines changed

10 files changed

+86
-41
lines changed

examples/dreambooth/train_dreambooth_lora_flux_kontext.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1614,7 +1614,7 @@ def load_model_hook(models, input_dir):
16141614
)
16151615
if args.cond_image_column is not None:
16161616
logger.info("I2I fine-tuning enabled.")
1617-
batch_sampler = BucketBatchSampler(train_dataset, batch_size=args.train_batch_size, drop_last=False)
1617+
batch_sampler = BucketBatchSampler(train_dataset, batch_size=args.train_batch_size, drop_last=True)
16181618
train_dataloader = torch.utils.data.DataLoader(
16191619
train_dataset,
16201620
batch_sampler=batch_sampler,

tests/pipelines/cosmos/test_cosmos.py

Lines changed: 8 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -153,11 +153,15 @@ def test_inference(self):
153153
inputs = self.get_dummy_inputs(device)
154154
video = pipe(**inputs).frames
155155
generated_video = video[0]
156-
157156
self.assertEqual(generated_video.shape, (9, 3, 32, 32))
158-
expected_video = torch.randn(9, 3, 32, 32)
159-
max_diff = np.abs(generated_video - expected_video).max()
160-
self.assertLessEqual(max_diff, 1e10)
157+
158+
# fmt: off
159+
expected_slice = torch.tensor([0.0, 0.9686, 0.8549, 0.8078, 0.0, 0.8431, 1.0, 0.4863, 0.7098, 0.1098, 0.8157, 0.4235, 0.6353, 0.2549, 0.5137, 0.5333])
160+
# fmt: on
161+
162+
generated_slice = generated_video.flatten()
163+
generated_slice = torch.cat([generated_slice[:8], generated_slice[-8:]])
164+
self.assertTrue(torch.allclose(generated_slice, expected_slice, atol=1e-3))
161165

162166
def test_callback_inputs(self):
163167
sig = inspect.signature(self.pipeline_class.__call__)

tests/pipelines/cosmos/test_cosmos2_text2image.py

Lines changed: 8 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -140,11 +140,15 @@ def test_inference(self):
140140
inputs = self.get_dummy_inputs(device)
141141
image = pipe(**inputs).images
142142
generated_image = image[0]
143-
144143
self.assertEqual(generated_image.shape, (3, 32, 32))
145-
expected_video = torch.randn(3, 32, 32)
146-
max_diff = np.abs(generated_image - expected_video).max()
147-
self.assertLessEqual(max_diff, 1e10)
144+
145+
# fmt: off
146+
expected_slice = torch.tensor([0.451, 0.451, 0.4471, 0.451, 0.451, 0.451, 0.451, 0.451, 0.4784, 0.4784, 0.4784, 0.4784, 0.4784, 0.4902, 0.4588, 0.5333])
147+
# fmt: on
148+
149+
generated_slice = generated_image.flatten()
150+
generated_slice = torch.cat([generated_slice[:8], generated_slice[-8:]])
151+
self.assertTrue(torch.allclose(generated_slice, expected_slice, atol=1e-3))
148152

149153
def test_callback_inputs(self):
150154
sig = inspect.signature(self.pipeline_class.__call__)

tests/pipelines/cosmos/test_cosmos2_video2world.py

Lines changed: 8 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -147,11 +147,15 @@ def test_inference(self):
147147
inputs = self.get_dummy_inputs(device)
148148
video = pipe(**inputs).frames
149149
generated_video = video[0]
150-
151150
self.assertEqual(generated_video.shape, (9, 3, 32, 32))
152-
expected_video = torch.randn(9, 3, 32, 32)
153-
max_diff = np.abs(generated_video - expected_video).max()
154-
self.assertLessEqual(max_diff, 1e10)
151+
152+
# fmt: off
153+
expected_slice = torch.tensor([0.451, 0.451, 0.4471, 0.451, 0.451, 0.451, 0.451, 0.451, 0.5098, 0.5137, 0.5176, 0.5098, 0.5255, 0.5412, 0.5098, 0.5059])
154+
# fmt: on
155+
156+
generated_slice = generated_video.flatten()
157+
generated_slice = torch.cat([generated_slice[:8], generated_slice[-8:]])
158+
self.assertTrue(torch.allclose(generated_slice, expected_slice, atol=1e-3))
155159

156160
def test_components_function(self):
157161
init_components = self.get_dummy_components()

tests/pipelines/cosmos/test_cosmos_video2world.py

Lines changed: 8 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -159,11 +159,15 @@ def test_inference(self):
159159
inputs = self.get_dummy_inputs(device)
160160
video = pipe(**inputs).frames
161161
generated_video = video[0]
162-
163162
self.assertEqual(generated_video.shape, (9, 3, 32, 32))
164-
expected_video = torch.randn(9, 3, 32, 32)
165-
max_diff = np.abs(generated_video - expected_video).max()
166-
self.assertLessEqual(max_diff, 1e10)
163+
164+
# fmt: off
165+
expected_slice = torch.tensor([0.0, 0.8275, 0.7529, 0.7294, 0.0, 0.6, 1.0, 0.3804, 0.6667, 0.0863, 0.8784, 0.5922, 0.6627, 0.2784, 0.5725, 0.7765])
166+
# fmt: on
167+
168+
generated_slice = generated_video.flatten()
169+
generated_slice = torch.cat([generated_slice[:8], generated_slice[-8:]])
170+
self.assertTrue(torch.allclose(generated_slice, expected_slice, atol=1e-3))
167171

168172
def test_components_function(self):
169173
init_components = self.get_dummy_components()

tests/pipelines/hidream_image/test_pipeline_hidream.py

Lines changed: 8 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -146,11 +146,15 @@ def test_inference(self):
146146
inputs = self.get_dummy_inputs(device)
147147
image = pipe(**inputs)[0]
148148
generated_image = image[0]
149-
150149
self.assertEqual(generated_image.shape, (128, 128, 3))
151-
expected_image = torch.randn(128, 128, 3).numpy()
152-
max_diff = np.abs(generated_image - expected_image).max()
153-
self.assertLessEqual(max_diff, 1e10)
150+
151+
# fmt: off
152+
expected_slice = np.array([0.4507, 0.5256, 0.4205, 0.5791, 0.4848, 0.4831, 0.4443, 0.5107, 0.6586, 0.3163, 0.7318, 0.5933, 0.6252, 0.5512, 0.5357, 0.5983])
153+
# fmt: on
154+
155+
generated_slice = generated_image.flatten()
156+
generated_slice = np.concatenate([generated_slice[:8], generated_slice[-8:]])
157+
self.assertTrue(np.allclose(generated_slice, expected_slice, atol=1e-3))
154158

155159
def test_inference_batch_single_identical(self):
156160
super().test_inference_batch_single_identical(expected_max_diff=3e-4)

tests/pipelines/hunyuan_video/test_hunyuan_image2video.py

Lines changed: 11 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -229,12 +229,19 @@ def test_inference(self):
229229
inputs = self.get_dummy_inputs(device)
230230
video = pipe(**inputs).frames
231231
generated_video = video[0]
232-
233232
# NOTE: The expected video has 4 lesser frames because they are dropped in the pipeline
234233
self.assertEqual(generated_video.shape, (5, 3, 16, 16))
235-
expected_video = torch.randn(5, 3, 16, 16)
236-
max_diff = np.abs(generated_video - expected_video).max()
237-
self.assertLessEqual(max_diff, 1e10)
234+
235+
# fmt: off
236+
expected_slice = torch.tensor([0.444, 0.479, 0.4485, 0.5752, 0.3539, 0.1548, 0.2706, 0.3593, 0.5323, 0.6635, 0.6795, 0.5255, 0.5091, 0.345, 0.4276, 0.4128])
237+
# fmt: on
238+
239+
generated_slice = generated_video.flatten()
240+
generated_slice = torch.cat([generated_slice[:8], generated_slice[-8:]])
241+
self.assertTrue(
242+
torch.allclose(generated_slice, expected_slice, atol=1e-3),
243+
"The generated video does not match the expected slice.",
244+
)
238245

239246
def test_callback_inputs(self):
240247
sig = inspect.signature(self.pipeline_class.__call__)

tests/pipelines/hunyuan_video/test_hunyuan_skyreels_image2video.py

Lines changed: 11 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -192,11 +192,18 @@ def test_inference(self):
192192
inputs = self.get_dummy_inputs(device)
193193
video = pipe(**inputs).frames
194194
generated_video = video[0]
195-
196195
self.assertEqual(generated_video.shape, (9, 3, 16, 16))
197-
expected_video = torch.randn(9, 3, 16, 16)
198-
max_diff = np.abs(generated_video - expected_video).max()
199-
self.assertLessEqual(max_diff, 1e10)
196+
197+
# fmt: off
198+
expected_slice = torch.tensor([0.5832, 0.5498, 0.4839, 0.4744, 0.4515, 0.4832, 0.496, 0.563, 0.5918, 0.5979, 0.5101, 0.6168, 0.6613, 0.536, 0.55, 0.5775])
199+
# fmt: on
200+
201+
generated_slice = generated_video.flatten()
202+
generated_slice = torch.cat([generated_slice[:8], generated_slice[-8:]])
203+
self.assertTrue(
204+
torch.allclose(generated_slice, expected_slice, atol=1e-3),
205+
"The generated video does not match the expected slice.",
206+
)
200207

201208
def test_callback_inputs(self):
202209
sig = inspect.signature(self.pipeline_class.__call__)

tests/pipelines/hunyuan_video/test_hunyuan_video.py

Lines changed: 12 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -26,10 +26,7 @@
2626
HunyuanVideoPipeline,
2727
HunyuanVideoTransformer3DModel,
2828
)
29-
from diffusers.utils.testing_utils import (
30-
enable_full_determinism,
31-
torch_device,
32-
)
29+
from diffusers.utils.testing_utils import enable_full_determinism, torch_device
3330

3431
from ..test_pipelines_common import (
3532
FasterCacheTesterMixin,
@@ -206,11 +203,18 @@ def test_inference(self):
206203
inputs = self.get_dummy_inputs(device)
207204
video = pipe(**inputs).frames
208205
generated_video = video[0]
209-
210206
self.assertEqual(generated_video.shape, (9, 3, 16, 16))
211-
expected_video = torch.randn(9, 3, 16, 16)
212-
max_diff = np.abs(generated_video - expected_video).max()
213-
self.assertLessEqual(max_diff, 1e10)
207+
208+
# fmt: off
209+
expected_slice = torch.tensor([0.3946, 0.4649, 0.3196, 0.4569, 0.3312, 0.3687, 0.3216, 0.3972, 0.4469, 0.3888, 0.3929, 0.3802, 0.3479, 0.3888, 0.3825, 0.3542])
210+
# fmt: on
211+
212+
generated_slice = generated_video.flatten()
213+
generated_slice = torch.cat([generated_slice[:8], generated_slice[-8:]])
214+
self.assertTrue(
215+
torch.allclose(generated_slice, expected_slice, atol=1e-3),
216+
"The generated video does not match the expected slice.",
217+
)
214218

215219
def test_callback_inputs(self):
216220
sig = inspect.signature(self.pipeline_class.__call__)

tests/pipelines/hunyuan_video/test_hunyuan_video_framepack.py

Lines changed: 11 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -227,11 +227,18 @@ def test_inference(self):
227227
inputs = self.get_dummy_inputs(device)
228228
video = pipe(**inputs).frames
229229
generated_video = video[0]
230-
231230
self.assertEqual(generated_video.shape, (13, 3, 32, 32))
232-
expected_video = torch.randn(13, 3, 32, 32)
233-
max_diff = np.abs(generated_video - expected_video).max()
234-
self.assertLessEqual(max_diff, 1e10)
231+
232+
# fmt: off
233+
expected_slice = torch.tensor([0.363, 0.3384, 0.3426, 0.3512, 0.3372, 0.3276, 0.417, 0.4061, 0.5221, 0.467, 0.4813, 0.4556, 0.4107, 0.3945, 0.4049, 0.4551])
234+
# fmt: on
235+
236+
generated_slice = generated_video.flatten()
237+
generated_slice = torch.cat([generated_slice[:8], generated_slice[-8:]])
238+
self.assertTrue(
239+
torch.allclose(generated_slice, expected_slice, atol=1e-3),
240+
"The generated video does not match the expected slice.",
241+
)
235242

236243
def test_callback_inputs(self):
237244
sig = inspect.signature(self.pipeline_class.__call__)

0 commit comments

Comments
 (0)