Skip to content

Commit

Permalink
Merge branch 'huggingface:main' into xpu-enabling
Browse files Browse the repository at this point in the history
  • Loading branch information
faaany authored Jan 9, 2025
2 parents 286fa53 + daf9d0f commit a2ee718
Show file tree
Hide file tree
Showing 15 changed files with 16 additions and 40 deletions.
2 changes: 0 additions & 2 deletions tests/models/transformers/test_models_prior.py
Original file line number Diff line number Diff line change
Expand Up @@ -132,7 +132,6 @@ def test_output_pretrained(self):
output = model(**input)[0]

output_slice = output[0, :5].flatten().cpu()
print(output_slice)

# Since the VAE Gaussian prior's generator is seeded on the appropriate device,
# the expected output slices are not the same for CPU and GPU.
Expand Down Expand Up @@ -182,7 +181,6 @@ def test_kandinsky_prior(self, seed, expected_slice):
assert list(sample.shape) == [1, 768]

output_slice = sample[0, :8].flatten().cpu()
print(output_slice)
expected_output_slice = torch.tensor(expected_slice)

assert torch_all_close(output_slice, expected_output_slice, atol=1e-3)
3 changes: 1 addition & 2 deletions tests/models/unets/test_models_unet_2d_condition.py
Original file line number Diff line number Diff line change
Expand Up @@ -175,8 +175,7 @@ def create_ip_adapter_plus_state_dict(model):
)

ip_image_projection_state_dict = OrderedDict()
keys = [k for k in image_projection.state_dict() if "layers." in k]
print(keys)

for k, v in image_projection.state_dict().items():
if "2.to" in k:
k = k.replace("2.to", "0.to")
Expand Down
4 changes: 2 additions & 2 deletions tests/pipelines/controlnet/test_flax_controlnet.py
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ def test_canny(self):
expected_slice = jnp.array(
[0.167969, 0.116699, 0.081543, 0.154297, 0.132812, 0.108887, 0.169922, 0.169922, 0.205078]
)
print(f"output_slice: {output_slice}")

assert jnp.abs(output_slice - expected_slice).max() < 1e-2

def test_pose(self):
Expand Down Expand Up @@ -123,5 +123,5 @@ def test_pose(self):
expected_slice = jnp.array(
[[0.271484, 0.261719, 0.275391, 0.277344, 0.279297, 0.291016, 0.294922, 0.302734, 0.302734]]
)
print(f"output_slice: {output_slice}")

assert jnp.abs(output_slice - expected_slice).max() < 1e-2
2 changes: 0 additions & 2 deletions tests/pipelines/kandinsky/test_kandinsky_combined.py
Original file line number Diff line number Diff line change
Expand Up @@ -308,8 +308,6 @@ def test_kandinsky(self):

image_from_tuple_slice = image_from_tuple[0, -3:, -3:, -1]

print(image_from_tuple_slice)

assert image.shape == (1, 64, 64, 3)

expected_slice = np.array([0.0320, 0.0860, 0.4013, 0.0518, 0.2484, 0.5847, 0.4411, 0.2321, 0.4593])
Expand Down
6 changes: 3 additions & 3 deletions tests/pipelines/ledits_pp/test_ledits_pp_stable_diffusion.py
Original file line number Diff line number Diff line change
Expand Up @@ -146,7 +146,7 @@ def test_ledits_pp_inversion(self):
)

latent_slice = sd_pipe.init_latents[0, -1, -3:, -3:].to(device)
print(latent_slice.flatten())

expected_slice = np.array([-0.9084, -0.0367, 0.2940, 0.0839, 0.6890, 0.2651, -0.7104, 2.1090, -0.7822])
assert np.abs(latent_slice.flatten() - expected_slice).max() < 1e-3

Expand All @@ -167,12 +167,12 @@ def test_ledits_pp_inversion_batch(self):
)

latent_slice = sd_pipe.init_latents[0, -1, -3:, -3:].to(device)
print(latent_slice.flatten())

expected_slice = np.array([0.2528, 0.1458, -0.2166, 0.4565, -0.5657, -1.0286, -0.9961, 0.5933, 1.1173])
assert np.abs(latent_slice.flatten() - expected_slice).max() < 1e-3

latent_slice = sd_pipe.init_latents[1, -1, -3:, -3:].to(device)
print(latent_slice.flatten())

expected_slice = np.array([-0.0796, 2.0583, 0.5501, 0.5358, 0.0282, -0.2803, -1.0470, 0.7023, -0.0072])
assert np.abs(latent_slice.flatten() - expected_slice).max() < 1e-3

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -216,14 +216,14 @@ def test_ledits_pp_inversion_batch(self):
)

latent_slice = sd_pipe.init_latents[0, -1, -3:, -3:].to(device)
print(latent_slice.flatten())

expected_slice = np.array([0.2528, 0.1458, -0.2166, 0.4565, -0.5656, -1.0286, -0.9961, 0.5933, 1.1172])
assert np.abs(latent_slice.flatten() - expected_slice).max() < 1e-3

latent_slice = sd_pipe.init_latents[1, -1, -3:, -3:].to(device)
print(latent_slice.flatten())

expected_slice = np.array([-0.0796, 2.0583, 0.5500, 0.5358, 0.0282, -0.2803, -1.0470, 0.7024, -0.0072])
print(latent_slice.flatten())

assert np.abs(latent_slice.flatten() - expected_slice).max() < 1e-3

def test_ledits_pp_warmup_steps(self):
Expand Down
3 changes: 1 addition & 2 deletions tests/pipelines/pag/test_pag_sd.py
Original file line number Diff line number Diff line change
Expand Up @@ -318,7 +318,7 @@ def test_pag_cfg(self):

image_slice = image[0, -3:, -3:, -1].flatten()
assert image.shape == (1, 512, 512, 3)
print(image_slice.flatten())

expected_slice = np.array(
[0.58251953, 0.5722656, 0.5683594, 0.55029297, 0.52001953, 0.52001953, 0.49951172, 0.45410156, 0.50146484]
)
Expand All @@ -339,7 +339,6 @@ def test_pag_uncond(self):
expected_slice = np.array(
[0.5986328, 0.52441406, 0.3972168, 0.4741211, 0.34985352, 0.22705078, 0.4128418, 0.2866211, 0.31713867]
)
print(image_slice.flatten())
assert (
np.abs(image_slice.flatten() - expected_slice).max() < 1e-3
), f"output is different from expected, {image_slice.flatten()}"
4 changes: 2 additions & 2 deletions tests/pipelines/pag/test_pag_sd_img2img.py
Original file line number Diff line number Diff line change
Expand Up @@ -255,7 +255,7 @@ def test_pag_cfg(self):

image_slice = image[0, -3:, -3:, -1].flatten()
assert image.shape == (1, 512, 512, 3)
print(image_slice.flatten())

expected_slice = np.array(
[0.58251953, 0.5722656, 0.5683594, 0.55029297, 0.52001953, 0.52001953, 0.49951172, 0.45410156, 0.50146484]
)
Expand All @@ -276,7 +276,7 @@ def test_pag_uncond(self):
expected_slice = np.array(
[0.5986328, 0.52441406, 0.3972168, 0.4741211, 0.34985352, 0.22705078, 0.4128418, 0.2866211, 0.31713867]
)
print(image_slice.flatten())

assert (
np.abs(image_slice.flatten() - expected_slice).max() < 1e-3
), f"output is different from expected, {image_slice.flatten()}"
2 changes: 1 addition & 1 deletion tests/pipelines/pag/test_pag_sd_inpaint.py
Original file line number Diff line number Diff line change
Expand Up @@ -292,7 +292,7 @@ def test_pag_cfg(self):

image_slice = image[0, -3:, -3:, -1].flatten()
assert image.shape == (1, 512, 512, 3)
print(image_slice.flatten())

expected_slice = np.array(
[0.38793945, 0.4111328, 0.47924805, 0.39208984, 0.4165039, 0.41674805, 0.37060547, 0.36791992, 0.40625]
)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -206,9 +206,6 @@ def test_stable_diffusion_pix2pix_euler(self):
image = sd_pipe(**inputs).images
image_slice = image[0, -3:, -3:, -1]

slice = [round(x, 4) for x in image_slice.flatten().tolist()]
print(",".join([str(x) for x in slice]))

assert image.shape == (1, 32, 32, 3)
expected_slice = np.array([0.7417, 0.3842, 0.4732, 0.5776, 0.5891, 0.5139, 0.4052, 0.5673, 0.4986])

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,7 @@ def test_stable_diffusion_flax(self):

output_slice = jnp.asarray(jax.device_get(image_slice.flatten()))
expected_slice = jnp.array([0.4238, 0.4414, 0.4395, 0.4453, 0.4629, 0.4590, 0.4531, 0.45508, 0.4512])
print(f"output_slice: {output_slice}")

assert jnp.abs(output_slice - expected_slice).max() < 1e-2


Expand Down Expand Up @@ -104,5 +104,5 @@ def test_stable_diffusion_dpm_flax(self):

output_slice = jnp.asarray(jax.device_get(image_slice.flatten()))
expected_slice = jnp.array([0.4336, 0.42969, 0.4453, 0.4199, 0.4297, 0.4531, 0.4434, 0.4434, 0.4297])
print(f"output_slice: {output_slice}")

assert jnp.abs(output_slice - expected_slice).max() < 1e-2
Original file line number Diff line number Diff line change
Expand Up @@ -78,5 +78,5 @@ def test_stable_diffusion_inpaint_pipeline(self):
expected_slice = jnp.array(
[0.3611307, 0.37649736, 0.3757408, 0.38213953, 0.39295167, 0.3841631, 0.41554978, 0.4137475, 0.4217084]
)
print(f"output_slice: {output_slice}")

assert jnp.abs(output_slice - expected_slice).max() < 1e-2
Original file line number Diff line number Diff line change
Expand Up @@ -642,9 +642,6 @@ def test_adapter_sdxl_lcm(self):
assert image.shape == (1, 64, 64, 3)
expected_slice = np.array([0.5313, 0.5375, 0.4942, 0.5021, 0.6142, 0.4968, 0.5434, 0.5311, 0.5448])

debug = [str(round(i, 4)) for i in image_slice.flatten().tolist()]
print(",".join(debug))

assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2

def test_adapter_sdxl_lcm_custom_timesteps(self):
Expand All @@ -667,7 +664,4 @@ def test_adapter_sdxl_lcm_custom_timesteps(self):
assert image.shape == (1, 64, 64, 3)
expected_slice = np.array([0.5313, 0.5375, 0.4942, 0.5021, 0.6142, 0.4968, 0.5434, 0.5311, 0.5448])

debug = [str(round(i, 4)) for i in image_slice.flatten().tolist()]
print(",".join(debug))

assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2
1 change: 0 additions & 1 deletion tests/pipelines/test_pipelines_common.py
Original file line number Diff line number Diff line change
Expand Up @@ -1192,7 +1192,6 @@ def _test_inference_batch_consistent(

logger.setLevel(level=diffusers.logging.WARNING)
for batch_size, batched_input in zip(batch_sizes, batched_inputs):
print(batch_size, batched_input)
output = pipe(**batched_input)
assert len(output[0]) == batch_size

Expand Down
8 changes: 0 additions & 8 deletions tests/schedulers/test_scheduler_sasolver.py
Original file line number Diff line number Diff line change
Expand Up @@ -103,8 +103,6 @@ def test_full_loop_no_noise(self):
elif torch_device in ["cuda"]:
assert abs(result_sum.item() - 329.1999816894531) < 1e-2
assert abs(result_mean.item() - 0.4286458194255829) < 1e-3
else:
print("None")

def test_full_loop_with_v_prediction(self):
scheduler_class = self.scheduler_classes[0]
Expand Down Expand Up @@ -135,8 +133,6 @@ def test_full_loop_with_v_prediction(self):
elif torch_device in ["cuda"]:
assert abs(result_sum.item() - 193.4154052734375) < 1e-2
assert abs(result_mean.item() - 0.2518429756164551) < 1e-3
else:
print("None")

def test_full_loop_device(self):
scheduler_class = self.scheduler_classes[0]
Expand Down Expand Up @@ -166,8 +162,6 @@ def test_full_loop_device(self):
elif torch_device in ["cuda"]:
assert abs(result_sum.item() - 337.394287109375) < 1e-2
assert abs(result_mean.item() - 0.4393154978752136) < 1e-3
else:
print("None")

def test_full_loop_device_karras_sigmas(self):
scheduler_class = self.scheduler_classes[0]
Expand Down Expand Up @@ -198,8 +192,6 @@ def test_full_loop_device_karras_sigmas(self):
elif torch_device in ["cuda"]:
assert abs(result_sum.item() - 837.25537109375) < 1e-2
assert abs(result_mean.item() - 1.0901763439178467) < 1e-2
else:
print("None")

def test_beta_sigmas(self):
self.check_over_configs(use_beta_sigmas=True)
Expand Down

0 comments on commit a2ee718

Please sign in to comment.