Skip to content

Commit cd34c0b

Browse files
committed
remove debug
1 parent 593da36 commit cd34c0b

File tree

1 file changed

+0
-11
lines changed

1 file changed

+0
-11
lines changed

examples/research_projects/lpl/lpl_loss.py

Lines changed: 0 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -66,12 +66,6 @@ def __init__(
6666
self.scale = torch.tensor(1.0 if scale_factor is None else scale_factor, device=device)
6767
self.shift = torch.tensor(0.0 if shift_factor is None else shift_factor, device=device)
6868

69-
# Debug logging
70-
print(f"VAE config: {self.vae.config}")
71-
print(f"Raw scale factor: {scale_factor}, shift factor: {shift_factor}")
72-
print(f"Final scale tensor: {self.scale}, shift tensor: {self.shift}")
73-
print(f"Device: {device}")
74-
7569
self.gradient_checkpointing = grad_ckpt
7670
self.pow_law = pow_law
7771
self.norm_type = norm_type.lower()
@@ -144,11 +138,6 @@ def custom_forward(*inputs):
144138
return features
145139

146140
def get_loss(self, input, target, get_hist=False):
147-
# Debug logging for each call
148-
print(f"Current scale: {self.scale}, shift: {self.shift}")
149-
print(f"Input shape: {input.shape}, dtype: {input.dtype}")
150-
print(f"Scale type: {type(self.scale)}, shift type: {type(self.shift)}")
151-
152141
if self.feature_type == "feature":
153142
inp_f = self.get_features(self.shift + input / self.scale)
154143
tar_f = self.get_features(self.shift + target / self.scale, disable_grads=True)

0 commit comments

Comments
 (0)