We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 1431967 commit 033a8a4Copy full SHA for 033a8a4
1 file changed
backend/app/services/music_service.py
@@ -804,7 +804,9 @@ def _pad_audio_token(token):
804
progress = int((i + 1) / max_audio_frames * 100)
805
callback(progress, f"Generating audio... {i + 1}/{max_audio_frames} frames")
806
807
- frames = torch.stack(frames).permute(1, 2, 0).squeeze(0).cpu() # Move to CPU immediately
+ # Stack frames and explicitly preserve torch.long dtype (critical for MPS compatibility)
808
+ # torch.stack may promote dtype to float on MPS, so we explicitly convert to long before CPU
809
+ frames = torch.stack(frames).permute(1, 2, 0).squeeze(0).to(dtype=torch.long).cpu()
810
811
# Sequential offload: Move HeartMuLa to CPU before loading HeartCodec
812
# This allows fitting on smaller GPUs (12GB) by never having both models in VRAM
0 commit comments