Skip to content

v2.5.19: Merge pull request #390 from AInVFX/main

Choose a tag to compare

@adrientoupet adrientoupet released this 10 Dec 06:56
· 58 commits to main since this release
2006fa3
  • 🎨 New header logo design - Refreshed ASCII art banner (thanks @naxci1)
  • 🧹 Remove dead flash attention wrapper - Removed legacy code from FP8CompatibleDiT; FlashAttentionVarlen already handles backend switching via its attention_mode attribute
  • 🛡️ Fix graceful fallback from flash-attn - Add compatibility shims for corrupted flash_attn/xformers DLLs, preventing startup crashes when CUDA extensions are broken
  • 📊 Improved VRAM tracking - Separate allocated vs reserved memory tracking, Windows-only overflow detection (WDDM paging behavior)
  • ♻️ Centralize backend detection - Unified is_mps_available(), is_cuda_available(), get_gpu_backend() helpers across codebase
  • 🔄 Revert 2.5.14 VRAM limit enforcement - Removed set_per_process_memory_fraction call; Overflow detection and warnings remain.