@@ -870,32 +870,45 @@ void CheckAndDoCompact(const std::vector<phi::MetaTensor*>& meta_tensors,
870870 auto NeedCompact = [&](const std::vector<phi::MetaTensor*>& meta_tensors) {
871871 if (max_reserved < FLAGS_max_reserved_threshold_in_gb << 30 ) return false ;
872872 if (cur_allocated < FLAGS_cur_allocated_threshold_in_gb << 30 ) return false ;
873- const auto [max_free_size, total_free_size ] =
873+ const auto [max_free_size, large_N_free_size ] =
874874 paddle::memory::VmmMaxFreeSize (phi::GPUPlace (current_device_id),
875875 meta_tensors.size ());
876876 const auto & [req_total_size, size_vec] = CalTensorSize (meta_tensors);
877+ VLOG (10 ) << " run api: " << api << " req_total_size: " << req_total_size
878+ << " , max_free_size: " << max_free_size
879+ << " , large_N_free_size: " << large_N_free_size
880+ << " , max_reserved: " << max_reserved
881+ << " , max_allocated: " << max_allocated
882+ << " , cur_allocated: " << cur_allocated;
877883 if (req_total_size < max_free_size) return false ;
878- if (req_total_size > total_free_size ) {
884+ if (req_total_size > large_N_free_size ) {
879885 VLOG (1 ) << " Need Compact req_total_size: " << req_total_size
880- << " , total_free_size: " << total_free_size
881- << " , max_free_size: " << max_free_size;
886+ << " , large_N_free_size: " << large_N_free_size
887+ << " , max_free_size: " << max_free_size
888+ << " , max_reserved: " << max_reserved
889+ << " , max_allocated: " << max_allocated
890+ << " , cur_allocated: " << cur_allocated;
882891 return true ;
883892 }
884893 if (FLAGS_try_allocate) {
885894 auto alloc_succ = paddle::memory::TryAllocBatch (
886895 phi::GPUPlace (current_device_id), size_vec);
887- VLOG (1 ) << " TryAllocBatch ret: " << ! alloc_succ
896+ VLOG (1 ) << " TryAllocBatch ret: " << alloc_succ
888897 << " , req_total_size: " << req_total_size
889- << " , total_free_size: " << total_free_size
890- << " , max_free_size: " << max_free_size;
898+ << " , large_N_free_size: " << large_N_free_size
899+ << " , max_free_size: " << max_free_size
900+ << " , max_reserved: " << max_reserved
901+ << " , max_allocated: " << max_allocated
902+ << " , cur_allocated: " << cur_allocated;
891903 return !alloc_succ;
892904 }
893905 return false ;
894906 };
895907
896908 if (NeedCompact (meta_tensors)) {
897909 VLOG (1 ) << " Before Compact max_reserved: " << max_reserved / divisor
898- << " , max_allocated: " << max_allocated / divisor;
910+ << " GB, max_allocated: " << max_allocated / divisor
911+ << " GB, cur_allocated: " << cur_allocated / divisor << " GB" ;
899912 paddle::memory::Compact (phi::GPUPlace (current_device_id));
900913 }
901914#endif
0 commit comments