File tree Expand file tree Collapse file tree 1 file changed +3
-3
lines changed
tensorrt_llm/_torch/modules Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Original file line number Diff line number Diff line change @@ -75,9 +75,9 @@ def load_weight_shard(
7575 # For integrated GPU systems (e.g., DGX Spark), CPU and GPU share limited physical memory.
7676 # Avoiding device transfers reduces memory consumption and unnecessary data copies,
7777 # enabling support for larger models on memory-constrained systems.
78- logger .debug (
79- f"[load_weight_shard] Skipping device transfer from { weight .device } to { device } on integrated GPU to conserve shared memory."
80- )
78+ logger .warning_once (
79+ f"[load_weight_shard] Skipping device transfer from { weight .device } to { device } on integrated GPU to conserve shared memory." ,
80+ key = "load_weight_shard_skip_device_transfer_with_integrated_gpu" )
8181 device = weight .device
8282 if isinstance (weight , torch .Tensor ):
8383 tensor_shape = weight .shape
You can’t perform that action at this time.
0 commit comments