Skip to content

Commit e69eaf5

Browse files
committed
Fix runtime bug
Signed-off-by: Jiayu Chang <jiayuc@nvidia.com>
1 parent 7a9730a commit e69eaf5

File tree

1 file changed

+1
-3
lines changed

1 file changed

+1
-3
lines changed

tensorrt_llm/_torch/pyexecutor/model_engine.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2667,7 +2667,7 @@ def previous_seq_slots_device():
26672667

26682668
attn_metadata.prepare()
26692669

2670-
peft_cache_manager = resource_manager.get_resource_manager(
2670+
peft_cache_manager = resource_manager and resource_manager.get_resource_manager(
26712671
ResourceManagerType.PEFT_CACHE_MANAGER)
26722672
lora_params = self._get_lora_params_from_requests(
26732673
scheduled_requests, attn_metadata, peft_cache_manager, maybe_graph)
@@ -3307,8 +3307,6 @@ def forward(self,
33073307
req_id_to_old_request: Optional[Dict[int, LlmRequest]] = None):
33083308
kv_cache_manager = resource_manager.get_resource_manager(
33093309
self.kv_cache_manager_key)
3310-
peft_cache_manager = resource_manager.get_resource_manager(
3311-
ResourceManagerType.PEFT_CACHE_MANAGER)
33123310

33133311
attn_metadata = self._set_up_attn_metadata(kv_cache_manager)
33143312
if self.enable_spec_decode:

0 commit comments

Comments
 (0)