Skip to content

Commit 1a514e3

Browse files
committed
Log attn_implementation setting in hf_transformers
1 parent 866491a commit 1a514e3

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

server/text_generation_server/inference_engine/hf_transformers.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -28,6 +28,7 @@ def __init__(
2828

2929
# TODO: consider if Flash Attention should be enabled based on FLASH_ATTENTION=True
3030
if attn_impl := os.getenv("TRANSFORMERS_ATTN_IMPL"):
31+
logger.info(f"Setting attn_implementation to {attn_impl}")
3132
kwargs["attn_implementation"] = attn_impl
3233

3334
if model_config.model_type == "mpt":

0 commit comments

Comments
 (0)