Skip to content

Commit 98772b9

Browse files
committed
default attn_implementaion to eager to avoid issues
Signed-off-by: Zhiyu Cheng <[email protected]>
1 parent 8d0e40f commit 98772b9

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

examples/llm_ptq/hf_ptq.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -758,10 +758,10 @@ def output_decode(generated_ids, input_shape):
758758
parser.add_argument(
759759
"--attn_implementation",
760760
help=(
761-
"Specify the attention implementation to use."
761+
"Specify the attention implementation to use. "
762762
"This arg will be passed to the HF model loading if specified."
763763
),
764-
default=None,
764+
default="eager",
765765
type=str,
766766
)
767767

0 commit comments

Comments
 (0)