Skip to content

Commit a938550

Browse files
committed
adding flash attention to example config
1 parent 3837366 commit a938550

File tree

1 file changed

+3
-1
lines changed

1 file changed

+3
-1
lines changed

config.yml

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,8 @@ data:
2424
# Model Definition -------------------
2525
model:
2626
hf_model_ckpt: "NousResearch/Llama-2-7b-hf"
27+
torch_dtype: "bfloat16"
28+
attn_implementation: "flash_attention_2"
2729
quantize: true
2830
bitsandbytes:
2931
load_in_4bit: true
@@ -80,4 +82,4 @@ qa:
8082
- verb_percent
8183
- adjective_percent
8284
- noun_percent
83-
- summary_length
85+
- summary_length

0 commit comments

Comments
 (0)