Skip to content

Commit 46ab69b

Browse files
authored
Update README.md
1 parent 698fd4b commit 46ab69b

File tree

1 file changed

+3
-0
lines changed

1 file changed

+3
-0
lines changed

llama_adapter_v2_chat65b/README.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -75,6 +75,9 @@ conda env create -f environment.yml
7575

7676
* Use Ctrl+C to exit the demo at any time.
7777

78+
## Demo
79+
<img src="../docs/chat_demo.png" width="80%" />
80+
7881
## Known issues
7982

8083
* Some users may experience the error `RuntimeError: Expected is_sm80 to be true, but got false.` (Mostly sm_86 GPU users, including A6000, A5000 and 3090). This is because we changed the attention module to use `torch.nn.functional.scaled_dot_product_attention` if it exists, but a [dispatch logic error](https://github.com/pytorch/pytorch/issues/94883) in PyTorch = 2.0.0 causes failure on some GPU architectures. The affected users can upgrade to PyTorch >= 2.1.0 or the nightly build, in which the bug is fixed.

0 commit comments

Comments
 (0)