-
Notifications
You must be signed in to change notification settings - Fork 4
Open
Description
Hi,
Thanks for the work on chatting interface, the chat.py and running command works good for 7b model(llama-7b-chat).
But when I am about to change it into 13b model, it seems the nature that 13b model run on distributed gpu influenced the running of gradio. the chat.py cannot run correctly after i change the --nproc_per_node to 2.
Can you kindly provide ideas about how to change the chat.py and run it for 13b model(llama-13b-chat)?
Thanks!
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels