-
Notifications
You must be signed in to change notification settings - Fork 248
OpenAI API: Changes to enable multi-modal for 3.2 11B #1211
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/torchchat/1211
Note: Links to docs will display an error until the docs builds have been completed. ❌ 7 New Failures, 6 Unrelated FailuresAs of commit 8c14c55 with merge base c454026 ( NEW FAILURES - The following jobs have failed:
BROKEN TRUNK - The following jobs failed but were present on the merge base:👉 Rebase onto the `viable/strict` branch to avoid these failures
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
it was defaulting to too small a number. This will fix some things for now
| ) | ||
| except: | ||
| # can not find max_seq_length in model config, use default value | ||
| self.max_seq_length = 128 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
haha now i can see where 128 comes from
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It happens to be the same as the head_dim, very tricky to trace it LOL
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
yep.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks to @iseeyuan for spotting it.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM. Thanks!
| + self.speculative_builder_args.speculate_k | ||
| + 1 | ||
| if self.draft_model is not None | ||
| else self.model.text_transformer_args.max_seq_length |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
One of the things on our list is having a unification configuration system for both tune-backend model and chat-backend models to get rid of the try .. except here.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for hopping in and fixing this
We'll have a separate PR for updating the MM README
|
This is irrelevant to the failures, I'm forcing it through |
Start the server by either specifying the model
or by specifying paths for the checkpoint
python3 torchchat.py server
Use this curl command to test
This is the image
