fix lmi/vllm virtual envs, update to vllm 0.7.1#2703
Merged
siddvenk merged 1 commit intodeepjavalibrary:masterfrom Feb 3, 2025
Merged
fix lmi/vllm virtual envs, update to vllm 0.7.1#2703siddvenk merged 1 commit intodeepjavalibrary:masterfrom
siddvenk merged 1 commit intodeepjavalibrary:masterfrom
Conversation
siddvenk
commented
Feb 3, 2025
| resolve_chat_template_content_format) | ||
|
|
||
|
|
||
| def is_chat_completions_request(inputs: Dict) -> bool: |
Contributor
Author
There was a problem hiding this comment.
deleted because it's not used
siddvenk
commented
Feb 3, 2025
| "You must enable rolling batch to use the chat completions format." | ||
| ) | ||
|
|
||
| if not is_mistral_tokenizer and not hasattr(tokenizer, |
Contributor
Author
There was a problem hiding this comment.
deleted because the vllm utils do this validation for us already
| git reset --hard 4b2092c | ||
| $venv_pip install . | ||
| cd .. | ||
| rm -rf AutoFP8 |
Contributor
There was a problem hiding this comment.
Do we not need FP8 installation?
Contributor
Author
There was a problem hiding this comment.
not anymore! we're using llm compressor now #2701
d43e822 to
16cc16a
Compare
sindhuvahinis
approved these changes
Feb 3, 2025
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Description
This changes updates to vllm 0.7.1, which involves shuffling around some dependencies and being less strict with dependency versions.
Additionally, it updates the chat processing for vllm to be functional. There is still a good amount we need to implement for chat processing, that i'll take up in a follow up PR:
I have tested this with (single test for each):
I also added a chat test for mistral with vllm.
Type of change
Please delete options that are not relevant.
Checklist:
pytest tests.py -k "TestCorrectnessLmiDist" -m "lmi_dist"Feature/Issue validation/testing
Please describe the Unit or Integration tests that you ran to verify your changes and relevant result summary. Provide instructions so it can be reproduced.
Please also list any relevant details for your test configuration.
Test A
Logs for Test A
Test B
Logs for Test B