Skip to content

Conversation

ckhordiasma
Copy link

@ckhordiasma ckhordiasma commented Mar 21, 2025

  • Update Dockerfile.ubi to install vllm-cuda using wheel from RHEL AI team

the install script is located in payload/run.sh. An args file was also added with the custom parameters, and is referenced in the tekton pipeline.

  • update payload/run.sh to use bot token

  • add trap to guarantee run.sh deletion

…d-hat-data-services#85)

* Update Dockerfile.ubi to install vllm-cuda using wheel from RHEL AI team

the install script is located in payload/run.sh. An args file was also
added with the custom parameters, and is referenced in the tekton
pipeline.

* update payload/run.sh to use bot token

* add trap to guarantee run.sh deletion
@ckhordiasma ckhordiasma requested a review from dtrifiro March 21, 2025 14:49
@dtrifiro
Copy link

dtrifiro commented Mar 24, 2025

Let's hold on merging this until we figure out if we can get provide a vLLM v0.7.3 tag in the midstream and a build from the AIPCC team.

Context: https://issues.redhat.com/browse/INFERENG-477

@ckhordiasma
Copy link
Author

@dtrifiro should I leave the 2.19 branch the way it is?

@dtrifiro
Copy link

Yeah, let's wait until we come up with a definitive strategy

@ckhordiasma ckhordiasma marked this pull request as draft March 24, 2025 13:55
@ckhordiasma
Copy link
Author

sounds good, I will draft this for now and will not make any changes to 2.19 (which already has this change merged in there)

WHEEL_RELEASE_ARTIFACTS="https://gitlab.com/api/v4/projects/68045055/packages/generic/rhelai-wheels/${WHEEL_RELEASE}/wheels-${WHEEL_RELEASE}.tar.gz"


# NOTE - ensure that flashinfer is included in wheel thing

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks like it's included? https://issues.redhat.com/browse/AIPCC-49

@dtrifiro
Copy link

@tarukumar fyi


## CUDA Base ###################################################################
FROM python-install as cuda-base
FROM python-install AS cuda-base
Copy link

@dtrifiro dtrifiro Apr 14, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Does this cuda-base layer include all of the rpm dependencies required by the wheel set? Should be enough to check with ldd (?)

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

If I was able to get a green build, does that mean that all the required dependencies are there? Or do I need to do additional checks?

@ckhordiasma ckhordiasma changed the title Build VLLM CUDA from RHEL AI wheels, add audio and video packages (#85) Build VLLM CUDA from AIPCC wheels Apr 14, 2025
@ckhordiasma
Copy link
Author

/build

@ckhordiasma
Copy link
Author

/build

@ckhordiasma
Copy link
Author

/build

@ckhordiasma
Copy link
Author

/build

@ckhordiasma
Copy link
Author

/build-from-rhelai-wheels

@ckhordiasma
Copy link
Author

/build-from-rhelai-wheels

@ckhordiasma
Copy link
Author

/build-from-rhelai-wheels

@ckhordiasma
Copy link
Author

/build-from-rhelai-wheels


## Base Layer ##################################################################
FROM registry.access.redhat.com/ubi9/ubi-minimal:${BASE_UBI_IMAGE_TAG} as base
FROM registry.access.redhat.com/ubi9/ubi-minimal:${BASE_UBI_IMAGE_TAG} AS base
Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

could you please try with a base image built by AIPCC: http://quay.io/aipcc/base-images/cuda
if your builder doesn't have access to that image, you can ping me on Slack and I'll send you the procedure to get access.

Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@dtrifiro
Copy link

dtrifiro commented Oct 9, 2025

outdated

@dtrifiro dtrifiro closed this Oct 9, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants