-
Notifications
You must be signed in to change notification settings - Fork 60
[CI] Auto detect the available GPU devices and distribute them with CPUs #2173
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Closed
+61
−45
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
…2171) https://github.com/intel/torch-xpu-ops/actions/runs/18370317029/job/52569740992?pr=1992#step:4:1135 pytest does not run but the result is pass disable_e2e disable_ut
clean skip list in skip_list_common.py, removed passed cases. disable_e2e disable_distributed Signed-off-by: chunhuanMeng <[email protected]> Co-authored-by: chunhuanMeng <[email protected]> Co-authored-by: Daisy Deng <[email protected]>
1. all huggingface models list file for nightly 2. all timm models list file for nightly 3. all torchbench models file for nightly 4. remove huggingface CamemBert due to removed in [pytorch/pytorch/pull/164815](https://github.com/pytorch/pytorch/pull/164815/files#diff-004303ad6116d64ab2a8356469ccb11b32d8caca702e1ad65cc0538600a76d2dL170) 5. add '--disable-cudagraphs' to reduce the impact of CUDA for XPU tests 6. align models list with pytorch for CI tests disable_build disable_ut disable_distributed
Stack from [ghstack](https://github.com/ezyang/ghstack) (oldest at bottom): * __->__ #2179
# Motivation I would like to clarify that, according to the [FP64 Partial Emulation Proposal](https://intel.sharepoint.com/:w:/s/MLTSHdGPU/EaroFZY371hOqNL9182g2_EBOe83qGYTriAavPB6WTWXYg?e=XSRnKt), the SYCL compiler and IGC only perform FP64 conversion on the DG2 and ATS-M architectures, and this is only available when AOT is enabled. If AOT is not enabled, many warnings like the following will be emitted: ```bash icx: warning: '-fsycl-fp64-conv-emu' option is supported only for AOT compilation of Intel GPUs. It will be ignored for other targets ``` To avoid these warnings, the `-fsycl-fp64-conv-emu` flag should only be added when AOT is enabled for the specific target architectures.
Since more FP8 Ops will be supported on XPU recently, basic FP8 cases should be activated. This PR will remove the following cases from skip list: ``` TestCommonXPU::test_compare_cpu_torch__scaled_mm_xpu_float8_e4m3fn SKIPPED (test doesn't work on XPU backend) TestCommonXPU::test_out_torch__scaled_mm_xpu_float8_e4m3fn SKIPPED (Skipped!) TestCommonXPU::test_python_ref__refs_eye_xpu_float8_e4m3fn SKIPPED (test doesn't work on XPU backend) TestCommonXPU::test_python_ref__refs_eye_xpu_float8_e4m3fnuz SKIPPED (test doesn't work on XPU backend) TestCommonXPU::test_python_ref__refs_eye_xpu_float8_e5m2 SKIPPED (test doesn't work on XPU backend) TestCommonXPU::test_python_ref__refs_eye_xpu_float8_e5m2fnuz SKIPPED (test doesn't work on XPU backend) TestCommonXPU::test_python_ref_executor__refs_eye_executor_aten_xpu_float8_e4m3fn PASSED TestCommonXPU::test_python_ref_executor__refs_eye_executor_aten_xpu_float8_e4m3fnuz PASSED TestCommonXPU::test_python_ref_executor__refs_eye_executor_aten_xpu_float8_e5m2 PASSED TestCommonXPU::test_python_ref_executor__refs_eye_executor_aten_xpu_float8_e5m2fnuz PASSED TestCommonXPU::test_python_ref_meta__refs_eye_xpu_float8_e4m3fn SKIPPED (test doesn't work on XPU backend) TestCommonXPU::test_python_ref_meta__refs_eye_xpu_float8_e4m3fnuz SKIPPED (test doesn't work on XPU backend) TestCommonXPU::test_python_ref_meta__refs_eye_xpu_float8_e5m2 SKIPPED (test doesn't work on XPU backend) TestCommonXPU::test_python_ref_meta__refs_eye_xpu_float8_e5m2fnuz SKIPPED (test doesn't work on XPU backend) TestCommonXPU::test_python_ref_torch_fallback__refs_eye_xpu_float8_e4m3fn SKIPPED (test doesn't work on XPU backend) TestCommonXPU::test_python_ref_torch_fallback__refs_eye_xpu_float8_e4m3fnuz SKIPPED (test doesn't work on XPU backend) TestCommonXPU::test_python_ref_torch_fallback__refs_eye_xpu_float8_e5m2 SKIPPED (test doesn't work on XPU backend) TestCommonXPU::test_python_ref_torch_fallback__refs_eye_xpu_float8_e5m2fnuz SKIPPED (test doesn't work on XPU backend) ```
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
disable_build