Skip to content

Conversation

@yao-matrix
Copy link
Contributor

test cases

  1. tests/quantization/bnb/test_mixed_int8.py::SlowBnb8bitTests::test_pipeline_cuda_placement_works_with_mixed_int8 PASS
  2. tests/quantization/bnb/test_4bit.py::BnB4BitBasicTests::test_model_memory_usage FAIL, but it's ipex issue, validated success when ipex fixed, so still enable and once ipex fixed, it will naturally pass w/o modify test case.

questions

  1. for tests/quantization/bnb/test_mixed_int8.py, I changed the hard-coding "cuda" to device. I tried directly it's OK to directly use torch_device, but I guess the reason you didn't use it is that maybe "rocm" system cannot work under "rocm" to(), but work on "cuda" to, so I added + device = torch_device if torch_device != "rocm" else "cuda" to workaround. This is just a guess, @hlky, let me know if you think it's not correct.

@hlky , pls help review, thx.

@DN6
Copy link
Collaborator

DN6 commented Apr 16, 2025

@bot /style

@github-actions
Copy link
Contributor

Style fixes have been applied. View the workflow run here.

@yao-matrix
Copy link
Contributor Author

@hlky , pls help review, thx.

@yiyixuxu yiyixuxu merged commit eef3d65 into huggingface:main Apr 17, 2025
8 checks passed
@yao-matrix yao-matrix deleted the issue234 branch April 18, 2025 00:49
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants