Precision of dot operation #18938
-
Hello Team,
But I did not observe any difference when running with these different values. When I use Could you please tell me if there is a way to perform high-precision matmul with |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 2 replies
-
What platform are you running your computations on? The |
Beta Was this translation helpful? Give feedback.
-
Hi @jakevdp, I found this behaviour on an H100 GPU. I took a look at the optimized HLO, I found that custom call with CUBLAS op is created with the correct precision parameters, but has no effect on the result. I'm interested if this is the correct behaviour?
|
Beta Was this translation helpful? Give feedback.
Can you say more about your goal here? You're doing bf16 dot products, which will always be done at bf16 precision.
If you want to accumulate in float32, you could pass
preferred_element_type='float32'
. Is that what you have in mind?