-
-
Notifications
You must be signed in to change notification settings - Fork 33.2k
gh-127022: Simplify PyStackRef_FromPyObjectSteal
#127024
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
This gets rid of the immortal check in `PyStackRef_FromPyObjectSteal()`. Overall, this improves performance about 2% in the free threading build. This also renames `PyStackRef_Is()` to `PyStackRef_IsExactly()` because the macro requires that the tag bits of the arguments match, which is only true in certain special cases.
2c43ad0 to
5583ac0
Compare
|
Said benchmark: https://github.com/facebookexperimental/free-threading-benchmarking/tree/main/results/bm-20241118-3.14.0a1+-ed7085a-NOGIL I was thinking of how this breaks the nice encapsulation we have :(, but 2% speedup is too good to give up. |
Co-authored-by: Pieter Eendebak <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There are a number of cases where we've gone from checking bitwise equality with PyStackRef_Is to checking equality after masking out the deferred bit with one of PyStackRef_Is{None,True,False}. Were the previous checks wrong?
Python/bytecodes.c
Outdated
| replaced op(_POP_JUMP_IF_TRUE, (cond -- )) { | ||
| assert(PyStackRef_BoolCheck(cond)); | ||
| int flag = PyStackRef_Is(cond, PyStackRef_True); | ||
| int flag = PyStackRef_IsExactly(cond, PyStackRef_True); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Why do we use PyStackRef_IsExactly here (which doesn't mask out the deferred bit) but use PyStackRef_IsFalse (which does mask out the deferred bit) in _POP_JUMP_IF_FALSE above? Is this the rare case where it's safe?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Our codegen ensures that these ops only see True or False. That's often by adding a TO_BOOL immediately before, which may be folded into COMPARE_OP. The preceding TO_BOOL, including in COMPARE_OP, ensures the canonical representation of PyStackRef_False or PyStackRef_True with the deferred bit set.
However, there are two places in codegen.c that omit the TO_BOOL because they have other reasons to know that the result is exactly a boolean:
Lines 678 to 682 in 09c240f
| ADDOP_I(c, loc, LOAD_FAST, 0); | |
| ADDOP_LOAD_CONST(c, loc, _PyLong_GetOne()); | |
| ADDOP_I(c, loc, COMPARE_OP, (Py_NE << 5) | compare_masks[Py_NE]); | |
| NEW_JUMP_TARGET_LABEL(c, body); | |
| ADDOP_JUMP(c, loc, POP_JUMP_IF_FALSE, body); |
Lines 5746 to 5749 in 09c240f
| ADDOP(c, LOC(p), GET_LEN); | |
| ADDOP_LOAD_CONST_NEW(c, LOC(p), PyLong_FromSsize_t(size)); | |
| ADDOP_COMPARE(c, LOC(p), GtE); | |
| RETURN_IF_ERROR(jump_to_fail_pop(c, LOC(p), pc, POP_JUMP_IF_FALSE)); |
The COMPARE_OPs here still generate bools, but not always in the canonical representation. So we can either:
- Modify
COMPARE_OPto ensure the canonical representation like https://github.com/colesbury/cpython/blob/5583ac0c311132e36ef458842e087945898ffdec/Python/bytecodes.c#L2409-L2416 - Use
PyStackRef_IsFalse(instead ofPyStackRef_IsExactly) in theJUMP_IF_FALSE - Modify the codegen by inserting
TO_BOOLin those two spots.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
That makes sense, thanks for the explanation. Since using PyStackRef_IsExactly safely is sensitive to code generation changes, I might suggest using it only when we're sure it actually matters for performance, and default to using the variants that mask out the deferred bits everywhere by default since those are always safe. I'd guess that this wouldn't affect the performance improvement of this change much, since it should come from avoiding the tagging in _PyStackRef_FromPyObjectSteal. I don't feel super strongly though.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'll switch to using PyStackRef_IsFalse and PyStackRef_IsTrue.
I'm no longer convinced that PyStackRef_IsExactly is actually a performance win (and I didn't see it in measurements). I think we have issues with code generation quality that we'll need to address later. Things like POP_JUMP_IF_NONE are composed of _IS_NONE and _POP_JUMP_IF_TRUE and we pack the intermediate result in a tagged _PyStackRef. Clang does a pretty good job of optimizing through it. GCC less so: https://gcc.godbolt.org/z/Ejs8c78qd.
No, the previous checks were okay when |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice!
Python/bytecodes.c
Outdated
| replaced op(_POP_JUMP_IF_TRUE, (cond -- )) { | ||
| assert(PyStackRef_BoolCheck(cond)); | ||
| int flag = PyStackRef_Is(cond, PyStackRef_True); | ||
| int flag = PyStackRef_IsExactly(cond, PyStackRef_True); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
That makes sense, thanks for the explanation. Since using PyStackRef_IsExactly safely is sensitive to code generation changes, I might suggest using it only when we're sure it actually matters for performance, and default to using the variants that mask out the deferred bits everywhere by default since those are always safe. I'd guess that this wouldn't affect the performance improvement of this change much, since it should come from avoiding the tagging in _PyStackRef_FromPyObjectSteal. I don't feel super strongly though.
|
Benchmark on most recent changes: https://github.com/facebookexperimental/free-threading-benchmarking/tree/main/results/bm-20241122-3.14.0a1+-a9e4872-NOGIL#vs-base
|
This gets rid of the immortal check in `PyStackRef_FromPyObjectSteal()`. Overall, this improves performance about 2% in the free threading build. This also renames `PyStackRef_Is()` to `PyStackRef_IsExactly()` because the macro requires that the tag bits of the arguments match, which is only true in certain special cases.
This gets rid of the immortal check in
PyStackRef_FromPyObjectSteal(). Overall, this improves performance about 1-2% in the free threading build.This also renames
PyStackRef_Is()toPyStackRef_IsExactly()because the macro requires that the tag bits of the arguments match, which is only true in certain special cases.PyStackRef_FromPyObjectSteal#127022