Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

gh-117139: Set up the tagged evaluation stack #117186

Closed

Conversation

Fidget-Spinner
Copy link
Member

@Fidget-Spinner Fidget-Spinner commented Mar 23, 2024

This PR is up mostly just for discussion. It compiles (with warnings) and passes (Ubuntu) tests as of now.

The main point of contention is how do we deal with bytecode that expect arrays of untagged objects? E.g. vectorcall takes an array from the stack of objects. There are two general approaches I can think of:

  1. Untag all values used by the bytecode on the stack, then let them do the usual. This has really bad performance.
  2. Secretly check for tagged values inside API. That is -- make them support tagged values internally, but don't change their signature on the outside. This way we don't break C API. This is kind of bad from a type safety standpoint (and needs us to cast things around), but it's the one with the least performance loss.

For example I currently have a _PyList_FromArraySteal and a _PyList_FromTaggedArraySteal. With the 2nd approach, I just need some bitwise operations in _PyList_FromArraySteal and it should be good to go. Then I can remove _PyList_FromTaggedArraySteal. What do y'all think?

However, we still have to untag everything if we call escaping functions that call 3rd party code (e.g. vectorcall). This is only safe right now for CPython's own C API.

@colesbury
Copy link
Contributor

colesbury commented Mar 26, 2024

I'm a confused by your description of approach 2:

Secretly check for tagged values inside API. That is -- make them support tagged values internally, but don't change their signature on the outside.

That does not look like the approach in the PR, which adds new functions that take tagged values. The approach in the PR seems correct. We can change the signatures of internal-only APIs as needed.

However, we still have to untag everything if we call escaping functions that call 3rd party code (e.g. vectorcall)

If you untag in-place on the stack, you need to Py_INCREF() deferred references: if you mark them as non-deferred on the evaluation stack then they need to reflect that.

I think it's better to do the untagging off the evaluation stack. In the common case, you can have some fixed size PyObject *buf[N] on the C stack and untag into it, for some reasonable small fixed N. In the general case,

Include/object.h Outdated
Comment on lines 231 to 247
typedef union {
PyObject *obj;
uintptr_t bits;
} _Py_TaggedObject;

#define Py_OBJECT_TAG (0b0)

#ifdef Py_GIL_DISABLED
#define Py_CLEAR_TAG(tagged) ((PyObject *)((tagged).bits & ~(Py_OBJECT_TAG)))
#else
#define Py_CLEAR_TAG(tagged) ((PyObject *)(uintptr_t)((tagged).bits))
#endif

#define Py_OBJ_PACK(obj) ((_Py_TaggedObject){.bits = (uintptr_t)(obj)})

#define Py_TAG_CAST(o) ((_Py_TaggedObject){.obj = (o)})

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We should not make any of these macros or data structures part of the public C API.

@Fidget-Spinner
Copy link
Member Author

I'm a confused by your description of approach 2:

Secretly check for tagged values inside API. That is -- make them support tagged values internally, but don't change their signature on the outside.

That does not look like the approach in the PR, which adds new functions that take tagged values. The approach in the PR seems correct for internal-only APIs -- we can change their signatures as needed.

I did a mix of both -- for internal functions I converted them to tagged form. For stuff that might be used by other people eg. _PyList_FromArraySteal, I made them internally support tagged pointers.

I think it's better to do the untagging off the evaluation stack. In the common case, you can have some fixed size PyObject *buf[N] on the C stack and untag into it, for some reasonable small fixed N. In the general case,

Ok I will think about that. That does make sense -- to pass to a scratch buffer for now. If we exceed the buffer I assume we need to untag the C stack then. One thing we might want to be wary of is overflowing the C recursion stack with the current recursion limit. Since this will make _Py_EvalEvalFrameDefault a little larger.

@Fidget-Spinner
Copy link
Member Author

Fidget-Spinner commented Mar 26, 2024

Also to note: every Py_DECREF(Py_CLEAR_TAG(x)) and Py_INCREF(Py_CLEAR_TAG(x)) and so on is an opportunity to swap out with Py_DECREF_TAGGED(x) and so on in a future PR, when we implement the actual deferring. To keep this PR small and easier to review, I didn't do them for now. Though I think I might as well since this PR is quite bulky already, and to reduce code churn in the future.

@Fidget-Spinner Fidget-Spinner changed the title gh-117139: Tagged evaluation stack gh-117139: Set up the tagged evaluation stack Mar 26, 2024
@colesbury
Copy link
Contributor

For stuff that might be used by other people eg. _PyList_FromArraySteal, I made them internally support tagged pointers.

That's undefined behavior in C -- let's avoid it if possible. If we're concerned about external users, add a new function that takes tagged references (and keep the old one unused, if desired).

If we exceed the buffer I assume we need to untag the C stack then...

If we exceed the fixed size buffer we should allocate enough space for a temporary buffer using PyMem_Malloc().

One thing we might want to be wary of is overflowing the C recursion stack...

I would structure this so that the small, temporary buffer is only used when we perform a vectorcall (and not to a Python function). For example, in pseudo-code:

#define N 5 

PyObject *
PyObject_VectorcallTagged(PyObject *callable, const _Py_TaggedObject *tagged, size_t nargs, PyObject *kwnames)
{
  PyObject *args[N];
  if (nargs >= N) { 
    return PyObject_VectorcallTaggedSlow(callable, tagged, nargs, kwnames);
  }
  untag(args, tagged, nargs);
  return PyObject_Vectorcall(callable, args, nargs, kwnames);
}

PyObject *
PyObject_VectorcallTaggedSlow(PyObject *callable, const _Py_TaggedObject *tagged, size_t nargs, PyObject *kwnames)
{
  PyObject *args = PyMem_Malloc(nargs * sizeof(PyObject*));
  if (args == NULL) ...
  untag(args, tagged, nargs);
  PyObject *res = PyObject_Vectorcall(callable, args, nargs, kwnames);
  PyMem_Free(args);
  return res;
}

@Fidget-Spinner
Copy link
Member Author

Thanks for the clear explanation! I will address them tomorrow.

@gvanrossum
Copy link
Member

Hm, I now recall that the problem of untagging arrays of arguments sunk my attempt (I got it to work, but it was too slow). Hope you fare better!

@nineteendo
Copy link
Contributor

Could you also configure pre-commit? https://devguide.python.org/getting-started/setup-building/#install-pre-commit

index 2bd9c40..3aa7dea 100644
--- a/Tools/cases_generator/analyzer.py
+++ b/Tools/cases_generator/analyzer.py
@@ -359,7 +359,7 @@ def has_error_without_pop(op: parser.InstDef) -> bool:
     "Py_XDECREF_STACKREF",
     "Py_INCREF_STACKREF",
     "Py_XINCREF_TAGGED",
-    "Py_NewRef_StackRef",    
+    "Py_NewRef_StackRef",
     "Py_INCREF",
     "_PyManagedDictPointer_IsValues",
     "_PyObject_GetManagedDict",

@Fidget-Spinner
Copy link
Member Author

!buildbot nogil

@bedevere-bot
Copy link

🤖 New build scheduled with the buildbot fleet by @Fidget-Spinner for commit d59145b 🤖

The command will test the builders whose names match following regular expression: nogil

The builders matched are:

  • x86-64 MacOS Intel ASAN NoGIL PR
  • AMD64 Ubuntu NoGIL Refleaks PR
  • ARM64 MacOS M1 NoGIL PR
  • AMD64 Windows Server 2022 NoGIL PR
  • AMD64 Ubuntu NoGIL PR
  • ARM64 MacOS M1 Refleaks NoGIL PR
  • x86-64 MacOS Intel NoGIL PR

@Fidget-Spinner
Copy link
Member Author

Fidget-Spinner commented Apr 26, 2024

Here are results from Sam's microbenchmark suite testing scalability. Note that I've only implemented deferred refcounting for method and function calls.

Specs:
20 virtual cores (hyper threading)
pyperf system tune

Main:
object_cfunction MEGA FAILED: 1.1x slower
cmodule_function MEGA FAILED: 2.1x slower
generator FAILED: 3.1x faster
pymethod MEGA FAILED: 1.4x slower
pyfunction MEGA FAILED: 2.7x slower
module_function MEGA FAILED: 1.9x slower
load_string_const MEGA FAILED: 2.1x slower
load_tuple_const MEGA FAILED: 1.9x slower
create_closure MEGA FAILED: 3.6x slower
create_pyobject MEGA FAILED: 2.9x slower

My branch:
object_cfunction FAILED: 9.3x faster
cmodule_function MEGA FAILED: 2.0x slower
generator FAILED: 2.9x faster
pymethod MEGA FAILED: 1.1x slower
pyfunction MEGA FAILED: 2.4x slower
module_function MEGA FAILED: 1.9x slower
load_string_const MEGA FAILED: 2.2x slower
load_tuple_const MEGA FAILED: 2.1x slower
create_closure MEGA FAILED: 3.6x slower
create_pyobject MEGA FAILED: 3.1x slower

Notice that object_cfunction went from 1.1x slower to 9.3x faster on a 20 thread workload!

All tests pass except ASAN. I am now happy with the state this PR is in. So I will close this and start upstreaming things in peices.

@Fidget-Spinner
Copy link
Member Author

Fidget-Spinner commented Apr 26, 2024

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

8 participants