gh-124218: Refactor per-thread reference counting (#124844)

Currently, we only use per-thread reference counting for heap type objects and
the naming reflects that. We will extend it to a few additional types in an
upcoming change to avoid scaling bottlenecks when creating nested functions.

Rename some of the files and functions in preparation for this change.
This commit is contained in:
Sam Gross 2024-10-01 13:05:42 -04:00 committed by GitHub
parent 5aa91c56bf
commit b482538523
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
15 changed files with 168 additions and 167 deletions

View file

@ -1,200 +0,0 @@
#include "Python.h"
#include "pycore_lock.h" // PyMutex_LockFlags()
#include "pycore_pystate.h" // _PyThreadState_GET()
#include "pycore_object.h" // _Py_IncRefTotal
#include "pycore_typeid.h"
// This contains code for allocating unique ids to heap type objects
// and re-using those ids when the type is deallocated.
//
// See Include/internal/pycore_typeid.h for more details.
#ifdef Py_GIL_DISABLED
#define POOL_MIN_SIZE 8
#define LOCK_POOL(pool) PyMutex_LockFlags(&pool->mutex, _Py_LOCK_DONT_DETACH)
#define UNLOCK_POOL(pool) PyMutex_Unlock(&pool->mutex)
static int
resize_interp_type_id_pool(struct _Py_type_id_pool *pool)
{
if ((size_t)pool->size > PY_SSIZE_T_MAX / (2 * sizeof(*pool->table))) {
return -1;
}
Py_ssize_t new_size = pool->size * 2;
if (new_size < POOL_MIN_SIZE) {
new_size = POOL_MIN_SIZE;
}
_Py_type_id_entry *table = PyMem_Realloc(pool->table,
new_size * sizeof(*pool->table));
if (table == NULL) {
return -1;
}
Py_ssize_t start = pool->size;
for (Py_ssize_t i = start; i < new_size - 1; i++) {
table[i].next = &table[i + 1];
}
table[new_size - 1].next = NULL;
pool->table = table;
pool->freelist = &table[start];
_Py_atomic_store_ssize(&pool->size, new_size);
return 0;
}
static int
resize_local_refcounts(_PyThreadStateImpl *tstate)
{
if (tstate->types.is_finalized) {
return -1;
}
struct _Py_type_id_pool *pool = &tstate->base.interp->type_ids;
Py_ssize_t size = _Py_atomic_load_ssize(&pool->size);
Py_ssize_t *refcnts = PyMem_Realloc(tstate->types.refcounts,
size * sizeof(Py_ssize_t));
if (refcnts == NULL) {
return -1;
}
Py_ssize_t old_size = tstate->types.size;
if (old_size < size) {
memset(refcnts + old_size, 0, (size - old_size) * sizeof(Py_ssize_t));
}
tstate->types.refcounts = refcnts;
tstate->types.size = size;
return 0;
}
void
_PyType_AssignId(PyHeapTypeObject *type)
{
PyInterpreterState *interp = _PyInterpreterState_GET();
struct _Py_type_id_pool *pool = &interp->type_ids;
LOCK_POOL(pool);
if (pool->freelist == NULL) {
if (resize_interp_type_id_pool(pool) < 0) {
type->unique_id = -1;
UNLOCK_POOL(pool);
return;
}
}
_Py_type_id_entry *entry = pool->freelist;
pool->freelist = entry->next;
entry->type = type;
_PyObject_SetDeferredRefcount((PyObject *)type);
type->unique_id = (entry - pool->table);
UNLOCK_POOL(pool);
}
void
_PyType_ReleaseId(PyHeapTypeObject *type)
{
PyInterpreterState *interp = _PyInterpreterState_GET();
struct _Py_type_id_pool *pool = &interp->type_ids;
if (type->unique_id < 0) {
// The type doesn't have an id assigned.
return;
}
LOCK_POOL(pool);
_Py_type_id_entry *entry = &pool->table[type->unique_id];
assert(entry->type == type);
entry->next = pool->freelist;
pool->freelist = entry;
type->unique_id = -1;
UNLOCK_POOL(pool);
}
void
_PyType_IncrefSlow(PyHeapTypeObject *type)
{
_PyThreadStateImpl *tstate = (_PyThreadStateImpl *)_PyThreadState_GET();
if (type->unique_id < 0 || resize_local_refcounts(tstate) < 0) {
// just incref the type directly.
Py_INCREF(type);
return;
}
assert(type->unique_id < tstate->types.size);
tstate->types.refcounts[type->unique_id]++;
#ifdef Py_REF_DEBUG
_Py_IncRefTotal((PyThreadState *)tstate);
#endif
_Py_INCREF_STAT_INC();
}
void
_PyType_MergeThreadLocalRefcounts(_PyThreadStateImpl *tstate)
{
if (tstate->types.refcounts == NULL) {
return;
}
struct _Py_type_id_pool *pool = &tstate->base.interp->type_ids;
LOCK_POOL(pool);
for (Py_ssize_t i = 0, n = tstate->types.size; i < n; i++) {
Py_ssize_t refcnt = tstate->types.refcounts[i];
if (refcnt != 0) {
PyObject *type = (PyObject *)pool->table[i].type;
assert(PyType_Check(type));
_Py_atomic_add_ssize(&type->ob_ref_shared,
refcnt << _Py_REF_SHARED_SHIFT);
tstate->types.refcounts[i] = 0;
}
}
UNLOCK_POOL(pool);
}
void
_PyType_FinalizeThreadLocalRefcounts(_PyThreadStateImpl *tstate)
{
_PyType_MergeThreadLocalRefcounts(tstate);
PyMem_Free(tstate->types.refcounts);
tstate->types.refcounts = NULL;
tstate->types.size = 0;
tstate->types.is_finalized = 1;
}
void
_PyType_FinalizeIdPool(PyInterpreterState *interp)
{
struct _Py_type_id_pool *pool = &interp->type_ids;
// First, set the free-list to NULL values
while (pool->freelist) {
_Py_type_id_entry *next = pool->freelist->next;
pool->freelist->type = NULL;
pool->freelist = next;
}
// Now everything non-NULL is a type. Set the type's id to -1 in case it
// outlives the interpreter.
for (Py_ssize_t i = 0; i < pool->size; i++) {
PyHeapTypeObject *ht = pool->table[i].type;
if (ht) {
ht->unique_id = -1;
pool->table[i].type = NULL;
}
}
PyMem_Free(pool->table);
pool->table = NULL;
pool->freelist = NULL;
pool->size = 0;
}
#endif /* Py_GIL_DISABLED */