mirror of
				https://github.com/python/cpython.git
				synced 2025-11-03 23:21:29 +00:00 
			
		
		
		
	In the free threading build, the per thread reference counting uses a unique id for some objects to index into the local reference count table. Use 0 instead of -1 to indicate that the id is not assigned. This avoids bugs where zero-initialized heap type objects look like they have a unique id assigned.
		
			
				
	
	
		
			232 lines
		
	
	
	
		
			6.4 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			232 lines
		
	
	
	
		
			6.4 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
#include "Python.h"
 | 
						|
 | 
						|
#include "pycore_dict.h"        // _PyDict_UniqueId()
 | 
						|
#include "pycore_lock.h"        // PyMutex_LockFlags()
 | 
						|
#include "pycore_pystate.h"     // _PyThreadState_GET()
 | 
						|
#include "pycore_object.h"      // _Py_IncRefTotal
 | 
						|
#include "pycore_uniqueid.h"
 | 
						|
 | 
						|
// This contains code for allocating unique ids for per-thread reference
 | 
						|
// counting and re-using those ids when an object is deallocated.
 | 
						|
//
 | 
						|
// Currently, per-thread reference counting is only used for heap types.
 | 
						|
//
 | 
						|
// See Include/internal/pycore_uniqueid.h for more details.
 | 
						|
 | 
						|
#ifdef Py_GIL_DISABLED
 | 
						|
 | 
						|
#define POOL_MIN_SIZE 8
 | 
						|
 | 
						|
#define LOCK_POOL(pool) PyMutex_LockFlags(&pool->mutex, _Py_LOCK_DONT_DETACH)
 | 
						|
#define UNLOCK_POOL(pool) PyMutex_Unlock(&pool->mutex)
 | 
						|
 | 
						|
static int
 | 
						|
resize_interp_type_id_pool(struct _Py_unique_id_pool *pool)
 | 
						|
{
 | 
						|
    if ((size_t)pool->size > PY_SSIZE_T_MAX / (2 * sizeof(*pool->table))) {
 | 
						|
        return -1;
 | 
						|
    }
 | 
						|
 | 
						|
    Py_ssize_t new_size = pool->size * 2;
 | 
						|
    if (new_size < POOL_MIN_SIZE) {
 | 
						|
        new_size = POOL_MIN_SIZE;
 | 
						|
    }
 | 
						|
 | 
						|
    _Py_unique_id_entry *table = PyMem_Realloc(pool->table,
 | 
						|
                                               new_size * sizeof(*pool->table));
 | 
						|
    if (table == NULL) {
 | 
						|
        return -1;
 | 
						|
    }
 | 
						|
 | 
						|
    Py_ssize_t start = pool->size;
 | 
						|
    for (Py_ssize_t i = start; i < new_size - 1; i++) {
 | 
						|
        table[i].next = &table[i + 1];
 | 
						|
    }
 | 
						|
    table[new_size - 1].next = NULL;
 | 
						|
 | 
						|
    pool->table = table;
 | 
						|
    pool->freelist = &table[start];
 | 
						|
    _Py_atomic_store_ssize(&pool->size, new_size);
 | 
						|
    return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int
 | 
						|
resize_local_refcounts(_PyThreadStateImpl *tstate)
 | 
						|
{
 | 
						|
    if (tstate->refcounts.is_finalized) {
 | 
						|
        return -1;
 | 
						|
    }
 | 
						|
 | 
						|
    struct _Py_unique_id_pool *pool = &tstate->base.interp->unique_ids;
 | 
						|
    Py_ssize_t size = _Py_atomic_load_ssize(&pool->size);
 | 
						|
 | 
						|
    Py_ssize_t *refcnts = PyMem_Realloc(tstate->refcounts.values,
 | 
						|
                                        size * sizeof(Py_ssize_t));
 | 
						|
    if (refcnts == NULL) {
 | 
						|
        return -1;
 | 
						|
    }
 | 
						|
 | 
						|
    Py_ssize_t old_size = tstate->refcounts.size;
 | 
						|
    if (old_size < size) {
 | 
						|
       memset(refcnts + old_size, 0, (size - old_size) * sizeof(Py_ssize_t));
 | 
						|
    }
 | 
						|
 | 
						|
    tstate->refcounts.values = refcnts;
 | 
						|
    tstate->refcounts.size = size;
 | 
						|
    return 0;
 | 
						|
}
 | 
						|
 | 
						|
Py_ssize_t
 | 
						|
_PyObject_AssignUniqueId(PyObject *obj)
 | 
						|
{
 | 
						|
    PyInterpreterState *interp = _PyInterpreterState_GET();
 | 
						|
    struct _Py_unique_id_pool *pool = &interp->unique_ids;
 | 
						|
 | 
						|
    LOCK_POOL(pool);
 | 
						|
    if (pool->freelist == NULL) {
 | 
						|
        if (resize_interp_type_id_pool(pool) < 0) {
 | 
						|
            UNLOCK_POOL(pool);
 | 
						|
            return _Py_INVALID_UNIQUE_ID;
 | 
						|
        }
 | 
						|
    }
 | 
						|
 | 
						|
    _Py_unique_id_entry *entry = pool->freelist;
 | 
						|
    pool->freelist = entry->next;
 | 
						|
    entry->obj = obj;
 | 
						|
    _PyObject_SetDeferredRefcount(obj);
 | 
						|
    // The unique id is one plus the index of the entry in the table.
 | 
						|
    Py_ssize_t unique_id = (entry - pool->table) + 1;
 | 
						|
    assert(unique_id > 0);
 | 
						|
    UNLOCK_POOL(pool);
 | 
						|
    return unique_id;
 | 
						|
}
 | 
						|
 | 
						|
void
 | 
						|
_PyObject_ReleaseUniqueId(Py_ssize_t unique_id)
 | 
						|
{
 | 
						|
    PyInterpreterState *interp = _PyInterpreterState_GET();
 | 
						|
    struct _Py_unique_id_pool *pool = &interp->unique_ids;
 | 
						|
 | 
						|
    LOCK_POOL(pool);
 | 
						|
    assert(unique_id > 0 && unique_id <= pool->size);
 | 
						|
    Py_ssize_t idx = unique_id - 1;
 | 
						|
    _Py_unique_id_entry *entry = &pool->table[idx];
 | 
						|
    entry->next = pool->freelist;
 | 
						|
    pool->freelist = entry;
 | 
						|
    UNLOCK_POOL(pool);
 | 
						|
}
 | 
						|
 | 
						|
static Py_ssize_t
 | 
						|
clear_unique_id(PyObject *obj)
 | 
						|
{
 | 
						|
    Py_ssize_t id = _Py_INVALID_UNIQUE_ID;
 | 
						|
    if (PyType_Check(obj)) {
 | 
						|
        if (PyType_HasFeature((PyTypeObject *)obj, Py_TPFLAGS_HEAPTYPE)) {
 | 
						|
            PyHeapTypeObject *ht = (PyHeapTypeObject *)obj;
 | 
						|
            id = ht->unique_id;
 | 
						|
            ht->unique_id = _Py_INVALID_UNIQUE_ID;
 | 
						|
        }
 | 
						|
    }
 | 
						|
    else if (PyCode_Check(obj)) {
 | 
						|
        PyCodeObject *co = (PyCodeObject *)obj;
 | 
						|
        id = co->_co_unique_id;
 | 
						|
        co->_co_unique_id = _Py_INVALID_UNIQUE_ID;
 | 
						|
    }
 | 
						|
    else if (PyDict_Check(obj)) {
 | 
						|
        PyDictObject *mp = (PyDictObject *)obj;
 | 
						|
        id = _PyDict_UniqueId(mp);
 | 
						|
        mp->_ma_watcher_tag &= ~(UINT64_MAX << DICT_UNIQUE_ID_SHIFT);
 | 
						|
    }
 | 
						|
    return id;
 | 
						|
}
 | 
						|
 | 
						|
void
 | 
						|
_PyObject_DisablePerThreadRefcounting(PyObject *obj)
 | 
						|
{
 | 
						|
    Py_ssize_t id = clear_unique_id(obj);
 | 
						|
    if (id != _Py_INVALID_UNIQUE_ID) {
 | 
						|
        _PyObject_ReleaseUniqueId(id);
 | 
						|
    }
 | 
						|
}
 | 
						|
 | 
						|
void
 | 
						|
_PyObject_ThreadIncrefSlow(PyObject *obj, size_t idx)
 | 
						|
{
 | 
						|
    _PyThreadStateImpl *tstate = (_PyThreadStateImpl *)_PyThreadState_GET();
 | 
						|
    if (((Py_ssize_t)idx) < 0 || resize_local_refcounts(tstate) < 0) {
 | 
						|
        // just incref the object directly.
 | 
						|
        Py_INCREF(obj);
 | 
						|
        return;
 | 
						|
    }
 | 
						|
 | 
						|
    assert(idx < (size_t)tstate->refcounts.size);
 | 
						|
    tstate->refcounts.values[idx]++;
 | 
						|
#ifdef Py_REF_DEBUG
 | 
						|
    _Py_IncRefTotal((PyThreadState *)tstate);
 | 
						|
#endif
 | 
						|
    _Py_INCREF_STAT_INC();
 | 
						|
}
 | 
						|
 | 
						|
void
 | 
						|
_PyObject_MergePerThreadRefcounts(_PyThreadStateImpl *tstate)
 | 
						|
{
 | 
						|
    if (tstate->refcounts.values == NULL) {
 | 
						|
        return;
 | 
						|
    }
 | 
						|
 | 
						|
    struct _Py_unique_id_pool *pool = &tstate->base.interp->unique_ids;
 | 
						|
 | 
						|
    LOCK_POOL(pool);
 | 
						|
    for (Py_ssize_t i = 0, n = tstate->refcounts.size; i < n; i++) {
 | 
						|
        Py_ssize_t refcnt = tstate->refcounts.values[i];
 | 
						|
        if (refcnt != 0) {
 | 
						|
            PyObject *obj = pool->table[i].obj;
 | 
						|
            _Py_atomic_add_ssize(&obj->ob_ref_shared,
 | 
						|
                                 refcnt << _Py_REF_SHARED_SHIFT);
 | 
						|
            tstate->refcounts.values[i] = 0;
 | 
						|
        }
 | 
						|
    }
 | 
						|
    UNLOCK_POOL(pool);
 | 
						|
}
 | 
						|
 | 
						|
void
 | 
						|
_PyObject_FinalizePerThreadRefcounts(_PyThreadStateImpl *tstate)
 | 
						|
{
 | 
						|
    _PyObject_MergePerThreadRefcounts(tstate);
 | 
						|
 | 
						|
    PyMem_Free(tstate->refcounts.values);
 | 
						|
    tstate->refcounts.values = NULL;
 | 
						|
    tstate->refcounts.size = 0;
 | 
						|
    tstate->refcounts.is_finalized = 1;
 | 
						|
}
 | 
						|
 | 
						|
void
 | 
						|
_PyObject_FinalizeUniqueIdPool(PyInterpreterState *interp)
 | 
						|
{
 | 
						|
    struct _Py_unique_id_pool *pool = &interp->unique_ids;
 | 
						|
 | 
						|
    // First, set the free-list to NULL values
 | 
						|
    while (pool->freelist) {
 | 
						|
        _Py_unique_id_entry *next = pool->freelist->next;
 | 
						|
        pool->freelist->obj = NULL;
 | 
						|
        pool->freelist = next;
 | 
						|
    }
 | 
						|
 | 
						|
    // Now everything non-NULL is a object. Clear their unique ids as the
 | 
						|
    // object outlives the interpreter.
 | 
						|
    for (Py_ssize_t i = 0; i < pool->size; i++) {
 | 
						|
        PyObject *obj = pool->table[i].obj;
 | 
						|
        pool->table[i].obj = NULL;
 | 
						|
        if (obj != NULL) {
 | 
						|
            Py_ssize_t id = clear_unique_id(obj);
 | 
						|
            (void)id;
 | 
						|
            assert(id == i + 1);
 | 
						|
        }
 | 
						|
    }
 | 
						|
    PyMem_Free(pool->table);
 | 
						|
    pool->table = NULL;
 | 
						|
    pool->freelist = NULL;
 | 
						|
    pool->size = 0;
 | 
						|
}
 | 
						|
 | 
						|
#endif   /* Py_GIL_DISABLED */
 |