| 
									
										
										
										
											2023-11-08 17:39:29 -05:00
										 |  |  | #ifndef Py_INTERNAL_CRITICAL_SECTION_H
 | 
					
						
							|  |  |  | #define Py_INTERNAL_CRITICAL_SECTION_H
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #ifndef Py_BUILD_CORE
 | 
					
						
							|  |  |  | #  error "this header requires Py_BUILD_CORE define"
 | 
					
						
							|  |  |  | #endif
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #include "pycore_lock.h"        // PyMutex
 | 
					
						
							|  |  |  | #include "pycore_pystate.h"     // _PyThreadState_GET()
 | 
					
						
							|  |  |  | #include <stdint.h>
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #ifdef __cplusplus
 | 
					
						
							|  |  |  | extern "C" { | 
					
						
							|  |  |  | #endif
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | // Implementation of Python critical sections
 | 
					
						
							|  |  |  | //
 | 
					
						
							|  |  |  | // Conceptually, critical sections are a deadlock avoidance layer on top of
 | 
					
						
							|  |  |  | // per-object locks. These helpers, in combination with those locks, replace
 | 
					
						
							|  |  |  | // our usage of the global interpreter lock to provide thread-safety for
 | 
					
						
							|  |  |  | // otherwise thread-unsafe objects, such as dict.
 | 
					
						
							|  |  |  | //
 | 
					
						
							|  |  |  | // NOTE: These APIs are no-ops in non-free-threaded builds.
 | 
					
						
							|  |  |  | //
 | 
					
						
							|  |  |  | // Straightforward per-object locking could introduce deadlocks that were not
 | 
					
						
							|  |  |  | // present when running with the GIL. Threads may hold locks for multiple
 | 
					
						
							|  |  |  | // objects simultaneously because Python operations can nest. If threads were
 | 
					
						
							|  |  |  | // to acquire the same locks in different orders, they would deadlock.
 | 
					
						
							|  |  |  | //
 | 
					
						
							|  |  |  | // One way to avoid deadlocks is to allow threads to hold only the lock (or
 | 
					
						
							|  |  |  | // locks) for a single operation at a time (typically a single lock, but some
 | 
					
						
							|  |  |  | // operations involve two locks). When a thread begins a nested operation it
 | 
					
						
							|  |  |  | // could suspend the locks for any outer operation: before beginning the nested
 | 
					
						
							|  |  |  | // operation, the locks for the outer operation are released and when the
 | 
					
						
							|  |  |  | // nested operation completes, the locks for the outer operation are
 | 
					
						
							|  |  |  | // reacquired.
 | 
					
						
							|  |  |  | //
 | 
					
						
							|  |  |  | // To improve performance, this API uses a variation of the above scheme.
 | 
					
						
							|  |  |  | // Instead of immediately suspending locks any time a nested operation begins,
 | 
					
						
							|  |  |  | // locks are only suspended if the thread would block. This reduces the number
 | 
					
						
							|  |  |  | // of lock acquisitions and releases for nested operations, while still
 | 
					
						
							|  |  |  | // avoiding deadlocks.
 | 
					
						
							|  |  |  | //
 | 
					
						
							|  |  |  | // Additionally, the locks for any active operation are suspended around
 | 
					
						
							|  |  |  | // other potentially blocking operations, such as I/O. This is because the
 | 
					
						
							|  |  |  | // interaction between locks and blocking operations can lead to deadlocks in
 | 
					
						
							|  |  |  | // the same way as the interaction between multiple locks.
 | 
					
						
							|  |  |  | //
 | 
					
						
							|  |  |  | // Each thread's critical sections and their corresponding locks are tracked in
 | 
					
						
							|  |  |  | // a stack in `PyThreadState.critical_section`. When a thread calls
 | 
					
						
							|  |  |  | // `_PyThreadState_Detach()`, such as before a blocking I/O operation or when
 | 
					
						
							|  |  |  | // waiting to acquire a lock, the thread suspends all of its active critical
 | 
					
						
							|  |  |  | // sections, temporarily releasing the associated locks. When the thread calls
 | 
					
						
							|  |  |  | // `_PyThreadState_Attach()`, it resumes the top-most (i.e., most recent)
 | 
					
						
							|  |  |  | // critical section by reacquiring the associated lock or locks.  See
 | 
					
						
							|  |  |  | // `_PyCriticalSection_Resume()`.
 | 
					
						
							|  |  |  | //
 | 
					
						
							|  |  |  | // NOTE: Only the top-most critical section is guaranteed to be active.
 | 
					
						
							|  |  |  | // Operations that need to lock two objects at once must use
 | 
					
						
							|  |  |  | // `Py_BEGIN_CRITICAL_SECTION2()`. You *CANNOT* use nested critical sections
 | 
					
						
							|  |  |  | // to lock more than one object at once, because the inner critical section
 | 
					
						
							|  |  |  | // may  suspend the outer critical sections. This API does not provide a way
 | 
					
						
							|  |  |  | // to lock more than two objects at once (though it could be added later
 | 
					
						
							|  |  |  | // if actually needed).
 | 
					
						
							|  |  |  | //
 | 
					
						
							|  |  |  | // NOTE: Critical sections implicitly behave like reentrant locks because
 | 
					
						
							|  |  |  | // attempting to acquire the same lock will suspend any outer (earlier)
 | 
					
						
							|  |  |  | // critical sections. However, they are less efficient for this use case than
 | 
					
						
							|  |  |  | // purposefully designed reentrant locks.
 | 
					
						
							|  |  |  | //
 | 
					
						
							|  |  |  | // Example usage:
 | 
					
						
							|  |  |  | //  Py_BEGIN_CRITICAL_SECTION(op);
 | 
					
						
							|  |  |  | //  ...
 | 
					
						
							|  |  |  | //  Py_END_CRITICAL_SECTION();
 | 
					
						
							|  |  |  | //
 | 
					
						
							|  |  |  | // To lock two objects at once:
 | 
					
						
							|  |  |  | //  Py_BEGIN_CRITICAL_SECTION2(op1, op2);
 | 
					
						
							|  |  |  | //  ...
 | 
					
						
							|  |  |  | //  Py_END_CRITICAL_SECTION2();
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | // Tagged pointers to critical sections use the two least significant bits to
 | 
					
						
							|  |  |  | // mark if the pointed-to critical section is inactive and whether it is a
 | 
					
						
							|  |  |  | // _PyCriticalSection2 object.
 | 
					
						
							|  |  |  | #define _Py_CRITICAL_SECTION_INACTIVE       0x1
 | 
					
						
							|  |  |  | #define _Py_CRITICAL_SECTION_TWO_MUTEXES    0x2
 | 
					
						
							|  |  |  | #define _Py_CRITICAL_SECTION_MASK           0x3
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-11-20 15:52:00 +02:00
										 |  |  | #ifdef Py_GIL_DISABLED
 | 
					
						
							| 
									
										
										
										
											2024-03-28 16:05:08 +01:00
										 |  |  | # define Py_BEGIN_CRITICAL_SECTION_MUT(mutex)                           \
 | 
					
						
							| 
									
										
										
										
											2023-11-08 17:39:29 -05:00
										 |  |  |     {                                                                   \ | 
					
						
							|  |  |  |         _PyCriticalSection _cs;                                         \ | 
					
						
							| 
									
										
										
										
											2024-03-28 16:05:08 +01:00
										 |  |  |         _PyCriticalSection_Begin(&_cs, mutex) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | # define Py_BEGIN_CRITICAL_SECTION(op)                                  \
 | 
					
						
							|  |  |  |         Py_BEGIN_CRITICAL_SECTION_MUT(&_PyObject_CAST(op)->ob_mutex) | 
					
						
							| 
									
										
										
										
											2023-11-08 17:39:29 -05:00
										 |  |  | 
 | 
					
						
							|  |  |  | # define Py_END_CRITICAL_SECTION()                                      \
 | 
					
						
							|  |  |  |         _PyCriticalSection_End(&_cs);                                   \ | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | # define Py_BEGIN_CRITICAL_SECTION2(a, b)                               \
 | 
					
						
							|  |  |  |     {                                                                   \ | 
					
						
							|  |  |  |         _PyCriticalSection2 _cs2;                                       \ | 
					
						
							|  |  |  |         _PyCriticalSection2_Begin(&_cs2, &_PyObject_CAST(a)->ob_mutex, &_PyObject_CAST(b)->ob_mutex) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | # define Py_END_CRITICAL_SECTION2()                                     \
 | 
					
						
							|  |  |  |         _PyCriticalSection2_End(&_cs2);                                 \ | 
					
						
							|  |  |  |     } | 
					
						
							| 
									
										
										
										
											2024-02-06 14:03:43 -08:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-05-22 17:45:34 +00:00
										 |  |  | // Specialized version of critical section locking to safely use
 | 
					
						
							|  |  |  | // PySequence_Fast APIs without the GIL. For performance, the argument *to*
 | 
					
						
							|  |  |  | // PySequence_Fast() is provided to the macro, not the *result* of
 | 
					
						
							|  |  |  | // PySequence_Fast(), which would require an extra test to determine if the
 | 
					
						
							|  |  |  | // lock must be acquired.
 | 
					
						
							|  |  |  | # define Py_BEGIN_CRITICAL_SECTION_SEQUENCE_FAST(original)              \
 | 
					
						
							|  |  |  |     {                                                                   \ | 
					
						
							|  |  |  |         PyObject *_orig_seq = _PyObject_CAST(original);                 \ | 
					
						
							|  |  |  |         const bool _should_lock_cs = PyList_CheckExact(_orig_seq);      \ | 
					
						
							|  |  |  |         _PyCriticalSection _cs;                                         \ | 
					
						
							|  |  |  |         if (_should_lock_cs) {                                          \ | 
					
						
							|  |  |  |             _PyCriticalSection_Begin(&_cs, &_orig_seq->ob_mutex);       \ | 
					
						
							|  |  |  |         } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | # define Py_END_CRITICAL_SECTION_SEQUENCE_FAST()                        \
 | 
					
						
							|  |  |  |         if (_should_lock_cs) {                                          \ | 
					
						
							|  |  |  |             _PyCriticalSection_End(&_cs);                               \ | 
					
						
							|  |  |  |         }                                                               \ | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-02-06 14:03:43 -08:00
										 |  |  | // Asserts that the mutex is locked.  The mutex must be held by the
 | 
					
						
							|  |  |  | // top-most critical section otherwise there's the possibility
 | 
					
						
							|  |  |  | // that the mutex would be swalled out in some code paths.
 | 
					
						
							|  |  |  | #define  _Py_CRITICAL_SECTION_ASSERT_MUTEX_LOCKED(mutex) \
 | 
					
						
							|  |  |  |     _PyCriticalSection_AssertHeld(mutex) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | // Asserts that the mutex for the given object is locked. The mutex must
 | 
					
						
							|  |  |  | // be held by the top-most critical section otherwise there's the
 | 
					
						
							|  |  |  | // possibility that the mutex would be swalled out in some code paths.
 | 
					
						
							|  |  |  | #ifdef Py_DEBUG
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #define  _Py_CRITICAL_SECTION_ASSERT_OBJECT_LOCKED(op)                           \
 | 
					
						
							|  |  |  |     if (Py_REFCNT(op) != 1) {                                                    \ | 
					
						
							|  |  |  |         _Py_CRITICAL_SECTION_ASSERT_MUTEX_LOCKED(&_PyObject_CAST(op)->ob_mutex); \ | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #else   /* Py_DEBUG */
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #define  _Py_CRITICAL_SECTION_ASSERT_OBJECT_LOCKED(op)
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #endif  /* Py_DEBUG */
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-11-20 15:52:00 +02:00
										 |  |  | #else  /* !Py_GIL_DISABLED */
 | 
					
						
							| 
									
										
										
										
											2023-11-08 17:39:29 -05:00
										 |  |  | // The critical section APIs are no-ops with the GIL.
 | 
					
						
							| 
									
										
										
										
											2024-03-28 16:05:08 +01:00
										 |  |  | # define Py_BEGIN_CRITICAL_SECTION_MUT(mut)
 | 
					
						
							| 
									
										
										
										
											2023-11-08 17:39:29 -05:00
										 |  |  | # define Py_BEGIN_CRITICAL_SECTION(op)
 | 
					
						
							|  |  |  | # define Py_END_CRITICAL_SECTION()
 | 
					
						
							|  |  |  | # define Py_BEGIN_CRITICAL_SECTION2(a, b)
 | 
					
						
							|  |  |  | # define Py_END_CRITICAL_SECTION2()
 | 
					
						
							| 
									
										
										
										
											2024-05-22 17:45:34 +00:00
										 |  |  | # define Py_BEGIN_CRITICAL_SECTION_SEQUENCE_FAST(original)
 | 
					
						
							|  |  |  | # define Py_END_CRITICAL_SECTION_SEQUENCE_FAST()
 | 
					
						
							| 
									
										
										
										
											2024-02-06 14:03:43 -08:00
										 |  |  | # define _Py_CRITICAL_SECTION_ASSERT_MUTEX_LOCKED(mutex)
 | 
					
						
							|  |  |  | # define _Py_CRITICAL_SECTION_ASSERT_OBJECT_LOCKED(op)
 | 
					
						
							| 
									
										
										
										
											2023-11-20 15:52:00 +02:00
										 |  |  | #endif  /* !Py_GIL_DISABLED */
 | 
					
						
							| 
									
										
										
										
											2023-11-08 17:39:29 -05:00
										 |  |  | 
 | 
					
						
							|  |  |  | typedef struct { | 
					
						
							|  |  |  |     // Tagged pointer to an outer active critical section (or 0).
 | 
					
						
							|  |  |  |     // The two least-significant-bits indicate whether the pointed-to critical
 | 
					
						
							|  |  |  |     // section is inactive and whether it is a _PyCriticalSection2 object.
 | 
					
						
							|  |  |  |     uintptr_t prev; | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     // Mutex used to protect critical section
 | 
					
						
							|  |  |  |     PyMutex *mutex; | 
					
						
							|  |  |  | } _PyCriticalSection; | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | // A critical section protected by two mutexes. Use
 | 
					
						
							|  |  |  | // _PyCriticalSection2_Begin and _PyCriticalSection2_End.
 | 
					
						
							|  |  |  | typedef struct { | 
					
						
							|  |  |  |     _PyCriticalSection base; | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     PyMutex *mutex2; | 
					
						
							|  |  |  | } _PyCriticalSection2; | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline int | 
					
						
							|  |  |  | _PyCriticalSection_IsActive(uintptr_t tag) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  |     return tag != 0 && (tag & _Py_CRITICAL_SECTION_INACTIVE) == 0; | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | // Resumes the top-most critical section.
 | 
					
						
							|  |  |  | PyAPI_FUNC(void) | 
					
						
							|  |  |  | _PyCriticalSection_Resume(PyThreadState *tstate); | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | // (private) slow path for locking the mutex
 | 
					
						
							|  |  |  | PyAPI_FUNC(void) | 
					
						
							|  |  |  | _PyCriticalSection_BeginSlow(_PyCriticalSection *c, PyMutex *m); | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | PyAPI_FUNC(void) | 
					
						
							|  |  |  | _PyCriticalSection2_BeginSlow(_PyCriticalSection2 *c, PyMutex *m1, PyMutex *m2, | 
					
						
							|  |  |  |                              int is_m1_locked); | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline void | 
					
						
							|  |  |  | _PyCriticalSection_Begin(_PyCriticalSection *c, PyMutex *m) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  |     if (PyMutex_LockFast(&m->v)) { | 
					
						
							|  |  |  |         PyThreadState *tstate = _PyThreadState_GET(); | 
					
						
							|  |  |  |         c->mutex = m; | 
					
						
							|  |  |  |         c->prev = tstate->critical_section; | 
					
						
							|  |  |  |         tstate->critical_section = (uintptr_t)c; | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  |     else { | 
					
						
							|  |  |  |         _PyCriticalSection_BeginSlow(c, m); | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | // Removes the top-most critical section from the thread's stack of critical
 | 
					
						
							|  |  |  | // sections. If the new top-most critical section is inactive, then it is
 | 
					
						
							|  |  |  | // resumed.
 | 
					
						
							|  |  |  | static inline void | 
					
						
							|  |  |  | _PyCriticalSection_Pop(_PyCriticalSection *c) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  |     PyThreadState *tstate = _PyThreadState_GET(); | 
					
						
							|  |  |  |     uintptr_t prev = c->prev; | 
					
						
							|  |  |  |     tstate->critical_section = prev; | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     if ((prev & _Py_CRITICAL_SECTION_INACTIVE) != 0) { | 
					
						
							|  |  |  |         _PyCriticalSection_Resume(tstate); | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline void | 
					
						
							|  |  |  | _PyCriticalSection_End(_PyCriticalSection *c) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  |     PyMutex_Unlock(c->mutex); | 
					
						
							|  |  |  |     _PyCriticalSection_Pop(c); | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline void | 
					
						
							|  |  |  | _PyCriticalSection2_Begin(_PyCriticalSection2 *c, PyMutex *m1, PyMutex *m2) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  |     if (m1 == m2) { | 
					
						
							|  |  |  |         // If the two mutex arguments are the same, treat this as a critical
 | 
					
						
							|  |  |  |         // section with a single mutex.
 | 
					
						
							|  |  |  |         c->mutex2 = NULL; | 
					
						
							|  |  |  |         _PyCriticalSection_Begin(&c->base, m1); | 
					
						
							|  |  |  |         return; | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     if ((uintptr_t)m2 < (uintptr_t)m1) { | 
					
						
							|  |  |  |         // Sort the mutexes so that the lower address is locked first.
 | 
					
						
							|  |  |  |         // The exact order does not matter, but we need to acquire the mutexes
 | 
					
						
							|  |  |  |         // in a consistent order to avoid lock ordering deadlocks.
 | 
					
						
							|  |  |  |         PyMutex *tmp = m1; | 
					
						
							|  |  |  |         m1 = m2; | 
					
						
							|  |  |  |         m2 = tmp; | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     if (PyMutex_LockFast(&m1->v)) { | 
					
						
							|  |  |  |         if (PyMutex_LockFast(&m2->v)) { | 
					
						
							|  |  |  |             PyThreadState *tstate = _PyThreadState_GET(); | 
					
						
							|  |  |  |             c->base.mutex = m1; | 
					
						
							|  |  |  |             c->mutex2 = m2; | 
					
						
							|  |  |  |             c->base.prev = tstate->critical_section; | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |             uintptr_t p = (uintptr_t)c | _Py_CRITICAL_SECTION_TWO_MUTEXES; | 
					
						
							|  |  |  |             tstate->critical_section = p; | 
					
						
							|  |  |  |         } | 
					
						
							|  |  |  |         else { | 
					
						
							|  |  |  |             _PyCriticalSection2_BeginSlow(c, m1, m2, 1); | 
					
						
							|  |  |  |         } | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  |     else { | 
					
						
							|  |  |  |         _PyCriticalSection2_BeginSlow(c, m1, m2, 0); | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline void | 
					
						
							|  |  |  | _PyCriticalSection2_End(_PyCriticalSection2 *c) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  |     if (c->mutex2) { | 
					
						
							|  |  |  |         PyMutex_Unlock(c->mutex2); | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  |     PyMutex_Unlock(c->base.mutex); | 
					
						
							|  |  |  |     _PyCriticalSection_Pop(&c->base); | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | PyAPI_FUNC(void) | 
					
						
							|  |  |  | _PyCriticalSection_SuspendAll(PyThreadState *tstate); | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-02-06 14:03:43 -08:00
										 |  |  | #ifdef Py_GIL_DISABLED
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline void | 
					
						
							| 
									
										
										
										
											2024-02-15 10:54:57 -08:00
										 |  |  | _PyCriticalSection_AssertHeld(PyMutex *mutex) | 
					
						
							|  |  |  | { | 
					
						
							| 
									
										
										
										
											2024-02-06 14:03:43 -08:00
										 |  |  | #ifdef Py_DEBUG
 | 
					
						
							|  |  |  |     PyThreadState *tstate = _PyThreadState_GET(); | 
					
						
							|  |  |  |     uintptr_t prev = tstate->critical_section; | 
					
						
							|  |  |  |     if (prev & _Py_CRITICAL_SECTION_TWO_MUTEXES) { | 
					
						
							|  |  |  |         _PyCriticalSection2 *cs = (_PyCriticalSection2 *)(prev & ~_Py_CRITICAL_SECTION_MASK); | 
					
						
							|  |  |  |         assert(cs != NULL && (cs->base.mutex == mutex || cs->mutex2 == mutex)); | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  |     else { | 
					
						
							|  |  |  |         _PyCriticalSection *cs = (_PyCriticalSection *)(tstate->critical_section & ~_Py_CRITICAL_SECTION_MASK); | 
					
						
							|  |  |  |         assert(cs != NULL && cs->mutex == mutex); | 
					
						
							|  |  |  |     } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #endif
 | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #endif
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-11-08 17:39:29 -05:00
										 |  |  | #ifdef __cplusplus
 | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | #endif
 | 
					
						
							|  |  |  | #endif /* !Py_INTERNAL_CRITICAL_SECTION_H */
 |