Coverage Report

Created: 2026-04-20 06:11

next uncovered line (L), next uncovered region (R), next uncovered branch (B)
/src/cpython/Python/ceval_macros.h
Line
Count
Source
1
// Macros and other things needed by ceval.c, and bytecodes.c
2
3
/* Computed GOTOs, or
4
       the-optimization-commonly-but-improperly-known-as-"threaded code"
5
   using gcc's labels-as-values extension
6
   (http://gcc.gnu.org/onlinedocs/gcc/Labels-as-Values.html).
7
8
   The traditional bytecode evaluation loop uses a "switch" statement, which
9
   decent compilers will optimize as a single indirect branch instruction
10
   combined with a lookup table of jump addresses. However, since the
11
   indirect jump instruction is shared by all opcodes, the CPU will have a
12
   hard time making the right prediction for where to jump next (actually,
13
   it will be always wrong except in the uncommon case of a sequence of
14
   several identical opcodes).
15
16
   "Threaded code" in contrast, uses an explicit jump table and an explicit
17
   indirect jump instruction at the end of each opcode. Since the jump
18
   instruction is at a different address for each opcode, the CPU will make a
19
   separate prediction for each of these instructions, which is equivalent to
20
   predicting the second opcode of each opcode pair. These predictions have
21
   a much better chance to turn out valid, especially in small bytecode loops.
22
23
   A mispredicted branch on a modern CPU flushes the whole pipeline and
24
   can cost several CPU cycles (depending on the pipeline depth),
25
   and potentially many more instructions (depending on the pipeline width).
26
   A correctly predicted branch, however, is nearly free.
27
28
   At the time of this writing, the "threaded code" version is up to 15-20%
29
   faster than the normal "switch" version, depending on the compiler and the
30
   CPU architecture.
31
32
   NOTE: care must be taken that the compiler doesn't try to "optimize" the
33
   indirect jumps by sharing them between all opcodes. Such optimizations
34
   can be disabled on gcc by using the -fno-gcse flag (or possibly
35
   -fno-crossjumping).
36
*/
37
38
/* Use macros rather than inline functions, to make it as clear as possible
39
 * to the C compiler that the tracing check is a simple test then branch.
40
 * We want to be sure that the compiler knows this before it generates
41
 * the CFG.
42
 */
43
44
#ifdef WITH_DTRACE
45
#define OR_DTRACE_LINE | (PyDTrace_LINE_ENABLED() ? 255 : 0)
46
#else
47
#define OR_DTRACE_LINE
48
#endif
49
50
#ifdef HAVE_COMPUTED_GOTOS
51
    #ifndef USE_COMPUTED_GOTOS
52
    #define USE_COMPUTED_GOTOS 1
53
    #endif
54
#else
55
    #if defined(USE_COMPUTED_GOTOS) && USE_COMPUTED_GOTOS
56
    #error "Computed gotos are not supported on this compiler."
57
    #endif
58
    #undef USE_COMPUTED_GOTOS
59
    #define USE_COMPUTED_GOTOS 0
60
#endif
61
62
#ifdef Py_STATS
63
#define INSTRUCTION_STATS(op) \
64
    do { \
65
        PyStats *s = _PyStats_GET(); \
66
        OPCODE_EXE_INC(op); \
67
        if (s) s->opcode_stats[lastopcode].pair_count[op]++; \
68
        lastopcode = op; \
69
    } while (0)
70
#else
71
44.3G
#define INSTRUCTION_STATS(op) ((void)0)
72
#endif
73
74
#ifdef Py_STATS
75
#   define TAIL_CALL_PARAMS _PyInterpreterFrame *frame, _PyStackRef *stack_pointer, PyThreadState *tstate, _Py_CODEUNIT *next_instr, const void *instruction_funcptr_table, int oparg, int lastopcode
76
#   define TAIL_CALL_ARGS frame, stack_pointer, tstate, next_instr, instruction_funcptr_table, oparg, lastopcode
77
#else
78
#   define TAIL_CALL_PARAMS _PyInterpreterFrame *frame, _PyStackRef *stack_pointer, PyThreadState *tstate, _Py_CODEUNIT *next_instr, const void *instruction_funcptr_table, int oparg
79
#   define TAIL_CALL_ARGS frame, stack_pointer, tstate, next_instr, instruction_funcptr_table, oparg
80
#endif
81
82
#if _Py_TAIL_CALL_INTERP
83
#   if defined(__clang__) || defined(__GNUC__)
84
#       if !_Py__has_attribute(preserve_none) || !_Py__has_attribute(musttail)
85
#           error "This compiler does not have support for efficient tail calling."
86
#       endif
87
#   elif defined(_MSC_VER) && (_MSC_VER < 1950)
88
#       error "You need at least VS 2026 / PlatformToolset v145 for tail calling."
89
#   endif
90
#   if defined(_MSC_VER) && !defined(__clang__)
91
#      define Py_MUSTTAIL [[msvc::musttail]]
92
#      define Py_PRESERVE_NONE_CC __preserve_none
93
#   else
94
#       define Py_MUSTTAIL __attribute__((musttail))
95
#       define Py_PRESERVE_NONE_CC __attribute__((preserve_none))
96
#   endif
97
    typedef PyObject *(Py_PRESERVE_NONE_CC *py_tail_call_funcptr)(TAIL_CALL_PARAMS);
98
99
#   define DISPATCH_TABLE_VAR instruction_funcptr_table
100
#   define DISPATCH_TABLE instruction_funcptr_handler_table
101
#   define TRACING_DISPATCH_TABLE instruction_funcptr_tracing_table
102
#   define TARGET(op) Py_NO_INLINE PyObject *Py_PRESERVE_NONE_CC _TAIL_CALL_##op(TAIL_CALL_PARAMS)
103
104
#   define DISPATCH_GOTO() \
105
        do { \
106
            Py_MUSTTAIL return (((py_tail_call_funcptr *)instruction_funcptr_table)[opcode])(TAIL_CALL_ARGS); \
107
        } while (0)
108
#   define DISPATCH_GOTO_NON_TRACING() \
109
        do { \
110
            Py_MUSTTAIL return (((py_tail_call_funcptr *)DISPATCH_TABLE)[opcode])(TAIL_CALL_ARGS); \
111
        } while (0)
112
#   define JUMP_TO_LABEL(name) \
113
        do { \
114
            Py_MUSTTAIL return (_TAIL_CALL_##name)(TAIL_CALL_ARGS); \
115
        } while (0)
116
#   ifdef Py_STATS
117
#       define JUMP_TO_PREDICTED(name) \
118
            do { \
119
                Py_MUSTTAIL return (_TAIL_CALL_##name)(frame, stack_pointer, tstate, this_instr, instruction_funcptr_table, oparg, lastopcode); \
120
            } while (0)
121
#   else
122
#       define JUMP_TO_PREDICTED(name) \
123
            do { \
124
                Py_MUSTTAIL return (_TAIL_CALL_##name)(frame, stack_pointer, tstate, this_instr, instruction_funcptr_table, oparg); \
125
            } while (0)
126
#   endif
127
#    define LABEL(name) TARGET(name)
128
#elif USE_COMPUTED_GOTOS
129
#  define DISPATCH_TABLE_VAR opcode_targets
130
6.60M
#  define DISPATCH_TABLE opcode_targets_table
131
#  define TRACING_DISPATCH_TABLE opcode_tracing_targets_table
132
44.3G
#  define TARGET(op) TARGET_##op:
133
44.5G
#  define DISPATCH_GOTO() goto *opcode_targets[opcode]
134
6.60M
#  define DISPATCH_GOTO_NON_TRACING() goto *DISPATCH_TABLE[opcode];
135
116M
#  define JUMP_TO_LABEL(name) goto name;
136
246M
#  define JUMP_TO_PREDICTED(name) goto PREDICTED_##name;
137
434M
#  define LABEL(name) name:
138
#else
139
#  define TARGET(op) case op: TARGET_##op:
140
#  define DISPATCH_GOTO() dispatch_code = opcode | tracing_mode ; goto dispatch_opcode
141
#  define DISPATCH_GOTO_NON_TRACING() dispatch_code = opcode; goto dispatch_opcode
142
#  define JUMP_TO_LABEL(name) goto name;
143
#  define JUMP_TO_PREDICTED(name) goto PREDICTED_##name;
144
#  define LABEL(name) name:
145
#endif
146
147
#if (_Py_TAIL_CALL_INTERP || USE_COMPUTED_GOTOS) && _Py_TIER2
148
#  define IS_JIT_TRACING() (DISPATCH_TABLE_VAR == TRACING_DISPATCH_TABLE)
149
#  define ENTER_TRACING() \
150
    DISPATCH_TABLE_VAR = TRACING_DISPATCH_TABLE;
151
#  define LEAVE_TRACING() \
152
    DISPATCH_TABLE_VAR = DISPATCH_TABLE;
153
#else
154
#  define IS_JIT_TRACING() (tracing_mode != 0)
155
#  define ENTER_TRACING() tracing_mode = 255
156
#  define LEAVE_TRACING() tracing_mode = 0
157
#endif
158
159
#if _Py_TIER2
160
#define STOP_TRACING() \
161
    do { \
162
        if (IS_JIT_TRACING()) { \
163
            LEAVE_TRACING(); \
164
            _PyJit_FinalizeTracing(tstate, 0); \
165
        } \
166
    } while (0);
167
#else
168
136M
#define STOP_TRACING() ((void)(0));
169
#endif
170
171
172
/* PRE_DISPATCH_GOTO() does lltrace if enabled. Normally a no-op */
173
#ifdef Py_DEBUG
174
#define PRE_DISPATCH_GOTO() if (frame->lltrace >= 5) { \
175
    lltrace_instruction(frame, stack_pointer, next_instr, opcode, oparg); }
176
#else
177
44.5G
#define PRE_DISPATCH_GOTO() ((void)0)
178
#endif
179
180
#ifdef Py_DEBUG
181
#define LLTRACE_RESUME_FRAME() \
182
do { \
183
    _PyFrame_SetStackPointer(frame, stack_pointer); \
184
    int lltrace = maybe_lltrace_resume_frame(frame, GLOBALS()); \
185
    stack_pointer = _PyFrame_GetStackPointer(frame); \
186
    frame->lltrace = lltrace; \
187
} while (0)
188
#else
189
2.50G
#define LLTRACE_RESUME_FRAME() ((void)0)
190
#endif
191
192
#ifdef Py_GIL_DISABLED
193
#define QSBR_QUIESCENT_STATE(tstate) _Py_qsbr_quiescent_state(((_PyThreadStateImpl *)tstate)->qsbr)
194
#else
195
#define QSBR_QUIESCENT_STATE(tstate)
196
#endif
197
198
199
/* Do interpreter dispatch accounting for tracing and instrumentation */
200
#define DISPATCH() \
201
44.2G
    { \
202
44.2G
        assert(frame->stackpointer == NULL); \
203
44.2G
        NEXTOPARG(); \
204
44.2G
        PRE_DISPATCH_GOTO(); \
205
44.2G
        DISPATCH_GOTO(); \
206
44.2G
    }
207
208
#define DISPATCH_NON_TRACING() \
209
    { \
210
        assert(frame->stackpointer == NULL); \
211
        NEXTOPARG(); \
212
        PRE_DISPATCH_GOTO(); \
213
        DISPATCH_GOTO_NON_TRACING(); \
214
    }
215
216
#define DISPATCH_SAME_OPARG() \
217
6.60M
    { \
218
6.60M
        opcode = next_instr->op.code; \
219
6.60M
        PRE_DISPATCH_GOTO(); \
220
6.60M
        DISPATCH_GOTO_NON_TRACING(); \
221
6.60M
    }
222
223
#define DISPATCH_INLINED(NEW_FRAME)                              \
224
2.24M
    do {                                                         \
225
2.24M
        assert(!IS_PEP523_HOOKED(tstate));                       \
226
2.24M
        _PyFrame_SetStackPointer(frame, stack_pointer);          \
227
2.24M
        assert((NEW_FRAME)->previous == frame);                  \
228
2.24M
        frame = tstate->current_frame = (NEW_FRAME);             \
229
2.24M
        CALL_STAT_INC(inlined_py_calls);                         \
230
2.24M
        JUMP_TO_LABEL(start_frame);                              \
231
0
    } while (0)
232
233
/* Tuple access macros */
234
235
#ifndef Py_DEBUG
236
2.80G
#define GETITEM(v, i) PyTuple_GET_ITEM((v), (i))
237
#else
238
static inline PyObject *
239
GETITEM(PyObject *v, Py_ssize_t i) {
240
    assert(PyTuple_Check(v));
241
    assert(i >= 0);
242
    assert(i < PyTuple_GET_SIZE(v));
243
    return PyTuple_GET_ITEM(v, i);
244
}
245
#endif
246
247
/* Code access macros */
248
249
/* The integer overflow is checked by an assertion below. */
250
70.8M
#define INSTR_OFFSET() ((int)(next_instr - _PyFrame_GetBytecode(frame)))
251
44.2G
#define NEXTOPARG()  do { \
252
44.2G
        _Py_CODEUNIT word  = {.cache = FT_ATOMIC_LOAD_UINT16_RELAXED(*(uint16_t*)next_instr)}; \
253
44.2G
        opcode = word.op.code; \
254
44.2G
        oparg = word.op.arg; \
255
44.2G
    } while (0)
256
257
/* JUMPBY makes the generator identify the instruction as a jump. SKIP_OVER is
258
 * for advancing to the next instruction, taking into account cache entries
259
 * and skipped instructions.
260
 */
261
6.95G
#define JUMPBY(x)       (next_instr += (x))
262
#define SKIP_OVER(x)    (next_instr += (x))
263
264
#define STACK_LEVEL()     ((int)(stack_pointer - _PyFrame_Stackbase(frame)))
265
#define STACK_SIZE()      (_PyFrame_GetCode(frame)->co_stacksize)
266
267
#define WITHIN_STACK_BOUNDS() \
268
   (frame->owner == FRAME_OWNED_BY_INTERPRETER || (STACK_LEVEL() >= 0 && STACK_LEVEL() <= STACK_SIZE()))
269
270
#if defined(Py_DEBUG) && !defined(_Py_JIT)
271
// This allows temporary stack "overflows", provided it's all in the cache at any point of time.
272
#define WITHIN_STACK_BOUNDS_IGNORING_CACHE() \
273
   (frame->owner == FRAME_OWNED_BY_INTERPRETER || (STACK_LEVEL() >= 0 && (STACK_LEVEL()) <= STACK_SIZE()))
274
#else
275
#define WITHIN_STACK_BOUNDS_IGNORING_CACHE WITHIN_STACK_BOUNDS
276
#endif
277
278
/* Data access macros */
279
#define FRAME_CO_CONSTS (_PyFrame_GetCode(frame)->co_consts)
280
#define FRAME_CO_NAMES  (_PyFrame_GetCode(frame)->co_names)
281
282
/* Local variable macros */
283
284
1.21M
#define LOCALS_ARRAY    (frame->localsplus)
285
19.5G
#define GETLOCAL(i)     (frame->localsplus[i])
286
287
288
#ifdef Py_STATS
289
#define UPDATE_MISS_STATS(INSTNAME)                              \
290
    do {                                                         \
291
        STAT_INC(opcode, miss);                                  \
292
        STAT_INC((INSTNAME), miss);                              \
293
        /* The counter is always the first cache entry: */       \
294
        if (ADAPTIVE_COUNTER_TRIGGERS(next_instr->cache)) {       \
295
            STAT_INC((INSTNAME), deopt);                         \
296
        }                                                        \
297
    } while (0)
298
#else
299
246M
#define UPDATE_MISS_STATS(INSTNAME) ((void)0)
300
#endif
301
302
303
// Try to lock an object in the free threading build, if it's not already
304
// locked. Use with a DEOPT_IF() to deopt if the object is already locked.
305
// These are no-ops in the default GIL build. The general pattern is:
306
//
307
// DEOPT_IF(!LOCK_OBJECT(op));
308
// if (/* condition fails */) {
309
//     UNLOCK_OBJECT(op);
310
//     DEOPT_IF(true);
311
//  }
312
//  ...
313
//  UNLOCK_OBJECT(op);
314
//
315
// NOTE: The object must be unlocked on every exit code path and you should
316
// avoid any potentially escaping calls (like PyStackRef_CLOSE) while the
317
// object is locked.
318
#ifdef Py_GIL_DISABLED
319
#  define LOCK_OBJECT(op) PyMutex_LockFast(&(_PyObject_CAST(op))->ob_mutex)
320
#  define UNLOCK_OBJECT(op) PyMutex_Unlock(&(_PyObject_CAST(op))->ob_mutex)
321
#else
322
708M
#  define LOCK_OBJECT(op) (1)
323
708M
#  define UNLOCK_OBJECT(op) ((void)0)
324
#endif
325
326
1.81G
#define GLOBALS() frame->f_globals
327
679M
#define BUILTINS() frame->f_builtins
328
3.91M
#define LOCALS() frame->f_locals
329
#define CONSTS() _PyFrame_GetCode(frame)->co_consts
330
#define NAMES() _PyFrame_GetCode(frame)->co_names
331
332
#define DTRACE_FUNCTION_ENTRY()  \
333
    if (PyDTrace_FUNCTION_ENTRY_ENABLED()) { \
334
        dtrace_function_entry(frame); \
335
    }
336
337
/* This takes a uint16_t instead of a _Py_BackoffCounter,
338
 * because it is used directly on the cache entry in generated code,
339
 * which is always an integral type. */
340
// Force re-specialization when tracing a side exit to get good side exits.
341
#define ADAPTIVE_COUNTER_TRIGGERS(COUNTER) \
342
1.61G
    backoff_counter_triggers(forge_backoff_counter((COUNTER)))
343
344
#define ADVANCE_ADAPTIVE_COUNTER(COUNTER) \
345
1.61G
    do { \
346
1.61G
        (COUNTER) = advance_backoff_counter((COUNTER)); \
347
1.61G
    } while (0);
348
349
#define PAUSE_ADAPTIVE_COUNTER(COUNTER) \
350
0
    do { \
351
0
        (COUNTER) = pause_backoff_counter((COUNTER)); \
352
0
    } while (0);
353
354
#ifdef ENABLE_SPECIALIZATION
355
/* Multiple threads may execute these concurrently if thread-local bytecode is
356
 * disabled and they all execute the main copy of the bytecode. Specialization
357
 * is disabled in that case so the value is unused, but the RMW cycle should be
358
 * free of data races.
359
 */
360
#define RECORD_BRANCH_TAKEN(bitset, flag) \
361
3.47G
    FT_ATOMIC_STORE_UINT16_RELAXED(       \
362
3.47G
        bitset, (FT_ATOMIC_LOAD_UINT16_RELAXED(bitset) << 1) | (flag))
363
#else
364
#define RECORD_BRANCH_TAKEN(bitset, flag)
365
#endif
366
367
#define UNBOUNDLOCAL_ERROR_MSG \
368
0
    "cannot access local variable '%s' where it is not associated with a value"
369
#define UNBOUNDFREE_ERROR_MSG \
370
0
    "cannot access free variable '%s' where it is not associated with a value" \
371
0
    " in enclosing scope"
372
95
#define NAME_ERROR_MSG "name '%.200s' is not defined"
373
374
// If a trace function sets a new f_lineno and
375
// *then* raises, we use the destination when searching
376
// for an exception handler, displaying the traceback, and so on
377
0
#define INSTRUMENTED_JUMP(src, dest, event) \
378
0
do { \
379
0
    if (tstate->tracing) {\
380
0
        next_instr = dest; \
381
0
    } else { \
382
0
        _PyFrame_SetStackPointer(frame, stack_pointer); \
383
0
        next_instr = _Py_call_instrumentation_jump(this_instr, tstate, event, frame, src, dest); \
384
0
        stack_pointer = _PyFrame_GetStackPointer(frame); \
385
0
        if (next_instr == NULL) { \
386
0
            next_instr = (dest)+1; \
387
0
            JUMP_TO_LABEL(error); \
388
0
        } \
389
0
    } \
390
0
} while (0);
391
392
393
255M
static inline int _Py_EnterRecursivePy(PyThreadState *tstate) {
394
255M
    return (tstate->py_recursion_remaining-- <= 0) &&
395
311
        _Py_CheckRecursiveCallPy(tstate);
396
255M
}
397
398
1.28G
static inline void _Py_LeaveRecursiveCallPy(PyThreadState *tstate)  {
399
1.28G
    tstate->py_recursion_remaining++;
400
1.28G
}
401
402
/* Implementation of "macros" that modify the instruction pointer,
403
 * stack pointer, or frame pointer.
404
 * These need to treated differently by tier 1 and 2.
405
 * The Tier 1 version is here; Tier 2 is inlined in ceval.c. */
406
407
2.26G
#define LOAD_IP(OFFSET) do { \
408
2.26G
        next_instr = frame->instr_ptr + (OFFSET); \
409
2.26G
    } while (0)
410
411
/* There's no STORE_IP(), it's inlined by the code generator. */
412
413
1.02G
#define LOAD_SP() \
414
1.02G
stack_pointer = _PyFrame_GetStackPointer(frame)
415
416
#define SAVE_SP() \
417
_PyFrame_SetStackPointer(frame, stack_pointer)
418
419
/* Tier-switching macros. */
420
421
#define TIER1_TO_TIER2(EXECUTOR)                        \
422
do {                                                   \
423
    OPT_STAT_INC(traces_executed);                     \
424
    next_instr = _Py_jit_entry((EXECUTOR), frame, stack_pointer, tstate); \
425
    frame = tstate->current_frame;                     \
426
    stack_pointer = _PyFrame_GetStackPointer(frame);   \
427
    int keep_tracing_bit = (uintptr_t)next_instr & 1;   \
428
    next_instr = (_Py_CODEUNIT *)(((uintptr_t)next_instr) & (~1)); \
429
    if (next_instr == NULL) {                          \
430
        /* gh-140104: The exception handler expects frame->instr_ptr
431
            to after this_instr, not this_instr! */ \
432
        next_instr = frame->instr_ptr + 1;                 \
433
        JUMP_TO_LABEL(error);                          \
434
    }                                                  \
435
    if (keep_tracing_bit) { \
436
        assert(uop_buffer_length(&((_PyThreadStateImpl *)tstate)->jit_tracer_state->code_buffer)); \
437
        ENTER_TRACING(); \
438
        DISPATCH_NON_TRACING(); \
439
    } \
440
    DISPATCH();                                        \
441
} while (0)
442
443
#define TIER2_TO_TIER2(EXECUTOR) \
444
do {                                                   \
445
    OPT_STAT_INC(traces_executed);                     \
446
    current_executor = (EXECUTOR);                     \
447
    goto tier2_start;                                  \
448
} while (0)
449
450
#define GOTO_TIER_ONE_SETUP \
451
    tstate->current_executor = NULL;                              \
452
    OPT_HIST(trace_uop_execution_counter, trace_run_length_hist); \
453
    _PyFrame_SetStackPointer(frame, stack_pointer);
454
455
#define GOTO_TIER_ONE(TARGET) \
456
    do \
457
    { \
458
        GOTO_TIER_ONE_SETUP \
459
        return (_Py_CODEUNIT *)(TARGET); \
460
    } while (0)
461
462
#define GOTO_TIER_ONE_CONTINUE_TRACING(TARGET) \
463
    do \
464
    { \
465
        GOTO_TIER_ONE_SETUP \
466
        return (_Py_CODEUNIT *)(((uintptr_t)(TARGET))| 1); \
467
    } while (0)
468
469
#define CURRENT_OPARG()    (next_uop[-1].oparg)
470
#define CURRENT_OPERAND0_64() (next_uop[-1].operand0)
471
#define CURRENT_OPERAND1_64() (next_uop[-1].operand1)
472
#define CURRENT_OPERAND0_32() (next_uop[-1].operand0)
473
#define CURRENT_OPERAND1_32() (next_uop[-1].operand1)
474
#define CURRENT_OPERAND0_16() (next_uop[-1].operand0)
475
#define CURRENT_OPERAND1_16() (next_uop[-1].operand1)
476
#define CURRENT_TARGET()   (next_uop[-1].target)
477
478
#define JUMP_TO_JUMP_TARGET() goto jump_to_jump_target
479
#define JUMP_TO_ERROR() goto jump_to_error_target
480
481
/* Stackref macros */
482
483
/* How much scratch space to give stackref to PyObject* conversion. */
484
1.02G
#define MAX_STACKREF_SCRATCH 10
485
486
#define STACKREFS_TO_PYOBJECTS(ARGS, ARG_COUNT, NAME) \
487
    /* +1 because vectorcall might use -1 to write self */ \
488
1.02G
    PyObject *NAME##_temp[MAX_STACKREF_SCRATCH+1]; \
489
1.02G
    PyObject **NAME = _PyObjectArray_FromStackRefArray(ARGS, ARG_COUNT, NAME##_temp);
490
491
#define STACKREFS_TO_PYOBJECTS_CLEANUP(NAME) \
492
    /* +1 because we +1 previously */ \
493
1.02G
    _PyObjectArray_Free(NAME - 1, NAME##_temp);
494
495
1.02G
#define CONVERSION_FAILED(NAME) ((NAME) == NULL)
496
497
#if defined(Py_DEBUG) && !defined(_Py_JIT)
498
#define SET_CURRENT_CACHED_VALUES(N) current_cached_values = (N)
499
#define CHECK_CURRENT_CACHED_VALUES(N) assert(current_cached_values == (N))
500
#else
501
#define SET_CURRENT_CACHED_VALUES(N) ((void)0)
502
#define CHECK_CURRENT_CACHED_VALUES(N) ((void)0)
503
#endif
504
505
1.07G
#define IS_PEP523_HOOKED(tstate) (tstate->interp->eval_frame != NULL)
506
507
static inline int
508
2.66G
check_periodics(PyThreadState *tstate) {
509
2.66G
    _Py_CHECK_EMSCRIPTEN_SIGNALS_PERIODICALLY();
510
2.66G
    QSBR_QUIESCENT_STATE(tstate);
511
2.66G
    if (_Py_atomic_load_uintptr_relaxed(&tstate->eval_breaker) & _PY_EVAL_EVENTS_MASK) {
512
92.4k
        return _Py_HandlePending(tstate);
513
92.4k
    }
514
2.66G
    return 0;
515
2.66G
}
516
517
// Mark the generator as executing. Returns true if the state was changed,
518
// false if it was already executing or finished.
519
static inline bool
520
gen_try_set_executing(PyGenObject *gen)
521
28.2M
{
522
#ifdef Py_GIL_DISABLED
523
    if (!_PyObject_IsUniquelyReferenced((PyObject *)gen)) {
524
        int8_t frame_state = _Py_atomic_load_int8_relaxed(&gen->gi_frame_state);
525
        while (frame_state < FRAME_SUSPENDED_YIELD_FROM_LOCKED) {
526
            if (_Py_atomic_compare_exchange_int8(&gen->gi_frame_state,
527
                                                 &frame_state,
528
                                                 FRAME_EXECUTING)) {
529
                return true;
530
            }
531
        }
532
        // NB: We return false for FRAME_SUSPENDED_YIELD_FROM_LOCKED as well.
533
        // That case is rare enough that we can just handle it in the deopt.
534
        return false;
535
    }
536
#endif
537
    // Use faster non-atomic modifications in the GIL-enabled build and when
538
    // the object is uniquely referenced in the free-threaded build.
539
28.2M
    if (gen->gi_frame_state < FRAME_EXECUTING) {
540
28.2M
        assert(gen->gi_frame_state != FRAME_SUSPENDED_YIELD_FROM_LOCKED);
541
28.2M
        gen->gi_frame_state = FRAME_EXECUTING;
542
28.2M
        return true;
543
28.2M
    }
544
0
    return false;
545
28.2M
}
546
547
// Macro for inplace float binary ops (tier 2 only).
548
// Mutates the uniquely-referenced TARGET operand in place.
549
// TARGET must be either left or right.
550
#define FLOAT_INPLACE_OP(left, right, TARGET, OP)                        \
551
    do {                                                                 \
552
        PyObject *left_o = PyStackRef_AsPyObjectBorrow(left);            \
553
        PyObject *right_o = PyStackRef_AsPyObjectBorrow(right);          \
554
        assert(PyFloat_CheckExact(left_o));                              \
555
        assert(PyFloat_CheckExact(right_o));                             \
556
        assert(_PyObject_IsUniquelyReferenced(                           \
557
            PyStackRef_AsPyObjectBorrow(TARGET)));                       \
558
        STAT_INC(BINARY_OP, hit);                                        \
559
        double _dres =                                                   \
560
            ((PyFloatObject *)left_o)->ob_fval                           \
561
            OP ((PyFloatObject *)right_o)->ob_fval;                      \
562
        ((PyFloatObject *)PyStackRef_AsPyObjectBorrow(TARGET))           \
563
            ->ob_fval = _dres;                                           \
564
    } while (0)
565
566
// Inplace float true division. Sets _divop_err to 1 on zero division.
567
// Caller must check _divop_err and call ERROR_NO_POP() if set.
568
#define FLOAT_INPLACE_DIVOP(left, right, TARGET)                         \
569
    int _divop_err = 0;                                                  \
570
    do {                                                                 \
571
        PyObject *left_o = PyStackRef_AsPyObjectBorrow(left);            \
572
        PyObject *right_o = PyStackRef_AsPyObjectBorrow(right);          \
573
        assert(PyFloat_CheckExact(left_o));                              \
574
        assert(PyFloat_CheckExact(right_o));                             \
575
        assert(_PyObject_IsUniquelyReferenced(                           \
576
            PyStackRef_AsPyObjectBorrow(TARGET)));                       \
577
        STAT_INC(BINARY_OP, hit);                                        \
578
        double _divisor = ((PyFloatObject *)right_o)->ob_fval;           \
579
        if (_divisor == 0.0) {                                           \
580
            PyErr_SetString(PyExc_ZeroDivisionError,                     \
581
                            "float division by zero");                   \
582
            _divop_err = 1;                                              \
583
            break;                                                       \
584
        }                                                                \
585
        double _dres = ((PyFloatObject *)left_o)->ob_fval / _divisor;    \
586
        ((PyFloatObject *)PyStackRef_AsPyObjectBorrow(TARGET))           \
587
            ->ob_fval = _dres;                                           \
588
    } while (0)
589
590
// Inplace compact int operation. TARGET is expected to be uniquely
591
// referenced at the optimizer level, but at runtime it may be a
592
// cached small int singleton. We check _Py_IsImmortal on TARGET
593
// to decide whether inplace mutation is safe.
594
//
595
// After the macro, _int_inplace_res holds the result (may be NULL
596
// on allocation failure). On success, TARGET was mutated in place
597
// and _int_inplace_res is a DUP'd reference to it. On fallback
598
// (small int target, small int result, or overflow), _int_inplace_res
599
// is from FUNC (_PyCompactLong_Add etc.).
600
// FUNC is the fallback function (_PyCompactLong_Add etc.)
601
#define INT_INPLACE_OP(left, right, TARGET, OP, FUNC)                    \
602
    _PyStackRef _int_inplace_res = PyStackRef_NULL;                      \
603
    do {                                                                 \
604
        PyObject *target_o = PyStackRef_AsPyObjectBorrow(TARGET);        \
605
        if (_Py_IsImmortal(target_o)) {                                  \
606
            break;                                                       \
607
        }                                                                \
608
        assert(_PyObject_IsUniquelyReferenced(target_o));                \
609
        Py_ssize_t left_val = _PyLong_CompactValue(                      \
610
            (PyLongObject *)PyStackRef_AsPyObjectBorrow(left));          \
611
        Py_ssize_t right_val = _PyLong_CompactValue(                     \
612
            (PyLongObject *)PyStackRef_AsPyObjectBorrow(right));         \
613
        Py_ssize_t result = left_val OP right_val;                       \
614
        if (!_PY_IS_SMALL_INT(result)                                    \
615
            && ((twodigits)((stwodigits)result) + PyLong_MASK            \
616
                < (twodigits)PyLong_MASK + PyLong_BASE))                 \
617
        {                                                                \
618
            _PyLong_SetSignAndDigitCount(                                \
619
                (PyLongObject *)target_o, result < 0 ? -1 : 1, 1);       \
620
            ((PyLongObject *)target_o)->long_value.ob_digit[0] =         \
621
                (digit)(result < 0 ? -result : result);                  \
622
            _int_inplace_res = PyStackRef_DUP(TARGET);                   \
623
            break;                                                       \
624
        }                                                                \
625
    } while (0);                                                         \
626
    if (PyStackRef_IsNull(_int_inplace_res)) {                           \
627
        _int_inplace_res = FUNC(                                         \
628
            (PyLongObject *)PyStackRef_AsPyObjectBorrow(left),           \
629
            (PyLongObject *)PyStackRef_AsPyObjectBorrow(right));         \
630
    }