Searched refs:allocated_stack (Results 1 – 4 of 4) sorted by relevance
331 int allocated_stack; member464 (((slot < frame->allocated_stack / BPF_REG_SIZE) && \471 iter < frame->allocated_stack / BPF_REG_SIZE; \
277 __naked void allocated_stack(void) in allocated_stack() function
597 int allocated_slots = state->allocated_stack / BPF_REG_SIZE; in is_spi_bounds_valid()841 for (i = 1; i < state->allocated_stack / BPF_REG_SIZE; i++) { in unmark_stack_slots_dynptr()1417 size_t n = src->allocated_stack / BPF_REG_SIZE; in copy_stack_state()1424 dst->allocated_stack = src->allocated_stack; in copy_stack_state()1444 size_t old_n = state->allocated_stack / BPF_REG_SIZE, n; in grow_stack_state()1457 state->allocated_stack = size; in grow_stack_state()4400 for (j = 0; j < func->allocated_stack / BPF_REG_SIZE; j++) { in mark_all_scalars_precise()4430 for (j = 0; j < func->allocated_stack / BPF_REG_SIZE; j++) { in mark_all_scalars_imprecise()4659 if (i >= func->allocated_stack / BPF_REG_SIZE) { in __mark_chain_precision()4661 i, func->allocated_stack / BPF_REG_SIZE); in __mark_chain_precision()[all …]
777 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE; i++) { in print_verifier_state()