/linux-6.14.4/fs/befs/ |
D | datastream.c | 187 /* Size of indirect block */ in befs_count_blocks() 189 metablocks += ds->indirect.len; in befs_count_blocks() 192 * Double indir block, plus all the indirect blocks it maps. in befs_count_blocks() 193 * In the double-indirect range, all block runs of data are in befs_count_blocks() 195 * how many data block runs are in the double-indirect region, in befs_count_blocks() 196 * and from that we know how many indirect blocks it takes to in befs_count_blocks() 197 * map them. We assume that the indirect blocks are also in befs_count_blocks() 243 * as in the indirect region code). 291 * blockno is in the indirect region of the datastream. 297 * For each block in the indirect run of the datastream, read [all …]
|
/linux-6.14.4/tools/perf/pmu-events/arch/arm64/arm/cortex-a510/ |
D | branch.json | 24 …"PublicDescription": "Indirect branch mispredicted. This event counts when any indirect branch tha… 27 …"BriefDescription": "Indirect branch mispredicted. This event counts when any indirect branch that… 30 …"PublicDescription": "Indirect branch mispredicted due to address miscompare. This event counts wh… 33 …"BriefDescription": "Indirect branch mispredicted due to address miscompare. This event counts whe… 36 …hen branch prediction is disabled due to the MMU being off. Conditional indirect branches that cor… 39 …hen branch prediction is disabled due to the MMU being off. Conditional indirect branches that cor… 42 …"PublicDescription": "Indirect branch with predicted address executed. This event counts when any … 45 …"BriefDescription": "Indirect branch with predicted address executed. This event counts when any i…
|
/linux-6.14.4/tools/perf/pmu-events/arch/arm64/arm/cortex-a55/ |
D | branch.json | 24 …"PublicDescription": "Indirect branch mis-predicted.This event counts when any indirect branch whi… 27 …"BriefDescription": "Indirect branch mis-predicted.This event counts when any indirect branch whic… 30 …"PublicDescription": "Indirect branch mis-predicted due to address mis-compare.This event counts w… 33 …"BriefDescription": "Indirect branch mis-predicted due to address mis-compare.This event counts wh… 36 …hen branch prediction is disabled due to the MMU being off. Conditional indirect branches which co… 39 …hen branch prediction is disabled due to the MMU being off. Conditional indirect branches which co… 42 …"PublicDescription": "Indirect branch with predicted address executed.This event counts when any i… 45 …"BriefDescription": "Indirect branch with predicted address executed.This event counts when any in…
|
/linux-6.14.4/tools/perf/pmu-events/arch/arm64/ampere/ampereonex/ |
D | branch.json | 30 …"PublicDescription": "Instruction architecturally executed, indirect branch excluding return retir… 33 …"BriefDescription": "Instruction architecturally executed, indirect branch excluding return retire… 48 "PublicDescription": "Instruction architecturally executed, predicted indirect branch", 51 "BriefDescription": "Instruction architecturally executed, predicted indirect branch" 54 "PublicDescription": "Instruction architecturally executed, mispredicted indirect branch", 57 "BriefDescription": "Instruction architecturally executed, mispredicted indirect branch" 72 …"PublicDescription": "Instruction architecturally executed, predicted indirect branch excluding re… 75 …"BriefDescription": "Instruction architecturally executed, predicted indirect branch excluding ret… 78 …"PublicDescription": "Instruction architecturally executed, mispredicted indirect branch excluding… 81 …"BriefDescription": "Instruction architecturally executed, mispredicted indirect branch excluding … [all …]
|
/linux-6.14.4/fs/minix/ |
D | itree_common.c | 8 } Indirect; typedef 12 static inline void add_chain(Indirect *p, struct buffer_head *bh, block_t *v) in add_chain() 18 static inline int verify_chain(Indirect *from, Indirect *to) in verify_chain() 30 static inline Indirect *get_branch(struct inode *inode, in get_branch() 33 Indirect chain[DEPTH], in get_branch() 37 Indirect *p = chain; in get_branch() 73 Indirect *branch) in alloc_branch() 116 Indirect chain[DEPTH], in splice_branch() 117 Indirect *where, in splice_branch() 136 /* had we spliced it onto indirect block? */ in splice_branch() [all …]
|
/linux-6.14.4/fs/ext4/ |
D | indirect.c | 3 * linux/fs/ext4/indirect.c 35 } Indirect; typedef 37 static inline void add_chain(Indirect *p, struct buffer_head *bh, __le32 *v) in add_chain() 49 * followed (on disk) by an indirect block. 53 * data blocks at leaves and indirect blocks in intermediate nodes. 60 * we need to know is the capacity of indirect blocks (taken from the 66 * indirect block) is spelled differently, because otherwise on an 115 * ext4_get_branch - read the chain of indirect blocks leading to data 118 * @offsets: offsets of pointers in inode/indirect blocks 128 * for i>0) and chain[i].bh points to the buffer_head of i-th indirect [all …]
|
/linux-6.14.4/Documentation/admin-guide/hw-vuln/ |
D | spectre.rst | 62 execution of indirect branches to leak privileged memory. 93 execution of indirect branches :ref:`[3] <spec_ref3>`. The indirect 95 indirect branches can be influenced by an attacker, causing gadget code 102 In Spectre variant 2 attacks, the attacker can steer speculative indirect 104 buffer of a CPU used for predicting indirect branch addresses. Such 105 poisoning could be done by indirect branching into existing code, 106 with the address offset of the indirect branch under the attacker's 109 this could cause privileged code's indirect branch to jump to a gadget 130 steer its indirect branch speculations to gadget code, and measure the 135 Branch History Buffer (BHB) to speculatively steer an indirect branch [all …]
|
/linux-6.14.4/drivers/net/ethernet/intel/i40e/ |
D | i40e_adminq_cmd.h | 285 /* command structures and indirect data structures */ 289 * - _data for indirect sent data 290 * - _resp for indirect return data (data which is both will use _data) 325 /* Send driver version (indirect 0x0002) */ 368 /* Get function capabilities (indirect 0x000A) 369 * Get device capabilities (indirect 0x000B) 437 /* Set ARP Proxy command / response (indirect 0x0104) */ 449 /* Set NS Proxy Table Entry Command (indirect 0x0105) */ 476 /* Manage MAC Address Read Command (indirect 0x0107) */ 556 /* Used by many indirect commands that only pass an seid and a buffer in the [all …]
|
/linux-6.14.4/sound/mips/ |
D | hal2.h | 13 /* Indirect status register */ 28 /* Indirect address register */ 31 * Address of indirect internal register to be accessed. A write to this 32 * register initiates read or write access to the indirect registers in the 33 * HAL2. Note that there af four indirect data registers for write access to 44 /* blockin which the indirect */ 71 * The HAL2 has "indirect registers" (idr) which are accessed by writing to the 72 * Indirect Data registers. Write the address to the Indirect Address register 78 * When we write to indirect registers which are larger than one word (16 bit) 79 * we have to fill more than one indirect register before writing. When we read [all …]
|
/linux-6.14.4/fs/sysv/ |
D | itree.c | 5 * Handling of indirect blocks' trees. 15 enum {DIRECT = 10, DEPTH = 4}; /* Have triple indirect */ 64 } Indirect; typedef 68 static inline void add_chain(Indirect *p, struct buffer_head *bh, sysv_zone_t *v) in add_chain() 74 static inline int verify_chain(Indirect *from, Indirect *to) in verify_chain() 86 static Indirect *get_branch(struct inode *inode, in get_branch() 89 Indirect chain[], in get_branch() 93 Indirect *p = chain; in get_branch() 129 Indirect *branch) in alloc_branch() 174 Indirect chain[], in splice_branch() [all …]
|
/linux-6.14.4/drivers/gpu/drm/amd/amdgpu/ |
D | vcn_v5_0_1.c | 324 * @indirect: indirectly write sram 328 static void vcn_v5_0_1_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in vcn_v5_0_1_mc_resume_dpg_mode() argument 338 if (!indirect) { in vcn_v5_0_1_mc_resume_dpg_mode() 342 inst_idx].tmr_mc_addr_lo), 0, indirect); in vcn_v5_0_1_mc_resume_dpg_mode() 346 inst_idx].tmr_mc_addr_hi), 0, indirect); in vcn_v5_0_1_mc_resume_dpg_mode() 348 VCN, 0, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v5_0_1_mc_resume_dpg_mode() 351 VCN, 0, regUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v5_0_1_mc_resume_dpg_mode() 353 VCN, 0, regUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v5_0_1_mc_resume_dpg_mode() 355 VCN, 0, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v5_0_1_mc_resume_dpg_mode() 361 lower_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v5_0_1_mc_resume_dpg_mode() [all …]
|
D | vcn_v4_0_5.c | 425 * @indirect: indirectly write sram 429 static void vcn_v4_0_5_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in vcn_v4_0_5_mc_resume_dpg_mode() argument 439 if (!indirect) { in vcn_v4_0_5_mc_resume_dpg_mode() 443 0, indirect); in vcn_v4_0_5_mc_resume_dpg_mode() 447 0, indirect); in vcn_v4_0_5_mc_resume_dpg_mode() 449 VCN, inst_idx, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v4_0_5_mc_resume_dpg_mode() 452 VCN, inst_idx, regUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v4_0_5_mc_resume_dpg_mode() 454 VCN, inst_idx, regUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v4_0_5_mc_resume_dpg_mode() 456 VCN, inst_idx, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v4_0_5_mc_resume_dpg_mode() 462 lower_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v4_0_5_mc_resume_dpg_mode() [all …]
|
D | vcn_v4_0_3.c | 97 int inst_idx, bool indirect); 476 * @indirect: indirectly write sram 480 static void vcn_v4_0_3_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in vcn_v4_0_3_mc_resume_dpg_mode() argument 490 if (!indirect) { in vcn_v4_0_3_mc_resume_dpg_mode() 494 inst_idx].tmr_mc_addr_lo), 0, indirect); in vcn_v4_0_3_mc_resume_dpg_mode() 498 inst_idx].tmr_mc_addr_hi), 0, indirect); in vcn_v4_0_3_mc_resume_dpg_mode() 500 VCN, 0, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v4_0_3_mc_resume_dpg_mode() 503 VCN, 0, regUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v4_0_3_mc_resume_dpg_mode() 505 VCN, 0, regUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v4_0_3_mc_resume_dpg_mode() 507 VCN, 0, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v4_0_3_mc_resume_dpg_mode() [all …]
|
D | vcn_v5_0_0.c | 405 * @indirect: indirectly write sram 409 static void vcn_v5_0_0_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in vcn_v5_0_0_mc_resume_dpg_mode() argument 419 if (!indirect) { in vcn_v5_0_0_mc_resume_dpg_mode() 422 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_lo), 0, indirect); in vcn_v5_0_0_mc_resume_dpg_mode() 425 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_hi), 0, indirect); in vcn_v5_0_0_mc_resume_dpg_mode() 427 VCN, inst_idx, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v5_0_0_mc_resume_dpg_mode() 430 VCN, inst_idx, regUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v5_0_0_mc_resume_dpg_mode() 432 VCN, inst_idx, regUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v5_0_0_mc_resume_dpg_mode() 434 VCN, inst_idx, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v5_0_0_mc_resume_dpg_mode() 440 lower_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v5_0_0_mc_resume_dpg_mode() [all …]
|
D | vcn_v2_5.c | 515 static void vcn_v2_5_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in vcn_v2_5_mc_resume_dpg_mode() argument 522 if (!indirect) { in vcn_v2_5_mc_resume_dpg_mode() 525 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_lo), 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 528 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_hi), 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 530 VCN, 0, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 533 VCN, 0, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 535 VCN, 0, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 537 VCN, 0, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 543 lower_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 546 upper_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() [all …]
|
D | vcn_v2_0.c | 429 static void vcn_v2_0_mc_resume_dpg_mode(struct amdgpu_device *adev, bool indirect) in vcn_v2_0_mc_resume_dpg_mode() argument 436 if (!indirect) { in vcn_v2_0_mc_resume_dpg_mode() 439 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN].tmr_mc_addr_lo), 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 442 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN].tmr_mc_addr_hi), 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 444 UVD, 0, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 447 UVD, 0, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 449 UVD, 0, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 451 UVD, 0, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 457 lower_32_bits(adev->vcn.inst->gpu_addr), 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 460 upper_32_bits(adev->vcn.inst->gpu_addr), 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() [all …]
|
D | vcn_v4_0.c | 486 * @indirect: indirectly write sram 490 static void vcn_v4_0_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in vcn_v4_0_mc_resume_dpg_mode() argument 499 if (!indirect) { in vcn_v4_0_mc_resume_dpg_mode() 502 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_lo), 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 505 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_hi), 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 507 VCN, inst_idx, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 510 VCN, inst_idx, regUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 512 VCN, inst_idx, regUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 514 VCN, inst_idx, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 520 lower_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() [all …]
|
D | jpeg_v5_0_0.c | 284 int inst_idx, uint8_t indirect) in jpeg_engine_5_0_0_dpg_clock_gating_mode() argument 297 if (indirect) { in jpeg_engine_5_0_0_dpg_clock_gating_mode() 298 ADD_SOC24_JPEG_TO_DPG_SRAM(inst_idx, vcnipJPEG_CGC_CTRL, data, indirect); in jpeg_engine_5_0_0_dpg_clock_gating_mode() 302 ADD_SOC24_JPEG_TO_DPG_SRAM(inst_idx, vcnipJPEG_CGC_GATE, data, indirect); in jpeg_engine_5_0_0_dpg_clock_gating_mode() 304 WREG32_SOC24_JPEG_DPG_MODE(inst_idx, vcnipJPEG_CGC_CTRL, data, indirect); in jpeg_engine_5_0_0_dpg_clock_gating_mode() 308 WREG32_SOC24_JPEG_DPG_MODE(inst_idx, vcnipJPEG_CGC_GATE, data, indirect); in jpeg_engine_5_0_0_dpg_clock_gating_mode() 317 * @indirect: indirectly write sram 321 static int jpeg_v5_0_0_start_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in jpeg_v5_0_0_start_dpg_mode() argument 333 if (indirect) in jpeg_v5_0_0_start_dpg_mode() 337 jpeg_engine_5_0_0_dpg_clock_gating_mode(adev, inst_idx, indirect); in jpeg_v5_0_0_start_dpg_mode() [all …]
|
/linux-6.14.4/drivers/net/ethernet/intel/ice/ |
D | ice_adminq_cmd.h | 43 /* Send driver version (indirect 0x0002) */ 93 /* Get function capabilities (indirect 0x000A) 94 * Get device capabilities (indirect 0x000B) 149 /* Manage MAC address, read command - indirect (0x0107) 264 * Get Resource Allocation command (indirect 0x0204) 265 * Allocate Resources command (indirect 0x0208) 266 * Free Resources command (indirect 0x0209) 267 * Get Allocated Resource Descriptors Command (indirect 0x020A) 268 * Share Resource command (indirect 0x020B) 293 /* Allocate Resources command (indirect 0x0208) [all …]
|
/linux-6.14.4/arch/x86/kernel/ |
D | ksysfs.c | 95 struct setup_indirect *indirect; in get_setup_data_size() local 114 indirect = (struct setup_indirect *)data->data; in get_setup_data_size() 116 if (indirect->type != SETUP_INDIRECT) in get_setup_data_size() 117 *size = indirect->len; in get_setup_data_size() 138 struct setup_indirect *indirect; in type_show() local 162 indirect = (struct setup_indirect *)data->data; in type_show() 164 ret = sprintf(buf, "0x%x\n", indirect->type); in type_show() 179 struct setup_indirect *indirect; in setup_data_data_read() local 203 indirect = (struct setup_indirect *)data->data; in setup_data_data_read() 205 if (indirect->type != SETUP_INDIRECT) { in setup_data_data_read() [all …]
|
D | kdebugfs.c | 49 /* Is it direct data or invalid indirect one? */ in setup_data_read() 91 struct setup_indirect *indirect; in create_setup_data_nodes() local 129 indirect = (struct setup_indirect *)data->data; in create_setup_data_nodes() 131 if (indirect->type != SETUP_INDIRECT) { in create_setup_data_nodes() 132 node->paddr = indirect->addr; in create_setup_data_nodes() 133 node->type = indirect->type; in create_setup_data_nodes() 134 node->len = indirect->len; in create_setup_data_nodes()
|
/linux-6.14.4/arch/m68k/math-emu/ |
D | fp_decode.h | 29 * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES 53 * a0 - will point to source/dest operand for any indirect mode 121 | .long "addr register indirect" 122 | .long "addr register indirect postincrement" 123 | .long "addr register indirect predecrement" 184 | .long "no memory indirect action/reserved","null outer displacement" 196 | test if %pc is the base register for the indirect addr mode 220 | addressing mode: address register indirect 244 | addressing mode: address register indirect with postincrement 263 | addressing mode: address register indirect with predecrement [all …]
|
/linux-6.14.4/fs/ext2/ |
D | inode.c | 118 } Indirect; typedef 120 static inline void add_chain(Indirect *p, struct buffer_head *bh, __le32 *v) in add_chain() 126 static inline int verify_chain(Indirect *from, Indirect *to) in verify_chain() 139 * followed (on disk) by an indirect block. 142 * data blocks at leaves and indirect blocks in intermediate nodes. 149 * we need to know is the capacity of indirect blocks (taken from the 155 * indirect block) is spelled differently, because otherwise on an 206 * ext2_get_branch - read the chain of indirect blocks leading to data 209 * @offsets: offsets of pointers in inode/indirect blocks 219 * for i>0) and chain[i].bh points to the buffer_head of i-th indirect [all …]
|
/linux-6.14.4/drivers/ras/amd/atl/ |
D | access.c | 5 * access.c : DF Indirect Access functions 15 /* Protect the PCI config register pairs used for DF indirect access. */ 19 * Data Fabric Indirect Access uses FICAA/FICAD. 21 * Fabric Indirect Configuration Access Address (FICAA): constructed based 25 * Fabric Indirect Configuration Access Data (FICAD): there are FICAD 111 pr_warn("Error writing DF Indirect FICAA, FICAA=0x%x\n", ficaa); in __df_indirect_read() 117 pr_warn("Error reading DF Indirect FICAD LO, FICAA=0x%x.\n", ficaa); in __df_indirect_read()
|
/linux-6.14.4/tools/perf/pmu-events/arch/x86/silvermont/ |
D | pipeline.json | 8 …e following branch types: conditional branches, direct calls and jumps, indirect calls and jumps, … 17 …e following branch types: conditional branches, direct calls and jumps, indirect calls and jumps, … 27 …e following branch types: conditional branches, direct calls and jumps, indirect calls and jumps, … 37 …e following branch types: conditional branches, direct calls and jumps, indirect calls and jumps, … 42 "BriefDescription": "Counts the number of near indirect CALL branch instructions retired", 47 …indirect CALL branch instructions retired. Branch prediction predicts the branch target and enabl… 57 …e following branch types: conditional branches, direct calls and jumps, indirect calls and jumps, … 62 …"BriefDescription": "Counts the number of near indirect JMP and near indirect CALL branch instruct… 67 …indirect JMP and near indirect CALL branch instructions retired. Branch prediction predicts the b… 77 …e following branch types: conditional branches, direct calls and jumps, indirect calls and jumps, … [all …]
|