Lines Matching full:intid
76 static struct vgic_irq *vgic_add_lpi(struct kvm *kvm, u32 intid, in vgic_add_lpi() argument
80 struct vgic_irq *irq = vgic_get_irq(kvm, intid), *oldirq; in vgic_add_lpi()
92 ret = xa_reserve_irq(&dist->lpi_xa, intid, GFP_KERNEL_ACCOUNT); in vgic_add_lpi()
103 irq->intid = intid; in vgic_add_lpi()
113 oldirq = xa_load(&dist->lpi_xa, intid); in vgic_add_lpi()
122 ret = xa_err(__xa_store(&dist->lpi_xa, intid, irq, 0)); in vgic_add_lpi()
124 xa_release(&dist->lpi_xa, intid); in vgic_add_lpi()
321 ret = kvm_read_guest_lock(kvm, propbase + irq->intid - GIC_LPI_OFFSET, in update_lpi_config()
431 unsigned long intid, flags; in its_sync_lpi_pending_table() local
437 xa_for_each(&dist->lpi_xa, intid, irq) { in its_sync_lpi_pending_table()
440 byte_offset = intid / BITS_PER_BYTE; in its_sync_lpi_pending_table()
441 bit_nr = intid % BITS_PER_BYTE; in its_sync_lpi_pending_table()
457 irq = vgic_get_irq(vcpu->kvm, intid); in its_sync_lpi_pending_table()
1332 unsigned long intid; in vgic_its_invall() local
1334 xa_for_each(&dist->lpi_xa, intid, irq) { in vgic_its_invall()
1335 irq = vgic_get_irq(kvm, intid); in vgic_its_invall()
1388 unsigned long intid; in vgic_its_cmd_handle_movall() local
1400 xa_for_each(&dist->lpi_xa, intid, irq) { in vgic_its_cmd_handle_movall()
1401 irq = vgic_get_irq(kvm, intid); in vgic_its_cmd_handle_movall()
2144 ((u64)ite->irq->intid << KVM_ITS_ITE_PINTID_SHIFT) | in vgic_its_save_ite()