Lines Matching +full:cpu +full:- +full:crit
1 // SPDX-License-Identifier: GPL-2.0-only
7 * Kevin Wolf <mail@kevin-wolf.de>
101 if (is_kvmppc_hv_enabled(vcpu->kvm)) in kvmppc_update_int_pending()
113 bool crit; in kvmppc_critical_section() local
115 if (is_kvmppc_hv_enabled(vcpu->kvm)) in kvmppc_critical_section()
121 /* Truncate crit indicators in 32 bit mode */ in kvmppc_critical_section()
127 /* Critical section when crit == r1 */ in kvmppc_critical_section()
128 crit = (crit_raw == crit_r1); in kvmppc_critical_section()
130 crit = crit && !(kvmppc_get_msr(vcpu) & MSR_PR); in kvmppc_critical_section()
132 return crit; in kvmppc_critical_section()
137 vcpu->kvm->arch.kvm_ops->inject_interrupt(vcpu, vec, flags); in kvmppc_inject_interrupt()
170 unsigned long old_pending = vcpu->arch.pending_exceptions; in kvmppc_book3s_dequeue_irqprio()
173 &vcpu->arch.pending_exceptions); in kvmppc_book3s_dequeue_irqprio()
175 kvmppc_update_int_pending(vcpu, vcpu->arch.pending_exceptions, in kvmppc_book3s_dequeue_irqprio()
181 vcpu->stat.queue_intr++; in kvmppc_book3s_queue_irqprio()
184 &vcpu->arch.pending_exceptions); in kvmppc_book3s_queue_irqprio()
237 return test_bit(BOOK3S_IRQPRIO_DECREMENTER, &vcpu->arch.pending_exceptions); in kvmppc_core_pending_dec()
270 if (irq->irq == KVM_INTERRUPT_SET) in kvmppc_core_queue_external()
271 vcpu->arch.external_oneshot = 1; in kvmppc_core_queue_external()
301 bool crit = kvmppc_critical_section(vcpu); in kvmppc_book3s_irqprio_deliver() local
305 deliver = !kvmhv_is_nestedv2() && (kvmppc_get_msr(vcpu) & MSR_EE) && !crit; in kvmppc_book3s_irqprio_deliver()
309 deliver = !kvmhv_is_nestedv2() && (kvmppc_get_msr(vcpu) & MSR_EE) && !crit; in kvmppc_book3s_irqprio_deliver()
384 if (vcpu->arch.external_oneshot) { in clear_irqprio()
385 vcpu->arch.external_oneshot = 0; in clear_irqprio()
396 unsigned long *pending = &vcpu->arch.pending_exceptions; in kvmppc_core_prepare_to_enter()
397 unsigned long old_pending = vcpu->arch.pending_exceptions; in kvmppc_core_prepare_to_enter()
401 if (vcpu->arch.pending_exceptions) in kvmppc_core_prepare_to_enter()
402 printk(KERN_EMERG "KVM: Check pending: %lx\n", vcpu->arch.pending_exceptions); in kvmppc_core_prepare_to_enter()
408 clear_bit(priority, &vcpu->arch.pending_exceptions); in kvmppc_core_prepare_to_enter()
427 ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM; in kvmppc_gpa_to_pfn()
436 ulong shared_page = ((ulong)vcpu->arch.shared) & PAGE_MASK; in kvmppc_gpa_to_pfn()
460 r = vcpu->arch.mmu.xlate(vcpu, eaddr, pte, data, iswrite); in kvmppc_xlate()
462 pte->eaddr = eaddr; in kvmppc_xlate()
463 pte->raddr = eaddr & KVM_PAM; in kvmppc_xlate()
464 pte->vpage = VSID_REAL | eaddr >> 12; in kvmppc_xlate()
465 pte->may_read = true; in kvmppc_xlate()
466 pte->may_write = true; in kvmppc_xlate()
467 pte->may_execute = true; in kvmppc_xlate()
472 if ((vcpu->arch.hflags & BOOK3S_HFLAG_SPLIT_HACK) && in kvmppc_xlate()
474 pte->raddr &= ~SPLIT_HACK_MASK; in kvmppc_xlate()
484 * as used in HEIR, vcpu->arch.last_inst and vcpu->arch.emul_inst.
485 * Like vcpu->arch.last_inst but unlike vcpu->arch.emul_inst, each
486 * half of the value needs byte-swapping if the guest endianness is
497 pc -= 4; in kvmppc_load_last_inst()
535 ret = vcpu->kvm->arch.kvm_ops->get_sregs(vcpu, sregs); in kvm_arch_vcpu_ioctl_get_sregs()
547 ret = vcpu->kvm->arch.kvm_ops->set_sregs(vcpu, sregs); in kvm_arch_vcpu_ioctl_set_sregs()
557 regs->pc = kvmppc_get_pc(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
558 regs->cr = kvmppc_get_cr(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
559 regs->ctr = kvmppc_get_ctr(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
560 regs->lr = kvmppc_get_lr(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
561 regs->xer = kvmppc_get_xer(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
562 regs->msr = kvmppc_get_msr(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
563 regs->srr0 = kvmppc_get_srr0(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
564 regs->srr1 = kvmppc_get_srr1(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
565 regs->pid = kvmppc_get_pid(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
566 regs->sprg0 = kvmppc_get_sprg0(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
567 regs->sprg1 = kvmppc_get_sprg1(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
568 regs->sprg2 = kvmppc_get_sprg2(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
569 regs->sprg3 = kvmppc_get_sprg3(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
570 regs->sprg4 = kvmppc_get_sprg4(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
571 regs->sprg5 = kvmppc_get_sprg5(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
572 regs->sprg6 = kvmppc_get_sprg6(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
573 regs->sprg7 = kvmppc_get_sprg7(vcpu); in kvm_arch_vcpu_ioctl_get_regs()
575 for (i = 0; i < ARRAY_SIZE(regs->gpr); i++) in kvm_arch_vcpu_ioctl_get_regs()
576 regs->gpr[i] = kvmppc_get_gpr(vcpu, i); in kvm_arch_vcpu_ioctl_get_regs()
585 kvmppc_set_pc(vcpu, regs->pc); in kvm_arch_vcpu_ioctl_set_regs()
586 kvmppc_set_cr(vcpu, regs->cr); in kvm_arch_vcpu_ioctl_set_regs()
587 kvmppc_set_ctr(vcpu, regs->ctr); in kvm_arch_vcpu_ioctl_set_regs()
588 kvmppc_set_lr(vcpu, regs->lr); in kvm_arch_vcpu_ioctl_set_regs()
589 kvmppc_set_xer(vcpu, regs->xer); in kvm_arch_vcpu_ioctl_set_regs()
590 kvmppc_set_msr(vcpu, regs->msr); in kvm_arch_vcpu_ioctl_set_regs()
591 kvmppc_set_srr0(vcpu, regs->srr0); in kvm_arch_vcpu_ioctl_set_regs()
592 kvmppc_set_srr1(vcpu, regs->srr1); in kvm_arch_vcpu_ioctl_set_regs()
593 kvmppc_set_sprg0(vcpu, regs->sprg0); in kvm_arch_vcpu_ioctl_set_regs()
594 kvmppc_set_sprg1(vcpu, regs->sprg1); in kvm_arch_vcpu_ioctl_set_regs()
595 kvmppc_set_sprg2(vcpu, regs->sprg2); in kvm_arch_vcpu_ioctl_set_regs()
596 kvmppc_set_sprg3(vcpu, regs->sprg3); in kvm_arch_vcpu_ioctl_set_regs()
597 kvmppc_set_sprg4(vcpu, regs->sprg4); in kvm_arch_vcpu_ioctl_set_regs()
598 kvmppc_set_sprg5(vcpu, regs->sprg5); in kvm_arch_vcpu_ioctl_set_regs()
599 kvmppc_set_sprg6(vcpu, regs->sprg6); in kvm_arch_vcpu_ioctl_set_regs()
600 kvmppc_set_sprg7(vcpu, regs->sprg7); in kvm_arch_vcpu_ioctl_set_regs()
602 for (i = 0; i < ARRAY_SIZE(regs->gpr); i++) in kvm_arch_vcpu_ioctl_set_regs()
603 kvmppc_set_gpr(vcpu, i, regs->gpr[i]); in kvm_arch_vcpu_ioctl_set_regs()
610 return -EOPNOTSUPP; in kvm_arch_vcpu_ioctl_get_fpu()
615 return -EOPNOTSUPP; in kvm_arch_vcpu_ioctl_set_fpu()
624 r = vcpu->kvm->arch.kvm_ops->get_one_reg(vcpu, id, val); in kvmppc_get_one_reg()
625 if (r == -EINVAL) { in kvmppc_get_one_reg()
635 i = id - KVM_REG_PPC_FPR0; in kvmppc_get_one_reg()
644 i = id - KVM_REG_PPC_VSR0; in kvmppc_get_one_reg()
645 val->vsxval[0] = kvmppc_get_vsx_fpr(vcpu, i, 0); in kvmppc_get_one_reg()
646 val->vsxval[1] = kvmppc_get_vsx_fpr(vcpu, i, 1); in kvmppc_get_one_reg()
648 r = -ENXIO; in kvmppc_get_one_reg()
657 if (!vcpu->arch.icp && !vcpu->arch.xive_vcpu) { in kvmppc_get_one_reg()
658 r = -ENXIO; in kvmppc_get_one_reg()
669 if (!vcpu->arch.xive_vcpu) { in kvmppc_get_one_reg()
670 r = -ENXIO; in kvmppc_get_one_reg()
676 r = -ENXIO; in kvmppc_get_one_reg()
680 *val = get_reg_val(id, vcpu->arch.fscr); in kvmppc_get_one_reg()
698 r = -EINVAL; in kvmppc_get_one_reg()
712 r = vcpu->kvm->arch.kvm_ops->set_one_reg(vcpu, id, val); in kvmppc_set_one_reg()
713 if (r == -EINVAL) { in kvmppc_set_one_reg()
723 i = id - KVM_REG_PPC_FPR0; in kvmppc_set_one_reg()
727 vcpu->arch.fp.fpscr = set_reg_val(id, *val); in kvmppc_set_one_reg()
732 i = id - KVM_REG_PPC_VSR0; in kvmppc_set_one_reg()
733 kvmppc_set_vsx_fpr(vcpu, i, 0, val->vsxval[0]); in kvmppc_set_one_reg()
734 kvmppc_set_vsx_fpr(vcpu, i, 1, val->vsxval[1]); in kvmppc_set_one_reg()
736 r = -ENXIO; in kvmppc_set_one_reg()
742 if (!vcpu->arch.icp && !vcpu->arch.xive_vcpu) { in kvmppc_set_one_reg()
743 r = -ENXIO; in kvmppc_set_one_reg()
754 if (!vcpu->arch.xive_vcpu) { in kvmppc_set_one_reg()
755 r = -ENXIO; in kvmppc_set_one_reg()
761 r = -ENXIO; in kvmppc_set_one_reg()
783 r = -EINVAL; in kvmppc_set_one_reg()
791 void kvmppc_core_vcpu_load(struct kvm_vcpu *vcpu, int cpu) in kvmppc_core_vcpu_load() argument
793 vcpu->kvm->arch.kvm_ops->vcpu_load(vcpu, cpu); in kvmppc_core_vcpu_load()
798 vcpu->kvm->arch.kvm_ops->vcpu_put(vcpu); in kvmppc_core_vcpu_put()
803 vcpu->kvm->arch.kvm_ops->set_msr(vcpu, msr); in kvmppc_set_msr()
809 return vcpu->kvm->arch.kvm_ops->vcpu_run(vcpu); in kvmppc_vcpu_run()
822 vcpu->guest_debug = dbg->control; in kvm_arch_vcpu_ioctl_set_guest_debug()
835 return vcpu->kvm->arch.kvm_ops->vcpu_create(vcpu); in kvmppc_core_vcpu_create()
840 vcpu->kvm->arch.kvm_ops->vcpu_free(vcpu); in kvmppc_core_vcpu_free()
845 return vcpu->kvm->arch.kvm_ops->check_requests(vcpu); in kvmppc_core_check_requests()
855 return kvm->arch.kvm_ops->get_dirty_log(kvm, log); in kvm_vm_ioctl_get_dirty_log()
860 kvm->arch.kvm_ops->free_memslot(slot); in kvmppc_core_free_memslot()
865 kvm->arch.kvm_ops->flush_memslot(kvm, memslot); in kvmppc_core_flush_memslot()
873 return kvm->arch.kvm_ops->prepare_memory_region(kvm, old, new, change); in kvmppc_core_prepare_memory_region()
881 kvm->arch.kvm_ops->commit_memory_region(kvm, old, new, change); in kvmppc_core_commit_memory_region()
886 return kvm->arch.kvm_ops->unmap_gfn_range(kvm, range); in kvm_unmap_gfn_range()
891 return kvm->arch.kvm_ops->age_gfn(kvm, range); in kvm_age_gfn()
896 return kvm->arch.kvm_ops->test_age_gfn(kvm, range); in kvm_test_age_gfn()
903 INIT_LIST_HEAD_RCU(&kvm->arch.spapr_tce_tables); in kvmppc_core_init_vm()
904 INIT_LIST_HEAD(&kvm->arch.rtas_tokens); in kvmppc_core_init_vm()
905 mutex_init(&kvm->arch.rtas_token_lock); in kvmppc_core_init_vm()
908 return kvm->arch.kvm_ops->init_vm(kvm); in kvmppc_core_init_vm()
913 kvm->arch.kvm_ops->destroy_vm(kvm); in kvmppc_core_destroy_vm()
917 WARN_ON(!list_empty(&kvm->arch.spapr_tce_tables)); in kvmppc_core_destroy_vm()
925 kfree(kvm->arch.xive_devices.native); in kvmppc_core_destroy_vm()
926 kvm->arch.xive_devices.native = NULL; in kvmppc_core_destroy_vm()
927 kfree(kvm->arch.xive_devices.xics_on_xive); in kvmppc_core_destroy_vm()
928 kvm->arch.xive_devices.xics_on_xive = NULL; in kvmppc_core_destroy_vm()
929 kfree(kvm->arch.xics_device); in kvmppc_core_destroy_vm()
930 kvm->arch.xics_device = NULL; in kvmppc_core_destroy_vm()
945 srcu_idx = srcu_read_lock(&vcpu->kvm->srcu); in kvmppc_h_logical_ci_load()
947 srcu_read_unlock(&vcpu->kvm->srcu, srcu_idx); in kvmppc_h_logical_ci_load()
1006 srcu_idx = srcu_read_lock(&vcpu->kvm->srcu); in kvmppc_h_logical_ci_store()
1008 srcu_read_unlock(&vcpu->kvm->srcu, srcu_idx); in kvmppc_h_logical_ci_store()
1018 return kvm->arch.kvm_ops->hcall_implemented(hcall); in kvmppc_book3s_hcall_implemented()
1037 return kvm_set_irq(kvm, irq_source_id, irq_entry->gsi, in kvm_arch_set_irq_inatomic()
1044 return kvm_set_irq(kvm, irq_source_id, e->gsi, level, line_status); in kvmppc_book3s_set_irq()
1050 entries->gsi = gsi; in kvm_irq_map_gsi()
1051 entries->type = KVM_IRQ_ROUTING_IRQCHIP; in kvm_irq_map_gsi()
1052 entries->set = kvmppc_book3s_set_irq; in kvm_irq_map_gsi()
1053 entries->irqchip.irqchip = 0; in kvm_irq_map_gsi()
1054 entries->irqchip.pin = gsi; in kvm_irq_map_gsi()