/linux-6.14.4/arch/riscv/kernel/ |
D | asm-offsets.c | 135 OFFSET(KVM_ARCH_GUEST_ZERO, kvm_vcpu_arch, guest_context.zero); in asm_offsets() 136 OFFSET(KVM_ARCH_GUEST_RA, kvm_vcpu_arch, guest_context.ra); in asm_offsets() 137 OFFSET(KVM_ARCH_GUEST_SP, kvm_vcpu_arch, guest_context.sp); in asm_offsets() 138 OFFSET(KVM_ARCH_GUEST_GP, kvm_vcpu_arch, guest_context.gp); in asm_offsets() 139 OFFSET(KVM_ARCH_GUEST_TP, kvm_vcpu_arch, guest_context.tp); in asm_offsets() 140 OFFSET(KVM_ARCH_GUEST_T0, kvm_vcpu_arch, guest_context.t0); in asm_offsets() 141 OFFSET(KVM_ARCH_GUEST_T1, kvm_vcpu_arch, guest_context.t1); in asm_offsets() 142 OFFSET(KVM_ARCH_GUEST_T2, kvm_vcpu_arch, guest_context.t2); in asm_offsets() 143 OFFSET(KVM_ARCH_GUEST_S0, kvm_vcpu_arch, guest_context.s0); in asm_offsets() 144 OFFSET(KVM_ARCH_GUEST_S1, kvm_vcpu_arch, guest_context.s1); in asm_offsets() [all …]
|
/linux-6.14.4/arch/mips/kernel/ |
D | asm-offsets.c | 368 OFFSET(VCPU_FPR0, kvm_vcpu_arch, fpu.fpr[0]); in output_kvm_defines() 369 OFFSET(VCPU_FPR1, kvm_vcpu_arch, fpu.fpr[1]); in output_kvm_defines() 370 OFFSET(VCPU_FPR2, kvm_vcpu_arch, fpu.fpr[2]); in output_kvm_defines() 371 OFFSET(VCPU_FPR3, kvm_vcpu_arch, fpu.fpr[3]); in output_kvm_defines() 372 OFFSET(VCPU_FPR4, kvm_vcpu_arch, fpu.fpr[4]); in output_kvm_defines() 373 OFFSET(VCPU_FPR5, kvm_vcpu_arch, fpu.fpr[5]); in output_kvm_defines() 374 OFFSET(VCPU_FPR6, kvm_vcpu_arch, fpu.fpr[6]); in output_kvm_defines() 375 OFFSET(VCPU_FPR7, kvm_vcpu_arch, fpu.fpr[7]); in output_kvm_defines() 376 OFFSET(VCPU_FPR8, kvm_vcpu_arch, fpu.fpr[8]); in output_kvm_defines() 377 OFFSET(VCPU_FPR9, kvm_vcpu_arch, fpu.fpr[9]); in output_kvm_defines() [all …]
|
/linux-6.14.4/arch/mips/kvm/ |
D | entry.c | 190 UASM_i_SW(&p, GPR_SP, offsetof(struct kvm_vcpu_arch, host_stack), GPR_K1); in kvm_mips_build_vcpu_run() 193 UASM_i_SW(&p, GPR_GP, offsetof(struct kvm_vcpu_arch, host_gp), GPR_K1); in kvm_mips_build_vcpu_run() 204 UASM_i_LW(&p, GPR_K0, offsetof(struct kvm_vcpu_arch, guest_ebase), GPR_K1); in kvm_mips_build_vcpu_run() 246 UASM_i_LW(&p, GPR_T0, offsetof(struct kvm_vcpu_arch, pc), GPR_K1); in kvm_mips_build_enter_guest() 254 UASM_i_SW(&p, GPR_K0, offsetof(struct kvm_vcpu_arch, host_pgd), GPR_K1); in kvm_mips_build_enter_guest() 304 UASM_i_SW(&p, GPR_K0, offsetof(struct kvm_vcpu_arch, host_entryhi), in kvm_mips_build_enter_guest() 347 UASM_i_LW(&p, i, offsetof(struct kvm_vcpu_arch, gprs[i]), GPR_K1); in kvm_mips_build_enter_guest() 352 UASM_i_LW(&p, GPR_K0, offsetof(struct kvm_vcpu_arch, hi), GPR_K1); in kvm_mips_build_enter_guest() 355 UASM_i_LW(&p, GPR_K0, offsetof(struct kvm_vcpu_arch, lo), GPR_K1); in kvm_mips_build_enter_guest() 360 UASM_i_LW(&p, GPR_K0, offsetof(struct kvm_vcpu_arch, gprs[GPR_K0]), GPR_K1); in kvm_mips_build_enter_guest() [all …]
|
D | vz.c | 1081 struct kvm_vcpu_arch *arch = &vcpu->arch; in kvm_vz_gpsi_cache() 1216 struct kvm_vcpu_arch *arch = &vcpu->arch; in kvm_trap_vz_handle_gpsi() 1298 struct kvm_vcpu_arch *arch = &vcpu->arch; in kvm_trap_vz_handle_gsfc()
|
D | emulate.c | 45 struct kvm_vcpu_arch *arch = &vcpu->arch; in kvm_compute_return_epc()
|
/linux-6.14.4/arch/loongarch/include/asm/ |
D | kvm_host.h | 168 struct kvm_vcpu_arch { struct 262 static inline bool kvm_guest_has_fpu(struct kvm_vcpu_arch *arch) in kvm_guest_has_fpu() argument 267 static inline bool kvm_guest_has_lsx(struct kvm_vcpu_arch *arch) in kvm_guest_has_lsx() 272 static inline bool kvm_guest_has_lasx(struct kvm_vcpu_arch *arch) in kvm_guest_has_lasx() 277 static inline bool kvm_guest_has_lbt(struct kvm_vcpu_arch *arch) in kvm_guest_has_lbt() 282 static inline bool kvm_guest_has_pmu(struct kvm_vcpu_arch *arch) in kvm_guest_has_pmu() 287 static inline int kvm_get_pmu_num(struct kvm_vcpu_arch *arch) in kvm_get_pmu_num() 304 static inline void update_pc(struct kvm_vcpu_arch *arch) in update_pc() 316 static inline bool kvm_is_ifetch_fault(struct kvm_vcpu_arch *arch) in kvm_is_ifetch_fault()
|
/linux-6.14.4/arch/loongarch/kernel/ |
D | asm-offsets.c | 287 OFFSET(VCPU_FCC, kvm_vcpu_arch, fpu.fcc); in output_kvm_defines() 288 OFFSET(VCPU_FCSR0, kvm_vcpu_arch, fpu.fcsr); in output_kvm_defines() 296 OFFSET(KVM_ARCH_HSP, kvm_vcpu_arch, host_sp); in output_kvm_defines() 297 OFFSET(KVM_ARCH_HTP, kvm_vcpu_arch, host_tp); in output_kvm_defines() 298 OFFSET(KVM_ARCH_HPGD, kvm_vcpu_arch, host_pgd); in output_kvm_defines() 299 OFFSET(KVM_ARCH_HANDLE_EXIT, kvm_vcpu_arch, handle_exit); in output_kvm_defines() 300 OFFSET(KVM_ARCH_HEENTRY, kvm_vcpu_arch, host_eentry); in output_kvm_defines() 301 OFFSET(KVM_ARCH_GEENTRY, kvm_vcpu_arch, guest_eentry); in output_kvm_defines() 302 OFFSET(KVM_ARCH_GPC, kvm_vcpu_arch, pc); in output_kvm_defines() 303 OFFSET(KVM_ARCH_GGPR, kvm_vcpu_arch, gprs); in output_kvm_defines() [all …]
|
/linux-6.14.4/arch/mips/include/asm/ |
D | kvm_host.h | 290 struct kvm_vcpu_arch { struct 692 static inline bool kvm_mips_guest_can_have_fpu(struct kvm_vcpu_arch *vcpu) in kvm_mips_guest_can_have_fpu() 698 static inline bool kvm_mips_guest_has_fpu(struct kvm_vcpu_arch *vcpu) in kvm_mips_guest_has_fpu() 704 static inline bool kvm_mips_guest_can_have_msa(struct kvm_vcpu_arch *vcpu) in kvm_mips_guest_can_have_msa() 710 static inline bool kvm_mips_guest_has_msa(struct kvm_vcpu_arch *vcpu) in kvm_mips_guest_has_msa() 776 void __kvm_save_fpu(struct kvm_vcpu_arch *vcpu); 777 void __kvm_restore_fpu(struct kvm_vcpu_arch *vcpu); 778 void __kvm_restore_fcsr(struct kvm_vcpu_arch *vcpu); 779 void __kvm_save_msa(struct kvm_vcpu_arch *vcpu); 780 void __kvm_restore_msa(struct kvm_vcpu_arch *vcpu); [all …]
|
/linux-6.14.4/arch/riscv/include/asm/ |
D | kvm_nacl.h | 15 struct kvm_vcpu_arch; 48 void __kvm_riscv_nacl_switch_to(struct kvm_vcpu_arch *vcpu_arch,
|
D | kvm_host.h | 196 struct kvm_vcpu_arch { struct 390 void __kvm_riscv_switch_to(struct kvm_vcpu_arch *vcpu_arch);
|
/linux-6.14.4/arch/riscv/kvm/ |
D | tlb.c | 214 struct kvm_vcpu_arch *varch = &vcpu->arch; in vcpu_hfence_dequeue() 239 struct kvm_vcpu_arch *varch = &vcpu->arch; in vcpu_hfence_enqueue()
|
D | vcpu.c | 372 struct kvm_vcpu_arch *v = &vcpu->arch; in kvm_riscv_vcpu_sync_interrupts()
|
/linux-6.14.4/arch/arm64/include/asm/ |
D | kvm_host.h | 696 struct kvm_vcpu_arch { struct 1265 static inline void kvm_arm_pvtime_vcpu_init(struct kvm_vcpu_arch *vcpu_arch) in kvm_arm_pvtime_vcpu_init() 1270 static inline bool kvm_arm_is_pvtime_enabled(struct kvm_vcpu_arch *vcpu_arch) in kvm_arm_is_pvtime_enabled()
|
/linux-6.14.4/arch/powerpc/include/asm/ |
D | kvm_host.h | 528 struct kvm_vcpu_arch { struct
|
/linux-6.14.4/arch/s390/include/asm/ |
D | kvm_host.h | 751 struct kvm_vcpu_arch { struct
|
/linux-6.14.4/arch/x86/include/asm/ |
D | kvm_host.h | 762 struct kvm_vcpu_arch { struct
|
/linux-6.14.4/arch/powerpc/kvm/ |
D | book3s_hv.c | 4196 struct kvm_vcpu_arch *arch; in kvmhv_get_l1_to_l2_cs_time_vcpu() 4211 struct kvm_vcpu_arch *arch; in kvmhv_get_l2_to_l1_cs_time_vcpu() 4226 struct kvm_vcpu_arch *arch; in kvmhv_get_l2_runtime_agg_vcpu()
|
D | book3s_pr.c | 516 struct kvm_vcpu_arch *a = &vcpu->arch; in kvmppc_set_msr_pr()
|
/linux-6.14.4/include/linux/ |
D | kvm_host.h | 385 struct kvm_vcpu_arch arch;
|
/linux-6.14.4/arch/x86/kvm/ |
D | x86.c | 3124 struct kvm_vcpu_arch *vcpu = &v->arch; in kvm_setup_guest_pvclock() 3177 struct kvm_vcpu_arch *vcpu = &v->arch; in kvm_guest_time_update()
|