Lines Matching full:arch
40 struct gfn_to_pfn_cache *gpc = &kvm->arch.xen.shinfo_cache; in kvm_xen_shared_info_init()
67 BUILD_BUG_ON(offsetof(struct compat_shared_info, arch.wc_sec_hi) != 0x924); in kvm_xen_shared_info_init()
75 if (IS_ENABLED(CONFIG_64BIT) && kvm->arch.xen.long_mode) { in kvm_xen_shared_info_init()
85 wc_sec_hi = &shinfo->arch.wc_sec_hi; in kvm_xen_shared_info_init()
110 if (atomic_read(&vcpu->arch.xen.timer_pending) > 0) { in kvm_xen_inject_timer_irqs()
115 e.port = vcpu->arch.xen.timer_virq; in kvm_xen_inject_timer_irqs()
120 vcpu->arch.xen.timer_expires = 0; in kvm_xen_inject_timer_irqs()
121 atomic_set(&vcpu->arch.xen.timer_pending, 0); in kvm_xen_inject_timer_irqs()
128 arch.xen.timer); in xen_timer_callback()
132 if (atomic_read(&vcpu->arch.xen.timer_pending)) in xen_timer_callback()
137 e.port = vcpu->arch.xen.timer_virq; in xen_timer_callback()
142 vcpu->arch.xen.timer_expires = 0; in xen_timer_callback()
146 atomic_inc(&vcpu->arch.xen.timer_pending); in xen_timer_callback()
176 if (vcpu->arch.hv_clock.version && vcpu->kvm->arch.use_master_clock && in kvm_xen_start_timer()
200 guest_now = __pvclock_read_cycles(&vcpu->arch.hv_clock, in kvm_xen_start_timer()
245 if (vcpu->arch.xen.timer_expires) in kvm_xen_start_timer()
246 hrtimer_cancel(&vcpu->arch.xen.timer); in kvm_xen_start_timer()
248 atomic_set(&vcpu->arch.xen.timer_pending, 0); in kvm_xen_start_timer()
249 vcpu->arch.xen.timer_expires = guest_abs; in kvm_xen_start_timer()
252 xen_timer_callback(&vcpu->arch.xen.timer); in kvm_xen_start_timer()
254 hrtimer_start(&vcpu->arch.xen.timer, in kvm_xen_start_timer()
261 hrtimer_cancel(&vcpu->arch.xen.timer); in kvm_xen_stop_timer()
262 vcpu->arch.xen.timer_expires = 0; in kvm_xen_stop_timer()
263 atomic_set(&vcpu->arch.xen.timer_pending, 0); in kvm_xen_stop_timer()
268 struct kvm_vcpu_xen *vx = &v->arch.xen; in kvm_xen_update_runstate_guest()
338 if (IS_ENABLED(CONFIG_64BIT) && v->kvm->arch.xen.long_mode) { in kvm_xen_update_runstate_guest()
402 if (v->kvm->arch.xen.runstate_update_flag) in kvm_xen_update_runstate_guest()
466 if (v->kvm->arch.xen.runstate_update_flag) { in kvm_xen_update_runstate_guest()
534 struct kvm_vcpu_xen *vx = &v->arch.xen; in kvm_xen_update_runstate()
568 irq.vector = v->arch.xen.upcall_vector; in kvm_xen_inject_vcpu_vector()
579 * In that case, there are bits in vcpu->arch.xen.evtchn_pending_sel which
586 unsigned long evtchn_pending_sel = READ_ONCE(v->arch.xen.evtchn_pending_sel); in kvm_xen_inject_pending_events()
587 struct gfn_to_pfn_cache *gpc = &v->arch.xen.vcpu_info_cache; in kvm_xen_inject_pending_events()
609 if (IS_ENABLED(CONFIG_64BIT) && v->kvm->arch.xen.long_mode) { in kvm_xen_inject_pending_events()
617 "+m" (v->arch.xen.evtchn_pending_sel) in kvm_xen_inject_pending_events()
629 "+m" (v->arch.xen.evtchn_pending_sel) in kvm_xen_inject_pending_events()
638 if (v->arch.xen.upcall_vector) in kvm_xen_inject_pending_events()
644 struct gfn_to_pfn_cache *gpc = &v->arch.xen.vcpu_info_cache; in __kvm_xen_has_interrupt()
701 mutex_lock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_set_attr()
702 kvm->arch.xen.long_mode = !!data->u.long_mode; in kvm_xen_hvm_set_attr()
710 r = kvm->arch.xen.shinfo_cache.active ? in kvm_xen_hvm_set_attr()
712 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_set_attr()
720 mutex_lock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_set_attr()
728 kvm_gpc_deactivate(&kvm->arch.xen.shinfo_cache); in kvm_xen_hvm_set_attr()
731 r = kvm_gpc_activate(&kvm->arch.xen.shinfo_cache, in kvm_xen_hvm_set_attr()
740 kvm_gpc_deactivate(&kvm->arch.xen.shinfo_cache); in kvm_xen_hvm_set_attr()
743 r = kvm_gpc_activate_hva(&kvm->arch.xen.shinfo_cache, in kvm_xen_hvm_set_attr()
750 if (!r && kvm->arch.xen.shinfo_cache.active) in kvm_xen_hvm_set_attr()
753 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_set_attr()
760 mutex_lock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_set_attr()
761 kvm->arch.xen.upcall_vector = data->u.vector; in kvm_xen_hvm_set_attr()
762 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_set_attr()
772 mutex_lock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_set_attr()
773 kvm->arch.xen.xen_version = data->u.xen_version; in kvm_xen_hvm_set_attr()
774 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_set_attr()
783 mutex_lock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_set_attr()
784 kvm->arch.xen.runstate_update_flag = !!data->u.runstate_update_flag; in kvm_xen_hvm_set_attr()
785 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_set_attr()
800 mutex_lock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_get_attr()
804 data->u.long_mode = kvm->arch.xen.long_mode; in kvm_xen_hvm_get_attr()
809 if (kvm_gpc_is_gpa_active(&kvm->arch.xen.shinfo_cache)) in kvm_xen_hvm_get_attr()
810 data->u.shared_info.gfn = gpa_to_gfn(kvm->arch.xen.shinfo_cache.gpa); in kvm_xen_hvm_get_attr()
817 if (kvm_gpc_is_hva_active(&kvm->arch.xen.shinfo_cache)) in kvm_xen_hvm_get_attr()
818 data->u.shared_info.hva = kvm->arch.xen.shinfo_cache.uhva; in kvm_xen_hvm_get_attr()
825 data->u.vector = kvm->arch.xen.upcall_vector; in kvm_xen_hvm_get_attr()
830 data->u.xen_version = kvm->arch.xen.xen_version; in kvm_xen_hvm_get_attr()
839 data->u.runstate_update_flag = kvm->arch.xen.runstate_update_flag; in kvm_xen_hvm_get_attr()
847 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_get_attr()
855 mutex_lock(&vcpu->kvm->arch.xen.xen_lock); in kvm_xen_vcpu_set_attr()
869 kvm_gpc_deactivate(&vcpu->arch.xen.vcpu_info_cache); in kvm_xen_vcpu_set_attr()
874 r = kvm_gpc_activate(&vcpu->arch.xen.vcpu_info_cache, in kvm_xen_vcpu_set_attr()
878 kvm_gpc_deactivate(&vcpu->arch.xen.vcpu_info_cache); in kvm_xen_vcpu_set_attr()
883 r = kvm_gpc_activate_hva(&vcpu->arch.xen.vcpu_info_cache, in kvm_xen_vcpu_set_attr()
894 kvm_gpc_deactivate(&vcpu->arch.xen.vcpu_time_info_cache); in kvm_xen_vcpu_set_attr()
899 r = kvm_gpc_activate(&vcpu->arch.xen.vcpu_time_info_cache, in kvm_xen_vcpu_set_attr()
916 kvm_gpc_deactivate(&vcpu->arch.xen.runstate_cache); in kvm_xen_vcpu_set_attr()
917 kvm_gpc_deactivate(&vcpu->arch.xen.runstate2_cache); in kvm_xen_vcpu_set_attr()
926 if (IS_ENABLED(CONFIG_64BIT) && vcpu->kvm->arch.xen.long_mode) in kvm_xen_vcpu_set_attr()
933 r = kvm_gpc_activate(&vcpu->arch.xen.runstate_cache, in kvm_xen_vcpu_set_attr()
940 kvm_gpc_deactivate(&vcpu->arch.xen.runstate2_cache); in kvm_xen_vcpu_set_attr()
944 r = kvm_gpc_activate(&vcpu->arch.xen.runstate2_cache, in kvm_xen_vcpu_set_attr()
990 vcpu->arch.xen.current_runstate = data->u.runstate.state; in kvm_xen_vcpu_set_attr()
991 vcpu->arch.xen.runstate_entry_time = in kvm_xen_vcpu_set_attr()
993 vcpu->arch.xen.runstate_times[RUNSTATE_running] = in kvm_xen_vcpu_set_attr()
995 vcpu->arch.xen.runstate_times[RUNSTATE_runnable] = in kvm_xen_vcpu_set_attr()
997 vcpu->arch.xen.runstate_times[RUNSTATE_blocked] = in kvm_xen_vcpu_set_attr()
999 vcpu->arch.xen.runstate_times[RUNSTATE_offline] = in kvm_xen_vcpu_set_attr()
1001 vcpu->arch.xen.last_steal = current->sched_info.run_delay; in kvm_xen_vcpu_set_attr()
1026 (vcpu->arch.xen.runstate_entry_time + in kvm_xen_vcpu_set_attr()
1032 vcpu->arch.xen.runstate_entry_time += in kvm_xen_vcpu_set_attr()
1034 vcpu->arch.xen.runstate_times[RUNSTATE_running] += in kvm_xen_vcpu_set_attr()
1036 vcpu->arch.xen.runstate_times[RUNSTATE_runnable] += in kvm_xen_vcpu_set_attr()
1038 vcpu->arch.xen.runstate_times[RUNSTATE_blocked] += in kvm_xen_vcpu_set_attr()
1040 vcpu->arch.xen.runstate_times[RUNSTATE_offline] += in kvm_xen_vcpu_set_attr()
1045 else if (vcpu->arch.xen.runstate_cache.active) in kvm_xen_vcpu_set_attr()
1054 vcpu->arch.xen.vcpu_id = data->u.vcpu_id; in kvm_xen_vcpu_set_attr()
1068 vcpu->arch.xen.timer_virq = data->u.timer.port; in kvm_xen_vcpu_set_attr()
1081 vcpu->arch.xen.upcall_vector = data->u.vector; in kvm_xen_vcpu_set_attr()
1091 mutex_unlock(&vcpu->kvm->arch.xen.xen_lock); in kvm_xen_vcpu_set_attr()
1099 mutex_lock(&vcpu->kvm->arch.xen.xen_lock); in kvm_xen_vcpu_get_attr()
1103 if (kvm_gpc_is_gpa_active(&vcpu->arch.xen.vcpu_info_cache)) in kvm_xen_vcpu_get_attr()
1104 data->u.gpa = vcpu->arch.xen.vcpu_info_cache.gpa; in kvm_xen_vcpu_get_attr()
1111 if (kvm_gpc_is_hva_active(&vcpu->arch.xen.vcpu_info_cache)) in kvm_xen_vcpu_get_attr()
1112 data->u.hva = vcpu->arch.xen.vcpu_info_cache.uhva; in kvm_xen_vcpu_get_attr()
1119 if (vcpu->arch.xen.vcpu_time_info_cache.active) in kvm_xen_vcpu_get_attr()
1120 data->u.gpa = vcpu->arch.xen.vcpu_time_info_cache.gpa; in kvm_xen_vcpu_get_attr()
1131 if (vcpu->arch.xen.runstate_cache.active) { in kvm_xen_vcpu_get_attr()
1132 data->u.gpa = vcpu->arch.xen.runstate_cache.gpa; in kvm_xen_vcpu_get_attr()
1142 data->u.runstate.state = vcpu->arch.xen.current_runstate; in kvm_xen_vcpu_get_attr()
1151 data->u.runstate.state = vcpu->arch.xen.current_runstate; in kvm_xen_vcpu_get_attr()
1153 vcpu->arch.xen.runstate_entry_time; in kvm_xen_vcpu_get_attr()
1155 vcpu->arch.xen.runstate_times[RUNSTATE_running]; in kvm_xen_vcpu_get_attr()
1157 vcpu->arch.xen.runstate_times[RUNSTATE_runnable]; in kvm_xen_vcpu_get_attr()
1159 vcpu->arch.xen.runstate_times[RUNSTATE_blocked]; in kvm_xen_vcpu_get_attr()
1161 vcpu->arch.xen.runstate_times[RUNSTATE_offline]; in kvm_xen_vcpu_get_attr()
1170 data->u.vcpu_id = vcpu->arch.xen.vcpu_id; in kvm_xen_vcpu_get_attr()
1184 if (vcpu->arch.xen.timer_expires) { in kvm_xen_vcpu_get_attr()
1185 hrtimer_cancel(&vcpu->arch.xen.timer); in kvm_xen_vcpu_get_attr()
1189 data->u.timer.port = vcpu->arch.xen.timer_virq; in kvm_xen_vcpu_get_attr()
1191 data->u.timer.expires_ns = vcpu->arch.xen.timer_expires; in kvm_xen_vcpu_get_attr()
1201 if (vcpu->arch.xen.timer_expires) in kvm_xen_vcpu_get_attr()
1202 hrtimer_start_expires(&vcpu->arch.xen.timer, in kvm_xen_vcpu_get_attr()
1209 data->u.vector = vcpu->arch.xen.upcall_vector; in kvm_xen_vcpu_get_attr()
1217 mutex_unlock(&vcpu->kvm->arch.xen.xen_lock); in kvm_xen_vcpu_get_attr()
1229 mutex_lock(&kvm->arch.xen.xen_lock); in kvm_xen_write_hypercall_page()
1230 if (kvm->arch.xen.long_mode != lm) { in kvm_xen_write_hypercall_page()
1231 kvm->arch.xen.long_mode = lm; in kvm_xen_write_hypercall_page()
1237 if (kvm->arch.xen.shinfo_cache.active && in kvm_xen_write_hypercall_page()
1241 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_write_hypercall_page()
1283 hva_t blob_addr = lm ? kvm->arch.xen_hvm_config.blob_addr_64 in kvm_xen_write_hypercall_page()
1284 : kvm->arch.xen_hvm_config.blob_addr_32; in kvm_xen_write_hypercall_page()
1285 u8 blob_size = lm ? kvm->arch.xen_hvm_config.blob_size_64 in kvm_xen_write_hypercall_page()
1286 : kvm->arch.xen_hvm_config.blob_size_32; in kvm_xen_write_hypercall_page()
1327 mutex_lock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_config()
1329 if (xhc->msr && !kvm->arch.xen_hvm_config.msr) in kvm_xen_hvm_config()
1331 else if (!xhc->msr && kvm->arch.xen_hvm_config.msr) in kvm_xen_hvm_config()
1334 old_flags = kvm->arch.xen_hvm_config.flags; in kvm_xen_hvm_config()
1335 memcpy(&kvm->arch.xen_hvm_config, xhc, sizeof(*xhc)); in kvm_xen_hvm_config()
1337 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_hvm_config()
1355 if (unlikely(!kvm_is_linear_rip(vcpu, vcpu->arch.xen.hypercall_rip))) in kvm_xen_hypercall_complete_userspace()
1363 if (IS_ENABLED(CONFIG_64BIT) && kvm->arch.xen.long_mode) in max_evtchn_port()
1373 struct gfn_to_pfn_cache *gpc = &kvm->arch.xen.shinfo_cache; in wait_pending_event()
1385 if (IS_ENABLED(CONFIG_64BIT) && kvm->arch.xen.long_mode) { in wait_pending_event()
1416 !(vcpu->kvm->arch.xen_hvm_config.flags & KVM_XEN_HVM_CONFIG_EVTCHN_SEND)) in kvm_xen_schedop_poll()
1476 vcpu->arch.xen.poll_evtchn = port; in kvm_xen_schedop_poll()
1478 vcpu->arch.xen.poll_evtchn = -1; in kvm_xen_schedop_poll()
1480 set_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask); in kvm_xen_schedop_poll()
1483 vcpu->arch.mp_state = KVM_MP_STATE_HALTED; in kvm_xen_schedop_poll()
1486 mod_timer(&vcpu->arch.xen.poll_timer, in kvm_xen_schedop_poll()
1492 del_timer(&vcpu->arch.xen.poll_timer); in kvm_xen_schedop_poll()
1494 vcpu->arch.mp_state = KVM_MP_STATE_RUNNABLE; in kvm_xen_schedop_poll()
1497 vcpu->arch.xen.poll_evtchn = 0; in kvm_xen_schedop_poll()
1501 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask); in kvm_xen_schedop_poll()
1510 struct kvm_vcpu *vcpu = from_timer(vcpu, t, arch.xen.poll_timer); in cancel_evtchn_poll()
1551 if (vcpu->arch.xen.vcpu_id != vcpu_id) { in kvm_xen_hcall_vcpu_op()
1583 if (vcpu->arch.xen.vcpu_id != vcpu_id) { in kvm_xen_hcall_vcpu_op()
1656 if (params[0] == XENVER_version && vcpu->kvm->arch.xen.xen_version) { in kvm_xen_hypercall()
1657 r = vcpu->kvm->arch.xen.xen_version; in kvm_xen_hypercall()
1700 vcpu->arch.xen.hypercall_rip = kvm_get_linear_rip(vcpu); in kvm_xen_hypercall()
1701 vcpu->arch.complete_userspace_io = in kvm_xen_hypercall()
1709 int poll_evtchn = vcpu->arch.xen.poll_evtchn; in kvm_xen_check_poller()
1712 test_and_clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask)) { in kvm_xen_check_poller()
1730 struct gfn_to_pfn_cache *gpc = &kvm->arch.xen.shinfo_cache; in kvm_xen_set_evtchn_fast()
1759 if (IS_ENABLED(CONFIG_64BIT) && kvm->arch.xen.long_mode) { in kvm_xen_set_evtchn_fast()
1787 gpc = &vcpu->arch.xen.vcpu_info_cache; in kvm_xen_set_evtchn_fast()
1795 if (!test_and_set_bit(port_word_bit, &vcpu->arch.xen.evtchn_pending_sel)) in kvm_xen_set_evtchn_fast()
1800 if (IS_ENABLED(CONFIG_64BIT) && kvm->arch.xen.long_mode) { in kvm_xen_set_evtchn_fast()
1816 if (kick_vcpu && vcpu->arch.xen.upcall_vector) { in kvm_xen_set_evtchn_fast()
1870 struct gfn_to_pfn_cache *gpc = &kvm->arch.xen.shinfo_cache; in kvm_xen_set_evtchn()
1996 mutex_lock(&kvm->arch.xen.xen_lock); in kvm_xen_eventfd_update()
1997 evtchnfd = idr_find(&kvm->arch.xen.evtchn_ports, port); in kvm_xen_eventfd_update()
2027 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_eventfd_update()
2090 mutex_lock(&kvm->arch.xen.xen_lock); in kvm_xen_eventfd_assign()
2091 ret = idr_alloc(&kvm->arch.xen.evtchn_ports, evtchnfd, port, port + 1, in kvm_xen_eventfd_assign()
2093 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_eventfd_assign()
2111 mutex_lock(&kvm->arch.xen.xen_lock); in kvm_xen_eventfd_deassign()
2112 evtchnfd = idr_remove(&kvm->arch.xen.evtchn_ports, port); in kvm_xen_eventfd_deassign()
2113 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_eventfd_deassign()
2131 mutex_lock(&kvm->arch.xen.xen_lock); in kvm_xen_eventfd_reset()
2138 idr_for_each_entry(&kvm->arch.xen.evtchn_ports, evtchnfd, i) in kvm_xen_eventfd_reset()
2143 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_eventfd_reset()
2148 idr_for_each_entry(&kvm->arch.xen.evtchn_ports, evtchnfd, i) { in kvm_xen_eventfd_reset()
2150 idr_remove(&kvm->arch.xen.evtchn_ports, evtchnfd->send_port); in kvm_xen_eventfd_reset()
2152 mutex_unlock(&kvm->arch.xen.xen_lock); in kvm_xen_eventfd_reset()
2205 evtchnfd = idr_find(&vcpu->kvm->arch.xen.evtchn_ports, send.port); in kvm_xen_hcall_evtchn_send()
2224 vcpu->arch.xen.vcpu_id = vcpu->vcpu_idx; in kvm_xen_init_vcpu()
2225 vcpu->arch.xen.poll_evtchn = 0; in kvm_xen_init_vcpu()
2227 timer_setup(&vcpu->arch.xen.poll_timer, cancel_evtchn_poll, 0); in kvm_xen_init_vcpu()
2228 hrtimer_init(&vcpu->arch.xen.timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS_HARD); in kvm_xen_init_vcpu()
2229 vcpu->arch.xen.timer.function = xen_timer_callback; in kvm_xen_init_vcpu()
2231 kvm_gpc_init(&vcpu->arch.xen.runstate_cache, vcpu->kvm); in kvm_xen_init_vcpu()
2232 kvm_gpc_init(&vcpu->arch.xen.runstate2_cache, vcpu->kvm); in kvm_xen_init_vcpu()
2233 kvm_gpc_init(&vcpu->arch.xen.vcpu_info_cache, vcpu->kvm); in kvm_xen_init_vcpu()
2234 kvm_gpc_init(&vcpu->arch.xen.vcpu_time_info_cache, vcpu->kvm); in kvm_xen_init_vcpu()
2242 kvm_gpc_deactivate(&vcpu->arch.xen.runstate_cache); in kvm_xen_destroy_vcpu()
2243 kvm_gpc_deactivate(&vcpu->arch.xen.runstate2_cache); in kvm_xen_destroy_vcpu()
2244 kvm_gpc_deactivate(&vcpu->arch.xen.vcpu_info_cache); in kvm_xen_destroy_vcpu()
2245 kvm_gpc_deactivate(&vcpu->arch.xen.vcpu_time_info_cache); in kvm_xen_destroy_vcpu()
2247 del_timer_sync(&vcpu->arch.xen.poll_timer); in kvm_xen_destroy_vcpu()
2255 if (!vcpu->arch.xen.cpuid.base) in kvm_xen_update_tsc_info()
2258 function = vcpu->arch.xen.cpuid.base | XEN_CPUID_LEAF(3); in kvm_xen_update_tsc_info()
2259 if (function > vcpu->arch.xen.cpuid.limit) in kvm_xen_update_tsc_info()
2264 entry->ecx = vcpu->arch.hv_clock.tsc_to_system_mul; in kvm_xen_update_tsc_info()
2265 entry->edx = vcpu->arch.hv_clock.tsc_shift; in kvm_xen_update_tsc_info()
2270 entry->eax = vcpu->arch.hw_tsc_khz; in kvm_xen_update_tsc_info()
2275 mutex_init(&kvm->arch.xen.xen_lock); in kvm_xen_init_vm()
2276 idr_init(&kvm->arch.xen.evtchn_ports); in kvm_xen_init_vm()
2277 kvm_gpc_init(&kvm->arch.xen.shinfo_cache, kvm); in kvm_xen_init_vm()
2285 kvm_gpc_deactivate(&kvm->arch.xen.shinfo_cache); in kvm_xen_destroy_vm()
2287 idr_for_each_entry(&kvm->arch.xen.evtchn_ports, evtchnfd, i) { in kvm_xen_destroy_vm()
2292 idr_destroy(&kvm->arch.xen.evtchn_ports); in kvm_xen_destroy_vm()
2294 if (kvm->arch.xen_hvm_config.msr) in kvm_xen_destroy_vm()