Lines Matching +defs:access +defs:range
292 unsigned int access) in mark_mmio_spte()
676 gfn_t gfn, unsigned int access) in kvm_mmu_page_set_translation()
695 unsigned int access) in kvm_mmu_page_set_access()
1490 bool kvm_unmap_gfn_range(struct kvm *kvm, struct kvm_gfn_range *range) in kvm_unmap_gfn_range()
1525 u64 *spte, gfn_t gfn, unsigned int access) in __rmap_add()
1547 u64 *spte, gfn_t gfn, unsigned int access) in rmap_add()
1555 struct kvm_gfn_range *range, bool test_only) in kvm_rmap_age_gfn_range()
1591 bool kvm_age_gfn(struct kvm *kvm, struct kvm_gfn_range *range) in kvm_age_gfn()
1604 bool kvm_test_age_gfn(struct kvm *kvm, struct kvm_gfn_range *range) in kvm_test_age_gfn()
2214 unsigned int access) in kvm_mmu_child_role()
2261 bool direct, unsigned int access) in kvm_mmu_get_child_sp()
2898 int nr_pages, unsigned int access) in kvm_mmu_prefetch_sptes()
2939 unsigned int access = sp->role.access; in direct_pte_prefetch_many() local
3252 unsigned int access) in kvm_handle_noslot_fault()
4051 gpa_t vaddr, u64 access, in nonpaging_gva_to_gpa()
4171 unsigned int access = get_mmio_spte_access(spte); in handle_mmio_page_fault() local
4391 struct kvm_page_fault *fault, unsigned int access) in kvm_mmu_faultin_pfn()
4729 struct kvm_pre_fault_memory *range) in kvm_arch_vcpu_pre_fault_memory()
4918 unsigned int access) in sync_mmio_spte()
5187 #define BYTE_MASK(access) \ argument
6680 unsigned int access; in shadow_mmu_get_sp_for_split() local
7063 struct kvm_gfn_range range = { in kvm_mmu_zap_memslot() local
7500 struct kvm_gfn_range *range) in kvm_arch_pre_set_memory_attributes()
7561 struct kvm_gfn_range *range) in kvm_arch_post_set_memory_attributes()