Lines Matching full:asid
55 u16 asid) in arm_smmu_make_sva_cd() argument
77 FIELD_PREP(CTXDESC_CD_0_ASID, asid)); in arm_smmu_make_sva_cd()
160 arm_smmu_tlb_inv_asid(smmu_domain->smmu, smmu_domain->cd.asid); in arm_smmu_mm_arch_invalidate_secondary_tlbs()
162 arm_smmu_tlb_inv_range_asid(start, size, smmu_domain->cd.asid, in arm_smmu_mm_arch_invalidate_secondary_tlbs()
190 smmu_domain->cd.asid); in arm_smmu_mm_release()
196 arm_smmu_tlb_inv_asid(smmu_domain->smmu, smmu_domain->cd.asid); in arm_smmu_mm_release()
364 arm_smmu_make_sva_cd(&target, master, domain->mm, smmu_domain->cd.asid); in arm_smmu_sva_set_dev_pasid()
376 * Ensure the ASID is empty in the iommu cache before allowing reuse. in arm_smmu_sva_domain_free()
378 arm_smmu_tlb_inv_asid(smmu_domain->smmu, smmu_domain->cd.asid); in arm_smmu_sva_domain_free()
382 * still be called/running at this point. We allow the ASID to be in arm_smmu_sva_domain_free()
386 xa_erase(&arm_smmu_asid_xa, smmu_domain->cd.asid); in arm_smmu_sva_domain_free()
406 u32 asid; in arm_smmu_sva_domain_alloc() local
416 ret = xa_alloc(&arm_smmu_asid_xa, &asid, smmu_domain, in arm_smmu_sva_domain_alloc()
421 smmu_domain->cd.asid = asid; in arm_smmu_sva_domain_alloc()
430 xa_erase(&arm_smmu_asid_xa, smmu_domain->cd.asid); in arm_smmu_sva_domain_alloc()