/linux-6.14.4/drivers/clk/qcom/ |
D | clk-rcg.c | 106 static u32 md_to_m(struct mn *mn, u32 md) in md_to_m() argument 108 md >>= mn->m_val_shift; in md_to_m() 109 md &= BIT(mn->width) - 1; in md_to_m() 132 static u32 mn_to_md(struct mn *mn, u32 m, u32 n, u32 md) in mn_to_md() argument 136 mask_w = BIT(mn->width) - 1; in mn_to_md() 137 mask = (mask_w << mn->m_val_shift) | mask_w; in mn_to_md() 141 m <<= mn->m_val_shift; in mn_to_md() 149 static u32 ns_m_to_n(struct mn *mn, u32 ns, u32 m) in ns_m_to_n() argument 151 ns = ~ns >> mn->n_val_shift; in ns_m_to_n() 152 ns &= BIT(mn->width) - 1; in ns_m_to_n() [all …]
|
D | clk-rcg.h | 38 * struct mn - M/N:D counter 39 * @mnctr_en_bit: bit to enable mn counter 40 * @mnctr_reset_bit: bit to assert mn counter reset 41 * @mnctr_mode_shift: lowest bit of mn counter mode field 47 struct mn { struct 85 * @mn: mn counter 96 struct mn mn; member 122 * @mn: mn counter (banked) 135 struct mn mn[2]; member
|
D | gcc-msm8660.c | 106 .mn = { 157 .mn = { 208 .mn = { 259 .mn = { 310 .mn = { 361 .mn = { 412 .mn = { 463 .mn = { 514 .mn = { 565 .mn = { [all …]
|
D | gcc-mdm9615.c | 188 .mn = { 239 .mn = { 290 .mn = { 341 .mn = { 392 .mn = { 455 .mn = { 506 .mn = { 557 .mn = { 608 .mn = { 659 .mn = { [all …]
|
D | gcc-msm8960.c | 352 .mn = { 403 .mn = { 454 .mn = { 505 .mn = { 556 .mn = { 607 .mn = { 658 .mn = { 709 .mn = { 760 .mn = { 811 .mn = { [all …]
|
D | gcc-ipq806x.c | 414 .mn = { 465 .mn = { 516 .mn = { 567 .mn = { 618 .mn = { 669 .mn = { 733 .mn = { 784 .mn = { 835 .mn = { 886 .mn = { [all …]
|
D | mmcc-msm8960.c | 191 .mn = { 242 .mn = { 293 .mn = { 351 .mn = { 419 .mn = { 487 .mn = { 742 .mn = { 848 .mn[0] = { 856 .mn[1] = { 910 .mn[0] = { [all …]
|
/linux-6.14.4/drivers/iommu/amd/ |
D | pasid.c | 59 static void sva_arch_invalidate_secondary_tlbs(struct mmu_notifier *mn, in sva_arch_invalidate_secondary_tlbs() argument 67 sva_pdom = container_of(mn, struct protection_domain, mn); in sva_arch_invalidate_secondary_tlbs() 80 static void sva_mn_release(struct mmu_notifier *mn, struct mm_struct *mm) in sva_mn_release() argument 86 sva_pdom = container_of(mn, struct protection_domain, mn); in sva_mn_release() 171 if (sva_pdom->mn.ops) in iommu_sva_domain_free() 172 mmu_notifier_unregister(&sva_pdom->mn, domain->mm); in iommu_sva_domain_free() 193 pdom->mn.ops = &sva_mn; in amd_iommu_domain_alloc_sva() 196 ret = mmu_notifier_register(&pdom->mn, mm); in amd_iommu_domain_alloc_sva()
|
/linux-6.14.4/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
D | ram.c | 35 struct nvkm_mm_node *mn; member 52 .mem = vram->mn, in nvkm_vram_map() 61 return (u64)nvkm_mm_size(nvkm_vram(memory)->mn) << NVKM_RAM_MM_SHIFT; in nvkm_vram_size() 68 if (!nvkm_mm_contiguous(vram->mn)) in nvkm_vram_addr() 70 return (u64)nvkm_mm_addr(vram->mn) << NVKM_RAM_MM_SHIFT; in nvkm_vram_addr() 89 struct nvkm_mm_node *next = vram->mn; in nvkm_vram_dtor() 101 kfree(vram->mn); in nvkm_vram_dtor() 138 vram->mn = kzalloc(sizeof(*vram->mn), GFP_KERNEL); in nvkm_ram_wrap() 139 if (!vram->mn) in nvkm_ram_wrap() 142 vram->mn->offset = addr >> NVKM_RAM_MM_SHIFT; in nvkm_ram_wrap() [all …]
|
/linux-6.14.4/drivers/infiniband/hw/hfi1/ |
D | mmu_rb.c | 59 INIT_HLIST_NODE(&h->mn.hlist); in hfi1_mmu_rb_register() 61 h->mn.ops = &mn_opts; in hfi1_mmu_rb_register() 68 ret = mmu_notifier_register(&h->mn, current->mm); in hfi1_mmu_rb_register() 86 mmgrab(handler->mn.mm); in hfi1_mmu_rb_unregister() 89 mmu_notifier_unregister(&handler->mn, handler->mn.mm); in hfi1_mmu_rb_unregister() 115 mmdrop(handler->mn.mm); in hfi1_mmu_rb_unregister() 129 if (current->mm != handler->mn.mm) in hfi1_mmu_rb_insert() 233 if (current->mm != handler->mn.mm) in hfi1_mmu_rb_evict() 261 static int mmu_notifier_range_start(struct mmu_notifier *mn, in mmu_notifier_range_start() argument 265 container_of(mn, struct mmu_rb_handler, mn); in mmu_notifier_range_start()
|
D | mmu_rb.h | 34 * they fit together in one cache line. mn is relatively rarely 39 struct mmu_notifier mn; member
|
/linux-6.14.4/drivers/net/ethernet/mellanox/mlx5/core/diag/ |
D | en_rep_tracepoint.h | 24 TP_fast_assign(const struct mlx5e_neigh *mn = &nhe->m_neigh; 34 if (mn->family == AF_INET) { 35 *p32 = mn->dst_ip.v4; 37 } else if (mn->family == AF_INET6) { 38 *pin6 = mn->dst_ip.v6;
|
D | en_tc_tracepoint.h | 85 TP_fast_assign(const struct mlx5e_neigh *mn = &nhe->m_neigh; 94 if (mn->family == AF_INET) { 95 *p32 = mn->dst_ip.v4; 97 } else if (mn->family == AF_INET6) { 98 *pin6 = mn->dst_ip.v6;
|
/linux-6.14.4/drivers/gpu/drm/nouveau/nvkm/core/ |
D | memory.c | 38 nvkm_mm_free(&fb->tags.mm, &tags->mn); in nvkm_memory_tags_put() 61 if (tags->mn && tags->mn->length != nr) { in nvkm_memory_tags_get() 77 if (!nvkm_mm_head(&fb->tags.mm, 0, 1, nr, nr, 1, &tags->mn)) { in nvkm_memory_tags_get() 79 clr(device, tags->mn->offset, tags->mn->length); in nvkm_memory_tags_get() 90 tags->mn = NULL; in nvkm_memory_tags_get()
|
/linux-6.14.4/drivers/misc/sgi-gru/ |
D | grutlbpurge.c | 207 static int gru_invalidate_range_start(struct mmu_notifier *mn, in gru_invalidate_range_start() argument 210 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, in gru_invalidate_range_start() 222 static void gru_invalidate_range_end(struct mmu_notifier *mn, in gru_invalidate_range_end() argument 225 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, in gru_invalidate_range_end() 250 static void gru_free_notifier(struct mmu_notifier *mn) in gru_free_notifier() argument 252 kfree(container_of(mn, struct gru_mm_struct, ms_notifier)); in gru_free_notifier() 265 struct mmu_notifier *mn; in gru_register_mmu_notifier() local 267 mn = mmu_notifier_get_locked(&gru_mmuops, current->mm); in gru_register_mmu_notifier() 268 if (IS_ERR(mn)) in gru_register_mmu_notifier() 269 return ERR_CAST(mn); in gru_register_mmu_notifier() [all …]
|
/linux-6.14.4/drivers/iommu/intel/ |
D | svm.c | 51 static void intel_arch_invalidate_secondary_tlbs(struct mmu_notifier *mn, in intel_arch_invalidate_secondary_tlbs() argument 55 struct dmar_domain *domain = container_of(mn, struct dmar_domain, notifier); in intel_arch_invalidate_secondary_tlbs() 70 static void intel_mm_release(struct mmu_notifier *mn, struct mm_struct *mm) in intel_mm_release() argument 72 struct dmar_domain *domain = container_of(mn, struct dmar_domain, notifier); in intel_mm_release() 99 static void intel_mm_free_notifier(struct mmu_notifier *mn) in intel_mm_free_notifier() argument 101 struct dmar_domain *domain = container_of(mn, struct dmar_domain, notifier); in intel_mm_free_notifier()
|
/linux-6.14.4/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
D | gk20a.c | 53 struct nvkm_mm_node *mn; member 130 return (u64)gk20a_instobj(memory)->mn->offset << 12; in gk20a_instobj_addr() 136 return (u64)gk20a_instobj(memory)->mn->length << 12; in gk20a_instobj_size() 288 .mem = node->mn, in gk20a_instobj_map() 304 dma_free_attrs(dev, (u64)node->base.mn->length << PAGE_SHIFT, in gk20a_instobj_dtor_dma() 317 struct nvkm_mm_node *r = node->base.mn; in gk20a_instobj_dtor_iommu() 335 for (i = 0; i < node->base.mn->length; i++) { in gk20a_instobj_dtor_iommu() 416 node->base.mn = &node->r; in gk20a_instobj_ctor_dma() 494 node->base.mn = r; in gk20a_instobj_ctor_iommu() 543 size, align, (u64)node->mn->offset << 12); in gk20a_instobj_new()
|
/linux-6.14.4/drivers/media/tuners/ |
D | xc2028-types.h | 58 /* LCD firmwares exist only for MTS STD/MN (PAL or NTSC/M) 59 and for non-MTS STD/MN (PAL, NTSC/M or NTSC/Kr) 65 /* NOGD firmwares exist only for MTS STD/MN (PAL or NTSC/M) 66 and for non-MTS STD/MN (PAL, NTSC/M or NTSC/Kr)
|
/linux-6.14.4/mm/ |
D | page_ext.c | 433 struct memory_notify *mn = arg; in page_ext_callback() local 438 ret = online_page_ext(mn->start_pfn, in page_ext_callback() 439 mn->nr_pages, mn->status_change_nid); in page_ext_callback() 442 offline_page_ext(mn->start_pfn, in page_ext_callback() 443 mn->nr_pages); in page_ext_callback() 446 offline_page_ext(mn->start_pfn, in page_ext_callback() 447 mn->nr_pages); in page_ext_callback()
|
/linux-6.14.4/drivers/scsi/qla2xxx/ |
D | qla_bsg.c | 1019 struct verify_chip_entry_84xx *mn = NULL; in qla84xx_updatefw() local 1065 mn = dma_pool_zalloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma); in qla84xx_updatefw() 1066 if (!mn) { in qla84xx_updatefw() 1076 mn->entry_type = VERIFY_CHIP_IOCB_TYPE; in qla84xx_updatefw() 1077 mn->entry_count = 1; in qla84xx_updatefw() 1083 mn->options = cpu_to_le16(options); in qla84xx_updatefw() 1084 mn->fw_ver = cpu_to_le32(fw_ver); in qla84xx_updatefw() 1085 mn->fw_size = cpu_to_le32(data_len); in qla84xx_updatefw() 1086 mn->fw_seq_size = cpu_to_le32(data_len); in qla84xx_updatefw() 1087 put_unaligned_le64(fw_dma, &mn->dsd.address); in qla84xx_updatefw() [all …]
|
/linux-6.14.4/drivers/iommu/arm/arm-smmu-v3/ |
D | arm-smmu-v3-sva.c | 136 static void arm_smmu_mm_arch_invalidate_secondary_tlbs(struct mmu_notifier *mn, in arm_smmu_mm_arch_invalidate_secondary_tlbs() argument 142 container_of(mn, struct arm_smmu_domain, mmu_notifier); in arm_smmu_mm_arch_invalidate_secondary_tlbs() 168 static void arm_smmu_mm_release(struct mmu_notifier *mn, struct mm_struct *mm) in arm_smmu_mm_release() argument 171 container_of(mn, struct arm_smmu_domain, mmu_notifier); in arm_smmu_mm_release() 200 static void arm_smmu_mmu_notifier_free(struct mmu_notifier *mn) in arm_smmu_mmu_notifier_free() argument 202 kfree(container_of(mn, struct arm_smmu_domain, mmu_notifier)); in arm_smmu_mmu_notifier_free()
|
/linux-6.14.4/arch/powerpc/kexec/ |
D | crash.c | 423 * @mn: struct memory_notify data handler 425 static void update_crash_elfcorehdr(struct kimage *image, struct memory_notify *mn) in update_crash_elfcorehdr() argument 448 base_addr = PFN_PHYS(mn->start_pfn); in update_crash_elfcorehdr() 449 size = mn->nr_pages * PAGE_SIZE; in update_crash_elfcorehdr() 570 struct memory_notify *mn; in arch_crash_handle_hotplug_event() local 582 mn = (struct memory_notify *)arg; in arch_crash_handle_hotplug_event() 583 update_crash_elfcorehdr(image, mn); in arch_crash_handle_hotplug_event()
|
/linux-6.14.4/drivers/gpu/drm/radeon/ |
D | radeon_mn.c | 42 * @mn: our notifier 49 static bool radeon_mn_invalidate(struct mmu_interval_notifier *mn, in radeon_mn_invalidate() argument 53 struct radeon_bo *bo = container_of(mn, struct radeon_bo, notifier); in radeon_mn_invalidate()
|
/linux-6.14.4/arch/x86/kernel/apic/ |
D | x2apic_uv_x.c | 1160 struct mn { struct 1167 /* Initialize caller's MN struct and fill in values */ 1168 static void get_mn(struct mn *mnp) in get_mn() argument 1190 struct mn mn; in uv_init_hub_info() local 1192 get_mn(&mn); in uv_init_hub_info() 1193 hi->gpa_mask = mn.m_val ? in uv_init_hub_info() 1194 (1UL << (mn.m_val + mn.n_val)) - 1 : in uv_init_hub_info() 1197 hi->m_val = mn.m_val; in uv_init_hub_info() 1198 hi->n_val = mn.n_val; in uv_init_hub_info() 1199 hi->m_shift = mn.m_shift; in uv_init_hub_info() [all …]
|
/linux-6.14.4/Documentation/devicetree/bindings/arm/tegra/ |
D | nvidia,tegra234-cbb.yaml | 21 Root Master Node (MN) before it connects to any other element of the 22 fabric. Each Root MN contains a Error Monitor (EM) which detects and
|