Lines Matching +full:24 +full:m

145 		char ch = (char)((app_id >> (24 - j*8)) & 0xff);  in genwqe_read_app_id()
171 crc = i << 24; in genwqe_init_crc32()
204 i = ((crc >> 24) ^ *buff++) & 0xFF; in genwqe_crc32()
525 * @m: mapping params
545 int genwqe_user_vmap(struct genwqe_dev *cd, struct dma_mapping *m, void *uaddr, in genwqe_user_vmap() argument
553 m->size = 0; /* mark unused and not added */ in genwqe_user_vmap()
556 m->u_vaddr = uaddr; in genwqe_user_vmap()
557 m->size = size; in genwqe_user_vmap()
563 m->size = 0; /* mark unused and not added */ in genwqe_user_vmap()
566 m->nr_pages = DIV_ROUND_UP(offs + size, PAGE_SIZE); in genwqe_user_vmap()
568 m->page_list = kcalloc(m->nr_pages, in genwqe_user_vmap()
571 if (!m->page_list) { in genwqe_user_vmap()
573 m->nr_pages = 0; in genwqe_user_vmap()
574 m->u_vaddr = NULL; in genwqe_user_vmap()
575 m->size = 0; /* mark unused and not added */ in genwqe_user_vmap()
578 m->dma_list = (dma_addr_t *)(m->page_list + m->nr_pages); in genwqe_user_vmap()
582 m->nr_pages, in genwqe_user_vmap()
583 m->write ? FOLL_WRITE : 0, /* readable/writable */ in genwqe_user_vmap()
584 m->page_list); /* ptrs to pages */ in genwqe_user_vmap()
589 if (rc < m->nr_pages) { in genwqe_user_vmap()
590 unpin_user_pages_dirty_lock(m->page_list, rc, m->write); in genwqe_user_vmap()
595 rc = genwqe_map_pages(cd, m->page_list, m->nr_pages, m->dma_list); in genwqe_user_vmap()
602 unpin_user_pages_dirty_lock(m->page_list, m->nr_pages, m->write); in genwqe_user_vmap()
605 kfree(m->page_list); in genwqe_user_vmap()
606 m->page_list = NULL; in genwqe_user_vmap()
607 m->dma_list = NULL; in genwqe_user_vmap()
608 m->nr_pages = 0; in genwqe_user_vmap()
609 m->u_vaddr = NULL; in genwqe_user_vmap()
610 m->size = 0; /* mark unused and not added */ in genwqe_user_vmap()
618 * @m: mapping params
620 int genwqe_user_vunmap(struct genwqe_dev *cd, struct dma_mapping *m) in genwqe_user_vunmap() argument
624 if (!dma_mapping_used(m)) { in genwqe_user_vunmap()
626 __func__, m); in genwqe_user_vunmap()
630 if (m->dma_list) in genwqe_user_vunmap()
631 genwqe_unmap_pages(cd, m->dma_list, m->nr_pages); in genwqe_user_vunmap()
633 if (m->page_list) { in genwqe_user_vunmap()
634 unpin_user_pages_dirty_lock(m->page_list, m->nr_pages, in genwqe_user_vunmap()
635 m->write); in genwqe_user_vunmap()
636 kfree(m->page_list); in genwqe_user_vunmap()
637 m->page_list = NULL; in genwqe_user_vunmap()
638 m->dma_list = NULL; in genwqe_user_vunmap()
639 m->nr_pages = 0; in genwqe_user_vunmap()
642 m->u_vaddr = NULL; in genwqe_user_vunmap()
643 m->size = 0; /* mark as unused and not added */ in genwqe_user_vunmap()
686 * r-m-w may require a mask 0x3C to avoid hitting hard in genwqe_card_reset()
745 * @m: maximum possible entries
751 unsigned int *i, unsigned int m, u32 addr, u32 idx, in set_reg_idx() argument
754 if (WARN_ON_ONCE(*i >= m)) in set_reg_idx()
765 unsigned int *i, unsigned int m, u32 addr, u64 val) in set_reg() argument
767 return set_reg_idx(cd, r, i, m, addr, 0, val); in set_reg()
793 ufir_addr = (i << 24) | 0x008; in genwqe_read_ffdc_regs()
798 ufec_addr = (i << 24) | 0x018; in genwqe_read_ffdc_regs()
807 sfir_addr = (i << 24) | (0x100 + 8 * j); in genwqe_read_ffdc_regs()
811 sfec_addr = (i << 24) | (0x300 + 8 * j); in genwqe_read_ffdc_regs()
848 /* 38:24 */ in genwqe_ffdc_buff_size()
849 d_len = (val & 0x0000007fff000000ull) >> 24; in genwqe_ffdc_buff_size()
871 traps = (val >> 24) & 0xff; in genwqe_ffdc_buff_size()
906 d_len = (e & 0x0000007fff000000ull) >> 24; /* 38:24 */ in genwqe_ffdc_buff_read()
940 traps = (val >> 24) & 0xff; /* Number of Traps */ in genwqe_ffdc_buff_read()