/linux-6.14.4/drivers/gpu/drm/i915/ |
D | i915_query.c | 25 if (copy_from_user(query_hdr, u64_to_user_ptr(query_item->data_ptr), in copy_query_item() 68 if (copy_to_user(u64_to_user_ptr(query_item->data_ptr), in fill_topology_info() 72 if (copy_to_user(u64_to_user_ptr(query_item->data_ptr + sizeof(topo)), in fill_topology_info() 76 if (intel_sseu_copy_ssmask_to_user(u64_to_user_ptr(query_item->data_ptr + in fill_topology_info() 81 if (intel_sseu_copy_eumask_to_user(u64_to_user_ptr(query_item->data_ptr + in fill_topology_info() 132 u64_to_user_ptr(query_item->data_ptr); in query_engine_info() 200 u32 __user *p = u64_to_user_ptr(user_regs_ptr); in copy_perf_config_registers_or_number() 230 u64_to_user_ptr(query_item->data_ptr); in query_perf_config_data() 232 u64_to_user_ptr(query_item->data_ptr + in query_perf_config_data() 376 u64_to_user_ptr(query_item->data_ptr); in query_perf_config_list() [all …]
|
D | i915_user_extensions.c | 57 ext = u64_to_user_ptr(next); in i915_user_extensions()
|
/linux-6.14.4/io_uring/ |
D | fs.c | 61 oldf = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_renameat_prep() 62 newf = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_renameat_prep() 120 fname = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_unlinkat_prep() 167 fname = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_mkdirat_prep() 209 oldpath = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_symlinkat_prep() 210 newpath = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_symlinkat_prep() 253 oldf = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_linkat_prep() 254 newf = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_linkat_prep()
|
D | xattr.c | 53 name = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_getxattr_prep() 54 ix->ctx.value = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in __io_getxattr_prep() 94 path = u64_to_user_ptr(READ_ONCE(sqe->addr3)); in io_getxattr_prep() 136 name = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_setxattr_prep() 137 ix->ctx.cvalue = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in __io_setxattr_prep() 170 path = u64_to_user_ptr(READ_ONCE(sqe->addr3)); in io_setxattr_prep()
|
D | statx.c | 35 path = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_statx_prep() 36 sx->buffer = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_statx_prep()
|
D | rsrc.c | 173 u64 __user *tags = u64_to_user_ptr(up->tags); in __io_sqe_files_update() 174 __s32 __user *fds = u64_to_user_ptr(up->data); in __io_sqe_files_update() 238 u64 __user *tags = u64_to_user_ptr(up->tags); in __io_sqe_buffers_update() 255 uvec = u64_to_user_ptr(user_data); in __io_sqe_buffers_update() 361 return io_sqe_files_register(ctx, u64_to_user_ptr(rr.data), in io_register_rsrc() 362 rr.nr, u64_to_user_ptr(rr.tags)); in io_register_rsrc() 366 return io_sqe_buffers_register(ctx, u64_to_user_ptr(rr.data), in io_register_rsrc() 367 rr.nr, u64_to_user_ptr(rr.tags)); in io_register_rsrc() 393 __s32 __user *fds = u64_to_user_ptr(up->arg); in io_files_update_with_index_alloc()
|
D | epoll.c | 38 ev = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_epoll_ctl_prep()
|
/linux-6.14.4/drivers/gpu/drm/v3d/ |
D | v3d_submit.c | 178 struct drm_v3d_sem __user *handle = u64_to_user_ptr(se->in_syncs); in v3d_job_init() 341 post_deps = u64_to_user_ptr(handles); in v3d_get_multisync_post_deps() 484 offsets = u64_to_user_ptr(timestamp.offsets); in v3d_get_cpu_timestamp_query_params() 485 syncs = u64_to_user_ptr(timestamp.syncs); in v3d_get_cpu_timestamp_query_params() 550 syncs = u64_to_user_ptr(reset.syncs); in v3d_get_cpu_reset_timestamp_params() 614 offsets = u64_to_user_ptr(copy.offsets); in v3d_get_cpu_copy_query_results_params() 615 syncs = u64_to_user_ptr(copy.syncs); in v3d_get_cpu_copy_query_results_params() 690 ids_pointer = u64_to_user_ptr(ids); in v3d_copy_query_info() 751 u64_to_user_ptr(reset.syncs), in v3d_get_cpu_reset_performance_params() 752 u64_to_user_ptr(reset.kperfmon_ids), in v3d_get_cpu_reset_performance_params() [all …]
|
/linux-6.14.4/drivers/gpu/drm/xe/ |
D | xe_query.c | 130 query_ptr = u64_to_user_ptr(query->data); in query_engine_cycles() 185 u64_to_user_ptr(query->data); in query_engines() 247 u64_to_user_ptr(query->data); in query_mem_regions() 318 u64_to_user_ptr(query->data); in query_config() 359 u64_to_user_ptr(query->data); in query_gt_list() 428 void __user *query_ptr = u64_to_user_ptr(query->data); in query_hwconfig() 494 void __user *query_ptr = u64_to_user_ptr(query->data); in query_gt_topology() 553 struct drm_xe_query_uc_fw_version __user *query_ptr = u64_to_user_ptr(query->data); in query_uc_fw_version() 643 void __user *query_ptr = u64_to_user_ptr(query->data); in query_oa_units()
|
/linux-6.14.4/drivers/gpu/drm/imagination/ |
D | pvr_drv.c | 503 if (copy_to_user(u64_to_user_ptr(query.quirks), out, in pvr_dev_query_quirks_get() 586 if (copy_to_user(u64_to_user_ptr(query.enhancements), out, in pvr_dev_query_enhancements_get() 1168 return copy_struct_from_user(out, obj_size, u64_to_user_ptr(usr_ptr), usr_stride); in pvr_get_uobj() 1177 if (copy_to_user(u64_to_user_ptr(usr_ptr), in, min_t(u32, usr_stride, obj_size))) in pvr_set_uobj() 1181 clear_user(u64_to_user_ptr(usr_ptr + obj_size), usr_stride - obj_size)) { in pvr_set_uobj() 1205 if (copy_from_user(out_alloc, u64_to_user_ptr(in->array), in pvr_get_uobj_array() 1209 void __user *in_ptr = u64_to_user_ptr(in->array); in pvr_get_uobj_array() 1242 if (copy_to_user(u64_to_user_ptr(out->array), in, in pvr_set_uobj_array() 1247 void __user *out_ptr = u64_to_user_ptr(out->array); in pvr_set_uobj_array() 1259 clear_user(u64_to_user_ptr(out->array + obj_size), in pvr_set_uobj_array()
|
/linux-6.14.4/drivers/accel/amdxdna/ |
D | aie2_pci.c | 617 if (copy_from_user(&status, u64_to_user_ptr(args->buffer), sizeof(status))) { in aie2_get_aie_status() 628 ret = aie2_query_status(ndev, u64_to_user_ptr(status.buffer), in aie2_get_aie_status() 635 if (copy_to_user(u64_to_user_ptr(args->buffer), &status, sizeof(status))) { in aie2_get_aie_status() 681 if (copy_to_user(u64_to_user_ptr(args->buffer), meta, sizeof(*meta))) in aie2_get_aie_metadata() 699 if (copy_to_user(u64_to_user_ptr(args->buffer), &version, sizeof(version))) in aie2_get_aie_version() 716 if (copy_to_user(u64_to_user_ptr(args->buffer), &version, sizeof(version))) in aie2_get_firmware_version() 732 if (copy_to_user(u64_to_user_ptr(args->buffer), &mode, sizeof(mode))) in aie2_get_power_mode() 757 if (copy_to_user(u64_to_user_ptr(args->buffer), clock, sizeof(*clock))) in aie2_get_clock_metadata() 785 buf = u64_to_user_ptr(args->buffer); in aie2_get_hwctx_status() 873 if (copy_from_user(&power_state, u64_to_user_ptr(args->buffer), in aie2_set_power_mode()
|
/linux-6.14.4/drivers/gpu/drm/ |
D | drm_syncobj.c | 1248 u64_to_user_ptr(timeline_wait->points), in drm_syncobj_array_wait() 1341 u64_to_user_ptr(args->handles), in drm_syncobj_wait_ioctl() 1385 u64_to_user_ptr(args->handles), in drm_syncobj_timeline_wait_ioctl() 1526 u64_to_user_ptr(args->handles), in drm_syncobj_reset_ioctl() 1559 u64_to_user_ptr(args->handles), in drm_syncobj_signal_ioctl() 1597 u64_to_user_ptr(args->handles), in drm_syncobj_timeline_signal_ioctl() 1609 if (!u64_to_user_ptr(args->points)) { in drm_syncobj_timeline_signal_ioctl() 1611 } else if (copy_from_user(points, u64_to_user_ptr(args->points), in drm_syncobj_timeline_signal_ioctl() 1654 uint64_t __user *points = u64_to_user_ptr(args->points); in drm_syncobj_query_ioctl() 1668 u64_to_user_ptr(args->handles), in drm_syncobj_query_ioctl()
|
D | drm_property.c | 481 values_ptr = u64_to_user_ptr(out_resp->values_ptr); in drm_mode_getproperty_ioctl() 492 enum_ptr = u64_to_user_ptr(out_resp->enum_blob_ptr); in drm_mode_getproperty_ioctl() 828 if (copy_to_user(u64_to_user_ptr(out_resp->data), in drm_mode_getblob_ioctl() 857 u64_to_user_ptr(out_resp->data), in drm_mode_createblob_ioctl()
|
D | drm_mode_config.c | 111 fb_id = u64_to_user_ptr(card_res->fb_id_ptr); in drm_mode_getresources() 129 crtc_id = u64_to_user_ptr(card_res->crtc_id_ptr); in drm_mode_getresources() 141 encoder_id = u64_to_user_ptr(card_res->encoder_id_ptr); in drm_mode_getresources() 152 connector_id = u64_to_user_ptr(card_res->connector_id_ptr); in drm_mode_getresources()
|
/linux-6.14.4/drivers/gpu/drm/qxl/ |
D | qxl_ioctl.c | 165 if (!access_ok(u64_to_user_ptr(cmd->command), in qxl_process_single_command() 187 u64_to_user_ptr(cmd->command), cmd->command_size); in qxl_process_single_command() 205 struct drm_qxl_reloc __user *u = u64_to_user_ptr(cmd->relocs); in qxl_process_single_command() 281 u64_to_user_ptr(execbuffer->commands); in qxl_execbuffer_ioctl()
|
/linux-6.14.4/net/bpf/ |
D | bpf_dummy_struct_ops.c | 43 ctx_in = u64_to_user_ptr(kattr->test.ctx_in); in dummy_ops_init_args() 48 u_state = u64_to_user_ptr(args->args[0]); in dummy_ops_init_args() 63 u_state = u64_to_user_ptr(args->args[0]); in dummy_ops_copy_args()
|
/linux-6.14.4/drivers/infiniband/core/ |
D | uverbs_ioctl.c | 144 return ib_is_buffer_cleared(u64_to_user_ptr(uattr->data) + len, in uverbs_is_attr_cleared() 206 ret = copy_from_user(idr_vals, u64_to_user_ptr(uattr->data), in uverbs_process_idrs_array() 304 if (copy_from_user(p, u64_to_user_ptr(uattr->data), in uverbs_process_attr() 728 udata->inbuf = u64_to_user_ptr(in->ptr_attr.data); in uverbs_fill_udata() 735 udata->outbuf = u64_to_user_ptr(out->ptr_attr.data); in uverbs_fill_udata() 753 if (copy_to_user(u64_to_user_ptr(attr->ptr_attr.data), from, min_size)) in uverbs_copy_to() 831 if (clear_user(u64_to_user_ptr(attr->ptr_attr.data) + size, in uverbs_copy_to_struct_or_zero()
|
/linux-6.14.4/drivers/gpu/drm/virtio/ |
D | virtgpu_ioctl.c | 123 if (copy_to_user(u64_to_user_ptr(param->value), &value, sizeof(int))) in virtio_gpu_getparam_ioctl() 433 if (copy_to_user(u64_to_user_ptr(args->addr), ptr, size)) in virtio_gpu_get_caps_ioctl() 520 buf = memdup_user(u64_to_user_ptr(rc_blob->cmd), in virtio_gpu_resource_create_blob_ioctl() 596 ctx_set_params = memdup_user(u64_to_user_ptr(args->ctx_set_params), in virtio_gpu_context_init_ioctl() 662 u64_to_user_ptr(value), in virtio_gpu_context_init_ioctl()
|
D | virtgpu_submit.c | 118 u64_to_user_ptr(address), in virtio_gpu_parse_deps() 208 u64_to_user_ptr(address), in virtio_gpu_parse_post_deps() 311 if (copy_from_user(bo_handles, u64_to_user_ptr(exbuf->bo_handles), in virtio_gpu_init_submit_buflist() 416 submit->buf = vmemdup_user(u64_to_user_ptr(exbuf->command), exbuf->size); in virtio_gpu_init_submit()
|
/linux-6.14.4/kernel/bpf/ |
D | mprog.c | 420 uprog_id = u64_to_user_ptr(attr->query.prog_ids); in bpf_mprog_query() 421 uprog_flags = u64_to_user_ptr(attr->query.prog_attach_flags); in bpf_mprog_query() 422 ulink_id = u64_to_user_ptr(attr->query.link_ids); in bpf_mprog_query() 423 ulink_flags = u64_to_user_ptr(attr->query.link_attach_flags); in bpf_mprog_query()
|
D | syscall.c | 1636 void __user *ukey = u64_to_user_ptr(attr->key); in map_lookup_elem() 1637 void __user *uvalue = u64_to_user_ptr(attr->value); in map_lookup_elem() 1809 void __user *ukey = u64_to_user_ptr(attr->key); in map_get_next_key() 1810 void __user *unext_key = u64_to_user_ptr(attr->next_key); in map_get_next_key() 1867 void __user *keys = u64_to_user_ptr(attr->batch.keys); in generic_map_delete_batch() 1923 void __user *values = u64_to_user_ptr(attr->batch.values); in generic_map_update_batch() 1924 void __user *keys = u64_to_user_ptr(attr->batch.keys); in generic_map_update_batch() 1984 void __user *uobatch = u64_to_user_ptr(attr->batch.out_batch); in generic_map_lookup_batch() 1985 void __user *ubatch = u64_to_user_ptr(attr->batch.in_batch); in generic_map_lookup_batch() 1986 void __user *values = u64_to_user_ptr(attr->batch.values); in generic_map_lookup_batch() [all …]
|
/linux-6.14.4/arch/x86/kvm/svm/ |
D | sev.c | 514 if (copy_from_user(&data, u64_to_user_ptr(argp->data), sizeof(data))) in sev_guest_init2() 563 if (copy_from_user(¶ms, u64_to_user_ptr(argp->data), sizeof(params))) in sev_launch_start() 607 if (copy_to_user(u64_to_user_ptr(argp->data), ¶ms, sizeof(params))) { in sev_launch_start() 746 if (copy_from_user(¶ms, u64_to_user_ptr(argp->data), sizeof(params))) in sev_launch_update_data() 974 void __user *measure = u64_to_user_ptr(argp->data); in sev_launch_measure() 994 p = u64_to_user_ptr(params.uaddr); in sev_launch_measure() 1067 if (copy_to_user(u64_to_user_ptr(argp->data), ¶ms, sizeof(params))) in sev_guest_status() 1232 if (copy_from_user(&debug, u64_to_user_ptr(argp->data), sizeof(debug))) in sev_dbg_crypt() 1316 if (copy_from_user(¶ms, u64_to_user_ptr(argp->data), sizeof(params))) in sev_launch_secret() 1380 void __user *report = u64_to_user_ptr(argp->data); in sev_get_attestation_report() [all …]
|
/linux-6.14.4/drivers/misc/ |
D | nsm.c | 159 if (copy_from_user(req->data, u64_to_user_ptr(raw->request.addr), in fill_req_raw() 176 if (copy_to_user(u64_to_user_ptr(raw->response.addr), in parse_resp_raw() 356 void __user *argp = u64_to_user_ptr((u64)arg); in nsm_dev_ioctl()
|
/linux-6.14.4/drivers/gpu/drm/i915/gem/ |
D | i915_gem_phys.c | 143 char __user *user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_object_pwrite_phys() 174 char __user *user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_object_pread_phys()
|
/linux-6.14.4/drivers/gpu/drm/etnaviv/ |
D | etnaviv_gem_submit.c | 479 ret = copy_from_user(bos, u64_to_user_ptr(args->bos), in etnaviv_ioctl_gem_submit() 486 ret = copy_from_user(relocs, u64_to_user_ptr(args->relocs), in etnaviv_ioctl_gem_submit() 493 ret = copy_from_user(pmrs, u64_to_user_ptr(args->pmrs), in etnaviv_ioctl_gem_submit() 500 ret = copy_from_user(stream, u64_to_user_ptr(args->stream), in etnaviv_ioctl_gem_submit()
|