Home
last modified time | relevance | path

Searched refs:blk_rq_bytes (Results 1 – 25 of 32) sorted by relevance

12

/linux-6.14.4/block/
Dblk-crypto-internal.h39 return bio_crypt_ctx_mergeable(req->crypt_ctx, blk_rq_bytes(req), in bio_crypt_ctx_back_mergeable()
53 return bio_crypt_ctx_mergeable(req->crypt_ctx, blk_rq_bytes(req), in bio_crypt_ctx_merge_rq()
Dblk-mq.c821 rq->bio, rq->biotail, blk_rq_bytes(rq)); in blk_dump_rq_flags()
857 int total_bytes = blk_rq_bytes(req); in blk_complete_request()
942 if (blk_crypto_rq_has_keyslot(req) && nr_bytes >= blk_rq_bytes(req)) in blk_update_request()
1023 if (blk_rq_bytes(req) < blk_rq_cur_bytes(req)) { in blk_update_request()
1083 if (blk_rq_bytes(req) & (bdev_logical_block_size(bio->bi_bdev) - 1)) in blk_rq_passthrough_stats()
1145 if (blk_update_request(rq, error, blk_rq_bytes(rq))) in blk_mq_end_request()
1383 blk_rq_bytes(last) >= BLK_PLUG_FLUSH_SIZE)) { in blk_add_rq_to_plug()
3222 if (q->disk && should_fail_request(q->disk->part0, blk_rq_bytes(rq))) in blk_insert_cloned_request()
3311 rq->__data_len = blk_rq_bytes(rq_src); in blk_rq_prep_clone()
Dbsg-lib.c223 buf->payload_len = blk_rq_bytes(req); in bsg_map_buffer()
Dblk-merge.c885 req->__data_len += blk_rq_bytes(next); in attempt_merge()
Dbfq-cgroup.c350 blkg_rwstat_add(&bfqg->stats.bytes, rq->cmd_flags, blk_rq_bytes(rq)); in bfqg_stats_update_legacy_io()
/linux-6.14.4/drivers/scsi/
Dscsi_lib.c739 return blk_rq_bytes(rq); in scsi_rq_err_bytes()
755 BUG_ON(blk_rq_bytes(rq) && !bytes); in scsi_rq_err_bytes()
990 } else if (blk_rq_bytes(req) == 0 && sense_current) { in scsi_io_completion_nz_result()
1078 if (likely(blk_rq_bytes(req) > 0 || blk_stat == BLK_STS_OK)) { in scsi_io_completion()
1085 if (scsi_end_request(req, blk_stat, blk_rq_bytes(req))) in scsi_io_completion()
1154 if (blk_rq_bytes(rq) & rq->q->limits.dma_pad_mask) { in scsi_alloc_sgtables()
1156 (rq->q->limits.dma_pad_mask & ~blk_rq_bytes(rq)) + 1; in scsi_alloc_sgtables()
1293 BUG_ON(blk_rq_bytes(req)); in scsi_setup_scsi_cmnd()
1298 cmd->transfersize = blk_rq_bytes(req); in scsi_setup_scsi_cmnd()
1673 if (blk_rq_bytes(req)) in scsi_prepare_cmd()
Dsd.c2294 good_bytes = blk_rq_bytes(req); in sd_done()
2298 scsi_set_resid(SCpnt, blk_rq_bytes(req)); in sd_done()
/linux-6.14.4/include/linux/
Dblk-mq.h1075 static inline unsigned int blk_rq_bytes(const struct request *rq) in blk_rq_bytes() function
1091 return blk_rq_bytes(rq) >> SECTOR_SHIFT; in blk_rq_sectors()
1114 return blk_rq_bytes(rq); in blk_rq_payload_bytes()
/linux-6.14.4/kernel/trace/
Dblktrace.c838 blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_INSERT, in blk_add_trace_rq_insert()
844 blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_ISSUE, in blk_add_trace_rq_issue()
850 blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_BACKMERGE, in blk_add_trace_rq_merge()
856 blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_REQUEUE, in blk_add_trace_rq_requeue()
1044 __blk_add_trace(bt, blk_rq_pos(rq), blk_rq_bytes(rq), in blk_add_trace_rq_remap()
1071 __blk_add_trace(bt, blk_rq_trace_sector(rq), blk_rq_bytes(rq), 0, in blk_add_driver_data()
/linux-6.14.4/include/scsi/
Dscsi_cmnd.h239 return blk_rq_bytes(scsi_cmd_to_rq(scmd)) >> shift; in scsi_logical_block_count()
/linux-6.14.4/include/trace/events/
Dblock.h211 __entry->bytes = blk_rq_bytes(rq);
/linux-6.14.4/drivers/mtd/ubi/
Dblock.c185 int to_read = blk_rq_bytes(req); in ubiblock_read()
/linux-6.14.4/drivers/block/
Dloop.c275 ret = file->f_op->fallocate(file, mode, pos, blk_rq_bytes(rq)); in lo_fallocate()
303 if (cmd->ret < 0 || cmd->ret == blk_rq_bytes(rq) || in lo_complete_rq()
400 iov_iter_bvec(&iter, rw, bvec, nr_bvec, blk_rq_bytes(rq)); in lo_rw_aio()
Dublk_drv.c912 const unsigned int rq_bytes = blk_rq_bytes(req); in ublk_map_io()
936 const unsigned int rq_bytes = blk_rq_bytes(req); in ublk_unmap_io()
1222 if (unlikely(mapped_bytes != blk_rq_bytes(req))) { in __ublk_rq_task_work()
1937 if (offset > blk_rq_bytes(req)) in __ublk_check_and_get_req()
Dnbd.c523 blk_rq_bytes(req), (req->timeout / HZ) * cmd->retries); in nbd_xmit_timeout()
702 request.len = htonl(blk_rq_bytes(req)); in nbd_send_cmd()
711 (unsigned long long)blk_rq_pos(req) << 9, blk_rq_bytes(req)); in nbd_send_cmd()
Dataflop.c464 blk_rq_bytes(fd_request))); in fd_end_request_cur()
/linux-6.14.4/drivers/s390/block/
Dscm_blk.c190 aidaw = scm_aidaw_fetch(scmrq, blk_rq_bytes(req)); in scm_request_prepare()
/linux-6.14.4/drivers/memstick/core/
Dmspro_block.c651 count = blk_rq_bytes(msb->block_req); in mspro_block_issue_req()
695 t_len = blk_rq_bytes(msb->block_req); in mspro_block_complete_req()
Dms_block.c1916 blk_rq_bytes(req), &len); in msb_io_work()
1919 blk_rq_bytes(req), &len); in msb_io_work()
/linux-6.14.4/drivers/nvme/host/
Dcore.c345 blk_rq_bytes(req) >> ns->head->lba_shift, in nvme_log_error()
929 cpu_to_le16((blk_rq_bytes(req) >> ns->head->lba_shift) - 1); in nvme_setup_write_zeroes()
960 if (blk_rq_bytes(req) > queue_atomic_write_unit_max_bytes(q)) in nvme_valid_atomic_write()
966 u64 end = start + blk_rq_bytes(req) - 1; in nvme_valid_atomic_write()
969 if (blk_rq_bytes(req) > boundary_bytes) in nvme_valid_atomic_write()
1006 cpu_to_le16((blk_rq_bytes(req) >> ns->head->lba_shift) - 1); in nvme_setup_rw()
Dmultipath.c158 blk_rq_bytes(rq) >> SECTOR_SHIFT, in nvme_mpath_end_request()
/linux-6.14.4/arch/um/drivers/
Dubd_kern.c1208 io_req->io_desc[0].length = blk_rq_bytes(req); in ubd_map_req()
/linux-6.14.4/drivers/block/rnbd/
Drnbd-clt.c1004 msg.bi_size = cpu_to_le32(blk_rq_bytes(rq)); in rnbd_client_xfer_request()
/linux-6.14.4/drivers/md/
Ddm-mpath.c512 size_t nr_bytes = blk_rq_bytes(rq); in multipath_clone_and_map()
/linux-6.14.4/drivers/block/null_blk/
Dmain.c1287 if (atomic_long_sub_return(blk_rq_bytes(rq), &nullb->cur_bytes) < 0) { in null_handle_throttled()

12