Lines Matching full:disk

62 		if (h->disk)  in nvme_mpath_unfreeze()
63 blk_mq_unfreeze_queue_nomemrestore(h->disk->queue); in nvme_mpath_unfreeze()
72 if (h->disk) in nvme_mpath_wait_freeze()
73 blk_mq_freeze_queue_wait(h->disk->queue); in nvme_mpath_wait_freeze()
82 if (h->disk) in nvme_mpath_start_freeze()
83 blk_freeze_queue_start(h->disk->queue); in nvme_mpath_start_freeze()
107 bio_set_dev(bio, ns->head->disk->part0); in nvme_failover_req()
132 struct gendisk *disk = ns->head->disk; in nvme_mpath_start_request() local
139 if (!blk_queue_io_stat(disk->queue) || blk_rq_is_passthrough(rq)) in nvme_mpath_start_request()
143 nvme_req(rq)->start_time = bdev_start_io_acct(disk->part0, req_op(rq), in nvme_mpath_start_request()
157 bdev_end_io_acct(ns->head->disk->part0, req_op(rq), in nvme_mpath_end_request()
170 if (!ns->head->disk) in nvme_kick_requeue_lists()
174 disk_uevent(ns->head->disk, KOBJ_CHANGE); in nvme_kick_requeue_lists()
224 sector_t capacity = get_capacity(head->disk); in nvme_mpath_revalidate_paths()
231 if (capacity != get_capacity(ns->disk)) in nvme_mpath_revalidate_paths()
451 struct device *dev = disk_to_dev(head->disk); in nvme_ns_head_submit_bio()
467 bio_set_dev(bio, ns->disk->part0); in nvme_ns_head_submit_bio()
469 trace_block_bio_remap(bio, disk_devt(ns->head->disk), in nvme_ns_head_submit_bio()
487 static int nvme_ns_head_open(struct gendisk *disk, blk_mode_t mode) in nvme_ns_head_open() argument
489 if (!nvme_tryget_ns_head(disk->private_data)) in nvme_ns_head_open()
494 static void nvme_ns_head_release(struct gendisk *disk) in nvme_ns_head_release() argument
496 nvme_put_ns_head(disk->private_data); in nvme_ns_head_release()
499 static int nvme_ns_head_get_unique_id(struct gendisk *disk, u8 id[16], in nvme_ns_head_get_unique_id() argument
502 struct nvme_ns_head *head = disk->private_data; in nvme_ns_head_get_unique_id()
515 static int nvme_ns_head_report_zones(struct gendisk *disk, sector_t sector, in nvme_ns_head_report_zones() argument
518 struct nvme_ns_head *head = disk->private_data; in nvme_ns_head_report_zones()
594 &head->disk->state))) in nvme_partition_scan_work()
597 mutex_lock(&head->disk->open_mutex); in nvme_partition_scan_work()
598 bdev_disk_changed(head->disk, false); in nvme_partition_scan_work()
599 mutex_unlock(&head->disk->open_mutex); in nvme_partition_scan_work()
645 head->disk = blk_alloc_disk(&lim, ctrl->numa_node); in nvme_mpath_alloc_disk()
646 if (IS_ERR(head->disk)) in nvme_mpath_alloc_disk()
647 return PTR_ERR(head->disk); in nvme_mpath_alloc_disk()
648 head->disk->fops = &nvme_ns_head_ops; in nvme_mpath_alloc_disk()
649 head->disk->private_data = head; in nvme_mpath_alloc_disk()
659 set_bit(GD_SUPPRESS_PART_SCAN, &head->disk->state); in nvme_mpath_alloc_disk()
660 sprintf(head->disk->disk_name, "nvme%dn%d", in nvme_mpath_alloc_disk()
670 if (!head->disk) in nvme_mpath_set_live()
679 rc = device_add_disk(&head->subsys->dev, head->disk, in nvme_mpath_set_live()
996 if (blk_queue_is_zoned(ns->queue) && ns->head->disk) in nvme_mpath_add_disk()
997 ns->head->disk->nr_zones = ns->disk->nr_zones; in nvme_mpath_add_disk()
1003 if (!head->disk) in nvme_mpath_shutdown_disk()
1013 del_gendisk(head->disk); in nvme_mpath_shutdown_disk()
1019 if (!head->disk) in nvme_mpath_remove_disk()
1025 put_disk(head->disk); in nvme_mpath_remove_disk()