Searched refs:mq_hctx (Results 1 – 14 of 14) sorted by relevance
| /linux/block/ |
| A D | blk-mq-sched.c | 53 return rqa->mq_hctx > rqb->mq_hctx; in sched_rq_cmp() 59 list_first_entry(rq_list, struct request, queuelist)->mq_hctx; in blk_mq_dispatch_hctx_list() 65 if (rq->mq_hctx != hctx) { in blk_mq_dispatch_hctx_list() 141 if (rq->mq_hctx != hctx) in __blk_mq_do_dispatch_sched() 264 } while (blk_mq_dispatch_rq_list(rq->mq_hctx, &rq_list, 1)); in blk_mq_do_dispatch_ctx()
|
| A D | blk-mq.c | 357 rq->mq_hctx = hctx; in blk_mq_rq_ctx_init() 714 rq->mq_hctx = NULL; in __blk_mq_free_request() 1110 cur_hctx = rq->mq_hctx; in blk_mq_end_request_batch() 1201 if ((rq->mq_hctx->nr_ctx == 1 && in blk_mq_complete_request_remote() 1355 if (!rq->mq_hctx) in blk_rq_is_poll() 1749 blk_mq_tag_busy(rq->mq_hctx); in __blk_mq_alloc_driver_tag() 2671 if (hctx != rq->mq_hctx) { in blk_mq_plug_issue_direct() 2676 hctx = rq->mq_hctx; in blk_mq_plug_issue_direct() 2722 this_hctx = rq->mq_hctx; in blk_mq_dispatch_plug_list() 2890 if (type != rq->mq_hctx->type && in blk_mq_peek_cached_request() [all …]
|
| A D | blk-flush.c | 310 flush_rq->mq_hctx = first_rq->mq_hctx; in blk_kick_flush() 341 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in mq_flush_data_end_io()
|
| A D | blk-mq-tag.c | 292 if (rq->q == q && (!hctx || rq->mq_hctx == hctx)) in bt_iter() 681 return (rq->mq_hctx->queue_num << BLK_MQ_UNIQUE_TAG_BITS) | in blk_mq_unique_tag()
|
| A D | blk-mq.h | 349 __blk_mq_put_driver_tag(rq->mq_hctx, rq); in blk_mq_put_driver_tag()
|
| A D | blk-mq-debugfs.c | 359 if (rq->mq_hctx == params->hctx) in hctx_show_busy_rq()
|
| /linux/drivers/block/ |
| A D | virtio_blk.c | 339 struct virtio_blk *vblk = req->mq_hctx->queue->queuedata; in virtblk_request_done() 468 struct virtio_blk *vblk = req->mq_hctx->queue->queuedata; in virtblk_prep_rq_batch() 471 return virtblk_prep_rq(req->mq_hctx, vblk, req, vbr) == BLK_STS_OK; in virtblk_prep_rq_batch() 507 struct virtio_blk_vq *vq = get_virtio_blk_vq(req->mq_hctx); in virtio_queue_rqs() 517 if (!next || req->mq_hctx != next->mq_hctx) { in virtio_queue_rqs()
|
| A D | ublk_drv.c | 993 struct ublk_queue *ubq = req->mq_hctx->driver_data; in __ublk_complete_rq() 1103 struct ublk_queue *ubq = req->mq_hctx->driver_data; in __ublk_rq_task_work() 1211 struct ublk_queue *ubq = rq->mq_hctx->driver_data; in ublk_timeout() 1933 if (!req->mq_hctx || !req->mq_hctx->driver_data) in ublk_check_and_get_req() 1958 ubq = req->mq_hctx->driver_data; in ublk_ch_read_iter() 1976 ubq = req->mq_hctx->driver_data; in ublk_ch_write_iter()
|
| /linux/drivers/nvme/host/ |
| A D | pci.c | 510 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_pci_use_sgls() 777 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_map_data() 938 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_queue_rqs() 949 if (!next || req->mq_hctx != next->mq_hctx) { in nvme_queue_rqs() 964 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_pci_unmap_rq() 1256 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in abort_endio() 1319 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_timeout() 1348 nvme_poll(req->mq_hctx, NULL); in nvme_timeout()
|
| A D | nvme.h | 217 return req->mq_hctx->queue_num + 1; in nvme_req_qid()
|
| A D | core.c | 736 if (req->mq_hctx->type == HCTX_TYPE_POLL) in nvme_init_request()
|
| /linux/drivers/ufs/core/ |
| A D | ufs-mcq.c | 114 struct blk_mq_hw_ctx *hctx = READ_ONCE(req->mq_hctx); in ufshcd_mcq_req_to_hwq()
|
| /linux/include/linux/ |
| A D | blk-mq.h | 105 struct blk_mq_hw_ctx *mq_hctx; member
|
| /linux/drivers/block/null_blk/ |
| A D | main.c | 1555 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in null_timeout_rq() 1652 ret = null_queue_rq(rq->mq_hctx, &bd); in null_queue_rqs()
|
Completed in 52 milliseconds