Lines Matching refs:request_queue

49 static void blk_mq_poll_stats_start(struct request_queue *q);
72 static inline struct blk_mq_hw_ctx *blk_qc_to_hctx(struct request_queue *q, in blk_qc_to_hctx()
144 unsigned int blk_mq_in_flight(struct request_queue *q, in blk_mq_in_flight()
154 void blk_mq_in_flight_rw(struct request_queue *q, struct block_device *part, in blk_mq_in_flight_rw()
164 void blk_freeze_queue_start(struct request_queue *q) in blk_freeze_queue_start()
178 void blk_mq_freeze_queue_wait(struct request_queue *q) in blk_mq_freeze_queue_wait()
184 int blk_mq_freeze_queue_wait_timeout(struct request_queue *q, in blk_mq_freeze_queue_wait_timeout()
197 void blk_freeze_queue(struct request_queue *q) in blk_freeze_queue()
210 void blk_mq_freeze_queue(struct request_queue *q) in blk_mq_freeze_queue()
220 void __blk_mq_unfreeze_queue(struct request_queue *q, bool force_atomic) in __blk_mq_unfreeze_queue()
234 void blk_mq_unfreeze_queue(struct request_queue *q) in blk_mq_unfreeze_queue()
244 void blk_mq_quiesce_queue_nowait(struct request_queue *q) in blk_mq_quiesce_queue_nowait()
282 void blk_mq_quiesce_queue(struct request_queue *q) in blk_mq_quiesce_queue()
298 void blk_mq_unquiesce_queue(struct request_queue *q) in blk_mq_unquiesce_queue()
320 struct request_queue *q; in blk_mq_quiesce_tagset()
334 struct request_queue *q; in blk_mq_unquiesce_tagset()
345 void blk_mq_wake_waiters(struct request_queue *q) in blk_mq_wake_waiters()
355 void blk_rq_init(struct request_queue *q, struct request *rq) in blk_rq_init()
377 struct request_queue *q = data->q; in blk_mq_rq_ctx_init()
473 struct request_queue *q = data->q; in __blk_mq_alloc_requests()
544 static struct request *blk_mq_rq_cache_fill(struct request_queue *q, in blk_mq_rq_cache_fill()
569 static struct request *blk_mq_alloc_cached_request(struct request_queue *q, in blk_mq_alloc_cached_request()
603 struct request *blk_mq_alloc_request(struct request_queue *q, blk_opf_t opf, in blk_mq_alloc_request()
636 struct request *blk_mq_alloc_request_hctx(struct request_queue *q, in blk_mq_alloc_request_hctx()
712 struct request_queue *q = rq->q; in __blk_mq_free_request()
730 struct request_queue *q = rq->q; in blk_mq_free_request()
1065 struct request_queue *q = hctx->queue; in blk_mq_flush_tag_batch()
1251 struct request_queue *q = rq->q; in blk_mq_start_request()
1424 struct request_queue *q = rq->q; in __blk_mq_requeue_request()
1450 struct request_queue *q = in blk_mq_requeue_work()
1451 container_of(work, struct request_queue, requeue_work.work); in blk_mq_requeue_work()
1488 struct request_queue *q = rq->q; in blk_mq_add_to_requeue_list()
1510 void blk_mq_kick_requeue_list(struct request_queue *q) in blk_mq_kick_requeue_list()
1516 void blk_mq_delay_kick_requeue_list(struct request_queue *q, in blk_mq_delay_kick_requeue_list()
1541 bool blk_mq_queue_inflight(struct request_queue *q) in blk_mq_queue_inflight()
1630 struct request_queue *q = in blk_mq_timeout_work()
1631 container_of(work, struct request_queue, timeout_work); in blk_mq_timeout_work()
1987 static void blk_mq_release_budgets(struct request_queue *q, in blk_mq_release_budgets()
2024 struct request_queue *q = hctx->queue; in blk_mq_dispatch_rq_list()
2309 static struct blk_mq_hw_ctx *blk_mq_get_sq_hctx(struct request_queue *q) in blk_mq_get_sq_hctx()
2331 void blk_mq_run_hw_queues(struct request_queue *q, bool async) in blk_mq_run_hw_queues()
2359 void blk_mq_delay_run_hw_queues(struct request_queue *q, unsigned long msecs) in blk_mq_delay_run_hw_queues()
2416 void blk_mq_stop_hw_queues(struct request_queue *q) in blk_mq_stop_hw_queues()
2434 void blk_mq_start_hw_queues(struct request_queue *q) in blk_mq_start_hw_queues()
2454 void blk_mq_start_stopped_hw_queues(struct request_queue *q, bool async) in blk_mq_start_stopped_hw_queues()
2575 struct request_queue *q = rq->q; in __blk_mq_issue_directly()
2609 struct request_queue *q = rq->q; in __blk_mq_try_issue_directly()
2715 static void __blk_mq_flush_plug_list(struct request_queue *q, in __blk_mq_flush_plug_list()
2759 struct request_queue *q; in blk_mq_flush_plug_list()
2825 static bool blk_mq_attempt_bio_merge(struct request_queue *q, in blk_mq_attempt_bio_merge()
2837 static struct request *blk_mq_get_new_requests(struct request_queue *q, in blk_mq_get_new_requests()
2874 static inline struct request *blk_mq_get_cached_request(struct request_queue *q, in blk_mq_get_cached_request()
2936 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in blk_mq_submit_bio()
3001 struct request_queue *q = rq->q; in blk_insert_cloned_request()
3575 static void blk_mq_exit_hctx(struct request_queue *q, in blk_mq_exit_hctx()
3602 static void blk_mq_exit_hw_queues(struct request_queue *q, in blk_mq_exit_hw_queues()
3615 static int blk_mq_init_hctx(struct request_queue *q, in blk_mq_init_hctx()
3653 blk_mq_alloc_hctx(struct request_queue *q, struct blk_mq_tag_set *set, in blk_mq_alloc_hctx()
3717 static void blk_mq_init_cpu_queues(struct request_queue *q, in blk_mq_init_cpu_queues()
3801 static void blk_mq_map_swqueue(struct request_queue *q) in blk_mq_map_swqueue()
3908 static void queue_set_hctx_shared(struct request_queue *q, bool shared) in queue_set_hctx_shared()
3926 struct request_queue *q; in blk_mq_update_tag_set_shared()
3937 static void blk_mq_del_queue_tag_set(struct request_queue *q) in blk_mq_del_queue_tag_set()
3954 struct request_queue *q) in blk_mq_add_queue_tag_set()
3975 static int blk_mq_alloc_ctxs(struct request_queue *q) in blk_mq_alloc_ctxs()
4008 void blk_mq_release(struct request_queue *q) in blk_mq_release()
4031 static struct request_queue *blk_mq_init_queue_data(struct blk_mq_tag_set *set, in blk_mq_init_queue_data()
4034 struct request_queue *q; in blk_mq_init_queue_data()
4049 struct request_queue *blk_mq_init_queue(struct blk_mq_tag_set *set) in blk_mq_init_queue()
4065 void blk_mq_destroy_queue(struct request_queue *q) in blk_mq_destroy_queue()
4085 struct request_queue *q; in __blk_mq_alloc_disk()
4103 struct gendisk *blk_mq_alloc_disk_for_queue(struct request_queue *q, in blk_mq_alloc_disk_for_queue()
4118 struct blk_mq_tag_set *set, struct request_queue *q, in blk_mq_alloc_and_init_hctx()
4152 struct request_queue *q) in blk_mq_realloc_hw_ctxs()
4194 static void blk_mq_update_poll_flag(struct request_queue *q) in blk_mq_update_poll_flag()
4206 struct request_queue *q) in blk_mq_init_allocated_queue()
4268 void blk_mq_exit_queue(struct request_queue *q) in blk_mq_exit_queue()
4553 int blk_mq_update_nr_requests(struct request_queue *q, unsigned int nr) in blk_mq_update_nr_requests()
4612 struct request_queue *q;
4621 struct request_queue *q) in blk_mq_elv_switch_none()
4648 struct request_queue *q) in blk_lookup_qe_pair()
4660 struct request_queue *q) in blk_mq_elv_switch_back()
4682 struct request_queue *q; in __blk_mq_update_nr_hw_queues()
4758 static bool blk_poll_stats_enable(struct request_queue *q) in blk_poll_stats_enable()
4766 static void blk_mq_poll_stats_start(struct request_queue *q) in blk_mq_poll_stats_start()
4780 struct request_queue *q = cb->data; in blk_mq_poll_stats_fn()
4789 static unsigned long blk_mq_poll_nsecs(struct request_queue *q, in blk_mq_poll_nsecs()
4821 static bool blk_mq_poll_hybrid(struct request_queue *q, blk_qc_t qc) in blk_mq_poll_hybrid()
4887 static int blk_mq_poll_classic(struct request_queue *q, blk_qc_t cookie, in blk_mq_poll_classic()
4915 int blk_mq_poll(struct request_queue *q, blk_qc_t cookie, struct io_comp_batch *iob, in blk_mq_poll()
4932 void blk_mq_cancel_work_sync(struct request_queue *q) in blk_mq_cancel_work_sync()