| /drivers/mmc/core/ |
| A D | crypto.h | 13 struct request_queue; 19 void mmc_crypto_setup_queue(struct request_queue *q, struct mmc_host *host); 29 static inline void mmc_crypto_setup_queue(struct request_queue *q, in mmc_crypto_setup_queue()
|
| A D | queue.c | 86 struct request_queue *q = req->q; in mmc_cqe_recovery_notifier() 122 struct request_queue *q = req->q; in mmc_mq_timed_out() 140 struct request_queue *q = mq->queue; in mmc_mq_recovery_handler() 233 struct request_queue *q = req->q; in mmc_mq_queue_rq() 491 struct request_queue *q = mq->queue; in mmc_cleanup_queue()
|
| A D | crypto.c | 22 void mmc_crypto_setup_queue(struct request_queue *q, struct mmc_host *host) in mmc_crypto_setup_queue()
|
| /drivers/char/ipmi/ |
| A D | ipmb_dev_int.c | 60 struct list_head request_queue; member 86 while (list_empty(&ipmb_dev->request_queue)) { in ipmb_read() 93 !list_empty(&ipmb_dev->request_queue)); in ipmb_read() 100 queue_elem = list_first_entry(&ipmb_dev->request_queue, in ipmb_read() 217 list_add(&queue_elem->list, &ipmb_dev->request_queue); in ipmb_handle_request() 315 INIT_LIST_HEAD(&ipmb_dev->request_queue); in ipmb_probe()
|
| /drivers/scsi/ |
| A D | scsi_lib.c | 412 blk_mq_run_hw_queues(sdev->request_queue, true); in scsi_kick_sdev_queue() 487 struct request_queue *slq; in scsi_starved_list_run() 521 slq = sdev->request_queue; in scsi_starved_list_run() 558 struct request_queue *q; in scsi_requeue_run_queue() 561 q = sdev->request_queue; in scsi_requeue_run_queue() 570 scsi_run_queue(sdev->request_queue); in scsi_run_host_queues() 635 struct request_queue *q = sdev->request_queue; in scsi_end_request() 1811 struct request_queue *q = req->q; in scsi_queue_rq() 2743 struct request_queue *q = sdev->request_queue; in scsi_device_quiesce() 2801 blk_clear_pm_only(sdev->request_queue); in scsi_device_resume() [all …]
|
| A D | scsi_dh.c | 251 int scsi_dh_activate(struct request_queue *q, activate_complete fn, void *data) in scsi_dh_activate() 298 int scsi_dh_set_params(struct request_queue *q, const char *params) in scsi_dh_set_params() 320 int scsi_dh_attach(struct request_queue *q, const char *name) in scsi_dh_attach() 359 const char *scsi_dh_attached_handler_name(struct request_queue *q, gfp_t gfp) in scsi_dh_attached_handler_name()
|
| A D | scsi_pm.c | 151 err = blk_pre_runtime_suspend(sdev->request_queue); in sdev_runtime_suspend() 156 blk_post_runtime_suspend(sdev->request_queue, err); in sdev_runtime_suspend() 180 blk_pre_runtime_resume(sdev->request_queue); in sdev_runtime_resume() 183 blk_post_runtime_resume(sdev->request_queue); in sdev_runtime_resume()
|
| A D | scsi_ioctl.c | 204 queue_max_bytes(sdev->request_queue)); in sg_get_reserved_size() 220 queue_max_bytes(sdev->request_queue)); in sg_set_reserved_size() 228 static int sg_emulated_host(struct request_queue *q, int __user *p) in sg_emulated_host() 440 if (hdr->dxfer_len > (queue_max_hw_sectors(sdev->request_queue) << 9)) in sg_io() 457 rq = scsi_alloc_request(sdev->request_queue, writing ? in sg_io() 520 static int sg_scsi_ioctl(struct request_queue *q, bool open_for_write, in sg_scsi_ioctl() 884 struct request_queue *q = sdev->request_queue; in scsi_ioctl()
|
| A D | scsi_scan.c | 245 memflags = blk_mq_freeze_queue(sdev->request_queue); in scsi_realloc_sdev_budget_map() 251 sdev->request_queue->node, false, true); in scsi_realloc_sdev_budget_map() 261 blk_mq_unfreeze_queue(sdev->request_queue, memflags); in scsi_realloc_sdev_budget_map() 285 struct request_queue *q; in scsi_alloc_sdev() 348 sdev->request_queue = q; in scsi_alloc_sdev() 1074 lim = queue_limits_start_update(sdev->request_queue); in scsi_add_lun() 1083 queue_limits_cancel_update(sdev->request_queue); in scsi_add_lun() 1094 ret = queue_limits_commit_update(sdev->request_queue, &lim); in scsi_add_lun() 1685 blk_queue_pm_only(sdev->request_queue)) { in scsi_resume_device() 1719 blk_queue_pm_only(sdev->request_queue)) { in scsi_rescan_device()
|
| A D | scsi_bsg.c | 12 static int scsi_bsg_sg_io_fn(struct request_queue *q, struct sg_io_v4 *hdr, in scsi_bsg_sg_io_fn() 101 return bsg_register_queue(sdev->request_queue, &sdev->sdev_gendev, in scsi_bsg_register_queue()
|
| A D | sg.c | 287 struct request_queue *q; in sg_open() 345 q = device->request_queue; in sg_open() 843 static int max_sectors_bytes(struct request_queue *q) in max_sectors_bytes() 1025 max_sectors_bytes(sdp->device->request_queue)); in sg_ioctl_common() 1041 max_sectors_bytes(sdp->device->request_queue)); in sg_ioctl_common() 1431 struct request_queue *q = scsidp->request_queue; in sg_alloc() 1497 if (!blk_get_queue(scsidp->request_queue)) { in sg_add_device() 1559 blk_put_queue(scsidp->request_queue); in sg_add_device() 1567 struct request_queue *q = sdp->device->request_queue; in sg_device_destroy() 1731 struct request_queue *q = sfp->parentdp->device->request_queue; in sg_start_req() [all …]
|
| A D | sr.c | 628 disk = blk_mq_alloc_disk_for_queue(sdev->request_queue, in sr_probe() 654 blk_queue_rq_timeout(sdev->request_queue, SR_TIMEOUT); in sr_probe() 688 blk_pm_runtime_init(sdev->request_queue, dev); in sr_probe() 724 struct request_queue *q = cd->device->request_queue; in get_sectorsize()
|
| A D | st.c | 758 STp->device->request_queue->rq_timeout, in cross_eof() 803 STp->device->request_queue->rq_timeout, in st_flush_write_buffer() 1141 STp->device->request_queue->rq_timeout, in check_tape() 1168 STp->device->request_queue->rq_timeout, in check_tape() 1415 STp->device->request_queue->rq_timeout, in st_flush() 1583 STp->device->request_queue)) == 0) { in setup_buffering() 1803 STp->device->request_queue->rq_timeout, in st_write() 1973 STp->device->request_queue->rq_timeout, in read_tape() 2405 blk_queue_rq_timeout(STp->device->request_queue, 3154 STp->device->request_queue->rq_timeout, in get_location() [all …]
|
| A D | sd_zbc.c | 116 const int timeout = sdp->request_queue->rq_timeout; in sd_zbc_do_report_zones() 171 struct request_queue *q = sdkp->disk->queue; in sd_zbc_alloc_report_buffer() 550 struct request_queue *q = disk->queue; in sd_zbc_revalidate_zones()
|
| A D | scsi_sysfs.c | 477 blk_put_queue(sdev->request_queue); in scsi_device_dev_release() 479 sdev->request_queue = NULL; in scsi_device_dev_release() 706 return snprintf(buf, 20, "%d\n", sdev->request_queue->rq_timeout / HZ); in sdev_show_timeout() 717 blk_queue_rq_timeout(sdev->request_queue, timeout * HZ); in sdev_store_timeout() 847 blk_mq_run_hw_queues(sdev->request_queue, true); in store_state_field() 1138 err = scsi_dh_attach(sdev->request_queue, buf); in sdev_store_dh_state() 1512 blk_mq_destroy_queue(sdev->request_queue); in __scsi_remove_device()
|
| /drivers/md/ |
| A D | dm-rq.h | 37 void dm_start_queue(struct request_queue *q); 38 void dm_stop_queue(struct request_queue *q);
|
| A D | dm.h | 64 int dm_table_set_restrictions(struct dm_table *t, struct request_queue *q, 105 int dm_set_zones_restrictions(struct dm_table *t, struct request_queue *q, 107 int dm_revalidate_zones(struct dm_table *t, struct request_queue *q);
|
| A D | dm-zone.c | 139 struct request_queue *q = md->queue; in dm_is_zone_write() 159 int dm_revalidate_zones(struct dm_table *t, struct request_queue *q) in dm_revalidate_zones() 338 int dm_set_zones_restrictions(struct dm_table *t, struct request_queue *q, in dm_set_zones_restrictions()
|
| A D | dm-rq.c | 64 void dm_start_queue(struct request_queue *q) in dm_start_queue() 70 void dm_stop_queue(struct request_queue *q) in dm_stop_queue() 171 static void __dm_mq_kick_requeue_list(struct request_queue *q, unsigned long msecs) in __dm_mq_kick_requeue_list()
|
| /drivers/ufs/core/ |
| A D | ufshcd-crypto.h | 73 void ufshcd_crypto_register(struct ufs_hba *hba, struct request_queue *q); 106 struct request_queue *q) { } in ufshcd_crypto_register()
|
| /drivers/usb/storage/ |
| A D | scsiglue.c | 580 return sprintf(buf, "%u\n", queue_max_hw_sectors(sdev->request_queue)); in max_sectors_show() 595 lim = queue_limits_start_update(sdev->request_queue); in max_sectors_store() 597 ret = queue_limits_commit_update_frozen(sdev->request_queue, &lim); in max_sectors_store()
|
| /drivers/thunderbolt/ |
| A D | ctl.c | 47 struct list_head request_queue; member 143 list_add_tail(&req->list, &ctl->request_queue); in tb_cfg_request_enqueue() 177 list_for_each_entry(iter, &pkg->ctl->request_queue, list) { in tb_cfg_request_find() 661 INIT_LIST_HEAD(&ctl->request_queue); in tb_ctl_alloc() 753 if (!list_empty(&ctl->request_queue)) in tb_ctl_stop() 755 INIT_LIST_HEAD(&ctl->request_queue); in tb_ctl_stop()
|
| /drivers/nvme/host/ |
| A D | nvme.h | 288 struct request_queue *admin_q; 289 struct request_queue *connect_q; 290 struct request_queue *fabrics_q; 532 struct request_queue *queue; 898 int nvme_submit_sync_cmd(struct request_queue *q, struct nvme_command *cmd, 900 int __nvme_submit_sync_cmd(struct request_queue *q, struct nvme_command *cmd,
|
| /drivers/cdrom/ |
| A D | gdrom.c | 87 static DECLARE_WAIT_QUEUE_HEAD(request_queue); 103 struct request_queue *gdrom_rq; 541 wake_up_interruptible(&request_queue); in gdrom_dma_interrupt() 623 wait_event_interruptible_timeout(request_queue, in gdrom_readdisk_dma()
|
| /drivers/s390/block/ |
| A D | scm_blk.h | 18 struct request_queue *rq;
|