Lines Matching refs:request_queue

29 struct request_queue;
139 struct request_queue *queue;
396 struct request_queue { struct
582 void blk_queue_flag_set(unsigned int flag, struct request_queue *q);
583 void blk_queue_flag_clear(unsigned int flag, struct request_queue *q);
584 bool blk_queue_flag_test_and_set(unsigned int flag, struct request_queue *q);
619 extern void blk_set_pm_only(struct request_queue *q);
620 extern void blk_clear_pm_only(struct request_queue *q);
628 static inline bool queue_is_mq(struct request_queue *q) in queue_is_mq()
634 static inline enum rpm_status queue_rpm_status(struct request_queue *q) in queue_rpm_status()
639 static inline enum rpm_status queue_rpm_status(struct request_queue *q) in queue_rpm_status()
646 blk_queue_zoned_model(struct request_queue *q) in blk_queue_zoned_model()
653 static inline bool blk_queue_is_zoned(struct request_queue *q) in blk_queue_is_zoned()
732 static inline unsigned int blk_queue_depth(struct request_queue *q) in blk_queue_depth()
860 extern int blk_lld_busy(struct request_queue *q);
861 extern int blk_queue_enter(struct request_queue *q, blk_mq_req_flags_t flags);
862 extern void blk_queue_exit(struct request_queue *q);
863 extern void blk_sync_queue(struct request_queue *q);
879 static inline struct request_queue *bdev_get_queue(struct block_device *bdev) in bdev_get_queue()
911 void blk_queue_bounce_limit(struct request_queue *q, enum blk_bounce limit);
912 extern void blk_queue_max_hw_sectors(struct request_queue *, unsigned int);
913 extern void blk_queue_chunk_sectors(struct request_queue *, unsigned int);
914 extern void blk_queue_max_segments(struct request_queue *, unsigned short);
915 extern void blk_queue_max_discard_segments(struct request_queue *,
917 void blk_queue_max_secure_erase_sectors(struct request_queue *q,
919 extern void blk_queue_max_segment_size(struct request_queue *, unsigned int);
920 extern void blk_queue_max_discard_sectors(struct request_queue *q,
922 extern void blk_queue_max_write_zeroes_sectors(struct request_queue *q,
924 extern void blk_queue_logical_block_size(struct request_queue *, unsigned int);
925 extern void blk_queue_max_zone_append_sectors(struct request_queue *q,
927 extern void blk_queue_physical_block_size(struct request_queue *, unsigned int);
928 void blk_queue_zone_write_granularity(struct request_queue *q,
930 extern void blk_queue_alignment_offset(struct request_queue *q,
934 extern void blk_queue_io_min(struct request_queue *q, unsigned int min);
936 extern void blk_queue_io_opt(struct request_queue *q, unsigned int opt);
937 extern void blk_set_queue_depth(struct request_queue *q, unsigned int depth);
943 extern void blk_queue_update_dma_pad(struct request_queue *, unsigned int);
944 extern void blk_queue_segment_boundary(struct request_queue *, unsigned long);
945 extern void blk_queue_virt_boundary(struct request_queue *, unsigned long);
946 extern void blk_queue_dma_alignment(struct request_queue *, int);
947 extern void blk_queue_update_dma_alignment(struct request_queue *, int);
948 extern void blk_queue_rq_timeout(struct request_queue *, unsigned int);
949 extern void blk_queue_write_cache(struct request_queue *q, bool enabled, bool fua);
962 extern void blk_queue_required_elevator_features(struct request_queue *q,
964 extern bool blk_queue_can_use_dma_map_merging(struct request_queue *q,
967 bool __must_check blk_get_queue(struct request_queue *);
968 extern void blk_put_queue(struct request_queue *);
1107 static inline unsigned long queue_segment_boundary(const struct request_queue *q) in queue_segment_boundary()
1112 static inline unsigned long queue_virt_boundary(const struct request_queue *q) in queue_virt_boundary()
1117 static inline unsigned int queue_max_sectors(const struct request_queue *q) in queue_max_sectors()
1122 static inline unsigned int queue_max_bytes(struct request_queue *q) in queue_max_bytes()
1127 static inline unsigned int queue_max_hw_sectors(const struct request_queue *q) in queue_max_hw_sectors()
1132 static inline unsigned short queue_max_segments(const struct request_queue *q) in queue_max_segments()
1137 static inline unsigned short queue_max_discard_segments(const struct request_queue *q) in queue_max_discard_segments()
1142 static inline unsigned int queue_max_segment_size(const struct request_queue *q) in queue_max_segment_size()
1147 static inline unsigned int queue_max_zone_append_sectors(const struct request_queue *q) in queue_max_zone_append_sectors()
1166 static inline unsigned queue_logical_block_size(const struct request_queue *q) in queue_logical_block_size()
1181 static inline unsigned int queue_physical_block_size(const struct request_queue *q) in queue_physical_block_size()
1191 static inline unsigned int queue_io_min(const struct request_queue *q) in queue_io_min()
1201 static inline unsigned int queue_io_opt(const struct request_queue *q) in queue_io_opt()
1212 queue_zone_write_granularity(const struct request_queue *q) in queue_zone_write_granularity()
1244 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_zeroes_sectors()
1310 struct request_queue *q = bdev_get_queue(bdev); in bdev_zone_sectors()
1329 static inline int queue_dma_alignment(const struct request_queue *q) in queue_dma_alignment()
1346 static inline int blk_rq_aligned(struct request_queue *q, unsigned long addr, in blk_rq_aligned()
1375 struct request_queue *q);
1380 struct request_queue *q) in blk_crypto_register()