| /drivers/block/null_blk/ |
| A D | null_blk.h | 134 blk_status_t null_process_cmd(struct nullb_cmd *cmd, enum req_op op, 138 blk_status_t null_handle_memory_backed(struct nullb_cmd *cmd, enum req_op op, 147 blk_status_t null_process_zoned_cmd(struct nullb_cmd *cmd, enum req_op op, 166 enum req_op op, sector_t sector, sector_t nr_sectors) in null_process_zoned_cmd()
|
| A D | trace.h | 44 __field_struct(enum req_op, op) 49 __entry->op = req_op(blk_mq_rq_from_pdu(cmd));
|
| A D | main.c | 1291 op_is_write(req_op(rq)), sector, in null_handle_data_transfer() 1362 blk_status_t null_handle_memory_backed(struct nullb_cmd *cmd, enum req_op op, in null_handle_memory_backed() 1379 if (!dev->memory_backed && req_op(rq) == REQ_OP_READ) { in nullb_zero_read_cmd_buffer() 1413 blk_status_t null_process_cmd(struct nullb_cmd *cmd, enum req_op op, in null_process_cmd() 1433 sector_t nr_sectors, enum req_op op) in null_handle_cmd() 1598 cmd->error = null_process_cmd(cmd, req_op(req), blk_rq_pos(req), in null_poll() 1693 null_handle_cmd(cmd, sector, nr_sectors, req_op(rq)); in null_queue_rq()
|
| A D | zoned.c | 646 static blk_status_t null_zone_mgmt(struct nullb_cmd *cmd, enum req_op op, in null_zone_mgmt() 708 blk_status_t null_process_zoned_cmd(struct nullb_cmd *cmd, enum req_op op, in null_process_zoned_cmd()
|
| /drivers/block/rnbd/ |
| A D | rnbd-proto.h | 261 switch (req_op(rq)) { in rq_to_rnbd_flags() 282 (__force u32)req_op(rq), in rq_to_rnbd_flags()
|
| /drivers/block/ |
| A D | ublk_drv.c | 460 switch (req_op(req)) { in ublk_setup_iod_zoned() 995 (req_op(req) == REQ_OP_READ || req_op(req) == REQ_OP_DRV_IN); in ublk_need_unmap_req() 1074 enum req_op op = req_op(req); in ublk_setup_iod() 1081 switch (req_op(req)) { in ublk_setup_iod() 1127 if (!io->res && req_op(req) == REQ_OP_READ) in __ublk_complete_rq() 1141 if (req_op(req) != REQ_OP_READ && req_op(req) != REQ_OP_WRITE && in __ublk_complete_rq() 1142 req_op(req) != REQ_OP_DRV_IN) in __ublk_complete_rq() 2229 req_op(req) == REQ_OP_READ)) in ublk_check_commit_and_fetch() 2363 if (req_op(req) == REQ_OP_ZONE_APPEND) in __ublk_ch_uring_cmd() 2478 if ((req_op(req) == REQ_OP_READ || req_op(req) == REQ_OP_DRV_IN) && in ublk_check_ubuf_dir() [all …]
|
| A D | zloop.c | 365 bool is_append = req_op(rq) == REQ_OP_ZONE_APPEND; in zloop_rw() 366 bool is_write = req_op(rq) == REQ_OP_WRITE || is_append; in zloop_rw() 501 switch (req_op(rq)) { in zloop_handle_cmd() 535 pr_err("Unsupported operation %d\n", req_op(rq)); in zloop_handle_cmd() 561 switch (req_op(rq)) { in zloop_complete_rq() 580 req_op(rq) == REQ_OP_WRITE ? "" : "append ", in zloop_complete_rq() 598 if (req_op(rq) == REQ_OP_ZONE_APPEND) in zloop_complete_rq()
|
| A D | xen-blkfront.c | 566 if (req_op(req) == REQ_OP_SECURE_ERASE && info->feature_secdiscard) in blkif_queue_discard_req() 771 BUG_ON(req_op(req) == REQ_OP_FLUSH || req->cmd_flags & REQ_FUA); in blkif_queue_rw_req() 783 if (req_op(req) == REQ_OP_FLUSH || in blkif_queue_rw_req() 784 (req_op(req) == REQ_OP_WRITE && (req->cmd_flags & REQ_FUA))) { in blkif_queue_rw_req() 876 if (unlikely(req_op(req) == REQ_OP_DISCARD || in blkif_queue_request() 877 req_op(req) == REQ_OP_SECURE_ERASE)) in blkif_queue_request() 915 if (unlikely(req_op(qd->rq) == REQ_OP_FLUSH && !info->feature_flush)) in blkif_queue_rq() 2085 if (req_op(shadow[j].request) == REQ_OP_FLUSH || in blkfront_resume() 2086 req_op(shadow[j].request) == REQ_OP_DISCARD || in blkfront_resume() 2087 req_op(shadow[j].request) == REQ_OP_SECURE_ERASE || in blkfront_resume()
|
| A D | ps3disk.c | 168 switch (req_op(req)) { in ps3disk_do_request() 231 if (req_op(req) == REQ_OP_FLUSH) { in ps3disk_interrupt()
|
| A D | loop.c | 280 req_op(rq) != REQ_OP_READ) { in lo_complete_rq() 316 if (req_op(rq) == REQ_OP_WRITE) in lo_rw_aio_do_completion() 410 switch (req_op(rq)) { in do_req_filebacked() 1852 switch (req_op(rq)) { in loop_queue_rq() 1888 const bool write = op_is_write(req_op(rq)); in loop_handle_cmd()
|
| A D | virtio_blk.c | 247 if (!IS_ENABLED(CONFIG_BLK_DEV_ZONED) && op_is_zone_mgmt(req_op(req))) in virtblk_setup_cmd() 253 switch (req_op(req)) { in virtblk_setup_cmd() 343 if (req_op(req) == REQ_OP_ZONE_APPEND) in virtblk_request_done()
|
| /drivers/crypto/hisilicon/sec2/ |
| A D | sec_crypto.c | 249 ctx->req_op->buf_unmap(ctx, req); in sec_alg_send_backlog_soft() 332 ctx->req_op->buf_unmap(ctx, req); in sec_req_cb() 334 ctx->req_op->callback(ctx, req, err); in sec_req_cb() 1480 ret = ctx->req_op->buf_map(ctx, req); in sec_request_transfer() 1484 ctx->req_op->do_transfer(ctx, req); in sec_request_transfer() 1486 ret = ctx->req_op->bd_fill(ctx, req); in sec_request_transfer() 1493 ctx->req_op->buf_unmap(ctx, req); in sec_request_transfer() 1499 ctx->req_op->buf_unmap(ctx, req); in sec_request_untransfer() 1979 ret = ctx->req_op->bd_send(ctx, req); in sec_process() 2095 ctx->req_op = &sec_aead_req_ops; in sec_aead_init() [all …]
|
| A D | sec.h | 188 const struct sec_req_op *req_op; member
|
| /drivers/mmc/core/ |
| A D | queue.c | 46 switch (req_op(req)) { in mmc_cqe_issue_type() 67 if (req_op(req) == REQ_OP_READ || req_op(req) == REQ_OP_WRITE) in mmc_issue_type()
|
| /drivers/md/ |
| A D | dm-rq.c | 218 if (req_op(clone) == REQ_OP_DISCARD && in dm_done() 221 else if (req_op(clone) == REQ_OP_WRITE_ZEROES && in dm_done()
|
| A D | dm-ebs-target.c | 65 static int __ebs_rw_bvec(struct ebs_c *ec, enum req_op op, struct bio_vec *bv, in __ebs_rw_bvec() 122 static int __ebs_rw_bio(struct ebs_c *ec, enum req_op op, struct bio *bio) in __ebs_rw_bio()
|
| A D | dm-zoned.h | 249 unsigned int chunk, enum req_op op);
|
| A D | dm-io.c | 319 const enum req_op op = opf & REQ_OP_MASK; in do_region()
|
| /drivers/nvme/target/ |
| A D | passthru.c | 273 ARRAY_SIZE(req->inline_bvec), req_op(rq)); in nvmet_passthru_map_sg() 275 bio = bio_alloc(NULL, bio_max_segs(req->sg_cnt), req_op(rq), in nvmet_passthru_map_sg()
|
| A D | zns.c | 321 static inline enum req_op zsa_req_op(u8 zsa) in zsa_req_op() 476 enum req_op op = zsa_req_op(req->cmd->zms.zsa); in nvmet_bdev_zmgmt_send_work()
|
| /drivers/nvme/host/ |
| A D | zns.c | 244 if (req_op(req) == REQ_OP_ZONE_RESET_ALL) in nvme_setup_zone_mgmt_send()
|
| /drivers/s390/block/ |
| A D | dasd_fba.c | 556 if (req_op(req) == REQ_OP_DISCARD || req_op(req) == REQ_OP_WRITE_ZEROES) in dasd_fba_build_cp()
|
| /drivers/hwmon/ |
| A D | drivetemp.c | 167 enum req_op op; in drivetemp_scsi_command()
|
| /drivers/scsi/ |
| A D | sd_zbc.c | 375 if (op_is_zone_mgmt(req_op(rq)) && in sd_zbc_complete()
|
| /drivers/block/drbd/ |
| A D | drbd_actlog.c | 127 sector_t sector, enum req_op op) in _drbd_md_sync_page_io() 178 sector_t sector, enum req_op op) in drbd_md_sync_page_io()
|