Home
last modified time | relevance | path

Searched refs:io_req (Results 1 – 25 of 29) sorted by relevance

12

/linux/drivers/scsi/qedf/
A Dqedf_io.c86 if (io_req->cb_func && io_req->cb_arg) { in qedf_cmd_timeout()
87 io_req->cb_func(io_req->cb_arg); in qedf_cmd_timeout()
386 bd_tbl->io_req = io_req; in qedf_alloc_cmd()
443 io_req, io_req->xid); in qedf_release_cmd()
1441 io_req, io_req->xid); in qedf_process_warning_compl()
1708 io_req, io_req->xid); in qedf_flush_active_ios()
1728 io_req, io_req->xid); in qedf_flush_active_ios()
1828 io_req, io_req->xid, in qedf_flush_active_ios()
1914 io_req->xid, io_req->sc_cmd); in qedf_initiate_abts()
2211 io_req->xid, io_req->sc_cmd, io_req->cmd_type, io_req->flags, in qedf_initiate_cleanup()
[all …]
A Dqedf.h72 struct qedf_ioreq *io_req; member
426 struct qedf_ioreq *io_req; member
487 struct qedf_ioreq *io_req);
489 struct fcoe_cqe *cqe, struct qedf_ioreq *io_req);
491 struct fcoe_cqe *cqe, struct qedf_ioreq *io_req);
497 struct qedf_ioreq *io_req);
504 extern int qedf_init_mp_req(struct qedf_ioreq *io_req);
516 struct fcoe_cqe *cqe, struct qedf_ioreq *io_req);
519 struct qedf_ioreq *io_req);
533 struct qedf_ioreq *io_req);
[all …]
A Dqedf_main.c720 struct qedf_ioreq *io_req; in qedf_eh_abort() local
744 if (!io_req) { in qedf_eh_abort()
770 io_req, io_req->xid, sc_cmd, sc_cmd->cmnd[0], in qedf_eh_abort()
805 io_req, sc_cmd, io_req->xid, io_req->fp_idx, in qedf_eh_abort()
845 io_req->xid); in qedf_eh_abort()
848 io_req->xid); in qedf_eh_abort()
2256 if (!io_req) in qedf_process_completions()
2262 cpu = io_req->cpu; in qedf_process_completions()
2795 if (!io_req) { in qedf_process_cqe()
2801 fcport = io_req->fcport; in qedf_process_cqe()
[all …]
A Dqedf_els.c72 cb_arg->io_req = els_req; in qedf_initiate_els()
198 rrq_req = cb_arg->io_req; in qedf_rrq_compl()
430 els_req = cb_arg->io_req; in qedf_l2_els_compl()
552 srr_req = cb_arg->io_req; in qedf_srr_compl()
744 struct fcoe_cqe *cqe, struct qedf_ioreq *io_req) in qedf_process_seq_cleanup_compl() argument
749 cb_arg = io_req->cb_arg; in qedf_process_seq_cleanup_compl()
752 if (io_req->event == QEDF_IOREQ_EV_ELS_TMO || !cqe) { in qedf_process_seq_cleanup_compl()
759 cancel_delayed_work_sync(&io_req->timeout_work); in qedf_process_seq_cleanup_compl()
764 "abort, xid=0x%x.\n", io_req->xid); in qedf_process_seq_cleanup_compl()
767 kref_put(&io_req->refcount, qedf_release_cmd); in qedf_process_seq_cleanup_compl()
[all …]
/linux/drivers/scsi/bnx2fc/
A Dbnx2fc_io.c144 if ((io_req->cb_func) && (io_req->cb_arg)) { in bnx2fc_cmd_timeout()
145 io_req->cb_func(io_req->cb_arg); in bnx2fc_cmd_timeout()
275 io_req = kzalloc(sizeof(*io_req), GFP_KERNEL); in bnx2fc_cmd_mgr_alloc()
293 io_req++; in bnx2fc_cmd_mgr_alloc()
457 bd_tbl->io_req = io_req; in bnx2fc_elstm_alloc()
461 return io_req; in bnx2fc_elstm_alloc()
512 bd_tbl->io_req = io_req; in bnx2fc_cmd_alloc()
516 return io_req; in bnx2fc_cmd_alloc()
1327 kref_read(&io_req->refcount), io_req->cmd_type); in bnx2fc_process_cleanup_compl()
1365 kref_read(&io_req->refcount), io_req->cmd_type); in bnx2fc_process_abts_compl()
[all …]
A Dbnx2fc_tgt.c168 struct bnx2fc_cmd *io_req; in bnx2fc_flush_active_ios() local
181 io_req->on_active_queue = 0; in bnx2fc_flush_active_ios()
186 &io_req->req_flags)) { in bnx2fc_flush_active_ios()
192 kref_put(&io_req->refcount, in bnx2fc_flush_active_ios()
201 bnx2fc_process_cleanup_compl(io_req, io_req->task, 0); in bnx2fc_flush_active_ios()
211 io_req->on_tmf_queue = 0; in bnx2fc_flush_active_ios()
228 if ((io_req->cb_func) && (io_req->cb_arg)) { in bnx2fc_flush_active_ios()
229 io_req->cb_func(io_req->cb_arg); in bnx2fc_flush_active_ios()
230 io_req->cb_arg = NULL; in bnx2fc_flush_active_ios()
235 bnx2fc_process_cleanup_compl(io_req, io_req->task, 0); in bnx2fc_flush_active_ios()
[all …]
A Dbnx2fc_debug.c17 void BNX2FC_IO_DBG(const struct bnx2fc_cmd *io_req, const char *fmt, ...) in BNX2FC_IO_DBG() argument
30 if (io_req && io_req->port && io_req->port->lport && in BNX2FC_IO_DBG()
31 io_req->port->lport->host) in BNX2FC_IO_DBG()
32 shost_printk(KERN_INFO, io_req->port->lport->host, in BNX2FC_IO_DBG()
34 io_req->xid, &vaf); in BNX2FC_IO_DBG()
A Dbnx2fc_hwi.c711 if (!io_req) in bnx2fc_process_unsol_compl()
790 io_req->xid); in bnx2fc_process_unsol_compl()
832 if (!io_req) in bnx2fc_process_unsol_compl()
866 struct bnx2fc_cmd *io_req; in bnx2fc_process_cq_compl() local
877 if (io_req == NULL) { in bnx2fc_process_cq_compl()
884 cmd_type = io_req->cmd_type; in bnx2fc_process_cq_compl()
945 io_req->xid); in bnx2fc_process_cq_compl()
1630 io_req->task = task; in bnx2fc_init_mp_task()
1721 io_req->task = task; in bnx2fc_init_task()
1765 io_req->rec_retry = 0; in bnx2fc_init_task()
[all …]
A Dbnx2fc.h408 struct bnx2fc_cmd *io_req; member
475 struct bnx2fc_cmd *io_req; member
529 int bnx2fc_initiate_cleanup(struct bnx2fc_cmd *io_req);
530 int bnx2fc_initiate_abts(struct bnx2fc_cmd *io_req);
531 void bnx2fc_cmd_timer_set(struct bnx2fc_cmd *io_req,
533 int bnx2fc_init_mp_req(struct bnx2fc_cmd *io_req);
534 void bnx2fc_init_cleanup_task(struct bnx2fc_cmd *io_req,
541 void bnx2fc_init_mp_task(struct bnx2fc_cmd *io_req,
543 void bnx2fc_init_task(struct bnx2fc_cmd *io_req,
562 void bnx2fc_process_tm_compl(struct bnx2fc_cmd *io_req,
[all …]
A Dbnx2fc_debug.h41 void BNX2FC_IO_DBG(const struct bnx2fc_cmd *io_req, const char *fmt, ...);
/linux/drivers/scsi/fnic/
A Dfnic_scsi.c131 sizeof(io_req->sgl_list[0]) * io_req->sgl_cnt, in fnic_release_ioreq_buf()
510 memset(io_req, 0, sizeof(*io_req)); in fnic_queuecommand_lck()
540 io_req->sgl_list_alloc = io_req->sgl_list; in fnic_queuecommand_lck()
1381 if (io_req && io_req->dr_done) in fnic_cleanup_io_iter()
1383 else if (io_req && io_req->abts_done) in fnic_cleanup_io_iter()
1576 if (!io_req || io_req->port_id != iter_data->port_id) { in fnic_rport_abort_io_iter()
1844 if (io_req) in fnic_abort_cmd()
2094 if (io_req) in fnic_pending_aborts_iter()
2303 memset(io_req, 0, sizeof(*io_req)); in fnic_device_reset()
2434 if (io_req) in fnic_device_reset()
[all …]
/linux/arch/um/drivers/
A Dubd_kern.c488 blk_mq_end_request(io_req->req, io_req->error); in ubd_handler()
489 kfree(io_req); in ubd_handler()
1285 cowify_req(io_req, &io_req->io_desc[i], byte_offset, in ubd_map_req()
1300 io_req = kmalloc(sizeof(*io_req) + in ubd_alloc_req()
1303 if (!io_req) in ubd_alloc_req()
1306 io_req->req = req; in ubd_alloc_req()
1311 io_req->error = 0; in ubd_alloc_req()
1323 return io_req; in ubd_alloc_req()
1341 if (!io_req) in ubd_submit_request()
1348 ret = os_write_file(thread_fd, &io_req, sizeof(io_req)); in ubd_submit_request()
[all …]
/linux/drivers/md/
A Ddm-io.c500 switch (io_req->mem.type) { in dp_init()
502 list_dp_init(dp, io_req->mem.ptr.pl, io_req->mem.offset); in dp_init()
506 bio_dp_init(dp, io_req->mem.ptr.bio); in dp_init()
511 if (io_req->bi_op == REQ_OP_READ) { in dp_init()
515 vm_dp_init(dp, io_req->mem.ptr.vma); in dp_init()
519 km_dp_init(dp, io_req->mem.ptr.addr); in dp_init()
547 if (!io_req->notify.fn) in dm_io()
549 io_req->bi_op, io_req->bi_op_flags, &dp, in dm_io()
552 return async_io(io_req->client, num_regions, where, io_req->bi_op, in dm_io()
553 io_req->bi_op_flags, &dp, io_req->notify.fn, in dm_io()
[all …]
A Ddm-log.c240 struct dm_io_request io_req; member
296 lc->io_req.bi_op = op; in rw_header()
297 lc->io_req.bi_op_flags = 0; in rw_header()
310 lc->io_req.bi_op = REQ_OP_WRITE; in flush_header()
311 lc->io_req.bi_op_flags = REQ_PREFLUSH; in flush_header()
459 lc->io_req.mem.type = DM_IO_VMA; in create_log_context()
460 lc->io_req.notify.fn = NULL; in create_log_context()
462 if (IS_ERR(lc->io_req.client)) { in create_log_context()
463 r = PTR_ERR(lc->io_req.client); in create_log_context()
472 dm_io_client_destroy(lc->io_req.client); in create_log_context()
[all …]
A Ddm-integrity.c558 io_req.bi_op = op; in sync_rw_sb()
562 io_req.notify.fn = NULL; in sync_rw_sb()
563 io_req.client = ic->io; in sync_rw_sb()
1068 io_req.bi_op = op; in rw_journal_sectors()
1080 io_req.notify.fn = NULL; in rw_journal_sectors()
1082 io_req.client = ic->io; in rw_journal_sectors()
1190 io_req.bi_op_flags = 0; in copy_from_journal()
1194 io_req.notify.fn = fn; in copy_from_journal()
1196 io_req.client = ic->io; in copy_from_journal()
2707 io_req.bi_op_flags = 0; in integrity_recalc()
[all …]
A Ddm-snap-persistent.c214 struct dm_io_request *io_req; member
223 req->result = dm_io(req->io_req, 1, req->where, NULL); in do_metadata()
237 struct dm_io_request io_req = { in chunk_io() local
248 return dm_io(&io_req, 1, &where, NULL); in chunk_io()
251 req.io_req = &io_req; in chunk_io()
A Ddm-raid1.c262 struct dm_io_request io_req = { in mirror_flush() local
277 dm_io(&io_req, ms->nr_mirrors, io, &error_bits); in mirror_flush()
537 struct dm_io_request io_req = { in read_async_bio() local
549 BUG_ON(dm_io(&io_req, 1, &io, NULL)); in read_async_bio()
651 struct dm_io_request io_req = { in do_write() local
662 io_req.bi_op = REQ_OP_DISCARD; in do_write()
663 io_req.mem.type = DM_IO_KMEM; in do_write()
664 io_req.mem.ptr.addr = NULL; in do_write()
676 BUG_ON(dm_io(&io_req, ms->nr_mirrors, io, NULL)); in do_write()
A Ddm-bufio.c584 struct dm_io_request io_req = { in use_dmio() local
598 io_req.mem.type = DM_IO_KMEM; in use_dmio()
599 io_req.mem.ptr.addr = (char *)b->data + offset; in use_dmio()
601 io_req.mem.type = DM_IO_VMA; in use_dmio()
602 io_req.mem.ptr.vma = (char *)b->data + offset; in use_dmio()
605 r = dm_io(&io_req, 1, &region, NULL); in use_dmio()
1344 struct dm_io_request io_req = { in dm_bufio_issue_flush() local
1359 return dm_io(&io_req, 1, &io_reg, NULL); in dm_bufio_issue_flush()
1368 struct dm_io_request io_req = { in dm_bufio_issue_discard() local
1383 return dm_io(&io_req, 1, &io_reg, NULL); in dm_bufio_issue_discard()
/linux/samples/acrn/
A Dvm-sample.c53 struct acrn_io_request *io_req; in main() local
113 io_req = &io_req_buf[vcpu_id]; in main()
114 if ((__sync_add_and_fetch(&io_req->processed, 0) == ACRN_IOREQ_STATE_PROCESSING) in main()
115 && (!io_req->kernel_handled)) in main()
116 if (io_req->type == ACRN_IOREQ_TYPE_PORTIO) { in main()
119 port = io_req->reqs.pio_request.address; in main()
120 bytes = io_req->reqs.pio_request.size; in main()
121 in = (io_req->reqs.pio_request.direction == ACRN_IOREQ_DIR_READ); in main()
/linux/drivers/staging/rtl8723bs/include/
A Drtw_io.h109 struct io_req { struct
117 void (*_async_io_callback)(struct adapter *padater, struct io_req *pio_req, u8 *cnxt); argument
158 extern void sync_ioreq_enqueue(struct io_req *preq, struct io_queue *ioqueue);
162 extern uint free_ioreq(struct io_req *preq, struct io_queue *pio_queue);
163 extern struct io_req *alloc_ioreq(struct io_queue *pio_q);
193 void (*_async_io_callback)(struct adapter *padater, struct io_req *pio_req, u8 *cnxt), u8 *cnxt);
195 void (*_async_io_callback)(struct adapter *padater, struct io_req *pio_req, u8 *cnxt), u8 *cnxt);
197 void (*_async_io_callback)(struct adapter *padater, struct io_req *pio_req, u8 *cnxt), u8 *cnxt);
203 void (*_async_io_callback)(struct adapter *padater, struct io_req *pio_req, u8 *cnxt), u8 *cnxt);
205 void (*_async_io_callback)(struct adapter *padater, struct io_req *pio_req, u8 *cnxt), u8 *cnxt);
[all …]
/linux/drivers/staging/r8188eu/include/
A Drtw_io.h87 struct io_req { struct
97 struct io_req *pio_req, u8 *cnxt); argument
212 void sync_ioreq_enqueue(struct io_req *preq,struct io_queue *ioqueue);
214 uint free_ioreq(struct io_req *preq, struct io_queue *pio_queue);
215 struct io_req *alloc_ioreq(struct io_queue *pio_q);
251 struct io_req *pio_req,
255 struct io_req *pio_req,
259 struct io_req *pio_req,
267 struct io_req *pio_req,
271 struct io_req *pio_req,
[all …]
/linux/drivers/scsi/csiostor/
A Dcsio_lnode.c1435 struct csio_ioreq *io_req = NULL; in csio_ln_mgmt_wr_handler() local
1456 io_req); in csio_ln_mgmt_wr_handler()
1465 list_del_init(&io_req->sm.sm_list); in csio_ln_mgmt_wr_handler()
1470 if (io_req->io_cbfn) in csio_ln_mgmt_wr_handler()
1471 io_req->io_cbfn(hw, io_req); in csio_ln_mgmt_wr_handler()
1697 wr->cookie = io_req->fw_handle; in csio_ln_prep_ecwr()
1699 io_req->lnode->hwp, io_req->iq_idx)); in csio_ln_prep_ecwr()
1757 io_req, ret); in csio_ln_mgmt_submit_wr()
1817 io_req->fw_handle = (uintptr_t) (io_req); in csio_ln_mgmt_submit_req()
1818 io_req->eq_idx = mgmtm->eq_idx; in csio_ln_mgmt_submit_req()
[all …]
A Dcsio_hw.c4112 struct csio_ioreq *io_req; in csio_mgmt_tmo_handler() local
4119 io_req = (struct csio_ioreq *) tmp; in csio_mgmt_tmo_handler()
4120 io_req->tmo -= min_t(uint32_t, io_req->tmo, ECM_MIN_TMO); in csio_mgmt_tmo_handler()
4122 if (!io_req->tmo) { in csio_mgmt_tmo_handler()
4126 if (io_req->io_cbfn) { in csio_mgmt_tmo_handler()
4128 io_req->wr_status = -ETIMEDOUT; in csio_mgmt_tmo_handler()
4129 io_req->io_cbfn(mgmtm->hw, io_req); in csio_mgmt_tmo_handler()
4147 struct csio_ioreq *io_req; in csio_mgmtm_cleanup() local
4165 if (io_req->io_cbfn) { in csio_mgmtm_cleanup()
4167 io_req->wr_status = -ETIMEDOUT; in csio_mgmtm_cleanup()
[all …]
/linux/drivers/staging/rtl8712/
A Drtl871x_io.c101 struct io_req *pio_req; in r8712_alloc_io_queue()
111 (sizeof(struct io_req)) + 4, in r8712_alloc_io_queue()
118 pio_req = (struct io_req *)(pio_queue->free_ioreqs_buf); in r8712_alloc_io_queue()
/linux/include/linux/
A Ddm-io.h81 int dm_io(struct dm_io_request *io_req, unsigned num_regions,

Completed in 84 milliseconds

12