/linux-6.3-rc2/tools/io_uring/ |
A D | liburing.h | 112 memset(sqe, 0, sizeof(*sqe)); in io_uring_prep_rw() 113 sqe->opcode = op; in io_uring_prep_rw() 114 sqe->fd = fd; in io_uring_prep_rw() 115 sqe->off = offset; in io_uring_prep_rw() 117 sqe->len = len; in io_uring_prep_rw() 151 memset(sqe, 0, sizeof(*sqe)); in io_uring_prep_poll_add() 153 sqe->fd = fd; in io_uring_prep_poll_add() 163 memset(sqe, 0, sizeof(*sqe)); in io_uring_prep_poll_remove() 171 memset(sqe, 0, sizeof(*sqe)); in io_uring_prep_fsync() 173 sqe->fd = fd; in io_uring_prep_fsync() [all …]
|
A D | io_uring-cp.c | 71 struct io_uring_sqe *sqe; in queue_prepped() local 73 sqe = io_uring_get_sqe(ring); in queue_prepped() 74 assert(sqe); in queue_prepped() 77 io_uring_prep_readv(sqe, infd, &data->iov, 1, data->offset); in queue_prepped() 79 io_uring_prep_writev(sqe, outfd, &data->iov, 1, data->offset); in queue_prepped() 81 io_uring_sqe_set_data(sqe, data); in queue_prepped() 86 struct io_uring_sqe *sqe; in queue_read() local 93 sqe = io_uring_get_sqe(ring); in queue_read() 94 if (!sqe) { in queue_read() 106 io_uring_prep_readv(sqe, infd, &data->iov, 1, offset); in queue_read() [all …]
|
A D | io_uring-bench.c | 151 sqe->opcode = IORING_OP_NOP; in init_io() 172 sqe->flags = IOSQE_FIXED_FILE; in init_io() 173 sqe->fd = f->fixed_fd; in init_io() 175 sqe->flags = 0; in init_io() 176 sqe->fd = f->real_fd; in init_io() 181 sqe->len = BS; in init_io() 182 sqe->buf_index = index; in init_io() 186 sqe->len = 1; in init_io() 187 sqe->buf_index = 0; in init_io() 189 sqe->ioprio = 0; in init_io() [all …]
|
/linux-6.3-rc2/io_uring/ |
A D | fs.c | 55 if (sqe->buf_index || sqe->splice_fd_in) in io_renameat_prep() 60 ren->old_dfd = READ_ONCE(sqe->fd); in io_renameat_prep() 63 ren->new_dfd = READ_ONCE(sqe->len); in io_renameat_prep() 109 if (sqe->off || sqe->len || sqe->buf_index || sqe->splice_fd_in) in io_unlinkat_prep() 114 un->dfd = READ_ONCE(sqe->fd); in io_unlinkat_prep() 159 if (sqe->off || sqe->rw_flags || sqe->buf_index || sqe->splice_fd_in) in io_mkdirat_prep() 164 mkd->dfd = READ_ONCE(sqe->fd); in io_mkdirat_prep() 165 mkd->mode = READ_ONCE(sqe->len); in io_mkdirat_prep() 203 if (sqe->len || sqe->rw_flags || sqe->buf_index || sqe->splice_fd_in) in io_symlinkat_prep() 208 sl->new_dfd = READ_ONCE(sqe->fd); in io_symlinkat_prep() [all …]
|
A D | sync.c | 29 if (unlikely(sqe->addr || sqe->buf_index || sqe->splice_fd_in)) in io_sfr_prep() 32 sync->off = READ_ONCE(sqe->off); in io_sfr_prep() 33 sync->len = READ_ONCE(sqe->len); in io_sfr_prep() 57 if (unlikely(sqe->addr || sqe->buf_index || sqe->splice_fd_in)) in io_fsync_prep() 60 sync->flags = READ_ONCE(sqe->fsync_flags); in io_fsync_prep() 64 sync->off = READ_ONCE(sqe->off); in io_fsync_prep() 65 sync->len = READ_ONCE(sqe->len); in io_fsync_prep() 89 if (sqe->buf_index || sqe->rw_flags || sqe->splice_fd_in) in io_fallocate_prep() 92 sync->off = READ_ONCE(sqe->off); in io_fallocate_prep() 93 sync->len = READ_ONCE(sqe->addr); in io_fallocate_prep() [all …]
|
A D | advise.c | 31 int io_madvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_madvise_prep() argument 36 if (sqe->buf_index || sqe->off || sqe->splice_fd_in) in io_madvise_prep() 39 ma->addr = READ_ONCE(sqe->addr); in io_madvise_prep() 40 ma->len = READ_ONCE(sqe->len); in io_madvise_prep() 41 ma->advice = READ_ONCE(sqe->fadvise_advice); in io_madvise_prep() 77 int io_fadvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_fadvise_prep() argument 81 if (sqe->buf_index || sqe->addr || sqe->splice_fd_in) in io_fadvise_prep() 84 fa->offset = READ_ONCE(sqe->off); in io_fadvise_prep() 85 fa->len = READ_ONCE(sqe->len); in io_fadvise_prep() 86 fa->advice = READ_ONCE(sqe->fadvise_advice); in io_fadvise_prep()
|
A D | openclose.c | 49 if (unlikely(sqe->buf_index)) in __io_openat_prep() 58 open->dfd = READ_ONCE(sqe->fd); in __io_openat_prep() 59 fname = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_openat_prep() 67 open->file_slot = READ_ONCE(sqe->file_index); in __io_openat_prep() 81 u64 mode = READ_ONCE(sqe->len); in io_openat_prep() 82 u64 flags = READ_ONCE(sqe->open_flags); in io_openat_prep() 85 return __io_openat_prep(req, sqe); in io_openat_prep() 96 len = READ_ONCE(sqe->len); in io_openat2_prep() 104 return __io_openat_prep(req, sqe); in io_openat2_prep() 205 if (sqe->off || sqe->addr || sqe->len || sqe->rw_flags || sqe->buf_index) in io_close_prep() [all …]
|
A D | xattr.c | 45 const struct io_uring_sqe *sqe) in __io_getxattr_prep() argument 56 name = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_getxattr_prep() 58 ix->ctx.size = READ_ONCE(sqe->len); in __io_getxattr_prep() 59 ix->ctx.flags = READ_ONCE(sqe->xattr_flags); in __io_getxattr_prep() 84 return __io_getxattr_prep(req, sqe); in io_fgetxattr_prep() 93 ret = __io_getxattr_prep(req, sqe); in io_getxattr_prep() 149 const struct io_uring_sqe *sqe) in __io_setxattr_prep() argument 162 ix->ctx.size = READ_ONCE(sqe->len); in __io_setxattr_prep() 163 ix->ctx.flags = READ_ONCE(sqe->xattr_flags); in __io_setxattr_prep() 186 ret = __io_setxattr_prep(req, sqe); in io_setxattr_prep() [all …]
|
A D | splice.c | 27 const struct io_uring_sqe *sqe) in __io_splice_prep() argument 32 sp->len = READ_ONCE(sqe->len); in __io_splice_prep() 33 sp->flags = READ_ONCE(sqe->splice_flags); in __io_splice_prep() 36 sp->splice_fd_in = READ_ONCE(sqe->splice_fd_in); in __io_splice_prep() 41 int io_tee_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_tee_prep() argument 43 if (READ_ONCE(sqe->splice_off_in) || READ_ONCE(sqe->off)) in io_tee_prep() 45 return __io_splice_prep(req, sqe); in io_tee_prep() 79 int io_splice_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_splice_prep() argument 83 sp->off_in = READ_ONCE(sqe->splice_off_in); in io_splice_prep() 84 sp->off_out = READ_ONCE(sqe->off); in io_splice_prep() [all …]
|
A D | statx.c | 23 int io_statx_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_statx_prep() argument 28 if (sqe->buf_index || sqe->splice_fd_in) in io_statx_prep() 33 sx->dfd = READ_ONCE(sqe->fd); in io_statx_prep() 34 sx->mask = READ_ONCE(sqe->len); in io_statx_prep() 35 path = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_statx_prep() 36 sx->buffer = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_statx_prep() 37 sx->flags = READ_ONCE(sqe->statx_flags); in io_statx_prep()
|
A D | epoll.c | 24 int io_epoll_ctl_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_epoll_ctl_prep() argument 32 if (sqe->buf_index || sqe->splice_fd_in) in io_epoll_ctl_prep() 35 epoll->epfd = READ_ONCE(sqe->fd); in io_epoll_ctl_prep() 36 epoll->op = READ_ONCE(sqe->len); in io_epoll_ctl_prep() 37 epoll->fd = READ_ONCE(sqe->off); in io_epoll_ctl_prep() 42 ev = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_epoll_ctl_prep()
|
A D | net.h | 31 int io_shutdown_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 36 int io_sendmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 43 int io_recvmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 49 int io_accept_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 52 int io_socket_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 56 int io_connect_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 61 int io_send_zc_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
A D | net.c | 88 if (unlikely(sqe->off || sqe->addr || sqe->rw_flags || in io_shutdown_prep() 89 sqe->buf_index || sqe->splice_fd_in)) in io_shutdown_prep() 259 if (READ_ONCE(sqe->__pad3[0])) in io_sendmsg_prep() 263 } else if (sqe->addr2 || sqe->file_index) { in io_sendmsg_prep() 268 sr->len = READ_ONCE(sqe->len); in io_sendmsg_prep() 562 if (unlikely(sqe->file_index || sqe->addr2)) in io_recvmsg_prep() 959 if (unlikely(READ_ONCE(sqe->__pad2[0]) || READ_ONCE(sqe->addr3))) in io_send_zc_prep() 999 if (unlikely(sqe->addr2 || sqe->file_index)) in io_send_zc_prep() 1269 if (sqe->len || sqe->buf_index) in io_accept_prep() 1360 if (sqe->addr || sqe->rw_flags || sqe->buf_index) in io_socket_prep() [all …]
|
A D | fs.h | 3 int io_renameat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 7 int io_unlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 11 int io_mkdirat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 15 int io_symlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 18 int io_linkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
A D | fdinfo.c | 89 struct io_uring_sqe *sqe; in __io_uring_show_fdinfo() local 95 sqe = &ctx->sq_sqes[sq_idx << sq_shift]; in __io_uring_show_fdinfo() 99 sq_idx, io_uring_get_opcode(sqe->opcode), sqe->fd, in __io_uring_show_fdinfo() 100 sqe->flags, (unsigned long long) sqe->off, in __io_uring_show_fdinfo() 101 (unsigned long long) sqe->addr, sqe->rw_flags, in __io_uring_show_fdinfo() 102 sqe->buf_index, sqe->user_data); in __io_uring_show_fdinfo() 104 u64 *sqeb = (void *) (sqe + 1); in __io_uring_show_fdinfo()
|
A D | xattr.h | 5 int io_fsetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 8 int io_setxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 11 int io_fgetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 14 int io_getxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
A D | kbuf.c | 273 int io_remove_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_remove_buffers_prep() argument 278 if (sqe->rw_flags || sqe->addr || sqe->len || sqe->off || in io_remove_buffers_prep() 279 sqe->splice_fd_in) in io_remove_buffers_prep() 282 tmp = READ_ONCE(sqe->fd); in io_remove_buffers_prep() 288 p->bgid = READ_ONCE(sqe->buf_group); in io_remove_buffers_prep() 322 if (sqe->rw_flags || sqe->splice_fd_in) in io_provide_buffers_prep() 325 tmp = READ_ONCE(sqe->fd); in io_provide_buffers_prep() 329 p->addr = READ_ONCE(sqe->addr); in io_provide_buffers_prep() 330 p->len = READ_ONCE(sqe->len); in io_provide_buffers_prep() 342 p->bgid = READ_ONCE(sqe->buf_group); in io_provide_buffers_prep() [all …]
|
/linux-6.3-rc2/drivers/infiniband/sw/siw/ |
A D | siw_qp.c | 278 wqe->sqe.flags = 0; in siw_qp_mpa_rts() 279 wqe->sqe.num_sge = 1; in siw_qp_mpa_rts() 287 wqe->sqe.rkey = 1; in siw_qp_mpa_rts() 288 wqe->sqe.raddr = 0; in siw_qp_mpa_rts() 874 rreq->id = sqe->id; in siw_read_to_orq() 886 struct siw_sqe *sqe; in siw_activate_tx_from_sq() local 891 if (!sqe) in siw_activate_tx_from_sq() 898 memcpy(&wqe->sqe, sqe, sizeof(*sqe)); in siw_activate_tx_from_sq() 914 wqe->sqe.sge[0].laddr = (uintptr_t)&wqe->sqe.sge[1]; in siw_activate_tx_from_sq() 1006 wqe->sqe.flags = 0; in siw_activate_tx() [all …]
|
A D | siw_qp_tx.c | 43 struct siw_sge *sge = &wqe->sqe.sge[0]; in siw_try_1seg() 53 memcpy(paddr, &wqe->sqe.sge[1], bytes); in siw_try_1seg() 139 cpu_to_be64(wqe->sqe.sge[0].laddr); in siw_qp_prepare_tx() 246 cpu_to_be64(wqe->sqe.raddr); in siw_qp_prepare_tx() 660 sge = &wqe->sqe.sge[c_tx->sge_idx]; in siw_tx_hdt() 817 wqe->sqe.num_sge = 1; in siw_qp_sq_proc_tx() 845 wqe->sqe.sge[0].laddr = in siw_qp_sq_proc_tx() 846 (u64)(uintptr_t)&wqe->sqe.sge[1]; in siw_qp_sq_proc_tx() 868 wqe->sqe.id); in siw_qp_sq_proc_tx() 975 mem->stag = sqe->rkey; in siw_fastreg_mr() [all …]
|
A D | siw_verbs.c | 647 void *kbuf = &sqe->sge[1]; in siw_copy_inline_sgl() 651 sqe->sge[0].lkey = 0; in siw_copy_inline_sgl() 682 struct siw_sqe sqe = {}; in siw_sq_flush_wr() local 689 sqe.opcode = SIW_OP_READ; in siw_sq_flush_wr() 714 sqe.id = wr->wr_id; in siw_sq_flush_wr() 821 if (sqe->flags) { in siw_post_send() 831 sqe->id = wr->wr_id; in siw_post_send() 857 sqe->num_sge = 1; in siw_post_send() 886 sqe->num_sge = 1; in siw_post_send() 906 sqe->num_sge = 1; in siw_post_send() [all …]
|
/linux-6.3-rc2/drivers/crypto/hisilicon/zip/ |
A D | zip_crypto.c | 287 sqe->dw9 = val; in hisi_zip_fill_buf_type() 296 sqe->dw9 = val; in hisi_zip_fill_req_type() 301 sqe->dw13 = req->req_id; in hisi_zip_fill_tag_v1() 306 sqe->dw26 = req->req_id; in hisi_zip_fill_tag_v2() 315 sqe->dw7 = val; in hisi_zip_fill_sqe_type() 325 ops->fill_addr(sqe, req); in hisi_zip_fill_sqe() 329 ops->fill_tag(sqe, req); in hisi_zip_fill_sqe() 388 return sqe->dw13; in hisi_zip_get_tag_v1() 393 return sqe->dw26; in hisi_zip_get_tag_v2() 403 return sqe->produced; in hisi_zip_get_dstlen() [all …]
|
/linux-6.3-rc2/tools/testing/selftests/net/ |
A D | io_uring_zerocopy_tx.c | 274 memset(sqe, 0, sizeof(*sqe)); in io_uring_prep_send() 276 sqe->fd = sockfd; in io_uring_prep_send() 277 sqe->addr = (unsigned long) buf; in io_uring_prep_send() 278 sqe->len = len; in io_uring_prep_send() 279 sqe->msg_flags = (__u32) flags; in io_uring_prep_send() 288 sqe->ioprio = zc_flags; in io_uring_prep_sendzc() 360 struct io_uring_sqe *sqe; in do_tx() local 396 sqe = io_uring_get_sqe(&ring); in do_tx() 401 sqe->user_data = NONZC_TAG; in do_tx() 408 sqe->buf_index = buf_idx; in do_tx() [all …]
|
/linux-6.3-rc2/drivers/net/ethernet/qlogic/qed/ |
A D | qed_nvmetcp_fw_funcs.c | 68 if (!task_params->sqe) in init_sqe() 71 memset(task_params->sqe, 0, sizeof(*task_params->sqe)); in init_sqe() 72 task_params->sqe->task_id = cpu_to_le16(task_params->itid); in init_sqe() 79 SET_FIELD(task_params->sqe->contlen_cdbsize, in init_sqe() 81 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_WQE_TYPE, in init_sqe() 94 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_NUM_SGES, num_sges); in init_sqe() 99 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_WQE_TYPE, in init_sqe() 101 SET_FIELD(task_params->sqe->contlen_cdbsize, in init_sqe() 106 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_WQE_TYPE, in init_sqe() 112 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_NUM_SGES, in init_sqe() [all …]
|
/linux-6.3-rc2/include/trace/events/ |
A D | io_uring.h | 503 TP_ARGS(sqe, req, error), 529 __entry->user_data = sqe->user_data; 530 __entry->opcode = sqe->opcode; 531 __entry->flags = sqe->flags; 532 __entry->ioprio = sqe->ioprio; 533 __entry->off = sqe->off; 534 __entry->addr = sqe->addr; 535 __entry->len = sqe->len; 537 __entry->buf_index = sqe->buf_index; 540 __entry->pad1 = sqe->__pad2[0]; [all …]
|
/linux-6.3-rc2/drivers/scsi/qedf/ |
A D | drv_fcoe_fw_funcs.c | 13 memset(task_params->sqe, 0, sizeof(*(task_params->sqe))); in init_common_sqe() 14 SET_FIELD(task_params->sqe->flags, FCOE_WQE_REQ_TYPE, in init_common_sqe() 16 task_params->sqe->task_id = task_params->itid; in init_common_sqe() 167 task_params->sqe->additional_info_union.burst_length = in init_initiator_midpath_unsolicited_fcoe_task() 169 SET_FIELD(task_params->sqe->flags, in init_initiator_midpath_unsolicited_fcoe_task() 171 SET_FIELD(task_params->sqe->flags, FCOE_WQE_SGL_MODE, in init_initiator_midpath_unsolicited_fcoe_task() 193 task_params->sqe->additional_info_union.seq_rec_updated_offset = in init_initiator_sequence_recovery_fcoe_task()
|