| /drivers/net/ethernet/intel/idpf/ |
| A D | idpf_virtchnl_ptp.c | 30 .send_buf.iov_len = sizeof(send_ptp_caps_msg), in idpf_ptp_get_caps() 162 .send_buf.iov_len = sizeof(get_dev_clk_time_msg), in idpf_ptp_get_dev_clk_time() 164 .recv_buf.iov_len = sizeof(get_dev_clk_time_msg), in idpf_ptp_get_dev_clk_time() 199 .send_buf.iov_len = sizeof(cross_time_msg), in idpf_ptp_get_cross_time() 201 .recv_buf.iov_len = sizeof(cross_time_msg), in idpf_ptp_get_cross_time() 235 .send_buf.iov_len = sizeof(set_dev_clk_time_msg), in idpf_ptp_set_dev_clk_time() 237 .recv_buf.iov_len = sizeof(set_dev_clk_time_msg), in idpf_ptp_set_dev_clk_time() 268 .send_buf.iov_len = sizeof(adj_dev_clk_time_msg), in idpf_ptp_adj_dev_clk_time() 337 .send_buf.iov_len = sizeof(send_tx_tstamp_caps), in idpf_ptp_get_vport_tstamps_caps() 338 .recv_buf.iov_len = IDPF_CTLQ_MAX_BUF_LEN, in idpf_ptp_get_vport_tstamps_caps() [all …]
|
| A D | idpf_virtchnl.c | 316 xn->reply.iov_len = 0; in idpf_vc_xn_release_bufs() 1425 xn_params.send_buf.iov_len = buf_size; in idpf_send_create_vport_msg() 1667 xn_params.send_buf.iov_len = buf_sz; in idpf_send_config_tx_queues_msg() 1819 xn_params.send_buf.iov_len = buf_sz; in idpf_send_config_rx_queues_msg() 1965 xn_params.send_buf.iov_len = buf_sz; in idpf_send_ena_dis_queues_msg() 2088 xn_params.send_buf.iov_len = buf_sz; in idpf_send_map_unmap_queue_vector_msg() 2212 xn_params.send_buf.iov_len = buf_size; in idpf_send_delete_queues_msg() 2373 xn_params.send_buf.iov_len = buf_size; in idpf_send_dealloc_vectors_msg() 2502 xn_params.send_buf.iov_len = buf_size; in idpf_send_get_set_rss_lut_msg() 2575 xn_params.send_buf.iov_len = buf_size; in idpf_send_get_set_rss_key_msg() [all …]
|
| /drivers/usb/usbip/ |
| A D | stub_tx.c | 218 iov[iovnum].iov_len = sizeof(pdu_header); in stub_send_ret_submit() 232 iov[iovnum].iov_len = in stub_send_ret_submit() 254 iov[iovnum].iov_len = size; in stub_send_ret_submit() 261 iov[iovnum].iov_len = urb->actual_length; in stub_send_ret_submit() 280 iov[iovnum].iov_len = in stub_send_ret_submit() 311 iov[iovnum].iov_len = len; in stub_send_ret_submit() 388 iov[0].iov_len = sizeof(pdu_header); in stub_send_ret_unlink()
|
| A D | vhci_tx.c | 100 iov[iovnum].iov_len = sizeof(pdu_header); in vhci_send_cmd_submit() 110 iov[iovnum].iov_len = sg->length; in vhci_send_cmd_submit() 115 iov[iovnum].iov_len = in vhci_send_cmd_submit() 134 iov[iovnum].iov_len = len; in vhci_send_cmd_submit() 216 iov.iov_len = sizeof(pdu_header); in vhci_send_cmd_unlink()
|
| A D | vudc_tx.c | 57 iov[0].iov_len = sizeof(pdu_header); in v_send_ret_unlink() 115 iov[iovnum].iov_len = sizeof(pdu_header); in v_send_ret_submit() 123 iov[iovnum].iov_len = urb->actual_length; in v_send_ret_submit() 134 iov[iovnum].iov_len = in v_send_ret_submit() 161 iov[iovnum].iov_len = len; in v_send_ret_submit()
|
| /drivers/net/ethernet/google/gve/ |
| A D | gve_tx.c | 106 iov[0].iov_len = bytes; in gve_tx_alloc_fifo() 115 iov[0].iov_len -= overflow; in gve_tx_alloc_fifo() 117 iov[1].iov_len = overflow; in gve_tx_alloc_fifo() 151 info->iov[i].iov_len = 0; in gve_tx_clear_buffer_state() 552 u64 iov_offset, u64 iov_len) in gve_dma_sync_for_device() argument 621 info->iov[i].iov_len, in gve_tx_add_skb_copy() 626 info->iov[i].iov_len); in gve_tx_add_skb_copy() 629 info->iov[i].iov_len); in gve_tx_add_skb_copy() 804 info->iov[iovi].iov_len, in gve_tx_fill_xdp() 809 info->iov[iovi].iov_len, in gve_tx_fill_xdp() [all …]
|
| /drivers/xen/xenbus/ |
| A D | xenbus_xs.c | 298 vec->iov_len = msg->len; in xenbus_dev_request_and_reply() 342 msg.len += iovec[i].iov_len; in xs_talkv() 377 iovec.iov_len = strlen(string) + 1; in xs_single() 505 iovec[0].iov_len = strlen(path) + 1; in xenbus_write() 507 iovec[1].iov_len = strlen(string); in xenbus_write() 660 iov[0].iov_len = strlen(path) + 1; in xs_watch() 662 iov[1].iov_len = strlen(token) + 1; in xs_watch() 673 iov[0].iov_len = strlen(path) + 1; in xs_unwatch() 675 iov[1].iov_len = strlen(token) + 1; in xs_unwatch()
|
| /drivers/infiniband/sw/siw/ |
| A D | siw_qp.c | 403 iov[0].iov_len = sizeof(*term); in siw_send_terminate() 467 iov[1].iov_len = sizeof(*rreq); in siw_send_terminate() 478 iov[1].iov_len = in siw_send_terminate() 481 iov[1].iov_len = in siw_send_terminate() 496 iov[1].iov_len = in siw_send_terminate() 499 iov[1].iov_len = sizeof(struct iwarp_send); in siw_send_terminate() 504 term->ctrl.mpa_len = cpu_to_be16(iov[1].iov_len); in siw_send_terminate() 538 iov[2].iov_len = sizeof(crc); in siw_send_terminate() 543 iov[1].iov_len = sizeof(crc); in siw_send_terminate() 562 iov[0].iov_base, iov[0].iov_len); in siw_send_terminate() [all …]
|
| A D | siw_qp_tx.c | 299 .iov_len = c_tx->ctrl_len - c_tx->ctrl_sent }; in siw_tx_ctrl() 301 int rv = siw_sendmsg(s, flags, &iov, 1, iov.iov_len); in siw_tx_ctrl() 461 iov[0].iov_len = hdr_len = in siw_tx_hdt() 487 iov[seg].iov_len = sge_len; in siw_tx_hdt() 520 iov[seg].iov_len = plen; in siw_tx_hdt() 572 iov[seg].iov_len = trl_len = MAX_TRAILER - (4 - c_tx->pad); in siw_tx_hdt() 575 iov[seg].iov_len = trl_len = MAX_TRAILER - c_tx->ctrl_sent; in siw_tx_hdt()
|
| /drivers/vhost/ |
| A D | vringh.c | 89 iov->iov[iov->i].iov_len -= partlen; in vringh_kiov_advance() 92 if (!iov->iov[iov->i].iov_len) { in vringh_kiov_advance() 94 iov->iov[iov->i].iov_len = iov->consumed; in vringh_kiov_advance() 119 partlen = min(iov->iov[iov->i].iov_len, len); in vringh_iov_xfer() 127 iov->iov[iov->i].iov_len -= partlen; in vringh_iov_xfer() 130 if (!iov->iov[iov->i].iov_len) { in vringh_iov_xfer() 132 iov->iov[iov->i].iov_len = iov->consumed; in vringh_iov_xfer() 409 iov->iov[iov->used].iov_len = len; in __vringh_iov() 733 BUILD_BUG_ON(offsetof(struct iovec, iov_len) != in vringh_getdesc_user() 734 offsetof(struct kvec, iov_len)); in vringh_getdesc_user() [all …]
|
| A D | vsock.c | 111 size_t iov_len, payload_len; in vhost_transport_do_send_pkt() local 152 iov_len = iov_length(&vq->iov[out], in); in vhost_transport_do_send_pkt() 153 if (iov_len < sizeof(*hdr)) { in vhost_transport_do_send_pkt() 155 vq_err(vq, "Buffer len [%zu] too small\n", iov_len); in vhost_transport_do_send_pkt() 159 iov_iter_init(&iov_iter, ITER_DEST, &vq->iov[out], in, iov_len); in vhost_transport_do_send_pkt() 167 if (payload_len > iov_len - sizeof(*hdr)) { in vhost_transport_do_send_pkt() 168 payload_len = iov_len - sizeof(*hdr); in vhost_transport_do_send_pkt()
|
| /drivers/target/iscsi/ |
| A D | iscsi_target.c | 508 iov[niov++].iov_len = ISCSI_HDR_LEN; in iscsit_xmit_nondatain_pdu() 515 iov[0].iov_len += ISCSI_CRC_LEN; in iscsit_xmit_nondatain_pdu() 531 iov[niov++].iov_len = padding; in iscsit_xmit_nondatain_pdu() 585 iov[0].iov_len += ISCSI_CRC_LEN; in iscsit_xmit_datain_pdu() 915 iov[i].iov_len = cur_len; in iscsit_map_iovec() 1866 iov[niov++].iov_len = padding; in iscsit_handle_nop_out() 2266 iov[niov++].iov_len = rx_size; in iscsit_handle_text_cmd() 4030 iov.iov_len = ISCSI_HDR_LEN; in iscsit_get_rx_pdu() 4040 iov.iov_len = hdr->hlength * 4; in iscsit_get_rx_pdu() 4051 if (ret != iov.iov_len) { in iscsit_get_rx_pdu() [all …]
|
| /drivers/mtd/ |
| A D | mtdconcat.c | 197 total_len += vecs[i].iov_len; in concat_writev() 226 if (size <= vecs_copy[entry_high].iov_len) in concat_writev() 228 size -= vecs_copy[entry_high++].iov_len; in concat_writev() 231 old_iov_len = vecs_copy[entry_high].iov_len; in concat_writev() 232 vecs_copy[entry_high].iov_len = size; in concat_writev() 237 vecs_copy[entry_high].iov_len = old_iov_len - size; in concat_writev()
|
| /drivers/nvme/target/ |
| A D | tcp.c | 369 bvec_set_page(iov, sg_page(sg), iov_len, in nvmet_tcp_build_pdu_iovec() 372 length -= iov_len; in nvmet_tcp_build_pdu_iovec() 758 .iov_len = left in nvmet_try_send_ddgst() 906 iov.iov_len = sizeof(*icresp); in nvmet_tcp_handle_icreq() 1154 iov.iov_len = queue->left; in nvmet_tcp_try_recv_pdu() 1160 iov.iov_len, msg.msg_flags); in nvmet_tcp_try_recv_pdu() 1258 .iov_len = queue->left in nvmet_tcp_try_recv_ddgst() 1266 iov.iov_len, msg.msg_flags); in nvmet_tcp_try_recv_ddgst() 1719 .iov_len = sizeof(struct nvme_tcp_hdr), in nvmet_tcp_try_peek_pdu() 1732 iov.iov_len, msg.msg_flags); in nvmet_tcp_try_peek_pdu() [all …]
|
| /drivers/mtd/lpddr/ |
| A D | lpddr_cmds.c | 438 if (n > vec->iov_len - vec_seek) in do_write_buffer() 439 n = vec->iov_len - vec_seek; in do_write_buffer() 458 if (vec_seek == vec->iov_len) { in do_write_buffer() 636 vec.iov_len = len; in lpddr_write_buffers() 654 len += vecs[i].iov_len; in lpddr_writev()
|
| /drivers/xen/ |
| A D | pvcalls-back.c | 132 vec[0].iov_len = wanted; in pvcalls_conn_back_read() 136 vec[0].iov_len = array_size - masked_prod; in pvcalls_conn_back_read() 138 vec[1].iov_len = wanted - vec[0].iov_len; in pvcalls_conn_back_read() 193 vec[0].iov_len = size; in pvcalls_conn_back_write() 197 vec[0].iov_len = array_size - pvcalls_mask(cons, array_size); in pvcalls_conn_back_write() 199 vec[1].iov_len = size - vec[0].iov_len; in pvcalls_conn_back_write()
|
| /drivers/hv/ |
| A D | channel.c | 1046 bufferlist[0].iov_len = sizeof(struct vmpacket_descriptor); in vmbus_sendpacket_getid() 1048 bufferlist[1].iov_len = bufferlen; in vmbus_sendpacket_getid() 1050 bufferlist[2].iov_len = (packetlen_aligned - packetlen); in vmbus_sendpacket_getid() 1108 bufferlist[0].iov_len = desc_size; in vmbus_sendpacket_mpb_desc() 1110 bufferlist[1].iov_len = bufferlen; in vmbus_sendpacket_mpb_desc() 1112 bufferlist[2].iov_len = (packetlen_aligned - packetlen); in vmbus_sendpacket_mpb_desc()
|
| A D | ring_buffer.c | 295 totalbytes_towrite += kv_list[i].iov_len; in hv_ringbuffer_write() 329 kv_list[i].iov_len); in hv_ringbuffer_write()
|
| /drivers/infiniband/hw/hfi1/ |
| A D | user_sdma.c | 297 if (iovec[idx].iov_len < sizeof(info) + sizeof(req->hdr)) { in hfi1_user_sdma_process_request() 302 iovec[idx].iov_len, sizeof(info) + sizeof(req->hdr)); in hfi1_user_sdma_process_request() 464 if (req->iovs[i].iov.iov_len == 0) { in hfi1_user_sdma_process_request() 468 req->data_len += req->iovs[i].iov.iov_len; in hfi1_user_sdma_process_request() 483 u16 ntids = iovec[idx].iov_len / sizeof(*req->tids); in hfi1_user_sdma_process_request() 739 if (READ_ONCE(iovec->offset) == iovec->iov.iov_len) { in user_sdma_send_pkts()
|
| /drivers/soc/qcom/ |
| A D | qmi_interface.c | 534 iv.iov_len = qmi->recv_buf_size; in qmi_data_ready_work() 539 iv.iov_len, MSG_DONTWAIT); in qmi_data_ready_work() 758 iv.iov_len = len; in qmi_send_message()
|
| /drivers/fsi/ |
| A D | fsi-sbefifo.c | 663 ffdc_iov.iov_len = SBEFIFO_MAX_FFDC_SIZE; in sbefifo_collect_async_ffdc() 760 resp_iov.iov_len = rbytes; in sbefifo_submit() 844 resp_iov.iov_len = len; in sbefifo_user_read()
|
| /drivers/block/rnbd/ |
| A D | rnbd-clt.c | 452 .iov_len = sizeof(msg) in send_msg_close() 538 .iov_len = sizeof(msg) in send_msg_open() 584 .iov_len = sizeof(msg) in send_msg_sess_info() 1023 .iov_len = sizeof(msg) in rnbd_client_xfer_request() 1541 .iov_len = sizeof(msg) in rnbd_clt_map_device()
|
| /drivers/media/dvb-core/ |
| A D | dvb_net.c | 64 c = crc32_be( c, iov[j].iov_base, iov[j].iov_len ); in iov_crc32() 660 hexdump(iov[0].iov_base, iov[0].iov_len); in dvb_net_ule_check_crc() 661 hexdump(iov[1].iov_base, iov[1].iov_len); in dvb_net_ule_check_crc() 662 hexdump(iov[2].iov_base, iov[2].iov_len); in dvb_net_ule_check_crc()
|
| /drivers/s390/char/ |
| A D | zcore.c | 105 kvec.iov_len = count; in memcpy_hsa_kernel()
|
| /drivers/atm/ |
| A D | nicstar.c | 2113 iov->iov_len = ns_rsqe_cellcount(rsqe) * 48; in dequeue_rx() 2114 iovb->len += iov->iov_len; in dequeue_rx() 2149 (skb->data + iov->iov_len - 6); in dequeue_rx() 2294 iov->iov_len); in dequeue_rx() 2295 skb_put(hb, iov->iov_len); in dequeue_rx() 2296 remaining = len - iov->iov_len; in dequeue_rx() 2305 min_t(int, remaining, iov->iov_len); in dequeue_rx()
|