Lines Matching refs:iov

83 void vringh_kiov_advance(struct vringh_kiov *iov, size_t len)  in vringh_kiov_advance()  argument
85 while (len && iov->i < iov->used) { in vringh_kiov_advance()
86 size_t partlen = min(iov->iov[iov->i].iov_len, len); in vringh_kiov_advance()
88 iov->consumed += partlen; in vringh_kiov_advance()
89 iov->iov[iov->i].iov_len -= partlen; in vringh_kiov_advance()
90 iov->iov[iov->i].iov_base += partlen; in vringh_kiov_advance()
92 if (!iov->iov[iov->i].iov_len) { in vringh_kiov_advance()
94 iov->iov[iov->i].iov_len = iov->consumed; in vringh_kiov_advance()
95 iov->iov[iov->i].iov_base -= iov->consumed; in vringh_kiov_advance()
97 iov->consumed = 0; in vringh_kiov_advance()
98 iov->i++; in vringh_kiov_advance()
108 struct vringh_kiov *iov, in vringh_iov_xfer() argument
116 while (len && iov->i < iov->used) { in vringh_iov_xfer()
119 partlen = min(iov->iov[iov->i].iov_len, len); in vringh_iov_xfer()
120 err = xfer(vrh, iov->iov[iov->i].iov_base, ptr, partlen); in vringh_iov_xfer()
126 iov->consumed += partlen; in vringh_iov_xfer()
127 iov->iov[iov->i].iov_len -= partlen; in vringh_iov_xfer()
128 iov->iov[iov->i].iov_base += partlen; in vringh_iov_xfer()
130 if (!iov->iov[iov->i].iov_len) { in vringh_iov_xfer()
132 iov->iov[iov->i].iov_len = iov->consumed; in vringh_iov_xfer()
133 iov->iov[iov->i].iov_base -= iov->consumed; in vringh_iov_xfer()
135 iov->consumed = 0; in vringh_iov_xfer()
136 iov->i++; in vringh_iov_xfer()
218 static int resize_iovec(struct vringh_kiov *iov, gfp_t gfp) in resize_iovec() argument
221 unsigned int flag, new_num = (iov->max_num & ~VRINGH_IOV_ALLOCATED) * 2; in resize_iovec()
226 flag = (iov->max_num & VRINGH_IOV_ALLOCATED); in resize_iovec()
228 new = krealloc_array(iov->iov, new_num, sizeof(*new), gfp); in resize_iovec()
232 memcpy(new, iov->iov, in resize_iovec()
233 iov->max_num * sizeof(struct iovec)); in resize_iovec()
239 iov->iov = new; in resize_iovec()
240 iov->max_num = (new_num | flag); in resize_iovec()
325 struct vringh_kiov *iov; in __vringh_iov() local
373 iov = wiov; in __vringh_iov()
375 iov = riov; in __vringh_iov()
384 if (!iov) { in __vringh_iov()
402 if (unlikely(iov->used == (iov->max_num & ~VRINGH_IOV_ALLOCATED))) { in __vringh_iov()
403 err = resize_iovec(iov, gfp); in __vringh_iov()
408 iov->iov[iov->used].iov_base = addr; in __vringh_iov()
409 iov->iov[iov->used].iov_len = len; in __vringh_iov()
410 iov->used++; in __vringh_iov()
722 BUILD_BUG_ON(offsetof(struct vringh_kiov, iov) != in vringh_getdesc_user()
723 offsetof(struct vringh_iov, iov)); in vringh_getdesc_user()
1035 } iov; member
1074 struct iovec *iovec = ivec->iov.iovec; in iotlb_translate()
1080 struct bio_vec *bvec = ivec->iov.bvec; in iotlb_translate()
1108 } iov; in copy_from_iotlb() local
1111 ivec.iov.iovec = iov.iovec; in copy_from_iotlb()
1128 iov_iter_init(&iter, ITER_SOURCE, ivec.iov.iovec, ret, in copy_from_iotlb()
1131 iov_iter_bvec(&iter, ITER_SOURCE, ivec.iov.bvec, ret, in copy_from_iotlb()
1154 } iov; in copy_to_iotlb() local
1157 ivec.iov.iovec = iov.iovec; in copy_to_iotlb()
1174 iov_iter_init(&iter, ITER_DEST, ivec.iov.iovec, ret, in copy_to_iotlb()
1177 iov_iter_bvec(&iter, ITER_DEST, ivec.iov.bvec, ret, in copy_to_iotlb()
1200 } iov; in getu16_iotlb() local
1204 ivec.iov.iovec = iov.iovec; in getu16_iotlb()
1214 ret = __get_user(tmp, (__virtio16 __user *)ivec.iov.iovec[0].iov_base); in getu16_iotlb()
1218 __virtio16 *from = bvec_kmap_local(&ivec.iov.bvec[0]); in getu16_iotlb()
1236 } iov; in putu16_iotlb() local
1240 ivec.iov.iovec = &iov.iovec; in putu16_iotlb()
1252 ret = __put_user(tmp, (__virtio16 __user *)ivec.iov.iovec[0].iov_base); in putu16_iotlb()
1256 __virtio16 *to = bvec_kmap_local(&ivec.iov.bvec[0]); in putu16_iotlb()