| /drivers/nvme/target/ |
| A D | io-cmd-file.c | 94 iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count); in nvmet_file_submit_bvec() 108 if (req->f.bvec != req->inline_bvec) { in nvmet_file_io_done() 110 kfree(req->f.bvec); in nvmet_file_io_done() 112 mempool_free(req->f.bvec, req->ns->bvec_pool); in nvmet_file_io_done() 142 bvec_set_page(&req->f.bvec[bv_cnt], sg_page(sg), sg->length, in nvmet_file_execute_io() 144 len += req->f.bvec[bv_cnt].bv_len; in nvmet_file_execute_io() 145 total_len += req->f.bvec[bv_cnt].bv_len; in nvmet_file_execute_io() 231 req->f.bvec = kmalloc_array(nr_bvec, sizeof(struct bio_vec), in nvmet_file_execute_rw() 234 req->f.bvec = req->inline_bvec; in nvmet_file_execute_rw() 236 if (unlikely(!req->f.bvec)) { in nvmet_file_execute_rw() [all …]
|
| A D | tcp.c | 617 struct bio_vec bvec; in nvmet_try_send_data_pdu() local 623 iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bvec, 1, left); in nvmet_try_send_data_pdu() 649 struct bio_vec bvec; in nvmet_try_send_data() local 657 bvec_set_page(&bvec, page, left, cmd->offset); in nvmet_try_send_data() 658 iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bvec, 1, left); in nvmet_try_send_data() 696 struct bio_vec bvec; in nvmet_try_send_response() local 706 bvec_set_virt(&bvec, (void *)cmd->rsp_pdu + cmd->offset, left); in nvmet_try_send_response() 707 iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bvec, 1, left); in nvmet_try_send_response() 726 struct bio_vec bvec; in nvmet_try_send_r2t() local 736 bvec_set_virt(&bvec, (void *)cmd->r2t_pdu + cmd->offset, left); in nvmet_try_send_r2t() [all …]
|
| A D | nvmet.h | 462 struct bio_vec *bvec; member
|
| /drivers/block/ |
| A D | n64cart.c | 89 struct bio_vec bvec; in n64cart_submit_bio() local 94 bio_for_each_segment(bvec, bio, iter) { in n64cart_submit_bio() 95 if (!n64cart_do_bvec(dev, &bvec, pos)) { in n64cart_submit_bio() 99 pos += bvec.bv_len; in n64cart_submit_bio()
|
| A D | zloop.c | 132 struct bio_vec *bvec; member 344 kfree(cmd->bvec); in zloop_put_cmd() 345 cmd->bvec = NULL; in zloop_put_cmd() 444 struct bio_vec *bvec; in zloop_rw() local 446 cmd->bvec = kmalloc_array(nr_bvec, sizeof(*cmd->bvec), GFP_NOIO); in zloop_rw() 447 if (!cmd->bvec) { in zloop_rw() 458 bvec = cmd->bvec; in zloop_rw() 460 *bvec = tmp; in zloop_rw() 461 bvec++; in zloop_rw() 463 iov_iter_bvec(&iter, rw, cmd->bvec, nr_bvec, blk_rq_bytes(rq)); in zloop_rw()
|
| A D | loop.c | 85 struct bio_vec *bvec; member 314 kfree(cmd->bvec); in lo_rw_aio_do_completion() 315 cmd->bvec = NULL; in lo_rw_aio_do_completion() 335 struct bio_vec *bvec; in lo_rw_aio() local 349 bvec = kmalloc_array(nr_bvec, sizeof(struct bio_vec), in lo_rw_aio() 351 if (!bvec) in lo_rw_aio() 353 cmd->bvec = bvec; in lo_rw_aio() 362 *bvec = tmp; in lo_rw_aio() 363 bvec++; in lo_rw_aio() 365 bvec = cmd->bvec; in lo_rw_aio() [all …]
|
| A D | ps3disk.c | 85 struct bio_vec bvec; in ps3disk_scatter_gather() local 87 rq_for_each_segment(bvec, req, iter) { in ps3disk_scatter_gather() 89 memcpy_from_bvec(dev->bounce_buf + offset, &bvec); in ps3disk_scatter_gather() 91 memcpy_to_bvec(&bvec, dev->bounce_buf + offset); in ps3disk_scatter_gather()
|
| A D | nbd.c | 741 struct bio_vec bvec; in nbd_send_cmd() local 743 bio_for_each_segment(bvec, bio, iter) { in nbd_send_cmd() 744 bool is_last = !next && bio_iter_last(bvec, iter); in nbd_send_cmd() 748 req, bvec.bv_len); in nbd_send_cmd() 749 iov_iter_bvec(&from, ITER_SOURCE, &bvec, 1, bvec.bv_len); in nbd_send_cmd() 935 struct bio_vec bvec; in nbd_handle_reply() local 938 rq_for_each_segment(bvec, req, iter) { in nbd_handle_reply() 939 iov_iter_bvec(&to, ITER_DEST, &bvec, 1, bvec.bv_len); in nbd_handle_reply() 958 req, bvec.bv_len); in nbd_handle_reply()
|
| A D | ps3vram.c | 538 struct bio_vec bvec; in ps3vram_do_bio() local 542 bio_for_each_segment(bvec, bio, iter) { in ps3vram_do_bio() 544 char *ptr = bvec_virt(&bvec); in ps3vram_do_bio() 545 size_t len = bvec.bv_len, retlen; in ps3vram_do_bio()
|
| /drivers/target/ |
| A D | target_core_file.c | 318 struct bio_vec *bvec; in fd_do_rw() local 323 bvec = kcalloc(sgl_nents, sizeof(struct bio_vec), GFP_KERNEL); in fd_do_rw() 324 if (!bvec) { in fd_do_rw() 334 iov_iter_bvec(&iter, is_write, bvec, sgl_nents, len); in fd_do_rw() 378 kfree(bvec); in fd_do_rw() 435 struct bio_vec *bvec; in fd_execute_write_same() local 458 bvec = kcalloc(nolb, sizeof(struct bio_vec), GFP_KERNEL); in fd_execute_write_same() 459 if (!bvec) in fd_execute_write_same() 463 bvec_set_page(&bvec[i], sg_page(&cmd->t_data_sg[0]), in fd_execute_write_same() 469 iov_iter_bvec(&iter, ITER_SOURCE, bvec, nolb, len); in fd_execute_write_same() [all …]
|
| /drivers/vhost/ |
| A D | vringh.c | 1034 struct bio_vec *bvec; member 1080 struct bio_vec *bvec = ivec->iov.bvec; in iotlb_translate() local 1082 bvec_set_page(&bvec[ret], pfn_to_page(pfn), io_len, in iotlb_translate() 1107 struct bio_vec bvec[IOTLB_IOV_STRIDE]; in copy_from_iotlb() member 1131 iov_iter_bvec(&iter, ITER_SOURCE, ivec.iov.bvec, ret, in copy_from_iotlb() 1153 struct bio_vec bvec[IOTLB_IOV_STRIDE]; in copy_to_iotlb() member 1177 iov_iter_bvec(&iter, ITER_DEST, ivec.iov.bvec, ret, in copy_to_iotlb() 1199 struct bio_vec bvec[1]; in getu16_iotlb() member 1218 __virtio16 *from = bvec_kmap_local(&ivec.iov.bvec[0]); in getu16_iotlb() 1235 struct bio_vec bvec; in putu16_iotlb() member [all …]
|
| /drivers/nvdimm/ |
| A D | pmem.c | 206 struct bio_vec bvec; in pmem_submit_bio() local 217 bio_for_each_segment(bvec, bio, iter) { in pmem_submit_bio() 219 rc = pmem_do_write(pmem, bvec.bv_page, bvec.bv_offset, in pmem_submit_bio() 220 iter.bi_sector, bvec.bv_len); in pmem_submit_bio() 222 rc = pmem_do_read(pmem, bvec.bv_page, bvec.bv_offset, in pmem_submit_bio() 223 iter.bi_sector, bvec.bv_len); in pmem_submit_bio()
|
| A D | btt.c | 1442 struct bio_vec bvec; in btt_submit_bio() local 1452 bio_for_each_segment(bvec, bio, iter) { in btt_submit_bio() 1453 unsigned int len = bvec.bv_len; in btt_submit_bio() 1463 err = btt_do_bvec(btt, bip, bvec.bv_page, len, bvec.bv_offset, in btt_submit_bio()
|
| /drivers/md/ |
| A D | dm-io.c | 219 struct bio_vec bvec = bvec_iter_bvec((struct bio_vec *)dp->context_ptr, in bio_get_page() local 222 *p = bvec.bv_page; in bio_get_page() 223 *len = bvec.bv_len; in bio_get_page() 224 *offset = bvec.bv_offset; in bio_get_page() 227 dp->context_bi.bi_sector = (sector_t)bvec.bv_len; in bio_get_page()
|
| A D | dm-flakey.c | 365 struct bio_vec bvec; in corrupt_bio_common() local 371 __bio_for_each_segment(bvec, bio, iter, start) { in corrupt_bio_common() 373 unsigned char *segment = bvec_kmap_local(&bvec); in corrupt_bio_common() 475 struct bio_vec bvec = bvec_iter_bvec(bio->bi_io_vec, iter); in clone_bio() local 476 unsigned this_step = min(bvec.bv_len, to_copy); in clone_bio() 477 void *map = bvec_kmap_local(&bvec); in clone_bio()
|
| /drivers/mtd/ |
| A D | mtd_blkdevs.c | 50 struct bio_vec bvec; in do_blktrans_request() local 76 rq_for_each_segment(bvec, req, iter) in do_blktrans_request() 77 flush_dcache_page(bvec.bv_page); in do_blktrans_request() 83 rq_for_each_segment(bvec, req, iter) in do_blktrans_request() 84 flush_dcache_page(bvec.bv_page); in do_blktrans_request()
|
| /drivers/s390/block/ |
| A D | dcssblk.c | 864 struct bio_vec bvec; in dcssblk_submit_bio() local 895 bio_for_each_segment(bvec, bio, iter) { in dcssblk_submit_bio() 896 page_addr = bvec_virt(&bvec); in dcssblk_submit_bio() 899 !IS_ALIGNED(bvec.bv_len, PAGE_SIZE))) in dcssblk_submit_bio() 903 memcpy(page_addr, __va(source_addr), bvec.bv_len); in dcssblk_submit_bio() 905 memcpy(__va(source_addr), page_addr, bvec.bv_len); in dcssblk_submit_bio() 906 bytes_done += bvec.bv_len; in dcssblk_submit_bio()
|
| /drivers/block/zram/ |
| A D | zram_drv.c | 195 static inline bool is_partial_io(struct bio_vec *bvec) in is_partial_io() argument 197 return bvec->bv_len != PAGE_SIZE; in is_partial_io() 201 static inline bool is_partial_io(struct bio_vec *bvec) in is_partial_io() argument 1781 memcpy_to_bvec(bvec, page_address(page) + offset); in zram_bvec_read_partial() 1789 if (is_partial_io(bvec)) in zram_bvec_read() 1790 return zram_bvec_read_partial(zram, bvec, index, offset); in zram_bvec_read() 1791 return zram_read_page(zram, bvec->bv_page, index, bio); in zram_bvec_read() 1929 memcpy_from_bvec(page_address(page) + offset, bvec); in zram_bvec_write_partial() 1939 if (is_partial_io(bvec)) in zram_bvec_write() 1940 return zram_bvec_write_partial(zram, bvec, index, offset, bio); in zram_bvec_write() [all …]
|
| /drivers/mtd/ubi/ |
| A D | block.c | 191 struct bio_vec bvec; in ubiblock_read() local 222 rq_for_each_segment(bvec, req, iter) in ubiblock_read() 223 flush_dcache_page(bvec.bv_page); in ubiblock_read()
|
| /drivers/block/drbd/ |
| A D | drbd_main.c | 1519 struct bio_vec bvec; in _drbd_send_page() local 1536 bvec_set_page(&bvec, page, len, offset); in _drbd_send_page() 1566 struct bio_vec bvec; in _drbd_send_bio() local 1570 bio_for_each_segment(bvec, bio, iter) { in _drbd_send_bio() 1574 bvec.bv_offset, bvec.bv_len, in _drbd_send_bio() 1575 bio_iter_last(bvec, iter) in _drbd_send_bio() 1585 struct bio_vec bvec; in _drbd_send_zc_bio() local 1589 bio_for_each_segment(bvec, bio, iter) { in _drbd_send_zc_bio() 1592 err = _drbd_send_page(peer_device, bvec.bv_page, in _drbd_send_zc_bio() 1593 bvec.bv_offset, bvec.bv_len, in _drbd_send_zc_bio() [all …]
|
| A D | drbd_worker.c | 319 struct bio_vec bvec; in drbd_csum_bio() local 326 bio_for_each_segment(bvec, bio, iter) { in drbd_csum_bio() 329 src = bvec_kmap_local(&bvec); in drbd_csum_bio() 330 crypto_shash_update(desc, src, bvec.bv_len); in drbd_csum_bio()
|
| /drivers/nvme/host/ |
| A D | tcp.c | 312 return req->iter.bvec->bv_page; in nvme_tcp_req_cur_page() 317 return req->iter.bvec->bv_offset + req->iter.iov_offset; in nvme_tcp_req_cur_offset() 1142 struct bio_vec bvec; in nvme_tcp_try_send_data() local 1161 bvec_set_page(&bvec, page, len, offset); in nvme_tcp_try_send_data() 1162 iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bvec, 1, len); in nvme_tcp_try_send_data() 1202 struct bio_vec bvec; in nvme_tcp_try_send_cmd_pdu() local 1217 bvec_set_virt(&bvec, (void *)pdu + req->offset, len); in nvme_tcp_try_send_cmd_pdu() 1218 iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bvec, 1, len); in nvme_tcp_try_send_cmd_pdu() 1243 struct bio_vec bvec; in nvme_tcp_try_send_data_pdu() local 1255 bvec_set_virt(&bvec, (void *)pdu + req->offset, len); in nvme_tcp_try_send_data_pdu() [all …]
|
| /drivers/block/null_blk/ |
| A D | main.c | 1283 struct bio_vec bvec; in null_handle_data_transfer() local 1286 rq_for_each_segment(bvec, rq, iter) { in null_handle_data_transfer() 1287 len = bvec.bv_len; in null_handle_data_transfer() 1290 err = null_transfer(nullb, bvec.bv_page, len, bvec.bv_offset, in null_handle_data_transfer()
|
| /drivers/infiniband/sw/siw/ |
| A D | siw_qp_tx.c | 326 struct bio_vec bvec; in siw_tcp_sendpages() local 342 bvec_set_page(&bvec, page[i], bytes, offset); in siw_tcp_sendpages() 343 iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bvec, 1, bytes); in siw_tcp_sendpages()
|
| /drivers/target/iscsi/ |
| A D | iscsi_target_util.c | 1089 struct bio_vec bvec; in iscsit_fe_sendpage_sg() local 1134 bvec_set_page(&bvec, sg_page(sg), sub_len, sg->offset + offset); in iscsit_fe_sendpage_sg() 1135 iov_iter_bvec(&msghdr.msg_iter, ITER_SOURCE, &bvec, 1, sub_len); in iscsit_fe_sendpage_sg()
|