| /drivers/block/ |
| A D | n64cart.c | 69 (bv->bv_len & (MIN_ALIGNMENT - 1))); in n64cart_do_bvec() 79 n64cart_write_reg(PI_WRITE_REG, bv->bv_len - 1); in n64cart_do_bvec() 83 dma_unmap_page(dev, dma_addr, bv->bv_len, DMA_FROM_DEVICE); in n64cart_do_bvec() 99 pos += bvec.bv_len; in n64cart_submit_bio()
|
| A D | brd.c | 118 bv.bv_len = min_t(u32, bv.bv_len, PAGE_SIZE - offset); in brd_rw_bvec() 130 memcpy_to_page(page, offset, kaddr, bv.bv_len); in brd_rw_bvec() 133 memcpy_from_page(kaddr, page, offset, bv.bv_len); in brd_rw_bvec() 135 memset(kaddr, 0, bv.bv_len); in brd_rw_bvec() 140 bio_advance_iter_single(bio, &bio->bi_iter, bv.bv_len); in brd_rw_bvec()
|
| A D | nbd.c | 748 req, bvec.bv_len); in nbd_send_cmd() 749 iov_iter_bvec(&from, ITER_SOURCE, &bvec, 1, bvec.bv_len); in nbd_send_cmd() 939 iov_iter_bvec(&to, ITER_DEST, &bvec, 1, bvec.bv_len); in nbd_handle_reply() 958 req, bvec.bv_len); in nbd_handle_reply()
|
| /drivers/md/ |
| A D | dm-ebs-target.c | 71 unsigned int bv_len = bv->bv_len; in __ebs_rw_bvec() local 76 if (unlikely(!bv->bv_page || !bv_len)) in __ebs_rw_bvec() 82 while (bv_len) { in __ebs_rw_bvec() 83 cur_len = min(dm_bufio_get_block_size(ec->bufio) - buf_off, bv_len); in __ebs_rw_bvec() 86 if (op == REQ_OP_READ || buf_off || bv_len < dm_bufio_get_block_size(ec->bufio)) in __ebs_rw_bvec() 113 bv_len -= cur_len; in __ebs_rw_bvec()
|
| A D | dm-io-rewind.c | 27 while (idx >= 0 && bytes && bytes > bv[idx].bv_len) { in dm_bvec_iter_rewind() 28 bytes -= bv[idx].bv_len; in dm_bvec_iter_rewind() 41 iter->bi_bvec_done = bv[idx].bv_len - bytes; in dm_bvec_iter_rewind()
|
| A D | dm-log-writes.c | 367 block->vecs[i].bv_len, 0); in log_one_block() 368 if (ret != block->vecs[i].bv_len) { in log_one_block() 380 block->vecs[i].bv_len, 0); in log_one_block() 381 if (ret != block->vecs[i].bv_len) { in log_one_block() 387 sector += block->vecs[i].bv_len >> SECTOR_SHIFT; in log_one_block() 750 block->vecs[i].bv_len = bv.bv_len; in log_writes_map()
|
| A D | dm-io.c | 223 *len = bvec.bv_len; in bio_get_page() 227 dp->context_bi.bi_sector = (sector_t)bvec.bv_len; in bio_get_page()
|
| A D | dm-integrity.c | 1763 } while (pos < bv.bv_len); in integrity_recheck() 1867 if (unlikely(pos < bv_copy.bv_len)) { in integrity_metadata() 1869 bv_copy.bv_len -= pos; in integrity_metadata() 1893 this_len = min(biv.bv_len, data_to_process); in integrity_metadata() 1933 bv.bv_offset, bv.bv_len, ic->sectors_per_block); in dm_integrity_check_limits() 2039 if (unlikely(bv.bv_len >> SECTOR_SHIFT > n_sectors)) in __journal_read_write() 2040 bv.bv_len = n_sectors << SECTOR_SHIFT; in __journal_read_write() 2041 n_sectors -= bv.bv_len >> SECTOR_SHIFT; in __journal_read_write() 2042 bio_advance_iter(bio, &bio->bi_iter, bv.bv_len); in __journal_read_write() 2097 unsigned int tag_now = min(biv.bv_len, tag_todo); in __journal_read_write() [all …]
|
| A D | dm-flakey.c | 476 unsigned this_step = min(bvec.bv_len, to_copy); in clone_bio()
|
| /drivers/s390/block/ |
| A D | dasd_fba.c | 458 if (bv.bv_len & (blksize - 1)) in dasd_fba_build_cp_regular() 461 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_fba_build_cp_regular() 462 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_fba_build_cp_regular() 463 cidaw += bv.bv_len / blksize; in dasd_fba_build_cp_regular() 505 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_fba_build_cp_regular() 509 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_build_cp_regular() 583 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_free_cp() 594 memcpy(dst, cda, bv.bv_len); in dasd_fba_free_cp()
|
| A D | dasd_diag.c | 537 if (bv.bv_len & (blksize - 1)) in dasd_diag_build_cp() 540 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_diag_build_cp() 557 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_diag_build_cp()
|
| A D | dcssblk.c | 899 !IS_ALIGNED(bvec.bv_len, PAGE_SIZE))) in dcssblk_submit_bio() 903 memcpy(page_addr, __va(source_addr), bvec.bv_len); in dcssblk_submit_bio() 905 memcpy(__va(source_addr), page_addr, bvec.bv_len); in dcssblk_submit_bio() 906 bytes_done += bvec.bv_len; in dcssblk_submit_bio()
|
| A D | dasd_eckd.c | 3311 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_eckd_ese_read() 3993 if (bv.bv_len & (blksize - 1)) in dasd_eckd_build_cp_cmd_single() 3996 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_eckd_build_cp_cmd_single() 3998 cidaw += bv.bv_len >> (block->s2b_shift + 9); in dasd_eckd_build_cp_cmd_single() 4071 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_eckd_build_cp_cmd_single() 4075 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_eckd_build_cp_cmd_single() 4232 seg_len = bv.bv_len; in dasd_eckd_build_cp_cmd_track() 4575 seg_len = bv.bv_len; in dasd_eckd_build_cp_tpm_track() 4609 dst, bv.bv_len); in dasd_eckd_build_cp_tpm_track() 4841 seg_len = bv.bv_len; in dasd_eckd_build_cp_raw() [all …]
|
| A D | scm_blk.c | 203 msb->blk_count += bv.bv_len >> 12; in scm_request_prepare()
|
| /drivers/md/bcache/ |
| A D | util.c | 244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, in bch_bio_map() 251 base += bv->bv_len; in bch_bio_map() 254 size -= bv->bv_len; in bch_bio_map()
|
| A D | debug.c | 137 cache_set_err_on(memcmp(p1, p2, bv.bv_len), in bch_data_verify() 145 bio_advance_iter(check, &citer, bv.bv_len); in bch_data_verify()
|
| /drivers/xen/ |
| A D | biomerge.c | 15 return bfn1 + PFN_DOWN(vec1->bv_offset + vec1->bv_len) == bfn2; in xen_biovec_phys_mergeable()
|
| /drivers/nvme/target/ |
| A D | io-cmd-file.c | 144 len += req->f.bvec[bv_cnt].bv_len; in nvmet_file_execute_io() 145 total_len += req->f.bvec[bv_cnt].bv_len; in nvmet_file_execute_io()
|
| /drivers/block/zram/ |
| A D | zram_drv.c | 197 return bvec->bv_len != PAGE_SIZE; in is_partial_io() 2349 bv.bv_len = min_t(u32, bv.bv_len, PAGE_SIZE - offset); in zram_bio_read() 2362 bio_advance_iter_single(bio, &iter, bv.bv_len); in zram_bio_read() 2380 bv.bv_len = min_t(u32, bv.bv_len, PAGE_SIZE - offset); in zram_bio_write() 2392 bio_advance_iter_single(bio, &iter, bv.bv_len); in zram_bio_write()
|
| /drivers/nvdimm/ |
| A D | pmem.c | 220 iter.bi_sector, bvec.bv_len); in pmem_submit_bio() 223 iter.bi_sector, bvec.bv_len); in pmem_submit_bio()
|
| /drivers/nvme/host/ |
| A D | apple.c | 499 iod->dma_len = bv->bv_len; in apple_nvme_setup_prp_simple() 502 if (bv->bv_len > first_prp_len) in apple_nvme_setup_prp_simple() 518 if (bv.bv_offset + bv.bv_len <= NVME_CTRL_PAGE_SIZE * 2) in apple_nvme_map_data()
|
| A D | pci.c | 949 bool prp_possible = prp1_offset + bv.bv_len <= NVME_CTRL_PAGE_SIZE * 2; in nvme_pci_setup_data_simple() 960 iod->total_len = bv.bv_len; in nvme_pci_setup_data_simple() 966 iod->cmd.common.dptr.sgl.length = cpu_to_le32(bv.bv_len); in nvme_pci_setup_data_simple() 973 if (bv.bv_len > first_prp_len) in nvme_pci_setup_data_simple() 1226 rq_integrity_vec(req).bv_len, in nvme_unmap_metadata()
|
| /drivers/scsi/ |
| A D | aha1542.c | 271 buf += bv.bv_len; in aha1542_free_cmd() 454 buf += bv.bv_len; in aha1542_queuecommand()
|
| /drivers/block/aoe/ |
| A D | aoecmd.c | 303 bv.bv_offset, bv.bv_len); in skb_fillup() 1034 skb_copy_bits(skb, soff, p, bv.bv_len); in bvcpy() 1036 soff += bv.bv_len; in bvcpy()
|
| /drivers/md/dm-vdo/ |
| A D | data-vio.c | 300 data_ptr += biovec.bv_len; in copy_to_bio() 531 data_ptr += biovec.bv_len; in copy_from_bio()
|