Home
last modified time | relevance | path

Searched refs:bv (Results 1 – 25 of 125) sorted by relevance

12345

/linux/drivers/gpu/drm/i915/gt/
A Dgen7_renderclear.c52 return bv->max_threads; in num_primitives()
62 bv->max_threads = 70; in batch_get_defaults()
65 bv->max_threads = 140; in batch_get_defaults()
68 bv->max_threads = 280; in batch_get_defaults()
77 bv->max_threads = 36; in batch_get_defaults()
86 bv->state_start = round_up(SZ_1K + num_primitives(bv) * 64, SZ_4K); in batch_get_defaults()
87 bv->surface_start = bv->state_start + SZ_4K; in batch_get_defaults()
88 bv->size = bv->surface_start + bv->surface_height * bv->surface_width; in batch_get_defaults()
179 gen7_fill_surface_state(state, bv->surface_start, bv); in gen7_fill_binding_table()
439 return bv.size; in gen7_setup_clear_gpr_bb()
[all …]
/linux/include/linux/
A Dbvec.h47 bv->bv_page = page; in bvec_set_page()
48 bv->bv_len = len; in bvec_set_page()
49 bv->bv_offset = offset; in bvec_set_page()
89 struct bio_vec bv; member
155 bytes -= bv[idx].bv_len; in bvec_iter_advance()
201 return &iter_all->bv; in bvec_init_iter_all()
207 struct bio_vec *bv = &iter_all->bv; in bvec_advance() local
210 bv->bv_page++; in bvec_advance()
211 bv->bv_offset = 0; in bvec_advance()
216 bv->bv_len = min_t(unsigned int, PAGE_SIZE - bv->bv_offset, in bvec_advance()
[all …]
/linux/drivers/md/bcache/
A Dutil.c234 struct bio_vec *bv = bio->bi_io_vec; in bch_bio_map() local
242 for (; size; bio->bi_vcnt++, bv++) { in bch_bio_map()
243 bv->bv_offset = 0; in bch_bio_map()
244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, in bch_bio_map()
251 base += bv->bv_len; in bch_bio_map()
254 size -= bv->bv_len; in bch_bio_map()
271 struct bio_vec *bv; in bch_bio_alloc_pages() local
277 for (i = 0, bv = bio->bi_io_vec; i < bio->bi_vcnt; bv++, i++) { in bch_bio_alloc_pages()
279 if (!bv->bv_page) { in bch_bio_alloc_pages()
280 while (--bv >= bio->bi_io_vec) in bch_bio_alloc_pages()
[all …]
A Ddebug.c112 struct bio_vec bv, cbv; in bch_data_verify() local
130 bio_for_each_segment(bv, bio, iter) { in bch_data_verify()
131 void *p1 = bvec_kmap_local(&bv); in bch_data_verify()
137 cache_set_err_on(memcmp(p1, p2, bv.bv_len), in bch_data_verify()
145 bio_advance_iter(check, &citer, bv.bv_len); in bch_data_verify()
/linux/arch/parisc/kernel/
A Dentry.S161 bv,n 0(%r3)
772 bv %r0(%r2)
781 bv %r0(%r2)
940 bv %r0(%r20)
969 bv %r0(%r1)
1699 bv %r0(%r2)
1926 bv %r0(%rp)
1994 bv,n (%r1)
2090 bv,n (%r1)
2135 bv %r0(%rp)
[all …]
A Dreal2.S79 bv 0(%r31)
101 bv 0(%rp)
121 bv 0(%r2)
136 bv 0(%r2)
177 bv 0(%r2)
214 bv 0(%r2)
260 bv 0(%r31)
273 bv 0(%rp)
288 bv %r0(%r2)
A Dhpmc.S154 bv (r3) /* call pdce_proc */
169 bv (%r3) /* call pdce_proc */
199 bv (%r3) /* call pdce_proc */
221 bv (%r5)
269 bv (%r3) /* call pdce_proc */
A Dhead.S101 bv,n (%r1)
208 bv,n (%rp)
250 bv (%rp)
263 bv (%r3)
277 bv (%r3)
396 bv (%rp)
/linux/drivers/net/ethernet/netronome/nfp/bpf/
A Dmain.c66 struct nfp_bpf_vnic *bv; in nfp_bpf_vnic_alloc() local
79 bv = kzalloc(sizeof(*bv), GFP_KERNEL); in nfp_bpf_vnic_alloc()
80 if (!bv) in nfp_bpf_vnic_alloc()
82 nn->app_priv = bv; in nfp_bpf_vnic_alloc()
101 WARN_ON(bv->tc_prog); in nfp_bpf_vnic_free()
102 kfree(bv); in nfp_bpf_vnic_free()
111 struct nfp_bpf_vnic *bv; in nfp_bpf_setup_tc_block_cb() local
143 bv = nn->app_priv; in nfp_bpf_setup_tc_block_cb()
185 struct nfp_bpf_vnic *bv; in nfp_bpf_check_mtu() local
194 bv = nn->app_priv; in nfp_bpf_check_mtu()
[all …]
/linux/block/
A Dblk-merge.c31 bio_get_first_bvec(bio, bv); in bio_get_last_bvec()
32 if (bv->bv_len == bio->bi_iter.bi_size) in bio_get_last_bvec()
42 *bv = bio->bi_io_vec[idx]; in bio_get_last_bvec()
49 bv->bv_len = iter.bi_bvec_done; in bio_get_last_bvec()
318 bio_for_each_bvec(bv, bio, iter) { in bio_split_rw_at()
328 bv.bv_offset + bv.bv_len <= PAGE_SIZE) { in bio_split_rw_at()
330 bytes += bv.bv_len; in bio_split_rw_at()
337 bvprv = bv; in bio_split_rw_at()
426 struct bio_vec bv; in blk_recalc_rq_segments() local
448 rq_for_each_bvec(bv, rq, iter) in blk_recalc_rq_segments()
[all …]
A Dbio.c167 mempool_free(bv, pool); in bvec_free()
622 struct bio_vec bv; in zero_fill_bio_iter() local
626 memzero_bvec(&bv); in zero_fill_bio_iter()
642 struct bio_vec bv; in bio_truncate() local
654 if (done + bv.bv_len > new_size) { in bio_truncate()
661 zero_user(bv.bv_page, bv.bv_offset + offset, in bio_truncate()
662 bv.bv_len - offset); in bio_truncate()
665 done += bv.bv_len; in bio_truncate()
923 size_t bv_end = bv->bv_offset + bv->bv_len; in bvec_try_merge_page()
943 bv->bv_len += len; in bvec_try_merge_page()
[all …]
A Dblk-crypto-fallback.c164 struct bio_vec bv; in blk_crypto_fallback_clone_bio() local
179 bio_for_each_segment(bv, bio_src, iter) in blk_crypto_fallback_clone_bio()
180 bio->bi_io_vec[bio->bi_vcnt++] = bv; in blk_crypto_fallback_clone_bio()
216 struct bio_vec bv; in blk_crypto_fallback_split_bio_if_needed() local
219 bio_for_each_segment(bv, bio, iter) { in blk_crypto_fallback_split_bio_if_needed()
220 num_sectors += bv.bv_len >> SECTOR_SHIFT; in blk_crypto_fallback_split_bio_if_needed()
389 struct bio_vec bv; in blk_crypto_fallback_decrypt_bio() local
418 __bio_for_each_segment(bv, bio, iter, f_ctx->crypt_iter) { in blk_crypto_fallback_decrypt_bio()
419 struct page *page = bv.bv_page; in blk_crypto_fallback_decrypt_bio()
421 sg_set_page(&sg, page, data_unit_size, bv.bv_offset); in blk_crypto_fallback_decrypt_bio()
[all …]
A Dt10-pi.c381 struct bio_vec bv; in blk_integrity_generate() local
387 bio_for_each_segment(bv, bio, bviter) { in blk_integrity_generate()
388 void *kaddr = bvec_kmap_local(&bv); in blk_integrity_generate()
391 iter.data_size = bv.bv_len; in blk_integrity_generate()
413 struct bio_vec bv; in blk_integrity_verify() local
423 __bio_for_each_segment(bv, bio, bviter, bip->bio_iter) { in blk_integrity_verify()
424 void *kaddr = bvec_kmap_local(&bv); in blk_integrity_verify()
428 iter.data_size = bv.bv_len; in blk_integrity_verify()
A Dbio-integrity.c107 static void bio_integrity_unpin_bvec(struct bio_vec *bv, int nr_vecs, in bio_integrity_unpin_bvec() argument
113 if (dirty && !PageCompound(bv[i].bv_page)) in bio_integrity_unpin_bvec()
114 set_page_dirty_lock(bv[i].bv_page); in bio_integrity_unpin_bvec()
115 unpin_user_page(bv[i].bv_page); in bio_integrity_unpin_bvec()
171 struct bio_vec *bv = &bip->bip_vec[bip->bip_vcnt - 1]; in bio_integrity_add_page() local
174 if (bvec_try_merge_hw_page(q, bv, page, len, offset, in bio_integrity_add_page()
188 if (bvec_gap_to_prev(&q->limits, bv, offset)) in bio_integrity_add_page()
/linux/fs/bcachefs/
A Dchecksum.c262 struct bio_vec bv; in __bch2_checksum_bio() local
279 void *p = kmap_local_page(bv.bv_page) + bv.bv_offset; in __bch2_checksum_bio()
286 bch2_checksum_update(&state, page_address(bv.bv_page) + bv.bv_offset, in __bch2_checksum_bio()
287 bv.bv_len); in __bch2_checksum_bio()
302 void *p = kmap_local_page(bv.bv_page) + bv.bv_offset; in __bch2_checksum_bio()
310 page_address(bv.bv_page) + bv.bv_offset, in __bch2_checksum_bio()
311 bv.bv_len); in __bch2_checksum_bio()
334 struct bio_vec bv; in __bch2_encrypt_bio() local
345 bio_for_each_segment(bv, bio, iter) { in __bch2_encrypt_bio()
348 .offset = bv.bv_offset, in __bch2_encrypt_bio()
[all …]
A Dcompress.c43 struct bio_vec bv; in bio_phys_contig() local
47 __bio_for_each_bvec(bv, bio, iter, start) { in bio_phys_contig()
49 expected_start != page_address(bv.bv_page) + bv.bv_offset) in bio_phys_contig()
52 expected_start = page_address(bv.bv_page) + in bio_phys_contig()
53 bv.bv_offset + bv.bv_len; in bio_phys_contig()
63 struct bio_vec bv; in __bio_map_or_bounce() local
83 bv.bv_offset) in __bio_map_or_bounce()
86 if (bv.bv_len < iter.bi_size && in __bio_map_or_bounce()
87 bv.bv_offset + bv.bv_len < PAGE_SIZE) in __bio_map_or_bounce()
102 __bio_for_each_segment(bv, bio, iter, start) in __bio_map_or_bounce()
[all …]
A Dfs-io.h20 static inline struct folio_vec biovec_to_foliovec(struct bio_vec bv) in biovec_to_foliovec() argument
23 struct folio *folio = page_folio(bv.bv_page); in biovec_to_foliovec()
24 size_t offset = (folio_page_idx(folio, bv.bv_page) << PAGE_SHIFT) + in biovec_to_foliovec()
25 bv.bv_offset; in biovec_to_foliovec()
26 size_t len = min_t(size_t, folio_size(folio) - offset, bv.bv_len); in biovec_to_foliovec()
A Dutil.c673 struct bio_vec bv; in memcpy_to_bio() local
676 __bio_for_each_segment(bv, dst, iter, dst_iter) { in memcpy_to_bio()
677 void *dstp = kmap_local_page(bv.bv_page); in memcpy_to_bio()
679 memcpy(dstp + bv.bv_offset, src, bv.bv_len); in memcpy_to_bio()
682 src += bv.bv_len; in memcpy_to_bio()
688 struct bio_vec bv; in memcpy_from_bio() local
691 __bio_for_each_segment(bv, src, iter, src_iter) { in memcpy_from_bio()
692 void *srcp = kmap_local_page(bv.bv_page); in memcpy_from_bio()
694 memcpy(dst, srcp + bv.bv_offset, bv.bv_len); in memcpy_from_bio()
697 dst += bv.bv_len; in memcpy_from_bio()
/linux/drivers/block/
A Dn64cart.c62 static bool n64cart_do_bvec(struct device *dev, struct bio_vec *bv, u32 pos) in n64cart_do_bvec() argument
68 WARN_ON_ONCE((bv->bv_offset & (MIN_ALIGNMENT - 1)) || in n64cart_do_bvec()
69 (bv->bv_len & (MIN_ALIGNMENT - 1))); in n64cart_do_bvec()
71 dma_addr = dma_map_bvec(dev, bv, DMA_FROM_DEVICE, 0); in n64cart_do_bvec()
79 n64cart_write_reg(PI_WRITE_REG, bv->bv_len - 1); in n64cart_do_bvec()
83 dma_unmap_page(dev, dma_addr, bv->bv_len, DMA_FROM_DEVICE); in n64cart_do_bvec()
/linux/fs/netfs/
A Diterator.c40 struct bio_vec *bv = NULL; in netfs_extract_user_iter() local
54 bv_size = array_size(max_pages, sizeof(*bv)); in netfs_extract_user_iter()
55 bv = kvmalloc(bv_size, GFP_KERNEL); in netfs_extract_user_iter()
56 if (!bv) in netfs_extract_user_iter()
64 pages = (void *)bv + bv_size - pg_size; in netfs_extract_user_iter()
92 bvec_set_page(bv + npages + i, *pages++, len - offset, offset); in netfs_extract_user_iter()
100 iov_iter_bvec(new, orig->data_source, bv, npages, orig_len - count); in netfs_extract_user_iter()
/linux/drivers/s390/block/
A Ddasd_fba.c436 struct bio_vec bv; in dasd_fba_build_cp_regular() local
457 rq_for_each_segment(bv, req, iter) { in dasd_fba_build_cp_regular()
458 if (bv.bv_len & (blksize - 1)) in dasd_fba_build_cp_regular()
462 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_fba_build_cp_regular()
463 cidaw += bv.bv_len / blksize; in dasd_fba_build_cp_regular()
500 dst = bvec_virt(&bv); in dasd_fba_build_cp_regular()
505 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_fba_build_cp_regular()
507 dst = copy + bv.bv_offset; in dasd_fba_build_cp_regular()
568 struct bio_vec bv; in dasd_fba_free_cp() local
582 dst = bvec_virt(&bv); in dasd_fba_free_cp()
[all …]
/linux/net/ceph/
A Dmessenger_v2.c146 struct bio_vec bv; in do_try_sendpage() local
173 iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bv, 1, bv.bv_len); in do_try_sendpage()
897 struct bio_vec bv; in calc_sg_cnt_cursor() local
949 struct bio_vec bv; in init_sgs_cursor() local
956 sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset); in init_sgs_cursor()
1116 memcpy_page(bv.bv_page, bv.bv_offset, in process_v2_sparse_read()
1862 struct bio_vec bv; in prepare_read_data() local
1879 bv.bv_offset = 0; in prepare_read_data()
1888 struct bio_vec bv; in prepare_read_data_cont() local
1896 memcpy_to_page(bv.bv_page, bv.bv_offset, in prepare_read_data_cont()
[all …]
/linux/drivers/md/
A Ddm-io-rewind.c12 static inline bool dm_bvec_iter_rewind(const struct bio_vec *bv, in dm_bvec_iter_rewind() argument
27 while (idx >= 0 && bytes && bytes > bv[idx].bv_len) { in dm_bvec_iter_rewind()
28 bytes -= bv[idx].bv_len; in dm_bvec_iter_rewind()
41 iter->bi_bvec_done = bv[idx].bv_len - bytes; in dm_bvec_iter_rewind()
A Ddm-ebs-target.c65 static int __ebs_rw_bvec(struct ebs_c *ec, enum req_op op, struct bio_vec *bv, in __ebs_rw_bvec() argument
71 unsigned int bv_len = bv->bv_len; in __ebs_rw_bvec()
76 if (unlikely(!bv->bv_page || !bv_len)) in __ebs_rw_bvec()
79 pa = bvec_virt(bv); in __ebs_rw_bvec()
102 flush_dcache_page(bv->bv_page); in __ebs_rw_bvec()
104 flush_dcache_page(bv->bv_page); in __ebs_rw_bvec()
125 struct bio_vec bv; in __ebs_rw_bio() local
128 bio_for_each_bvec(bv, bio, iter) { in __ebs_rw_bio()
129 rr = __ebs_rw_bvec(ec, op, &bv, &iter); in __ebs_rw_bio()
/linux/fs/btrfs/
A Dbio.c170 struct bio_vec *bv = bio_first_bvec_all(&repair_bbio->bio); in btrfs_end_repair_bio() local
177 ASSERT(folio_order(page_folio(bv->bv_page)) == 0); in btrfs_end_repair_bio()
180 !btrfs_data_csum_ok(repair_bbio, dev, 0, bv)) { in btrfs_end_repair_bio()
200 page_folio(bv->bv_page), bv->bv_offset, mirror); in btrfs_end_repair_bio()
217 struct bio_vec *bv, in repair_one_sector() argument
251 __bio_add_page(repair_bio, bv->bv_page, bv->bv_len, bv->bv_offset); in repair_one_sector()
290 struct bio_vec bv = bio_iter_iovec(&bbio->bio, *iter); in btrfs_check_read_bio() local
292 bv.bv_len = min(bv.bv_len, sectorsize); in btrfs_check_read_bio()
293 if (status || !btrfs_data_csum_ok(bbio, dev, offset, &bv)) in btrfs_check_read_bio()
294 fbio = repair_one_sector(bbio, offset, &bv, fbio); in btrfs_check_read_bio()

Completed in 60 milliseconds

12345