Lines Matching refs:bdev

13 static sector_t bio_discard_limit(struct block_device *bdev, sector_t sector)  in bio_discard_limit()  argument
15 unsigned int discard_granularity = bdev_discard_granularity(bdev); in bio_discard_limit()
18 if (bdev_is_partition(bdev)) in bio_discard_limit()
19 sector += bdev->bd_start_sect; in bio_discard_limit()
38 struct bio *blk_alloc_discard_bio(struct block_device *bdev, in blk_alloc_discard_bio() argument
41 sector_t bio_sects = min(*nr_sects, bio_discard_limit(bdev, *sector)); in blk_alloc_discard_bio()
47 bio = bio_alloc(bdev, 0, REQ_OP_DISCARD, gfp_mask); in blk_alloc_discard_bio()
63 int __blkdev_issue_discard(struct block_device *bdev, sector_t sector, in __blkdev_issue_discard() argument
68 while ((bio = blk_alloc_discard_bio(bdev, &sector, &nr_sects, in __blkdev_issue_discard()
85 int blkdev_issue_discard(struct block_device *bdev, sector_t sector, in blkdev_issue_discard() argument
93 ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, &bio); in blkdev_issue_discard()
106 static sector_t bio_write_zeroes_limit(struct block_device *bdev) in bio_write_zeroes_limit() argument
108 sector_t bs_mask = (bdev_logical_block_size(bdev) >> 9) - 1; in bio_write_zeroes_limit()
110 return min(bdev_write_zeroes_sectors(bdev), in bio_write_zeroes_limit()
121 static void __blkdev_issue_write_zeroes(struct block_device *bdev, in __blkdev_issue_write_zeroes() argument
134 bio = bio_alloc(bdev, 0, REQ_OP_WRITE_ZEROES, gfp_mask); in __blkdev_issue_write_zeroes()
148 static int blkdev_issue_write_zeroes(struct block_device *bdev, sector_t sector, in blkdev_issue_write_zeroes() argument
151 sector_t limit = bio_write_zeroes_limit(bdev); in blkdev_issue_write_zeroes()
157 __blkdev_issue_write_zeroes(bdev, sector, nr_sects, gfp, &bio, in blkdev_issue_write_zeroes()
177 if (ret && !bdev_write_zeroes_sectors(bdev)) in blkdev_issue_write_zeroes()
195 static void __blkdev_issue_zero_pages(struct block_device *bdev, in __blkdev_issue_zero_pages() argument
203 bio = bio_alloc(bdev, nr_vecs, REQ_OP_WRITE, gfp_mask); in __blkdev_issue_zero_pages()
227 static int blkdev_issue_zero_pages(struct block_device *bdev, sector_t sector, in blkdev_issue_zero_pages() argument
238 __blkdev_issue_zero_pages(bdev, sector, nr_sects, gfp, &bio, flags); in blkdev_issue_zero_pages()
273 int __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, in __blkdev_issue_zeroout() argument
277 sector_t limit = bio_write_zeroes_limit(bdev); in __blkdev_issue_zeroout()
279 if (bdev_read_only(bdev)) in __blkdev_issue_zeroout()
283 __blkdev_issue_write_zeroes(bdev, sector, nr_sects, in __blkdev_issue_zeroout()
288 __blkdev_issue_zero_pages(bdev, sector, nr_sects, gfp_mask, in __blkdev_issue_zeroout()
308 int blkdev_issue_zeroout(struct block_device *bdev, sector_t sector, in blkdev_issue_zeroout() argument
313 if ((sector | nr_sects) & ((bdev_logical_block_size(bdev) >> 9) - 1)) in blkdev_issue_zeroout()
315 if (bdev_read_only(bdev)) in blkdev_issue_zeroout()
318 if (bdev_write_zeroes_sectors(bdev)) { in blkdev_issue_zeroout()
319 ret = blkdev_issue_write_zeroes(bdev, sector, nr_sects, in blkdev_issue_zeroout()
325 return blkdev_issue_zero_pages(bdev, sector, nr_sects, gfp_mask, flags); in blkdev_issue_zeroout()
329 int blkdev_issue_secure_erase(struct block_device *bdev, sector_t sector, in blkdev_issue_secure_erase() argument
332 sector_t bs_mask = (bdev_logical_block_size(bdev) >> 9) - 1; in blkdev_issue_secure_erase()
333 unsigned int max_sectors = bdev_max_secure_erase_sectors(bdev); in blkdev_issue_secure_erase()
347 if (bdev_read_only(bdev)) in blkdev_issue_secure_erase()
354 bio = blk_next_bio(bio, bdev, 0, REQ_OP_SECURE_ERASE, gfp); in blkdev_issue_secure_erase()