Lines Matching refs:bdev

180 __find_get_block_slow(struct block_device *bdev, sector_t block, bool atomic)  in __find_get_block_slow()  argument
182 struct address_space *bd_mapping = bdev->bd_mapping; in __find_get_block_slow()
243 bh->b_state, bh->b_size, bdev, in __find_get_block_slow()
665 void write_boundary_block(struct block_device *bdev, in write_boundary_block() argument
670 bh = __find_get_block_nonatomic(bdev, bblock + 1, blocksize); in write_boundary_block()
989 static sector_t blkdev_max_block(struct block_device *bdev, unsigned int size) in blkdev_max_block() argument
992 loff_t sz = bdev_nr_bytes(bdev); in blkdev_max_block()
1005 struct block_device *bdev, unsigned size) in folio_init_buffers() argument
1011 sector_t end_block = blkdev_max_block(bdev, size); in folio_init_buffers()
1017 bh->b_bdev = bdev; in folio_init_buffers()
1042 static bool grow_dev_folio(struct block_device *bdev, sector_t block, in grow_dev_folio() argument
1045 struct address_space *mapping = bdev->bd_mapping; in grow_dev_folio()
1058 end_block = folio_init_buffers(folio, bdev, size); in grow_dev_folio()
1086 end_block = folio_init_buffers(folio, bdev, size); in grow_dev_folio()
1099 static bool grow_buffers(struct block_device *bdev, sector_t block, in grow_buffers() argument
1111 bdev); in grow_buffers()
1116 return grow_dev_folio(bdev, block, pos / PAGE_SIZE, size, gfp); in grow_buffers()
1120 __getblk_slow(struct block_device *bdev, sector_t block, in __getblk_slow() argument
1125 if (WARN_ON_ONCE(!IS_ALIGNED(size, bdev_logical_block_size(bdev)))) { in __getblk_slow()
1127 size, bdev_logical_block_size(bdev)); in __getblk_slow()
1134 if (!grow_buffers(bdev, block, size, gfp)) in __getblk_slow()
1138 bh = __find_get_block_nonatomic(bdev, block, size); in __getblk_slow()
1140 bh = __find_get_block(bdev, block, size); in __getblk_slow()
1361 lookup_bh_lru(struct block_device *bdev, sector_t block, unsigned size) in lookup_bh_lru() argument
1375 if (bh && bh->b_blocknr == block && bh->b_bdev == bdev && in lookup_bh_lru()
1401 find_get_block_common(struct block_device *bdev, sector_t block, in find_get_block_common() argument
1404 struct buffer_head *bh = lookup_bh_lru(bdev, block, size); in find_get_block_common()
1408 bh = __find_get_block_slow(bdev, block, atomic); in find_get_block_common()
1418 __find_get_block(struct block_device *bdev, sector_t block, unsigned size) in __find_get_block() argument
1420 return find_get_block_common(bdev, block, size, true); in __find_get_block()
1426 __find_get_block_nonatomic(struct block_device *bdev, sector_t block, in __find_get_block_nonatomic() argument
1429 return find_get_block_common(bdev, block, size, false); in __find_get_block_nonatomic()
1447 struct buffer_head *bdev_getblk(struct block_device *bdev, sector_t block, in bdev_getblk() argument
1453 bh = __find_get_block_nonatomic(bdev, block, size); in bdev_getblk()
1455 bh = __find_get_block(bdev, block, size); in bdev_getblk()
1461 return __getblk_slow(bdev, block, size, gfp); in bdev_getblk()
1468 void __breadahead(struct block_device *bdev, sector_t block, unsigned size) in __breadahead() argument
1470 struct buffer_head *bh = bdev_getblk(bdev, block, size, in __breadahead()
1502 struct buffer_head *__bread_gfp(struct block_device *bdev, sector_t block, in __bread_gfp() argument
1507 gfp |= mapping_gfp_constraint(bdev->bd_mapping, ~__GFP_FS); in __bread_gfp()
1515 bh = bdev_getblk(bdev, block, size, gfp); in __bread_gfp()
1738 void clean_bdev_aliases(struct block_device *bdev, sector_t block, sector_t len) in clean_bdev_aliases() argument
1740 struct address_space *bd_mapping = bdev->bd_mapping; in clean_bdev_aliases()
2048 bh->b_bdev = iomap->bdev; in iomap_to_bh()