| /linux/fs/crypto/ |
| A D | bio.c | 59 bio = bio_alloc(inode->i_sb->s_bdev, BIO_MAX_VECS, REQ_OP_WRITE, in fscrypt_zeroout_range_inline_crypt() 85 bio_reset(bio, inode->i_sb->s_bdev, REQ_OP_WRITE); in fscrypt_zeroout_range_inline_crypt() 158 bio = bio_alloc(inode->i_sb->s_bdev, nr_pages, REQ_OP_WRITE, GFP_NOFS); in fscrypt_zeroout_range() 189 bio_reset(bio, inode->i_sb->s_bdev, REQ_OP_WRITE); in fscrypt_zeroout_range()
|
| /linux/drivers/block/rnbd/ |
| A D | rnbd-proto.h | 228 bio_opf = REQ_OP_WRITE; in rnbd_to_bio_flags() 231 bio_opf = REQ_OP_WRITE | REQ_PREFLUSH; in rnbd_to_bio_flags() 265 case REQ_OP_WRITE: in rq_to_rnbd_flags()
|
| /linux/fs/ |
| A D | direct-io.c | 175 if (ret < 0 && sdio->blocks_available && dio_op == REQ_OP_WRITE) { in dio_refill_pages() 299 ret > 0 && dio_op == REQ_OP_WRITE) in dio_complete() 312 if (ret > 0 && dio_op == REQ_OP_WRITE) in dio_complete() 361 (dio_op == REQ_OP_WRITE && in dio_bio_end_aio() 641 create = dio_op == REQ_OP_WRITE; in get_more_blocks() 793 if (dio_op == REQ_OP_WRITE) { in submit_page_section() 982 if (dio_op == REQ_OP_WRITE) { in do_direct_IO() 1182 dio->opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in __blockdev_direct_IO()
|
| /linux/drivers/md/ |
| A D | dm-ebs-target.c | 198 else if (bio_op(bio) == REQ_OP_WRITE && !(bio->bi_opf & REQ_PREFLUSH)) { in __ebs_process_bios() 211 else if (bio_op(bio) == REQ_OP_WRITE) { in __ebs_process_bios() 213 r = __ebs_rw_bio(ec, REQ_OP_WRITE, bio); in __ebs_process_bios() 231 if (unlikely(r && bio_op(bio) == REQ_OP_WRITE)) in __ebs_process_bios()
|
| A D | dm-zero.c | 48 case REQ_OP_WRITE: in zero_map()
|
| A D | dm-integrity.c | 545 if (op == REQ_OP_WRITE) { in sync_rw_sb() 1130 rw_journal(ic, REQ_OP_WRITE | REQ_FUA, in write_journal() 1178 io_req.bi_opf = REQ_OP_WRITE; in copy_from_journal() 2045 if (likely(dio->op == REQ_OP_WRITE)) in __journal_read_write() 2102 if (likely(dio->op == REQ_OP_WRITE)) in __journal_read_write() 2114 if (likely(dio->op == REQ_OP_WRITE)) { in __journal_read_write() 2157 if (likely(dio->op == REQ_OP_WRITE)) { in __journal_read_write() 2211 if (dio->op == REQ_OP_WRITE) { in dm_integrity_map_continue() 2504 if (dio->op == REQ_OP_WRITE) { in dm_integrity_map_inline() 2985 r = sync_rw_sb(ic, REQ_OP_WRITE); in recalc_write_super() [all …]
|
| A D | dm-snap-persistent.c | 303 REQ_OP_WRITE, 0); in zero_disk_area() 395 return chunk_io(ps, ps->header_area, 0, REQ_OP_WRITE, 1); in write_header() 742 if (ps->valid && area_io(ps, REQ_OP_WRITE | REQ_PREFLUSH | REQ_FUA | in persistent_commit_exception() 820 r = area_io(ps, REQ_OP_WRITE | REQ_PREFLUSH | REQ_FUA); in persistent_commit_merge()
|
| A D | dm-zoned-target.c | 90 bio_op(bio) == REQ_OP_WRITE && in dmz_bio_endio() 144 if (bio_op(bio) == REQ_OP_WRITE && dmz_is_seq(zone)) in dmz_submit_bio() 429 case REQ_OP_WRITE: in dmz_handle_bio() 645 if (!nr_sectors && bio_op(bio) != REQ_OP_WRITE) in dmz_map() 659 if (!nr_sectors && bio_op(bio) == REQ_OP_WRITE) { in dmz_map()
|
| A D | dm-log-writes.c | 221 bio = bio_alloc(lc->logdev->bdev, 1, REQ_OP_WRITE, GFP_KERNEL); in write_metadata() 273 bio = bio_alloc(lc->logdev->bdev, bio_pages, REQ_OP_WRITE, in write_inline_data() 355 REQ_OP_WRITE, GFP_KERNEL); in log_one_block() 373 REQ_OP_WRITE, GFP_KERNEL); in log_one_block()
|
| A D | dm-log.c | 314 lc->io_req.bi_opf = REQ_OP_WRITE | REQ_PREFLUSH; in flush_header() 638 r = rw_header(lc, REQ_OP_WRITE); in disk_resume() 710 r = rw_header(lc, REQ_OP_WRITE); in disk_flush()
|
| /linux/drivers/md/dm-vdo/ |
| A D | io-submitter.h | 44 REQ_OP_WRITE | REQ_PREFLUSH, NULL); in vdo_submit_flush_vio()
|
| /linux/drivers/block/drbd/ |
| A D | drbd_actlog.c | 138 if ((op == REQ_OP_WRITE) && !test_bit(MD_NO_FUA, &device->flags)) in _drbd_md_sync_page_io() 151 if (op != REQ_OP_WRITE && device->state.disk == D_DISKLESS && device->ldev == NULL) in _drbd_md_sync_page_io() 164 if (drbd_insert_fault(device, (op == REQ_OP_WRITE) ? DRBD_FAULT_MD_WR : DRBD_FAULT_MD_RD)) in _drbd_md_sync_page_io() 187 (unsigned long long)sector, (op == REQ_OP_WRITE) ? "WRITE" : "READ", in drbd_md_sync_page_io() 195 (op == REQ_OP_WRITE) ? "WRITE" : "READ"); in drbd_md_sync_page_io() 201 (op == REQ_OP_WRITE) ? "WRITE" : "READ", err); in drbd_md_sync_page_io() 389 if (drbd_md_sync_page_io(device, device->ldev, sector, REQ_OP_WRITE)) { in __al_write_transaction()
|
| /linux/fs/xfs/ |
| A D | xfs_bio_io.c | 26 if (is_vmalloc && op == REQ_OP_WRITE) in xfs_rw_bdev()
|
| /linux/fs/ocfs2/ |
| A D | buffer_head_io.c | 67 submit_bh(REQ_OP_WRITE, bh); in ocfs2_write_block() 452 submit_bh(REQ_OP_WRITE, bh); in ocfs2_write_super_or_backup()
|
| /linux/Documentation/block/ |
| A D | writeback_cache_control.rst | 88 When the BLK_FEAT_WRITE_CACHE flag is set, REQ_OP_WRITE | REQ_PREFLUSH requests 93 REQ_OP_WRITE request, else a REQ_OP_FLUSH request is sent by the block layer
|
| /linux/drivers/nvdimm/ |
| A D | nd_virtio.c | 118 REQ_OP_WRITE | REQ_PREFLUSH, in async_pmem_flush()
|
| /linux/drivers/block/null_blk/ |
| A D | zoned.c | 363 return null_process_cmd(cmd, REQ_OP_WRITE, sector, nr_sectors); in null_zone_write() 415 ret = null_process_cmd(cmd, REQ_OP_WRITE, sector, nr_sectors); in null_zone_write() 704 case REQ_OP_WRITE: in null_process_zoned_cmd()
|
| /linux/drivers/target/ |
| A D | target_core_iblock.c | 433 bio = bio_alloc(ib_dev->ibd_bd, 0, REQ_OP_WRITE | REQ_PREFLUSH, in iblock_execute_sync_cache() 536 bio = iblock_get_bio(cmd, block_lba, 1, REQ_OP_WRITE); in iblock_execute_write_same() 549 bio = iblock_get_bio(cmd, block_lba, 1, REQ_OP_WRITE); in iblock_execute_write_same() 753 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in iblock_execute_rw()
|
| /linux/drivers/nvme/target/ |
| A D | io-cmd-bdev.c | 266 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in nvmet_bdev_execute_rw() 348 ARRAY_SIZE(req->inline_bvec), REQ_OP_WRITE | REQ_PREFLUSH); in nvmet_bdev_execute_flush()
|
| /linux/fs/bcachefs/ |
| A D | fs-io-direct.c | 542 bio_reset(bio, NULL, REQ_OP_WRITE | REQ_SYNC | REQ_IDLE); in bch2_dio_write_loop() 559 bio_reset(&dio->op.wbio.bio, NULL, REQ_OP_WRITE); in bch2_dio_write_continue() 629 REQ_OP_WRITE | REQ_SYNC | REQ_IDLE, in bch2_direct_write()
|
| /linux/block/ |
| A D | blk-core.c | 782 if (WARN_ON_ONCE(bio_op(bio) != REQ_OP_WRITE && in submit_bio_noacct() 803 case REQ_OP_WRITE: in submit_bio_noacct() 891 } else if (bio_op(bio) == REQ_OP_WRITE) { in submit_bio()
|
| /linux/kernel/power/ |
| A D | swap.c | 331 error = hib_submit_io(REQ_OP_WRITE | REQ_SYNC, in mark_swapfiles() 411 return hib_submit_io(REQ_OP_WRITE | REQ_SYNC, offset, src, hb); in write_page() 1582 error = hib_submit_io(REQ_OP_WRITE | REQ_SYNC, in swsusp_check() 1637 error = hib_submit_io(REQ_OP_WRITE | REQ_SYNC, in swsusp_unmark()
|
| /linux/fs/hfsplus/ |
| A D | wrapper.c | 71 if (op != REQ_OP_WRITE && data) in hfsplus_submit_bio()
|
| /linux/fs/iomap/ |
| A D | direct-io.c | 256 bio = iomap_dio_alloc_bio(iter, dio, 1, REQ_OP_WRITE | REQ_SYNC | REQ_IDLE); in iomap_dio_zero() 281 opflags |= REQ_OP_WRITE; in iomap_dio_bio_opflags()
|
| /linux/mm/ |
| A D | page_io.c | 419 REQ_OP_WRITE | REQ_SWAP | wbc_to_write_flags(wbc)); in swap_writepage_bdev_sync() 439 REQ_OP_WRITE | REQ_SWAP | wbc_to_write_flags(wbc), in swap_writepage_bdev_async()
|