| /fs/erofs/ |
| A D | fscache.c | 50 io->cres.ops->end_operation(&io->cres); in erofs_fscache_io_put() 119 if (!io) in erofs_fscache_req_io_alloc() 160 io->end_io, io); in erofs_fscache_read_io_async() 188 io->bio.bi_end_io(&io->bio); in erofs_fscache_bio_endio() 190 erofs_fscache_io_put(&io->io); in erofs_fscache_bio_endio() 199 io->io.private = mdev->m_dif->fscache->cookie; in erofs_fscache_bio_alloc() 200 io->io.end_io = erofs_fscache_bio_endio; in erofs_fscache_bio_alloc() 201 refcount_set(&io->io.ref, 1); in erofs_fscache_bio_alloc() 211 iov_iter_bvec(&io->io.iter, ITER_DEST, io->bvecs, bio->bi_vcnt, in erofs_fscache_submit_bio() 214 bio->bi_iter.bi_sector << 9, &io->io); in erofs_fscache_submit_bio() [all …]
|
| A D | fileio.c | 129 if (io->rq && (map->m_pa + ofs != io->dev.m_pa || in erofs_fileio_scan_folio() 133 io->rq = NULL; in erofs_fileio_scan_folio() 136 if (!io->rq) { in erofs_fileio_scan_folio() 138 .m_pa = io->map.m_pa + ofs, in erofs_fileio_scan_folio() 144 io->rq = erofs_fileio_rq_alloc(&io->dev); in erofs_fileio_scan_folio() 146 (io->dev.m_dif->fsoff + io->dev.m_pa) >> 9; in erofs_fileio_scan_folio() 153 io->dev.m_pa += len; in erofs_fileio_scan_folio() 163 struct erofs_fileio io = {}; in erofs_fileio_read_folio() local 168 erofs_fileio_rq_submit(io.rq); in erofs_fileio_read_folio() 175 struct erofs_fileio io = {}; in erofs_fileio_readahead() local [all …]
|
| A D | zdata.c | 1399 .sb = io->sb, in z_erofs_decompress_queue() 1403 .pcl = io->head, in z_erofs_decompress_queue() 1406 int err = io->eio ? -EIO : 0; in z_erofs_decompress_queue() 1451 if (io->sync) { in z_erofs_decompress_kickoff() 1453 complete(&io->u.done); in z_erofs_decompress_kickoff() 1457 if (atomic_add_return(bios, &io->pending_bios)) in z_erofs_decompress_kickoff() 1468 queue_work(z_erofs_workqueue, &io->u.work); in z_erofs_decompress_kickoff() 1474 queue_work(z_erofs_workqueue, &io->u.work); in z_erofs_decompress_kickoff() 1481 z_erofs_decompressqueue_work(&io->u.work); in z_erofs_decompress_kickoff() 1797 struct z_erofs_decompressqueue io[NR_JOBQUEUES]; in z_erofs_runqueue() local [all …]
|
| /fs/ext4/ |
| A D | page-io.c | 406 io->io_bio = NULL; in ext4_io_submit() 412 io->io_wbc = wbc; in ext4_io_submit_init() 413 io->io_bio = NULL; in ext4_io_submit_init() 414 io->io_end = NULL; in ext4_io_submit_init() 431 io->io_bio = bio; in io_submit_init_bio() 442 if (io->io_bio && (bh->b_blocknr != io->io_next_block || in io_submit_add_bh() 445 ext4_io_submit(io); in io_submit_add_bh() 454 io->io_next_block++; in io_submit_add_bh() 549 if (io->io_bio) in ext4_bio_write_folio() 559 if (io->io_bio) in ext4_bio_write_folio() [all …]
|
| A D | Makefile | 11 mmp.o move_extent.o namei.o page-io.o readpage.o resize.o \
|
| /fs/bcachefs/ |
| A D | move.c | 128 if (io->b) in move_free() 139 bch2_bio_free_pages_pool(io->write.op.c, &io->write.op.wbio.bio); in move_free() 142 kfree(io); in move_free() 166 move_free(io); in move_write_done() 201 move_free(io); in move_write() 214 atomic_add(io->write_sectors, &io->write.ctxt->write_sectors); in move_write() 225 return io && io->read_completed ? io : NULL; in bch2_moving_ctxt_next_pending_write() 243 struct moving_io *io; in bch2_moving_ctxt_do_pending_writes() local 248 move_write(io); in bch2_moving_ctxt_do_pending_writes() 364 if (!io) in bch2_move_extent() [all …]
|
| A D | fs-io-buffered.c | 393 struct bch_writepage_io *io; member 424 if (io->op.error) { in bch2_writepage_io_done() 469 bch2_i_sectors_acct(c, io->inode, NULL, io->op.i_sectors_delta); in bch2_writepage_io_done() 478 bio_put(&io->op.wbio.bio); in bch2_writepage_io_done() 483 struct bch_writepage_io *io = w->io; in bch2_writepage_do_io() local 485 w->io = NULL; in bch2_writepage_do_io() 508 w->io->inode = inode; in bch2_writepage_io_alloc() 509 op = &w->io->op; in bch2_writepage_io_alloc() 629 if (w->io && in __bch2_writepage() 635 if (!w->io) in __bch2_writepage() [all …]
|
| A D | Makefile | 47 fs-io.o \ 48 fs-io-buffered.o \ 49 fs-io-direct.o \ 50 fs-io-pagecache.o \ 91 super-io.o \
|
| /fs/f2fs/ |
| A D | iostat.c | 90 int io, idx; in __record_iostat_latency() local 97 for (io = 0; io < NR_PAGE_TYPE; io++) { in __record_iostat_latency() 98 iostat_lat[idx][io].peak_lat = in __record_iostat_latency() 99 jiffies_to_msecs(io_lat->peak_lat[idx][io]); in __record_iostat_latency() 100 iostat_lat[idx][io].cnt = io_lat->bio_cnt[idx][io]; in __record_iostat_latency() 101 iostat_lat[idx][io].avg_lat = iostat_lat[idx][io].cnt ? in __record_iostat_latency() 102 jiffies_to_msecs(io_lat->sum_lat[idx][io]) / iostat_lat[idx][io].cnt : 0; in __record_iostat_latency() 103 io_lat->sum_lat[idx][io] = 0; in __record_iostat_latency() 104 io_lat->peak_lat[idx][io] = 0; in __record_iostat_latency() 105 io_lat->bio_cnt[idx][io] = 0; in __record_iostat_latency()
|
| A D | data.c | 532 if (!io->bio) in __submit_merged_bio() 536 trace_f2fs_prepare_read_bio(io->sbi->sb, fio->type, io->bio); in __submit_merged_bio() 537 f2fs_submit_read_bio(io->sbi, io->bio, fio->type); in __submit_merged_bio() 540 f2fs_submit_write_bio(io->sbi, io->bio, fio->type); in __submit_merged_bio() 542 io->bio = NULL; in __submit_merged_bio() 625 if (!io->bio) in __f2fs_submit_merged_write() 992 if (io->bio && in f2fs_submit_page_write() 993 (!io_is_mergeable(sbi, io->bio, io, fio, io->last_block_in_bio, in f2fs_submit_page_write() 1023 io->bi_private = io->bio->bi_private; in f2fs_submit_page_write() 1024 io->bio->bi_private = io; in f2fs_submit_page_write() [all …]
|
| /fs/fuse/ |
| A D | file.c | 605 if (io->err) in fuse_get_res_by_io() 608 if (io->bytes >= 0 && io->write) in fuse_get_res_by_io() 611 return io->bytes < 0 ? io->size : io->bytes; in fuse_get_res_by_io() 636 io->err = io->err ? : err; in fuse_aio_complete() 658 io->iocb->ki_complete(io->iocb, res); in fuse_aio_complete() 671 ia->io = io; in fuse_io_alloc() 692 struct fuse_io_priv *io = ia->io; in fuse_aio_complete_req() local 725 struct fuse_io_priv *io = ia->io; in fuse_async_req_send() local 730 io->reqs++; in fuse_async_req_send() 2767 if (!io) in fuse_direct_IO() [all …]
|
| A D | Kconfig | 69 bool "FUSE communication over io-uring" 74 This allows sending FUSE requests over the io-uring interface and 77 If you want to allow fuse server/client communication through io-uring,
|
| A D | cuse.c | 96 struct fuse_io_priv io = FUSE_IO_PRIV_SYNC(kiocb); in cuse_read_iter() local 99 return fuse_direct_io(&io, to, &pos, FUSE_DIO_CUSE); in cuse_read_iter() 104 struct fuse_io_priv io = FUSE_IO_PRIV_SYNC(kiocb); in cuse_write_iter() local 110 return fuse_direct_io(&io, from, &pos, in cuse_write_iter()
|
| /fs/orangefs/ |
| A D | file.c | 68 new_op->upcall.req.io.readahead_size = readahead_size; in wait_for_direct_io() 69 new_op->upcall.req.io.io_type = type; in wait_for_direct_io() 70 new_op->upcall.req.io.refn = orangefs_inode->refn; in wait_for_direct_io() 90 new_op->upcall.req.io.buf_index = buffer_index; in wait_for_direct_io() 91 new_op->upcall.req.io.count = total_size; in wait_for_direct_io() 92 new_op->upcall.req.io.offset = *offset; in wait_for_direct_io() 245 if (type == ORANGEFS_IO_READ && new_op->downcall.resp.io.amt_complete) { in wait_for_direct_io() 252 copy_amount = new_op->downcall.resp.io.amt_complete; in wait_for_direct_io() 267 (int)new_op->downcall.resp.io.amt_complete); in wait_for_direct_io() 269 ret = new_op->downcall.resp.io.amt_complete; in wait_for_direct_io()
|
| A D | downcall.h | 108 struct orangefs_io_response io; member
|
| /fs/btrfs/ |
| A D | Makefile | 26 file-item.o inode-item.o disk-io.o \ 35 subpage.o tree-mod-log.o extent-io-tree.o fs.o messages.o bio.o \ 36 lru_cache.o raid-stripe-tree.o fiemap.o direct-io.o 45 tests/extent-io-tests.o tests/inode-tests.o tests/qgroup-tests.o \
|
| A D | Kconfig | 34 https://btrfs.readthedocs.io 56 extent maps, extent io, extent buffers, inodes, qgroups and others,
|
| /fs/iomap/ |
| A D | Makefile | 13 buffered-io.o 14 iomap-$(CONFIG_BLOCK) += direct-io.o \
|
| /fs/smb/client/ |
| A D | smb2inode.c | 32 struct smb2_ioctl_rsp *io = iov->iov_base; in reparse_buf_ptr() local 35 count = le32_to_cpu(io->OutputCount); in reparse_buf_ptr() 36 off = le32_to_cpu(io->OutputOffset); in reparse_buf_ptr() 40 buf = (struct reparse_data_buffer *)((u8 *)io + off); in reparse_buf_ptr() 803 idata->reparse.io.iov = *iov; in smb2_compound_op() 804 idata->reparse.io.buftype = resp_buftype[i + 1]; in smb2_compound_op() 1433 *rsp = data.reparse.io.iov; in smb2_query_reparse_point() 1434 *rsp_buftype = data.reparse.io.buftype; in smb2_query_reparse_point() 1435 memset(&data.reparse.io.iov, 0, sizeof(data.reparse.io.iov)); in smb2_query_reparse_point() 1436 data.reparse.io.buftype = CIFS_NO_BUFFER; in smb2_query_reparse_point()
|
| /fs/befs/ |
| A D | Makefile | 8 befs-objs := datastream.o btree.o super.o inode.o debug.o io.o linuxvfs.o
|
| A D | ChangeLog | 135 * Implemented the abstracted read_block() in io.c [WD] 140 hanging disk io problem. It turns out you need to have the sync_pages 230 It doesn't chew up the buffer cache (it does page io instead), and seems 300 * Created io.c, for io related functions.
|
| /fs/cachefiles/ |
| A D | Makefile | 10 io.o \
|
| /fs/ceph/ |
| A D | Makefile | 9 export.o caps.o snap.o xattr.o quota.o io.o \
|
| /fs/ubifs/ |
| A D | Makefile | 4 ubifs-y += shrinker.o journal.o file.o dir.o super.o sb.o io.o
|
| /fs/nfs/ |
| A D | Makefile | 10 io.o direct.o pagelist.o read.o symlink.o unlink.o \
|