| /fs/nfs/ |
| A D | pagelist.c | 44 i->req = req; in nfs_page_iter_page_init() 50 const struct nfs_page *req = i->req; in nfs_page_iter_page_advance() local 53 i->count = (tmp < req->wb_bytes) ? tmp : req->wb_bytes; in nfs_page_iter_page_advance() 58 const struct nfs_page *req = i->req; in nfs_page_iter_page_get() local 239 if (ret || req->wb_head == req) in nfs_page_group_lock() 251 if (req != req->wb_head) in nfs_page_group_unlock() 278 tmp = req; in nfs_page_group_sync_on_bit_locked() 318 req->wb_head = req; in nfs_page_group_init() 319 req->wb_this_page = req; in nfs_page_group_init() 581 WARN_ON_ONCE(req->wb_this_page != req); in nfs_free_request() [all …]
|
| A D | write.c | 189 if (req) { in nfs_folio_find_head_request() 190 WARN_ON_ONCE(req->wb_head != req); in nfs_folio_find_head_request() 270 page_offset < (req->wb_pgbase + req->wb_bytes)) in nfs_page_group_search_locked() 273 req = req->wb_this_page; in nfs_page_group_search_locked() 644 if (!req) in nfs_do_writepage() 764 WARN_ON_ONCE(req->wb_this_page != req); in nfs_inode_add_request() 1077 rqend = req->wb_offset + req->wb_bytes; in nfs_try_to_update_request() 1093 req->wb_bytes = end - req->wb_offset; in nfs_try_to_update_request() 1095 req->wb_bytes = rqend - req->wb_offset; in nfs_try_to_update_request() 1713 if (lwb < (req_offset(req) + req->wb_bytes)) in nfs_get_lwb() [all …]
|
| A D | direct.c | 311 struct nfs_page *req; in nfs_read_sync_pgio_error() local 375 if (IS_ERR(req)) { in nfs_direct_read_schedule_iovec() 532 if (req->wb_head != req) { in nfs_direct_join_group() 533 nfs_direct_add_page_head(&req->wb_list, req); in nfs_direct_join_group() 537 if (subreq == req) in nfs_direct_join_group() 567 struct nfs_page *req; in nfs_direct_write_reschedule() local 586 req->wb_nio++; in nfs_direct_write_reschedule() 628 struct nfs_page *req; in nfs_direct_commit_complete() local 662 req->wb_nio = 0; in nfs_direct_commit_complete() 715 struct nfs_page *req; in nfs_direct_write_clear_reqs() local [all …]
|
| /fs/smb/server/ |
| A D | transport_ipc.c | 343 req->ifc_list_sz); in ipc_server_config_on_startup() 347 req->netbios_name, req->server_string, in ipc_server_config_on_startup() 685 snprintf(req->peer_addr, sizeof(req->peer_addr), "%pIS", peer_addr); in ksmbd_ipc_tree_connect_request() 779 req->handle = handle; in ksmbd_rpc_open() 782 req->payload_sz = 0; in ksmbd_rpc_open() 801 req->handle = handle; in ksmbd_rpc_close() 804 req->payload_sz = 0; in ksmbd_rpc_close() 827 req->handle = handle; in ksmbd_rpc_write() 851 req->handle = handle; in ksmbd_rpc_read() 855 req->payload_sz = 0; in ksmbd_rpc_read() [all …]
|
| /fs/fuse/ |
| A D | dev.c | 44 if (!req) in fuse_request_expired() 144 if (req) in fuse_request_alloc() 147 return req; in fuse_request_alloc() 227 if (!req) { in fuse_get_req() 259 return req; in fuse_get_req() 487 req->args->end(fm, req->args, req->out.h.error); in fuse_request_end() 742 if (!req) in fuse_simple_background() 791 if (req) { in lock_request() 1160 struct fuse_req *req = cs->req; in fuse_copy_folios() local 1475 cs->req = req; in fuse_dev_do_read() [all …]
|
| A D | dev_uring.c | 152 req = ent->fuse_req; in ent_list_request_expired() 338 if (req) { in fuse_uring_entry_teardown() 356 if (req) in fuse_uring_entry_teardown() 598 cs.req = req; in fuse_uring_copy_from_ring() 628 cs.req = req; in fuse_uring_args_to_ring() 799 if (req) in fuse_uring_ent_assign_req() 802 return req; in fuse_uring_ent_assign_req() 824 err = fuse_uring_out_header_has_err(&req->out.h, req, fc); in fuse_uring_commit() 851 if (req) { in fuse_uring_next_fuse_req() 910 if (!req) { in fuse_uring_commit_fetch() [all …]
|
| /fs/bcachefs/ |
| A D | alloc_foreground.c | 700 if (req->nr_effective >= req->nr_replicas) in add_new_bucket() 782 bch2_dev_alloc_list(c, &req->wp->stripe, &req->devs_may_alloc, &req->devs_sorted); in bucket_alloc_from_stripe() 819 (req->wp->data_type == BCH_DATA_btree || req->ec || req->have_cache)) in want_bucket() 843 req->wp->ptrs = req->scratch_ptrs; in bucket_alloc_set_writepoint() 901 req->devs_may_alloc = target_rw_devs(c, req->wp->data_type, req->target); in __open_bucket_add_buckets() 952 if (req->nr_effective >= req->nr_replicas) in open_bucket_add_buckets() 1228 req->ptrs = req->scratch_ptrs; in deallocate_extra_replicas() 1275 req->data_type = req->wp->data_type; in bch2_alloc_sectors_start_trans() 1326 if (req->nr_effective > req->nr_replicas) in bch2_alloc_sectors_start_trans() 1333 req->wp->ptrs = req->ptrs; in bch2_alloc_sectors_start_trans() [all …]
|
| /fs/cachefiles/ |
| A D | ondemand.c | 14 kfree(req); in cachefiles_req_put() 132 if (!req || req->msg.opcode != CACHEFILES_OP_READ || in cachefiles_ondemand_fd_ioctl() 194 if (!req || req->msg.opcode != CACHEFILES_OP_OPEN || in cachefiles_ondemand_copen() 384 return req; in cachefiles_ondemand_select_req() 394 return req; in cachefiles_ondemand_select_req() 405 if (xa_cmpxchg(xas->xa, xas->xa_index, req, NULL, 0) != req) in cachefiles_ondemand_finish_req() 434 if (!req) { in cachefiles_ondemand_daemon_read() 439 msg = &req->msg; in cachefiles_ondemand_daemon_read() 503 req = kzalloc(sizeof(*req) + data_len, GFP_KERNEL); in cachefiles_ondemand_send_req() 504 if (!req) { in cachefiles_ondemand_send_req() [all …]
|
| /fs/ubifs/ |
| A D | budget.c | 363 znodes = req->new_ino + (req->new_page << UBIFS_BLOCKS_PER_PAGE_SHIFT) + in calc_idx_growth() 364 req->new_dent; in calc_idx_growth() 380 if (req->new_page) in calc_data_growth() 382 if (req->new_dent) in calc_data_growth() 401 if (req->dirtied_ino) in calc_dd_growth() 403 if (req->mod_dent) in calc_dd_growth() 474 if (req->fast) { in ubifs_budget_space() 530 req->data_growth = calc_data_growth(c, req); in ubifs_release_budget() 531 req->dd_growth = calc_dd_growth(c, req); in ubifs_release_budget() 532 req->idx_growth = calc_idx_growth(c, req); in ubifs_release_budget() [all …]
|
| A D | compress.c | 104 err = crypto_acomp_compress(req); in ubifs_compress_common() 105 dlen = req->dlen; in ubifs_compress_common() 109 req = ACOMP_REQUEST_CLONE(req, GFP_NOFS | __GFP_NOWARN); in ubifs_compress_common() 112 err = crypto_acomp_compress(req); in ubifs_compress_common() 114 dlen = req->dlen; in ubifs_compress_common() 115 acomp_request_free(req); in ubifs_compress_common() 234 err = crypto_acomp_decompress(req); in ubifs_decompress_common() 235 dlen = req->dlen; in ubifs_decompress_common() 239 req = ACOMP_REQUEST_CLONE(req, GFP_NOFS | __GFP_NOWARN); in ubifs_decompress_common() 242 dlen = req->dlen; in ubifs_decompress_common() [all …]
|
| /fs/lockd/ |
| A D | clntproc.c | 140 lock->oh.len = snprintf(req->a_owner, sizeof(req->a_owner), "%u@%s", in nlmclnt_setlockargs() 350 .callback_data = req, in __nlm_async_call() 460 req->a_host->h_addrlen, req->a_res.status); in nlmclnt_test() 618 req->a_host->h_addrlen, req->a_res.status); in nlmclnt_lock() 625 req->a_host->h_addrlen, req->a_res.status); in nlmclnt_lock() 644 struct nlm_rqst *req) in nlmclnt_reclaim() argument 648 memset(req, 0, sizeof(*req)); in nlmclnt_reclaim() 651 req->a_host = host; in nlmclnt_reclaim() 725 req->a_host->h_addrlen, req->a_res.status); in nlmclnt_unlock() 788 struct nlm_rqst *req; in nlmclnt_cancel() local [all …]
|
| /fs/nilfs2/ |
| A D | dat.c | 65 brelse(req->pr_entry_bh); in nilfs_dat_commit_entry() 71 brelse(req->pr_entry_bh); in nilfs_dat_abort_entry() 95 req->pr_entry_bh); in nilfs_dat_commit_alloc() 119 req->pr_entry_bh); in nilfs_dat_commit_free() 128 if (unlikely(req->pr_desc_bh == NULL || req->pr_bitmap_bh == NULL)) { in nilfs_dat_commit_free() 149 req->pr_entry_bh); in nilfs_dat_commit_start() 171 req->pr_entry_bh); in nilfs_dat_prepare_end() 206 req->pr_entry_bh); in nilfs_dat_commit_end() 231 req->pr_entry_bh); in nilfs_dat_abort_end() 287 struct nilfs_palloc_req req; in nilfs_dat_mark_dirty() local [all …]
|
| A D | ifile.c | 55 struct nilfs_palloc_req req; in nilfs_ifile_create_inode() local 59 req.pr_entry_bh = NULL; in nilfs_ifile_create_inode() 64 &req.pr_entry_bh); in nilfs_ifile_create_inode() 69 brelse(req.pr_entry_bh); in nilfs_ifile_create_inode() 75 *out_ino = (ino_t)req.pr_entry_nr; in nilfs_ifile_create_inode() 76 *out_bh = req.pr_entry_bh; in nilfs_ifile_create_inode() 93 struct nilfs_palloc_req req = { in nilfs_ifile_delete_inode() local 103 &req.pr_entry_bh); in nilfs_ifile_delete_inode() 108 brelse(req.pr_entry_bh); in nilfs_ifile_delete_inode() 113 req.pr_entry_bh); in nilfs_ifile_delete_inode() [all …]
|
| /fs/ceph/ |
| A D | dir.c | 410 if (IS_ERR(req)) in ceph_readdir() 853 if (IS_ERR(req)) in ceph_lookup() 856 req->r_num_caps = 2; in ceph_lookup() 864 req->r_parent = dir; in ceph_lookup() 935 if (IS_ERR(req)) { in ceph_mknod() 1046 if (IS_ERR(req)) { in ceph_symlink() 1137 if (IS_ERR(req)) { in ceph_mkdir() 1212 if (IS_ERR(req)) { in ceph_link() 1250 int result = req->r_err ? req->r_err : in ceph_async_unlink_cb() 1391 if (IS_ERR(req)) { in ceph_unlink() [all …]
|
| A D | export.c | 151 if (IS_ERR(req)) in __lookup_inode() 159 req->r_ino1 = vino; in __lookup_inode() 251 if (IS_ERR(req)) in __snapfh_to_dentry() 268 req->r_ino1 = vino; in __snapfh_to_dentry() 269 req->r_num_caps = 1; in __snapfh_to_dentry() 334 if (IS_ERR(req)) in __get_parent() 481 if (IS_ERR(req)) { in __get_snap_name() 483 req = NULL; in __get_snap_name() 536 req = NULL; in __get_snap_name() 540 if (req) in __get_snap_name() [all …]
|
| A D | mds_client.c | 1151 if (req) in DEFINE_RB_FUNCS() 1182 doutc(cl, "%p tid %lld\n", req, req->r_tid); in __register_request() 2575 if (!req) in ceph_mdsc_create_request() 2987 req->r_parent, req->r_path1, req->r_ino1.ino, in create_request_message() 3002 req->r_path2, req->r_ino2.ino, in create_request_message() 3009 req->r_altname = get_fscrypt_altname(req, &req->r_altname_len); in create_request_message() 3038 (!!req->r_inode_drop + !!req->r_dentry_drop + in create_request_message() 3167 req->r_inode ? req->r_inode : d_inode(req->r_dentry), in create_request_message() 3172 req->r_parent, mds, req->r_dentry_drop, in create_request_message() 3244 req->r_callback(mdsc, req); in complete_request() [all …]
|
| A D | file.c | 197 return req; in prepare_open_request() 605 int result = req->r_err ? req->r_err : in ceph_async_create_cb() 644 req->r_err, req->r_deleg_ino, ino); in ceph_async_create_cb() 1356 aio_work->req = req; in ceph_aio_complete_req() 1440 if (!req) { in ceph_aio_retry_work() 1453 req->r_data_offset = req->r_ops[0].extent.offset; in ceph_aio_retry_work() 1468 ceph_osdc_start_request(req->r_osdc, req); in ceph_aio_retry_work() 1620 ceph_osdc_start_request(req->r_osdc, req); in ceph_direct_read_write() 1686 ceph_osdc_start_request(req->r_osdc, req); in ceph_direct_read_write() 2847 if (!req) in ceph_alloc_copyfrom_request() [all …]
|
| A D | locks.c | 81 struct ceph_mds_request *req; in ceph_lock_message() local 101 if (IS_ERR(req)) in ceph_lock_message() 102 return PTR_ERR(req); in ceph_lock_message() 103 req->r_inode = inode; in ceph_lock_message() 105 req->r_num_caps = 1; in ceph_lock_message() 151 ceph_mdsc_put_request(req); in ceph_lock_message() 164 struct inode *inode = req->r_inode; in ceph_lock_wait_for_completion() 191 mutex_lock(&req->r_fill_mutex); in ceph_lock_wait_for_completion() 192 req->r_err = err; in ceph_lock_wait_for_completion() 194 mutex_unlock(&req->r_fill_mutex); in ceph_lock_wait_for_completion() [all …]
|
| /fs/xfs/libxfs/ |
| A D | xfs_exchmaps.c | 694 if (check_add_overflow(req->ip1_bcount, bmbt_blocks, &req->ip1_bcount)) in xfs_exchmaps_estimate_overhead() 696 if (check_add_overflow(req->ip2_bcount, bmbt_blocks, &req->ip2_bcount)) in xfs_exchmaps_estimate_overhead() 717 req->resblks = resblks; in xfs_exchmaps_estimate_overhead() 964 if (req->startoff1 != 0 || req->startoff2 != 0) in xmi_can_exchange_reflink_flags() 966 if (req->blockcount != XFS_B_TO_FSB(mp, req->ip1->i_disk_size)) in xmi_can_exchange_reflink_flags() 968 if (req->blockcount != XFS_B_TO_FSB(mp, req->ip2->i_disk_size)) in xmi_can_exchange_reflink_flags() 1076 req->nr_exchanges++; in xfs_exchmaps_estimate() 1123 if (req->ip1 == req->ip2) { in xfs_exchmaps_estimate() 1124 error = xmi_ensure_delta_nextents(req, req->ip1, in xfs_exchmaps_estimate() 1127 error = xmi_ensure_delta_nextents(req, req->ip1, d_nexts1); in xfs_exchmaps_estimate() [all …]
|
| /fs/coda/ |
| A D | psdev.c | 151 req = tmp; in coda_psdev_write() 158 if (!req) { in coda_psdev_write() 174 wake_up(&req->uc_sleep); in coda_psdev_write() 180 req->uc_outSize = nbytes; in coda_psdev_write() 241 list_del(&req->uc_chain); in coda_psdev_read() 244 count = req->uc_inSize; in coda_psdev_read() 261 kvfree(req->uc_data); in coda_psdev_read() 262 kfree(req); in coda_psdev_read() 323 kvfree(req->uc_data); in coda_psdev_release() 324 kfree(req); in coda_psdev_release() [all …]
|
| /fs/erofs/ |
| A D | fscache.c | 60 pgoff_t last_page = ((req->start + req->len) / PAGE_SIZE) - 1; in erofs_fscache_req_complete() 80 kfree(req); in erofs_fscache_req_put() 88 if (!req) in erofs_fscache_req_alloc() 92 req->len = len; in erofs_fscache_req_alloc() 94 return req; in erofs_fscache_req_alloc() 231 if (!req) { in erofs_fscache_meta_read_folio() 262 loff_t pos = req->start + req->submitted; in erofs_fscache_data_read_slice() 293 count = req->len - req->submitted; in erofs_fscache_data_read_slice() 334 } while (!ret && req->submitted < req->len); in erofs_fscache_data_read() 345 if (!req) { in erofs_fscache_read_folio() [all …]
|
| /fs/xfs/scrub/ |
| A D | tempfile.c | 613 struct xfs_exchmaps_req *req = &tx->req; in xrep_tempexch_prep_request() local 661 struct xfs_exchmaps_req *req = &tx->req; in xrep_tempexch_estimate() local 746 const struct xfs_exchmaps_req *req = &tx->req; in xrep_tempexch_reserve_quota() local 754 if (!XFS_IS_QUOTA_ON(tp->t_mountp) || req->ip1 == req->ip2 || in xrep_tempexch_reserve_quota() 756 (req->ip1->i_udquot == req->ip2->i_udquot && in xrep_tempexch_reserve_quota() 757 req->ip1->i_gdquot == req->ip2->i_gdquot && in xrep_tempexch_reserve_quota() 758 req->ip1->i_pdquot == req->ip2->i_pdquot)) in xrep_tempexch_reserve_quota() 770 ddelta = max_t(int64_t, 0, req->ip2_bcount - req->ip1_bcount); in xrep_tempexch_reserve_quota() 773 ddelta + req->ip1_bcount, rdelta + req->ip1_rtbcount, in xrep_tempexch_reserve_quota() 778 ddelta = max_t(int64_t, 0, req->ip1_bcount - req->ip2_bcount); in xrep_tempexch_reserve_quota() [all …]
|
| /fs/ |
| A D | aio.c | 602 ctx = req->ki_ctx; in kiocb_set_cancel_fn() 657 req->ki_cancel(&req->rw); in free_ioctx_users() 1072 return req; in aio_get_req() 1518 req->ki_flags = req->ki_filp->f_iocb_flags | IOCB_AIO_RW; in aio_prep_rw() 1577 req->ki_complete(req, ret); in aio_rw_done() 1603 aio_rw_done(req, file->f_op->read_iter(req, &iter)); in aio_read() 1634 aio_rw_done(req, file->f_op->write_iter(req, &iter)); in aio_write() 1662 if (!req->creds) in aio_fsync() 1733 mask = vfs_poll(req->file, &pt) & req->events; in aio_poll_complete_work() 1927 mask = vfs_poll(req->file, &apt.pt) & req->events; in aio_poll() [all …]
|
| /fs/ecryptfs/ |
| A D | kthread.c | 46 struct ecryptfs_open_req *req; in ecryptfs_threadfn() local 61 list_del(&req->kthread_ctl_list); in ecryptfs_threadfn() 62 *req->lower_file = dentry_open(&req->path, in ecryptfs_threadfn() 64 complete(&req->done); in ecryptfs_threadfn() 99 complete(&req->done); in ecryptfs_destroy_kthread() 122 struct ecryptfs_open_req req; in ecryptfs_privileged_open() local 126 init_completion(&req.done); in ecryptfs_privileged_open() 127 req.lower_file = lower_file; in ecryptfs_privileged_open() 128 req.path.dentry = lower_dentry; in ecryptfs_privileged_open() 129 req.path.mnt = lower_mnt; in ecryptfs_privileged_open() [all …]
|
| /fs/ocfs2/ |
| A D | ioctl.c | 149 if (o2info_from_user(oib, req)) in ocfs2_info_handle_blocksize() 156 if (o2info_to_user(oib, req)) in ocfs2_info_handle_blocksize() 168 if (o2info_from_user(oic, req)) in ocfs2_info_handle_clustersize() 175 if (o2info_to_user(oic, req)) in ocfs2_info_handle_clustersize() 187 if (o2info_from_user(oim, req)) in ocfs2_info_handle_maxslots() 194 if (o2info_to_user(oim, req)) in ocfs2_info_handle_maxslots() 206 if (o2info_from_user(oil, req)) in ocfs2_info_handle_label() 213 if (o2info_to_user(oil, req)) in ocfs2_info_handle_label() 232 if (o2info_to_user(oiu, req)) in ocfs2_info_handle_uuid() 253 if (o2info_to_user(oif, req)) in ocfs2_info_handle_fs_features() [all …]
|