| /fs/nfs/blocklayout/ |
| A D | dev.c | 301 dev = bl_resolve_deviceid(server, v, gfp_mask); in bl_parse_simple() 463 sizeof(struct pnfs_block_dev), gfp_mask); in bl_parse_concat() 469 volumes, v->concat.volumes[i], gfp_mask); in bl_parse_concat() 492 sizeof(struct pnfs_block_dev), gfp_mask); in bl_parse_stripe() 498 volumes, v->stripe.volumes[i], gfp_mask); in bl_parse_stripe() 522 return bl_parse_slice(server, d, volumes, idx, gfp_mask); in bl_parse_deviceid() 528 return bl_parse_scsi(server, d, volumes, idx, gfp_mask); in bl_parse_deviceid() 537 gfp_t gfp_mask) in bl_alloc_deviceid_node() argument 548 scratch = alloc_page(gfp_mask); in bl_alloc_deviceid_node() 561 gfp_mask); in bl_alloc_deviceid_node() [all …]
|
| A D | blocklayout.c | 564 gfp_t gfp_mask) in bl_find_get_deviceid() argument 570 node = nfs4_find_get_deviceid(server, id, cred, gfp_mask); in bl_find_get_deviceid() 611 gfp_t gfp_mask) in bl_alloc_extent() argument 630 lo->plh_lc_cred, gfp_mask); in bl_alloc_extent() 667 gfp_t gfp_mask) in bl_alloc_lseg() argument 687 lseg = kzalloc(sizeof(*lseg), gfp_mask); in bl_alloc_lseg() 692 scratch = alloc_page(gfp_mask); in bl_alloc_lseg() 713 status = bl_alloc_extent(&xdr, lo, &lv, &extents, gfp_mask); in bl_alloc_lseg()
|
| A D | blocklayout.h | 183 struct pnfs_device *pdev, gfp_t gfp_mask); 200 struct pnfs_block_volume *b, gfp_t gfp_mask);
|
| A D | rpc_pipefs.c | 53 gfp_t gfp_mask) in bl_resolve_deviceid() argument 76 msg->data = kzalloc(msg->len, gfp_mask); in bl_resolve_deviceid()
|
| /fs/btrfs/ |
| A D | ulist.h | 50 struct ulist *ulist_alloc(gfp_t gfp_mask); 53 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask); 55 u64 *old_aux, gfp_t gfp_mask); 60 void **old_aux, gfp_t gfp_mask) in ulist_add_merge_ptr() argument 64 int ret = ulist_add_merge(ulist, val, (uintptr_t)aux, &old64, gfp_mask); in ulist_add_merge_ptr() 68 return ulist_add_merge(ulist, val, (u64)aux, (u64 *)old_aux, gfp_mask); in ulist_add_merge_ptr()
|
| A D | ulist.c | 99 struct ulist *ulist_alloc(gfp_t gfp_mask) in ulist_alloc() argument 101 struct ulist *ulist = kmalloc(sizeof(*ulist), gfp_mask); in ulist_alloc() 111 void ulist_prealloc(struct ulist *ulist, gfp_t gfp_mask) in ulist_prealloc() argument 114 ulist->prealloc = kzalloc(sizeof(*ulist->prealloc), gfp_mask); in ulist_prealloc() 200 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask) in ulist_add() argument 202 return ulist_add_merge(ulist, val, aux, NULL, gfp_mask); in ulist_add() 206 u64 *old_aux, gfp_t gfp_mask) in ulist_add_merge() argument 222 node = kmalloc(sizeof(*node), gfp_mask); in ulist_add_merge()
|
| A D | extent_io.h | 212 static inline void extent_changeset_prealloc(struct extent_changeset *changeset, gfp_t gfp_mask) in extent_changeset_prealloc() argument 214 ulist_prealloc(&changeset->range_changed, gfp_mask); in extent_changeset_prealloc()
|
| A D | backref.c | 391 struct share_check *sc, gfp_t gfp_mask) in add_prelim_ref() argument 398 ref = kmem_cache_alloc(btrfs_prelim_ref_cache, gfp_mask); in add_prelim_ref() 421 struct share_check *sc, gfp_t gfp_mask) in add_direct_ref() argument 424 parent, wanted_disk_byte, count, sc, gfp_mask); in add_direct_ref() 432 struct share_check *sc, gfp_t gfp_mask) in add_indirect_ref() argument 439 wanted_disk_byte, count, sc, gfp_mask); in add_indirect_ref()
|
| /fs/crypto/ |
| A D | inline_crypt.c | 303 u64 first_lblk, gfp_t gfp_mask) in fscrypt_set_bio_crypt_ctx() argument 313 bio_crypt_set_ctx(bio, ci->ci_enc_key.blk_key, dun, gfp_mask); in fscrypt_set_bio_crypt_ctx() 353 gfp_t gfp_mask) in fscrypt_set_bio_crypt_ctx_bh() argument 359 fscrypt_set_bio_crypt_ctx(bio, inode, first_lblk, gfp_mask); in fscrypt_set_bio_crypt_ctx_bh()
|
| /fs/xfs/ |
| A D | xfs_buf.c | 130 gfp_t gfp_mask) in xfs_buf_alloc_kmem() argument 135 bp->b_addr = kmalloc(size, gfp_mask | __GFP_NOFAIL); in xfs_buf_alloc_kmem() 188 gfp_mask |= __GFP_ZERO; in xfs_buf_alloc_backing_mem() 191 gfp_mask |= __GFP_NORETRY; in xfs_buf_alloc_backing_mem() 200 return xfs_buf_alloc_kmem(bp, size, gfp_mask); in xfs_buf_alloc_backing_mem() 207 gfp_mask |= __GFP_NOFAIL; in xfs_buf_alloc_backing_mem() 223 gfp_mask &= ~__GFP_DIRECT_RECLAIM; in xfs_buf_alloc_backing_mem() 224 gfp_mask |= __GFP_NORETRY; in xfs_buf_alloc_backing_mem() 226 folio = folio_alloc(gfp_mask, get_order(size)); in xfs_buf_alloc_backing_mem() 239 bp->b_addr = __vmalloc(size, gfp_mask); in xfs_buf_alloc_backing_mem() [all …]
|
| A D | xfs_iops.c | 1419 gfp_t gfp_mask; in xfs_setup_inode() local 1461 gfp_mask = mapping_gfp_mask(inode->i_mapping); in xfs_setup_inode() 1462 mapping_set_gfp_mask(inode->i_mapping, (gfp_mask & ~(__GFP_FS))); in xfs_setup_inode()
|
| /fs/nfs/ |
| A D | nfs4session.c | 105 u32 slotid, u32 seq_init, gfp_t gfp_mask) in nfs4_new_slot() argument 109 slot = kzalloc(sizeof(*slot), gfp_mask); in nfs4_new_slot() 121 u32 slotid, u32 seq_init, gfp_t gfp_mask) in nfs4_find_or_create_slot() argument 129 seq_init, gfp_mask); in nfs4_find_or_create_slot()
|
| A D | pnfs_dev.c | 187 gfp_t gfp_mask) in nfs4_find_get_deviceid() argument 196 new = nfs4_get_device_info(server, id, cred, gfp_mask); in nfs4_find_get_deviceid()
|
| A D | nfs4_fs.h | 306 extern int nfs4_do_close(struct nfs4_state *state, gfp_t gfp_mask, int wait); 524 extern struct nfs_seqid *nfs_alloc_seqid(struct nfs_seqid_counter *counter, gfp_t gfp_mask);
|
| A D | nfs4state.c | 775 fmode_t fmode, gfp_t gfp_mask, int wait) in __nfs4_close() argument 816 nfs4_do_close(state, gfp_mask, wait); in __nfs4_close() 1065 struct nfs_seqid *nfs_alloc_seqid(struct nfs_seqid_counter *counter, gfp_t gfp_mask) in nfs_alloc_seqid() argument 1069 new = kmalloc(sizeof(*new), gfp_mask); in nfs_alloc_seqid()
|
| /fs/jbd2/ |
| A D | revoke.c | 141 gfp_t gfp_mask = GFP_NOFS; in insert_revoke_hash() local 144 gfp_mask |= __GFP_NOFAIL; in insert_revoke_hash() 145 record = kmem_cache_alloc(jbd2_revoke_record_cache, gfp_mask); in insert_revoke_hash()
|
| A D | transaction.c | 313 gfp_t gfp_mask) in start_this_handle() argument 349 if ((gfp_mask & __GFP_FS) == 0) in start_this_handle() 350 gfp_mask |= __GFP_NOFAIL; in start_this_handle() 352 gfp_mask); in start_this_handle() 467 int revoke_records, gfp_t gfp_mask, in jbd2__journal_start() argument 501 err = start_this_handle(journal, handle, gfp_mask); in jbd2__journal_start() 760 gfp_t gfp_mask) in jbd2__journal_restart() argument 796 ret = start_this_handle(journal, handle, gfp_mask); in jbd2__journal_restart()
|
| /fs/nilfs2/ |
| A D | mdt.h | 78 int nilfs_mdt_init(struct inode *inode, gfp_t gfp_mask, size_t objsz);
|
| A D | mdt.c | 452 int nilfs_mdt_init(struct inode *inode, gfp_t gfp_mask, size_t objsz) in nilfs_mdt_init() argument 464 mapping_set_gfp_mask(inode->i_mapping, gfp_mask); in nilfs_mdt_init()
|
| /fs/gfs2/ |
| A D | inode.h | 15 bool gfs2_release_folio(struct folio *folio, gfp_t gfp_mask);
|
| /fs/jfs/ |
| A D | jfs_metapage.c | 262 static inline struct metapage *alloc_metapage(gfp_t gfp_mask) in alloc_metapage() argument 264 struct metapage *mp = mempool_alloc(metapage_mempool, gfp_mask); in alloc_metapage() 612 static bool metapage_release_folio(struct folio *folio, gfp_t gfp_mask) in metapage_release_folio() argument
|
| /fs/bcachefs/ |
| A D | sysfs.c | 417 sc.gfp_mask = GFP_KERNEL; in STORE() 425 sc.gfp_mask = GFP_KERNEL; in STORE()
|
| A D | util.c | 633 int bch2_bio_alloc_pages(struct bio *bio, size_t size, gfp_t gfp_mask) in bch2_bio_alloc_pages() argument 636 struct page *page = alloc_pages(gfp_mask, 0); in bch2_bio_alloc_pages()
|
| /fs/ntfs3/ |
| A D | frecord.c | 2040 gfp_t gfp_mask; in ni_readpage_cmpr() local 2071 gfp_mask = mapping_gfp_mask(mapping); in ni_readpage_cmpr() 2077 pg = find_or_create_page(mapping, index, gfp_mask); in ni_readpage_cmpr() 2117 gfp_t gfp_mask = mapping_gfp_mask(mapping); in ni_decompress_file() local 2178 pg = find_or_create_page(mapping, index, gfp_mask); in ni_decompress_file()
|
| A D | file.c | 895 gfp_t gfp_mask = mapping_gfp_mask(mapping); in ntfs_get_frame_pages() local 905 gfp_mask); in ntfs_get_frame_pages()
|