Home
last modified time | relevance | path

Searched refs:gfp_mask (Results 1 – 25 of 37) sorted by relevance

12

/fs/nfs/blocklayout/
A Ddev.c301 dev = bl_resolve_deviceid(server, v, gfp_mask); in bl_parse_simple()
463 sizeof(struct pnfs_block_dev), gfp_mask); in bl_parse_concat()
469 volumes, v->concat.volumes[i], gfp_mask); in bl_parse_concat()
492 sizeof(struct pnfs_block_dev), gfp_mask); in bl_parse_stripe()
498 volumes, v->stripe.volumes[i], gfp_mask); in bl_parse_stripe()
522 return bl_parse_slice(server, d, volumes, idx, gfp_mask); in bl_parse_deviceid()
528 return bl_parse_scsi(server, d, volumes, idx, gfp_mask); in bl_parse_deviceid()
537 gfp_t gfp_mask) in bl_alloc_deviceid_node() argument
548 scratch = alloc_page(gfp_mask); in bl_alloc_deviceid_node()
561 gfp_mask); in bl_alloc_deviceid_node()
[all …]
A Dblocklayout.c564 gfp_t gfp_mask) in bl_find_get_deviceid() argument
570 node = nfs4_find_get_deviceid(server, id, cred, gfp_mask); in bl_find_get_deviceid()
611 gfp_t gfp_mask) in bl_alloc_extent() argument
630 lo->plh_lc_cred, gfp_mask); in bl_alloc_extent()
667 gfp_t gfp_mask) in bl_alloc_lseg() argument
687 lseg = kzalloc(sizeof(*lseg), gfp_mask); in bl_alloc_lseg()
692 scratch = alloc_page(gfp_mask); in bl_alloc_lseg()
713 status = bl_alloc_extent(&xdr, lo, &lv, &extents, gfp_mask); in bl_alloc_lseg()
A Dblocklayout.h183 struct pnfs_device *pdev, gfp_t gfp_mask);
200 struct pnfs_block_volume *b, gfp_t gfp_mask);
A Drpc_pipefs.c53 gfp_t gfp_mask) in bl_resolve_deviceid() argument
76 msg->data = kzalloc(msg->len, gfp_mask); in bl_resolve_deviceid()
/fs/btrfs/
A Dulist.h50 struct ulist *ulist_alloc(gfp_t gfp_mask);
53 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask);
55 u64 *old_aux, gfp_t gfp_mask);
60 void **old_aux, gfp_t gfp_mask) in ulist_add_merge_ptr() argument
64 int ret = ulist_add_merge(ulist, val, (uintptr_t)aux, &old64, gfp_mask); in ulist_add_merge_ptr()
68 return ulist_add_merge(ulist, val, (u64)aux, (u64 *)old_aux, gfp_mask); in ulist_add_merge_ptr()
A Dulist.c99 struct ulist *ulist_alloc(gfp_t gfp_mask) in ulist_alloc() argument
101 struct ulist *ulist = kmalloc(sizeof(*ulist), gfp_mask); in ulist_alloc()
111 void ulist_prealloc(struct ulist *ulist, gfp_t gfp_mask) in ulist_prealloc() argument
114 ulist->prealloc = kzalloc(sizeof(*ulist->prealloc), gfp_mask); in ulist_prealloc()
200 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask) in ulist_add() argument
202 return ulist_add_merge(ulist, val, aux, NULL, gfp_mask); in ulist_add()
206 u64 *old_aux, gfp_t gfp_mask) in ulist_add_merge() argument
222 node = kmalloc(sizeof(*node), gfp_mask); in ulist_add_merge()
A Dextent_io.h212 static inline void extent_changeset_prealloc(struct extent_changeset *changeset, gfp_t gfp_mask) in extent_changeset_prealloc() argument
214 ulist_prealloc(&changeset->range_changed, gfp_mask); in extent_changeset_prealloc()
A Dbackref.c391 struct share_check *sc, gfp_t gfp_mask) in add_prelim_ref() argument
398 ref = kmem_cache_alloc(btrfs_prelim_ref_cache, gfp_mask); in add_prelim_ref()
421 struct share_check *sc, gfp_t gfp_mask) in add_direct_ref() argument
424 parent, wanted_disk_byte, count, sc, gfp_mask); in add_direct_ref()
432 struct share_check *sc, gfp_t gfp_mask) in add_indirect_ref() argument
439 wanted_disk_byte, count, sc, gfp_mask); in add_indirect_ref()
/fs/crypto/
A Dinline_crypt.c303 u64 first_lblk, gfp_t gfp_mask) in fscrypt_set_bio_crypt_ctx() argument
313 bio_crypt_set_ctx(bio, ci->ci_enc_key.blk_key, dun, gfp_mask); in fscrypt_set_bio_crypt_ctx()
353 gfp_t gfp_mask) in fscrypt_set_bio_crypt_ctx_bh() argument
359 fscrypt_set_bio_crypt_ctx(bio, inode, first_lblk, gfp_mask); in fscrypt_set_bio_crypt_ctx_bh()
/fs/xfs/
A Dxfs_buf.c130 gfp_t gfp_mask) in xfs_buf_alloc_kmem() argument
135 bp->b_addr = kmalloc(size, gfp_mask | __GFP_NOFAIL); in xfs_buf_alloc_kmem()
188 gfp_mask |= __GFP_ZERO; in xfs_buf_alloc_backing_mem()
191 gfp_mask |= __GFP_NORETRY; in xfs_buf_alloc_backing_mem()
200 return xfs_buf_alloc_kmem(bp, size, gfp_mask); in xfs_buf_alloc_backing_mem()
207 gfp_mask |= __GFP_NOFAIL; in xfs_buf_alloc_backing_mem()
223 gfp_mask &= ~__GFP_DIRECT_RECLAIM; in xfs_buf_alloc_backing_mem()
224 gfp_mask |= __GFP_NORETRY; in xfs_buf_alloc_backing_mem()
226 folio = folio_alloc(gfp_mask, get_order(size)); in xfs_buf_alloc_backing_mem()
239 bp->b_addr = __vmalloc(size, gfp_mask); in xfs_buf_alloc_backing_mem()
[all …]
A Dxfs_iops.c1419 gfp_t gfp_mask; in xfs_setup_inode() local
1461 gfp_mask = mapping_gfp_mask(inode->i_mapping); in xfs_setup_inode()
1462 mapping_set_gfp_mask(inode->i_mapping, (gfp_mask & ~(__GFP_FS))); in xfs_setup_inode()
/fs/nfs/
A Dnfs4session.c105 u32 slotid, u32 seq_init, gfp_t gfp_mask) in nfs4_new_slot() argument
109 slot = kzalloc(sizeof(*slot), gfp_mask); in nfs4_new_slot()
121 u32 slotid, u32 seq_init, gfp_t gfp_mask) in nfs4_find_or_create_slot() argument
129 seq_init, gfp_mask); in nfs4_find_or_create_slot()
A Dpnfs_dev.c187 gfp_t gfp_mask) in nfs4_find_get_deviceid() argument
196 new = nfs4_get_device_info(server, id, cred, gfp_mask); in nfs4_find_get_deviceid()
A Dnfs4_fs.h306 extern int nfs4_do_close(struct nfs4_state *state, gfp_t gfp_mask, int wait);
524 extern struct nfs_seqid *nfs_alloc_seqid(struct nfs_seqid_counter *counter, gfp_t gfp_mask);
A Dnfs4state.c775 fmode_t fmode, gfp_t gfp_mask, int wait) in __nfs4_close() argument
816 nfs4_do_close(state, gfp_mask, wait); in __nfs4_close()
1065 struct nfs_seqid *nfs_alloc_seqid(struct nfs_seqid_counter *counter, gfp_t gfp_mask) in nfs_alloc_seqid() argument
1069 new = kmalloc(sizeof(*new), gfp_mask); in nfs_alloc_seqid()
/fs/jbd2/
A Drevoke.c141 gfp_t gfp_mask = GFP_NOFS; in insert_revoke_hash() local
144 gfp_mask |= __GFP_NOFAIL; in insert_revoke_hash()
145 record = kmem_cache_alloc(jbd2_revoke_record_cache, gfp_mask); in insert_revoke_hash()
A Dtransaction.c313 gfp_t gfp_mask) in start_this_handle() argument
349 if ((gfp_mask & __GFP_FS) == 0) in start_this_handle()
350 gfp_mask |= __GFP_NOFAIL; in start_this_handle()
352 gfp_mask); in start_this_handle()
467 int revoke_records, gfp_t gfp_mask, in jbd2__journal_start() argument
501 err = start_this_handle(journal, handle, gfp_mask); in jbd2__journal_start()
760 gfp_t gfp_mask) in jbd2__journal_restart() argument
796 ret = start_this_handle(journal, handle, gfp_mask); in jbd2__journal_restart()
/fs/nilfs2/
A Dmdt.h78 int nilfs_mdt_init(struct inode *inode, gfp_t gfp_mask, size_t objsz);
A Dmdt.c452 int nilfs_mdt_init(struct inode *inode, gfp_t gfp_mask, size_t objsz) in nilfs_mdt_init() argument
464 mapping_set_gfp_mask(inode->i_mapping, gfp_mask); in nilfs_mdt_init()
/fs/gfs2/
A Dinode.h15 bool gfs2_release_folio(struct folio *folio, gfp_t gfp_mask);
/fs/jfs/
A Djfs_metapage.c262 static inline struct metapage *alloc_metapage(gfp_t gfp_mask) in alloc_metapage() argument
264 struct metapage *mp = mempool_alloc(metapage_mempool, gfp_mask); in alloc_metapage()
612 static bool metapage_release_folio(struct folio *folio, gfp_t gfp_mask) in metapage_release_folio() argument
/fs/bcachefs/
A Dsysfs.c417 sc.gfp_mask = GFP_KERNEL; in STORE()
425 sc.gfp_mask = GFP_KERNEL; in STORE()
A Dutil.c633 int bch2_bio_alloc_pages(struct bio *bio, size_t size, gfp_t gfp_mask) in bch2_bio_alloc_pages() argument
636 struct page *page = alloc_pages(gfp_mask, 0); in bch2_bio_alloc_pages()
/fs/ntfs3/
A Dfrecord.c2040 gfp_t gfp_mask; in ni_readpage_cmpr() local
2071 gfp_mask = mapping_gfp_mask(mapping); in ni_readpage_cmpr()
2077 pg = find_or_create_page(mapping, index, gfp_mask); in ni_readpage_cmpr()
2117 gfp_t gfp_mask = mapping_gfp_mask(mapping); in ni_decompress_file() local
2178 pg = find_or_create_page(mapping, index, gfp_mask); in ni_decompress_file()
A Dfile.c895 gfp_t gfp_mask = mapping_gfp_mask(mapping); in ntfs_get_frame_pages() local
905 gfp_mask); in ntfs_get_frame_pages()

Completed in 67 milliseconds

12