Lines Matching refs:gfpflags

2877 static inline bool pfmemalloc_match(struct slab *slab, gfp_t gfpflags);
3582 slab_out_of_memory(struct kmem_cache *s, gfp_t gfpflags, int nid) in slab_out_of_memory() argument
3590 if ((gfpflags & __GFP_NOWARN) || !__ratelimit(&slub_oom_rs)) in slab_out_of_memory()
3594 cpu, cpu_to_node(cpu), nid, gfpflags, &gfpflags); in slab_out_of_memory()
3618 slab_out_of_memory(struct kmem_cache *s, gfp_t gfpflags, int nid) { } in slab_out_of_memory() argument
3621 static inline bool pfmemalloc_match(struct slab *slab, gfp_t gfpflags) in pfmemalloc_match() argument
3624 return gfp_pfmemalloc_allowed(gfpflags); in pfmemalloc_match()
3721 static void *___slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, in ___slab_alloc() argument
3764 if (unlikely(!pfmemalloc_match(slab, gfpflags))) in ___slab_alloc()
3837 pfmemalloc_match(slab, gfpflags))) { in ___slab_alloc()
3854 pc.flags = gfpflags; in ___slab_alloc()
3866 if (unlikely(node != NUMA_NO_NODE && !(gfpflags & __GFP_THISNODE) in ___slab_alloc()
3895 if (node != NUMA_NO_NODE && !(gfpflags & __GFP_THISNODE) in ___slab_alloc()
3900 slab_out_of_memory(s, gfpflags, node); in ___slab_alloc()
3929 if (unlikely(!pfmemalloc_match(slab, gfpflags))) { in ___slab_alloc()
3967 static void *__slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, in __slab_alloc() argument
3981 p = ___slab_alloc(s, gfpflags, node, addr, c, orig_size); in __slab_alloc()
3989 gfp_t gfpflags, int node, unsigned long addr, size_t orig_size) in __slab_alloc_node() argument
4056 object = __slab_alloc(s, gfpflags, node, addr, c, orig_size); in __slab_alloc_node()
4086 gfp_t gfpflags, int node, unsigned long addr, size_t orig_size) in __slab_alloc_node() argument
4092 pc.flags = gfpflags; in __slab_alloc_node()
4099 slab = new_slab(s, gfpflags, node); in __slab_alloc_node()
4101 slab_out_of_memory(s, gfpflags, node); in __slab_alloc_node()
4204 gfp_t gfpflags, int node, unsigned long addr, size_t orig_size) in slab_alloc_node() argument
4209 s = slab_pre_alloc_hook(s, gfpflags); in slab_alloc_node()
4213 object = kfence_alloc(s, orig_size, gfpflags); in slab_alloc_node()
4217 object = __slab_alloc_node(s, gfpflags, node, addr, orig_size); in slab_alloc_node()
4220 init = slab_want_init_on_alloc(gfpflags, s); in slab_alloc_node()
4229 slab_post_alloc_hook(s, lru, gfpflags, 1, &object, init, orig_size); in slab_alloc_node()
4234 void *kmem_cache_alloc_noprof(struct kmem_cache *s, gfp_t gfpflags) in kmem_cache_alloc_noprof() argument
4236 void *ret = slab_alloc_node(s, NULL, gfpflags, NUMA_NO_NODE, _RET_IP_, in kmem_cache_alloc_noprof()
4239 trace_kmem_cache_alloc(_RET_IP_, ret, s, gfpflags, NUMA_NO_NODE); in kmem_cache_alloc_noprof()
4246 gfp_t gfpflags) in kmem_cache_alloc_lru_noprof() argument
4248 void *ret = slab_alloc_node(s, lru, gfpflags, NUMA_NO_NODE, _RET_IP_, in kmem_cache_alloc_lru_noprof()
4251 trace_kmem_cache_alloc(_RET_IP_, ret, s, gfpflags, NUMA_NO_NODE); in kmem_cache_alloc_lru_noprof()
4257 bool kmem_cache_charge(void *objp, gfp_t gfpflags) in kmem_cache_charge() argument
4262 return memcg_slab_post_charge(objp, gfpflags); in kmem_cache_charge()
4279 void *kmem_cache_alloc_node_noprof(struct kmem_cache *s, gfp_t gfpflags, int node) in kmem_cache_alloc_node_noprof() argument
4281 void *ret = slab_alloc_node(s, NULL, gfpflags, node, _RET_IP_, s->object_size); in kmem_cache_alloc_node_noprof()
4283 trace_kmem_cache_alloc(_RET_IP_, ret, s, gfpflags, node); in kmem_cache_alloc_node_noprof()
4389 void *__kmalloc_cache_noprof(struct kmem_cache *s, gfp_t gfpflags, size_t size) in __kmalloc_cache_noprof() argument
4391 void *ret = slab_alloc_node(s, NULL, gfpflags, NUMA_NO_NODE, in __kmalloc_cache_noprof()
4394 trace_kmalloc(_RET_IP_, ret, size, s->size, gfpflags, NUMA_NO_NODE); in __kmalloc_cache_noprof()
4396 ret = kasan_kmalloc(s, ret, size, gfpflags); in __kmalloc_cache_noprof()
4401 void *__kmalloc_cache_node_noprof(struct kmem_cache *s, gfp_t gfpflags, in __kmalloc_cache_node_noprof() argument
4404 void *ret = slab_alloc_node(s, NULL, gfpflags, node, _RET_IP_, size); in __kmalloc_cache_node_noprof()
4406 trace_kmalloc(_RET_IP_, ret, size, s->size, gfpflags, node); in __kmalloc_cache_node_noprof()
4408 ret = kasan_kmalloc(s, ret, size, gfpflags); in __kmalloc_cache_node_noprof()