Home
last modified time | relevance | path

Searched refs:gfp (Results 1 – 25 of 208) sorted by relevance

123456789

/drivers/base/
A Ddevres.c111 size_t size, gfp_t gfp, int nid) in alloc_dr() argument
119 dr = kmalloc_node_track_caller(tot_size, gfp, nid); in alloc_dr()
124 if (!(gfp & __GFP_ZERO)) in alloc_dr()
557 grp = kmalloc(sizeof(*grp), gfp); in devres_open_group()
906 return devm_kmalloc(dev, new_size, gfp); in devm_krealloc()
939 new_size, gfp, dev_to_node(dev)); in devm_krealloc()
993 return devm_kmemdup(dev, s, strlen(s) + 1, gfp); in devm_kstrdup()
1015 return devm_kstrdup(dev, s, gfp); in devm_kstrdup_const()
1041 p = devm_kmalloc(dev, len+1, gfp); in devm_kvasprintf()
1068 p = devm_kvasprintf(dev, gfp, fmt, ap); in devm_kasprintf()
[all …]
A Ddevcoredump.c249 gfp_t gfp) in dev_coredumpv() argument
251 dev_coredumpm(dev, NULL, data, datalen, gfp, devcd_readv, devcd_freev); in dev_coredumpv()
343 void *data, size_t datalen, gfp_t gfp, in dev_coredumpm_timeout() argument
367 devcd = kzalloc(sizeof(*devcd), gfp); in dev_coredumpm_timeout()
432 size_t datalen, gfp_t gfp) in dev_coredumpsg() argument
434 dev_coredumpm(dev, NULL, table, datalen, gfp, devcd_read_from_sgtable, in dev_coredumpsg()
/drivers/iommu/
A Ds390-iommu.c320 dma_addr_t dma_addr, gfp_t gfp) in dma_walk_rs_table() argument
331 rto = dma_alloc_cpu_table(gfp); in dma_walk_rs_table()
350 dma_addr_t dma_addr, gfp_t gfp) in dma_walk_rf_table() argument
361 rso = dma_alloc_cpu_table(gfp); in dma_walk_rf_table()
392 sto = dma_alloc_cpu_table(gfp); in dma_get_seg_table_origin()
419 pto = dma_alloc_page_table(gfp); in dma_get_page_table_origin()
437 dma_addr_t dma_addr, gfp_t gfp) in dma_walk_region_tables() argument
452 dma_addr_t dma_addr, gfp_t gfp) in dma_walk_cpu_trans() argument
841 gfp_t gfp) in s390_iommu_validate_trans() argument
896 int prot, gfp_t gfp, size_t *mapped) in s390_iommu_map_pages() argument
[all …]
A Diommu-pages.h49 void *iommu_alloc_pages_node_sz(int nid, gfp_t gfp, size_t size);
96 static inline void *iommu_alloc_pages_sz(gfp_t gfp, size_t size) in iommu_alloc_pages_sz() argument
98 return iommu_alloc_pages_node_sz(NUMA_NO_NODE, gfp, size); in iommu_alloc_pages_sz()
A Diommu-pages.c37 void *iommu_alloc_pages_node_sz(int nid, gfp_t gfp, size_t size) in iommu_alloc_pages_node_sz() argument
44 if (WARN_ON(gfp & __GFP_HIGHMEM)) in iommu_alloc_pages_node_sz()
59 folio = __folio_alloc_node(gfp | __GFP_ZERO, order, nid); in iommu_alloc_pages_node_sz()
A Ddma-iommu.c883 gfp |= __GFP_NOWARN | __GFP_HIGHMEM; in __iommu_dma_alloc_pages()
897 gfp_t alloc_flags = gfp; in __iommu_dma_alloc_pages()
953 gfp); in __iommu_dma_alloc_noncontiguous()
981 gfp); in __iommu_dma_alloc_noncontiguous()
1043 sh = kmalloc(sizeof(*sh), gfp); in iommu_dma_alloc_noncontiguous()
1610 page = dma_alloc_contiguous(dev, alloc_size, gfp); in iommu_dma_alloc_pages()
1639 gfp_t gfp, unsigned long attrs) in iommu_dma_alloc() argument
1646 gfp |= __GFP_ZERO; in iommu_dma_alloc()
1648 if (gfpflags_allow_blocking(gfp) && in iommu_dma_alloc()
1654 !gfpflags_allow_blocking(gfp) && !coherent) in iommu_dma_alloc()
[all …]
/drivers/staging/rtl8723bs/include/
A Dioctl_cfg80211.h112 #define rtw_cfg80211_rx_mgmt(adapter, freq, sig_dbm, buf, len, gfp) cfg80211_rx_mgmt((adapter)->rtw… argument
114 …tatus(adapter, cookie, buf, len, ack, gfp) cfg80211_mgmt_tx_status((adapter)->rtw_wdev, cookie, bu… argument
115 …okie, chan, channel_type, duration, gfp) cfg80211_ready_on_channel((adapter)->rtw_wdev, cookie, c… argument
116 …ed(adapter, cookie, chan, chan_type, gfp) cfg80211_remain_on_channel_expired((adapter)->rtw_wdev, … argument
/drivers/gpu/drm/i915/
A Di915_sw_fence.c342 wait_queue_entry_t *wq, gfp_t gfp) in __i915_sw_fence_await_sw_fence() argument
348 might_sleep_if(gfpflags_allow_blocking(gfp)); in __i915_sw_fence_await_sw_fence()
363 wq = kmalloc(sizeof(*wq), gfp); in __i915_sw_fence_await_sw_fence()
365 if (!gfpflags_allow_blocking(gfp)) in __i915_sw_fence_await_sw_fence()
405 gfp_t gfp) in i915_sw_fence_await_sw_fence_gfp() argument
484 gfp_t gfp) in i915_sw_fence_await_dma_fence() argument
491 might_sleep_if(gfpflags_allow_blocking(gfp)); in i915_sw_fence_await_dma_fence()
501 gfp); in i915_sw_fence_await_dma_fence()
503 if (!gfpflags_allow_blocking(gfp)) in i915_sw_fence_await_dma_fence()
583 gfp_t gfp) in i915_sw_fence_await_reservation() argument
[all …]
A Di915_gpu_error.h283 i915_gpu_coredump_alloc(struct drm_i915_private *i915, gfp_t gfp);
286 intel_gt_coredump_alloc(struct intel_gt *gt, gfp_t gfp, u32 dump_flags);
289 intel_engine_coredump_alloc(struct intel_engine_cs *engine, gfp_t gfp, u32 dump_flags);
294 gfp_t gfp);
347 i915_gpu_coredump_alloc(struct drm_i915_private *i915, gfp_t gfp) in i915_gpu_coredump_alloc() argument
353 intel_gt_coredump_alloc(struct intel_gt *gt, gfp_t gfp, u32 dump_flags) in intel_gt_coredump_alloc() argument
359 intel_engine_coredump_alloc(struct intel_engine_cs *engine, gfp_t gfp, u32 dump_flags) in intel_engine_coredump_alloc() argument
367 gfp_t gfp) in intel_engine_coredump_add_request() argument
A Di915_deps.c50 void i915_deps_init(struct i915_deps *deps, gfp_t gfp) in i915_deps_init() argument
53 deps->gfp = gfp; in i915_deps_init()
85 new_fences = kmalloc_array(new_size, sizeof(*new_fences), deps->gfp); in i915_deps_grow()
A Di915_sw_fence.h77 gfp_t gfp);
90 gfp_t gfp);
96 gfp_t gfp);
A Di915_deps.h29 gfp_t gfp; member
32 void i915_deps_init(struct i915_deps *deps, gfp_t gfp);
A Di915_gpu_error.c201 folio = folio_alloc(gfp, 0); in pool_refill()
217 err = pool_refill(fbatch, gfp); in pool_init()
228 folio = folio_alloc(gfp, 0); in pool_alloc()
1435 c = kmalloc(sizeof(*c), gfp); in capture_vma_snapshot()
1455 gfp_t gfp) in capture_vma() argument
1476 gfp_t gfp) in capture_user() argument
1531 ee = kzalloc(sizeof(*ee), gfp); in intel_engine_coredump_alloc()
1548 gfp_t gfp) in engine_coredump_add_context() argument
1570 gfp_t gfp) in intel_engine_coredump_add_request() argument
1584 vma = capture_user(vma, rq, gfp); in intel_engine_coredump_add_request()
[all …]
/drivers/gpu/drm/
A Ddrm_managed.c88 size_t size, gfp_t gfp, int nid) in alloc_dr() argument
97 dr = kmalloc_node_track_caller(tot_size, gfp, nid); in alloc_dr()
229 void *drmm_kmalloc(struct drm_device *dev, size_t size, gfp_t gfp) in drmm_kmalloc() argument
233 dr = alloc_dr(NULL, size, gfp, dev_to_node(dev->dev)); in drmm_kmalloc()
236 size, gfp); in drmm_kmalloc()
239 dr->node.name = kstrdup_const("kmalloc", gfp); in drmm_kmalloc()
257 char *drmm_kstrdup(struct drm_device *dev, const char *s, gfp_t gfp) in drmm_kstrdup() argument
266 buf = drmm_kmalloc(dev, size, gfp); in drmm_kstrdup()
/drivers/firmware/qcom/
A Dqcom_tzmem.c153 size_t size, gfp_t gfp) in qcom_tzmem_pool_add_memory() argument
158 gfp); in qcom_tzmem_pool_add_memory()
165 &area->paddr, gfp); in qcom_tzmem_pool_add_memory()
334 size_t requested, gfp_t gfp) in qcom_tzmem_try_grow_pool() argument
351 return !qcom_tzmem_pool_add_memory(pool, requested, gfp); in qcom_tzmem_try_grow_pool()
366 void *qcom_tzmem_alloc(struct qcom_tzmem_pool *pool, size_t size, gfp_t gfp) in qcom_tzmem_alloc() argument
377 gfp); in qcom_tzmem_alloc()
384 if (qcom_tzmem_try_grow_pool(pool, size, gfp)) in qcom_tzmem_alloc()
/drivers/gpu/drm/i915/gem/
A Di915_gem_internal.c40 gfp_t gfp; in i915_gem_object_get_pages_internal() local
49 gfp = GFP_KERNEL | __GFP_HIGHMEM | __GFP_RECLAIMABLE; in i915_gem_object_get_pages_internal()
52 gfp &= ~__GFP_HIGHMEM; in i915_gem_object_get_pages_internal()
53 gfp |= __GFP_DMA32; in i915_gem_object_get_pages_internal()
74 page = alloc_pages(gfp | (order ? QUIET : MAYFAIL), in i915_gem_object_get_pages_internal()
/drivers/crypto/hisilicon/sec/
A Dsec_algs.c179 gfp_t gfp) in sec_alloc_and_fill_hw_sgl() argument
194 gfp, &sgl_next_dma); in sec_alloc_and_fill_hw_sgl()
549 int *steps, gfp_t gfp) in sec_alg_alloc_and_calc_split_sizes() argument
572 struct device *dev, gfp_t gfp) in sec_map_and_split_sg() argument
593 *splits, *splits_nents, gfp); in sec_map_and_split_sg()
640 el = kzalloc(sizeof(*el), gfp); in sec_alg_alloc_and_fill_el()
672 n_ents_in, info, gfp); in sec_alg_alloc_and_fill_el()
683 n_ents_out, info, gfp); in sec_alg_alloc_and_fill_el()
733 &steps, gfp); in sec_alg_skcipher_crypto()
739 info->dev, gfp); in sec_alg_skcipher_crypto()
[all …]
/drivers/iommu/amd/
A Dio_pgtable.c108 gfp_t gfp) in increase_address_space() argument
117 pte = iommu_alloc_pages_node_sz(cfg->amd.nid, gfp, SZ_4K); in increase_address_space()
151 gfp_t gfp, in alloc_pte() argument
168 PAGE_SIZE_LEVEL(page_size), gfp)) in alloc_pte()
209 page = iommu_alloc_pages_node_sz(cfg->amd.nid, gfp, in alloc_pte()
323 int prot, gfp_t gfp, size_t *mapped) in iommu_v1_map_pages() argument
342 pte = alloc_pte(pgtable, iova, pgsize, NULL, gfp, &updated); in iommu_v1_map_pages()
/drivers/md/bcache/
A Dutil.h39 #define init_heap(heap, _size, gfp) \ argument
45 (heap)->data = kvmalloc(_bytes, (gfp) & GFP_KERNEL); \
124 #define __init_fifo(fifo, gfp) \ argument
135 (fifo)->data = kvmalloc(_bytes, (gfp) & GFP_KERNEL); \
139 #define init_fifo_exact(fifo, _size, gfp) \ argument
142 __init_fifo(fifo, gfp); \
145 #define init_fifo(fifo, _size, gfp) \ argument
150 __init_fifo(fifo, gfp); \
/drivers/crypto/cavium/nitrox/
A Dnitrox_req.h70 gfp_t gfp; member
517 gfp_t gfp; member
555 static inline void *alloc_req_buf(int nents, int extralen, gfp_t gfp) in alloc_req_buf() argument
562 return kzalloc(size, gfp); in alloc_req_buf()
625 nkreq->src = alloc_req_buf(nents, ivsize, creq->gfp); in alloc_src_req_buf()
672 nkreq->dst = alloc_req_buf(nents, extralen, creq->gfp); in alloc_dst_req_buf()
/drivers/staging/media/ipu3/
A Dipu3-dmamap.c34 static struct page **imgu_dmamap_alloc_buffer(size_t size, gfp_t gfp) in imgu_dmamap_alloc_buffer() argument
47 gfp |= __GFP_HIGHMEM | __GFP_ZERO; in imgu_dmamap_alloc_buffer()
59 gfp | high_order_gfp : gfp, order); in imgu_dmamap_alloc_buffer()
/drivers/greybus/
A Doperation.c215 static int gb_message_send(struct gb_message *message, gfp_t gfp) in gb_message_send() argument
223 gfp); in gb_message_send()
470 size_t response_size, gfp_t gfp) in gb_operation_response_alloc() argument
478 response = gb_operation_message_alloc(hd, type, response_size, gfp); in gb_operation_response_alloc()
579 gfp_t gfp) in gb_operation_create_flags() argument
593 flags, gfp); in gb_operation_create_flags()
605 gfp_t gfp) in gb_operation_create_core() argument
613 flags, gfp); in gb_operation_create_core()
724 gfp_t gfp) in gb_operation_request_send() argument
769 ret = gb_message_send(operation->request, gfp); in gb_operation_request_send()
/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
A Dmem.c155 gfp_t gfp = GFP_USER | __GFP_ZERO; in nvkm_mem_new_host() local
200 gfp |= GFP_HIGHUSER; in nvkm_mem_new_host()
202 gfp |= GFP_DMA32; in nvkm_mem_new_host()
205 struct page *p = alloc_page(gfp); in nvkm_mem_new_host()
/drivers/virtio/
A Dvirtio_ring.c477 gfp_t gfp) in alloc_indirect_split() argument
537 gfp_t gfp) in virtqueue_add_split() argument
1330 gfp_t gfp) in virtqueue_add_indirect_packed() argument
1456 gfp_t gfp) in virtqueue_add_packed() argument
2276 gfp_t gfp) in virtqueue_add() argument
2309 gfp_t gfp) in virtqueue_add_sgs() argument
2341 gfp_t gfp) in virtqueue_add_outbuf() argument
2364 gfp_t gfp) in virtqueue_add_outbuf_premapped() argument
2386 gfp_t gfp) in virtqueue_add_inbuf() argument
2410 gfp_t gfp) in virtqueue_add_inbuf_ctx() argument
[all …]
/drivers/media/pci/intel/ipu6/
A Dipu6-dma.c58 static struct page **__alloc_buffer(size_t size, gfp_t gfp, unsigned long attrs) in __alloc_buffer() argument
69 gfp |= __GFP_NOWARN; in __alloc_buffer()
74 pages[i] = alloc_pages(gfp, order); in __alloc_buffer()
76 pages[i] = alloc_pages(gfp, --order); in __alloc_buffer()
153 dma_addr_t *dma_handle, gfp_t gfp, in ipu6_dma_alloc() argument
179 pages = __alloc_buffer(size, gfp, attrs); in ipu6_dma_alloc()

Completed in 87 milliseconds

123456789