| /drivers/base/ |
| A D | devres.c | 111 size_t size, gfp_t gfp, int nid) in alloc_dr() argument 119 dr = kmalloc_node_track_caller(tot_size, gfp, nid); in alloc_dr() 124 if (!(gfp & __GFP_ZERO)) in alloc_dr() 557 grp = kmalloc(sizeof(*grp), gfp); in devres_open_group() 906 return devm_kmalloc(dev, new_size, gfp); in devm_krealloc() 939 new_size, gfp, dev_to_node(dev)); in devm_krealloc() 993 return devm_kmemdup(dev, s, strlen(s) + 1, gfp); in devm_kstrdup() 1015 return devm_kstrdup(dev, s, gfp); in devm_kstrdup_const() 1041 p = devm_kmalloc(dev, len+1, gfp); in devm_kvasprintf() 1068 p = devm_kvasprintf(dev, gfp, fmt, ap); in devm_kasprintf() [all …]
|
| A D | devcoredump.c | 249 gfp_t gfp) in dev_coredumpv() argument 251 dev_coredumpm(dev, NULL, data, datalen, gfp, devcd_readv, devcd_freev); in dev_coredumpv() 343 void *data, size_t datalen, gfp_t gfp, in dev_coredumpm_timeout() argument 367 devcd = kzalloc(sizeof(*devcd), gfp); in dev_coredumpm_timeout() 432 size_t datalen, gfp_t gfp) in dev_coredumpsg() argument 434 dev_coredumpm(dev, NULL, table, datalen, gfp, devcd_read_from_sgtable, in dev_coredumpsg()
|
| /drivers/iommu/ |
| A D | s390-iommu.c | 320 dma_addr_t dma_addr, gfp_t gfp) in dma_walk_rs_table() argument 331 rto = dma_alloc_cpu_table(gfp); in dma_walk_rs_table() 350 dma_addr_t dma_addr, gfp_t gfp) in dma_walk_rf_table() argument 361 rso = dma_alloc_cpu_table(gfp); in dma_walk_rf_table() 392 sto = dma_alloc_cpu_table(gfp); in dma_get_seg_table_origin() 419 pto = dma_alloc_page_table(gfp); in dma_get_page_table_origin() 437 dma_addr_t dma_addr, gfp_t gfp) in dma_walk_region_tables() argument 452 dma_addr_t dma_addr, gfp_t gfp) in dma_walk_cpu_trans() argument 841 gfp_t gfp) in s390_iommu_validate_trans() argument 896 int prot, gfp_t gfp, size_t *mapped) in s390_iommu_map_pages() argument [all …]
|
| A D | iommu-pages.h | 49 void *iommu_alloc_pages_node_sz(int nid, gfp_t gfp, size_t size); 96 static inline void *iommu_alloc_pages_sz(gfp_t gfp, size_t size) in iommu_alloc_pages_sz() argument 98 return iommu_alloc_pages_node_sz(NUMA_NO_NODE, gfp, size); in iommu_alloc_pages_sz()
|
| A D | iommu-pages.c | 37 void *iommu_alloc_pages_node_sz(int nid, gfp_t gfp, size_t size) in iommu_alloc_pages_node_sz() argument 44 if (WARN_ON(gfp & __GFP_HIGHMEM)) in iommu_alloc_pages_node_sz() 59 folio = __folio_alloc_node(gfp | __GFP_ZERO, order, nid); in iommu_alloc_pages_node_sz()
|
| A D | dma-iommu.c | 883 gfp |= __GFP_NOWARN | __GFP_HIGHMEM; in __iommu_dma_alloc_pages() 897 gfp_t alloc_flags = gfp; in __iommu_dma_alloc_pages() 953 gfp); in __iommu_dma_alloc_noncontiguous() 981 gfp); in __iommu_dma_alloc_noncontiguous() 1043 sh = kmalloc(sizeof(*sh), gfp); in iommu_dma_alloc_noncontiguous() 1610 page = dma_alloc_contiguous(dev, alloc_size, gfp); in iommu_dma_alloc_pages() 1639 gfp_t gfp, unsigned long attrs) in iommu_dma_alloc() argument 1646 gfp |= __GFP_ZERO; in iommu_dma_alloc() 1648 if (gfpflags_allow_blocking(gfp) && in iommu_dma_alloc() 1654 !gfpflags_allow_blocking(gfp) && !coherent) in iommu_dma_alloc() [all …]
|
| /drivers/staging/rtl8723bs/include/ |
| A D | ioctl_cfg80211.h | 112 #define rtw_cfg80211_rx_mgmt(adapter, freq, sig_dbm, buf, len, gfp) cfg80211_rx_mgmt((adapter)->rtw… argument 114 …tatus(adapter, cookie, buf, len, ack, gfp) cfg80211_mgmt_tx_status((adapter)->rtw_wdev, cookie, bu… argument 115 …okie, chan, channel_type, duration, gfp) cfg80211_ready_on_channel((adapter)->rtw_wdev, cookie, c… argument 116 …ed(adapter, cookie, chan, chan_type, gfp) cfg80211_remain_on_channel_expired((adapter)->rtw_wdev, … argument
|
| /drivers/gpu/drm/i915/ |
| A D | i915_sw_fence.c | 342 wait_queue_entry_t *wq, gfp_t gfp) in __i915_sw_fence_await_sw_fence() argument 348 might_sleep_if(gfpflags_allow_blocking(gfp)); in __i915_sw_fence_await_sw_fence() 363 wq = kmalloc(sizeof(*wq), gfp); in __i915_sw_fence_await_sw_fence() 365 if (!gfpflags_allow_blocking(gfp)) in __i915_sw_fence_await_sw_fence() 405 gfp_t gfp) in i915_sw_fence_await_sw_fence_gfp() argument 484 gfp_t gfp) in i915_sw_fence_await_dma_fence() argument 491 might_sleep_if(gfpflags_allow_blocking(gfp)); in i915_sw_fence_await_dma_fence() 501 gfp); in i915_sw_fence_await_dma_fence() 503 if (!gfpflags_allow_blocking(gfp)) in i915_sw_fence_await_dma_fence() 583 gfp_t gfp) in i915_sw_fence_await_reservation() argument [all …]
|
| A D | i915_gpu_error.h | 283 i915_gpu_coredump_alloc(struct drm_i915_private *i915, gfp_t gfp); 286 intel_gt_coredump_alloc(struct intel_gt *gt, gfp_t gfp, u32 dump_flags); 289 intel_engine_coredump_alloc(struct intel_engine_cs *engine, gfp_t gfp, u32 dump_flags); 294 gfp_t gfp); 347 i915_gpu_coredump_alloc(struct drm_i915_private *i915, gfp_t gfp) in i915_gpu_coredump_alloc() argument 353 intel_gt_coredump_alloc(struct intel_gt *gt, gfp_t gfp, u32 dump_flags) in intel_gt_coredump_alloc() argument 359 intel_engine_coredump_alloc(struct intel_engine_cs *engine, gfp_t gfp, u32 dump_flags) in intel_engine_coredump_alloc() argument 367 gfp_t gfp) in intel_engine_coredump_add_request() argument
|
| A D | i915_deps.c | 50 void i915_deps_init(struct i915_deps *deps, gfp_t gfp) in i915_deps_init() argument 53 deps->gfp = gfp; in i915_deps_init() 85 new_fences = kmalloc_array(new_size, sizeof(*new_fences), deps->gfp); in i915_deps_grow()
|
| A D | i915_sw_fence.h | 77 gfp_t gfp); 90 gfp_t gfp); 96 gfp_t gfp);
|
| A D | i915_deps.h | 29 gfp_t gfp; member 32 void i915_deps_init(struct i915_deps *deps, gfp_t gfp);
|
| A D | i915_gpu_error.c | 201 folio = folio_alloc(gfp, 0); in pool_refill() 217 err = pool_refill(fbatch, gfp); in pool_init() 228 folio = folio_alloc(gfp, 0); in pool_alloc() 1435 c = kmalloc(sizeof(*c), gfp); in capture_vma_snapshot() 1455 gfp_t gfp) in capture_vma() argument 1476 gfp_t gfp) in capture_user() argument 1531 ee = kzalloc(sizeof(*ee), gfp); in intel_engine_coredump_alloc() 1548 gfp_t gfp) in engine_coredump_add_context() argument 1570 gfp_t gfp) in intel_engine_coredump_add_request() argument 1584 vma = capture_user(vma, rq, gfp); in intel_engine_coredump_add_request() [all …]
|
| /drivers/gpu/drm/ |
| A D | drm_managed.c | 88 size_t size, gfp_t gfp, int nid) in alloc_dr() argument 97 dr = kmalloc_node_track_caller(tot_size, gfp, nid); in alloc_dr() 229 void *drmm_kmalloc(struct drm_device *dev, size_t size, gfp_t gfp) in drmm_kmalloc() argument 233 dr = alloc_dr(NULL, size, gfp, dev_to_node(dev->dev)); in drmm_kmalloc() 236 size, gfp); in drmm_kmalloc() 239 dr->node.name = kstrdup_const("kmalloc", gfp); in drmm_kmalloc() 257 char *drmm_kstrdup(struct drm_device *dev, const char *s, gfp_t gfp) in drmm_kstrdup() argument 266 buf = drmm_kmalloc(dev, size, gfp); in drmm_kstrdup()
|
| /drivers/firmware/qcom/ |
| A D | qcom_tzmem.c | 153 size_t size, gfp_t gfp) in qcom_tzmem_pool_add_memory() argument 158 gfp); in qcom_tzmem_pool_add_memory() 165 &area->paddr, gfp); in qcom_tzmem_pool_add_memory() 334 size_t requested, gfp_t gfp) in qcom_tzmem_try_grow_pool() argument 351 return !qcom_tzmem_pool_add_memory(pool, requested, gfp); in qcom_tzmem_try_grow_pool() 366 void *qcom_tzmem_alloc(struct qcom_tzmem_pool *pool, size_t size, gfp_t gfp) in qcom_tzmem_alloc() argument 377 gfp); in qcom_tzmem_alloc() 384 if (qcom_tzmem_try_grow_pool(pool, size, gfp)) in qcom_tzmem_alloc()
|
| /drivers/gpu/drm/i915/gem/ |
| A D | i915_gem_internal.c | 40 gfp_t gfp; in i915_gem_object_get_pages_internal() local 49 gfp = GFP_KERNEL | __GFP_HIGHMEM | __GFP_RECLAIMABLE; in i915_gem_object_get_pages_internal() 52 gfp &= ~__GFP_HIGHMEM; in i915_gem_object_get_pages_internal() 53 gfp |= __GFP_DMA32; in i915_gem_object_get_pages_internal() 74 page = alloc_pages(gfp | (order ? QUIET : MAYFAIL), in i915_gem_object_get_pages_internal()
|
| /drivers/crypto/hisilicon/sec/ |
| A D | sec_algs.c | 179 gfp_t gfp) in sec_alloc_and_fill_hw_sgl() argument 194 gfp, &sgl_next_dma); in sec_alloc_and_fill_hw_sgl() 549 int *steps, gfp_t gfp) in sec_alg_alloc_and_calc_split_sizes() argument 572 struct device *dev, gfp_t gfp) in sec_map_and_split_sg() argument 593 *splits, *splits_nents, gfp); in sec_map_and_split_sg() 640 el = kzalloc(sizeof(*el), gfp); in sec_alg_alloc_and_fill_el() 672 n_ents_in, info, gfp); in sec_alg_alloc_and_fill_el() 683 n_ents_out, info, gfp); in sec_alg_alloc_and_fill_el() 733 &steps, gfp); in sec_alg_skcipher_crypto() 739 info->dev, gfp); in sec_alg_skcipher_crypto() [all …]
|
| /drivers/iommu/amd/ |
| A D | io_pgtable.c | 108 gfp_t gfp) in increase_address_space() argument 117 pte = iommu_alloc_pages_node_sz(cfg->amd.nid, gfp, SZ_4K); in increase_address_space() 151 gfp_t gfp, in alloc_pte() argument 168 PAGE_SIZE_LEVEL(page_size), gfp)) in alloc_pte() 209 page = iommu_alloc_pages_node_sz(cfg->amd.nid, gfp, in alloc_pte() 323 int prot, gfp_t gfp, size_t *mapped) in iommu_v1_map_pages() argument 342 pte = alloc_pte(pgtable, iova, pgsize, NULL, gfp, &updated); in iommu_v1_map_pages()
|
| /drivers/md/bcache/ |
| A D | util.h | 39 #define init_heap(heap, _size, gfp) \ argument 45 (heap)->data = kvmalloc(_bytes, (gfp) & GFP_KERNEL); \ 124 #define __init_fifo(fifo, gfp) \ argument 135 (fifo)->data = kvmalloc(_bytes, (gfp) & GFP_KERNEL); \ 139 #define init_fifo_exact(fifo, _size, gfp) \ argument 142 __init_fifo(fifo, gfp); \ 145 #define init_fifo(fifo, _size, gfp) \ argument 150 __init_fifo(fifo, gfp); \
|
| /drivers/crypto/cavium/nitrox/ |
| A D | nitrox_req.h | 70 gfp_t gfp; member 517 gfp_t gfp; member 555 static inline void *alloc_req_buf(int nents, int extralen, gfp_t gfp) in alloc_req_buf() argument 562 return kzalloc(size, gfp); in alloc_req_buf() 625 nkreq->src = alloc_req_buf(nents, ivsize, creq->gfp); in alloc_src_req_buf() 672 nkreq->dst = alloc_req_buf(nents, extralen, creq->gfp); in alloc_dst_req_buf()
|
| /drivers/staging/media/ipu3/ |
| A D | ipu3-dmamap.c | 34 static struct page **imgu_dmamap_alloc_buffer(size_t size, gfp_t gfp) in imgu_dmamap_alloc_buffer() argument 47 gfp |= __GFP_HIGHMEM | __GFP_ZERO; in imgu_dmamap_alloc_buffer() 59 gfp | high_order_gfp : gfp, order); in imgu_dmamap_alloc_buffer()
|
| /drivers/greybus/ |
| A D | operation.c | 215 static int gb_message_send(struct gb_message *message, gfp_t gfp) in gb_message_send() argument 223 gfp); in gb_message_send() 470 size_t response_size, gfp_t gfp) in gb_operation_response_alloc() argument 478 response = gb_operation_message_alloc(hd, type, response_size, gfp); in gb_operation_response_alloc() 579 gfp_t gfp) in gb_operation_create_flags() argument 593 flags, gfp); in gb_operation_create_flags() 605 gfp_t gfp) in gb_operation_create_core() argument 613 flags, gfp); in gb_operation_create_core() 724 gfp_t gfp) in gb_operation_request_send() argument 769 ret = gb_message_send(operation->request, gfp); in gb_operation_request_send()
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
| A D | mem.c | 155 gfp_t gfp = GFP_USER | __GFP_ZERO; in nvkm_mem_new_host() local 200 gfp |= GFP_HIGHUSER; in nvkm_mem_new_host() 202 gfp |= GFP_DMA32; in nvkm_mem_new_host() 205 struct page *p = alloc_page(gfp); in nvkm_mem_new_host()
|
| /drivers/virtio/ |
| A D | virtio_ring.c | 477 gfp_t gfp) in alloc_indirect_split() argument 537 gfp_t gfp) in virtqueue_add_split() argument 1330 gfp_t gfp) in virtqueue_add_indirect_packed() argument 1456 gfp_t gfp) in virtqueue_add_packed() argument 2276 gfp_t gfp) in virtqueue_add() argument 2309 gfp_t gfp) in virtqueue_add_sgs() argument 2341 gfp_t gfp) in virtqueue_add_outbuf() argument 2364 gfp_t gfp) in virtqueue_add_outbuf_premapped() argument 2386 gfp_t gfp) in virtqueue_add_inbuf() argument 2410 gfp_t gfp) in virtqueue_add_inbuf_ctx() argument [all …]
|
| /drivers/media/pci/intel/ipu6/ |
| A D | ipu6-dma.c | 58 static struct page **__alloc_buffer(size_t size, gfp_t gfp, unsigned long attrs) in __alloc_buffer() argument 69 gfp |= __GFP_NOWARN; in __alloc_buffer() 74 pages[i] = alloc_pages(gfp, order); in __alloc_buffer() 76 pages[i] = alloc_pages(gfp, --order); in __alloc_buffer() 153 dma_addr_t *dma_handle, gfp_t gfp, in ipu6_dma_alloc() argument 179 pages = __alloc_buffer(size, gfp, attrs); in ipu6_dma_alloc()
|