| /drivers/gpu/drm/i915/ |
| A D | i915_deps.c | 38 if (deps->fences != &deps->single) in i915_deps_reset_fences() 39 kfree(deps->fences); in i915_deps_reset_fences() 40 deps->num_deps = 0; in i915_deps_reset_fences() 41 deps->fences_size = 1; in i915_deps_reset_fences() 42 deps->fences = &deps->single; in i915_deps_reset_fences() 52 deps->fences = NULL; in i915_deps_init() 53 deps->gfp = gfp; in i915_deps_init() 71 if (deps->fences != &deps->single) in i915_deps_fini() 72 kfree(deps->fences); in i915_deps_fini() 80 if (deps->num_deps >= deps->fences_size) { in i915_deps_grow() [all …]
|
| A D | i915_deps.h | 32 void i915_deps_init(struct i915_deps *deps, gfp_t gfp); 34 void i915_deps_fini(struct i915_deps *deps); 36 int i915_deps_add_dependency(struct i915_deps *deps, 40 int i915_deps_add_resv(struct i915_deps *deps, struct dma_resv *resv, 43 int i915_deps_sync(const struct i915_deps *deps,
|
| A D | i915_request.h | 422 int i915_request_await_deps(struct i915_request *rq, const struct i915_deps *deps);
|
| A D | i915_request.c | 1568 int i915_request_await_deps(struct i915_request *rq, const struct i915_deps *deps) in i915_request_await_deps() argument 1572 for (i = 0; i < deps->num_deps; ++i) { in i915_request_await_deps() 1573 err = i915_request_await_dma_fence(rq, deps->fences[i]); in i915_request_await_deps()
|
| /drivers/platform/x86/intel/ |
| A D | vsec.c | 119 const struct vsec_feature_dependency *deps = priv->info->deps; in get_consumer_dependencies() local 122 if (!deps) in get_consumer_dependencies() 126 if (deps[consumer_id].feature == BIT(cap_id)) in get_consumer_dependencies() 127 return &deps[consumer_id]; in get_consumer_dependencies() 164 const struct vsec_feature_dependency *deps; in intel_vsec_link_devices() local 182 deps = get_consumer_dependencies(priv, consumer_id); in intel_vsec_link_devices() 183 if (!deps) in intel_vsec_link_devices() 611 const struct vsec_feature_dependency *deps = priv->info->deps; in intel_vsec_skip_missing_dependencies() local 617 deps = &priv->info->deps[consumer_id]; in intel_vsec_skip_missing_dependencies() 736 .deps = oobmsm_deps, [all …]
|
| /drivers/gpu/drm/i915/gem/ |
| A D | i915_gem_ttm_move.c | 192 const struct i915_deps *deps) in i915_ttm_accel_move() argument 231 deps, src_rsgt->table.sgl, in i915_ttm_accel_move() 639 struct i915_deps deps; in i915_ttm_move() local 642 ret = i915_deps_add_resv(&deps, bo->base.resv, ctx); in i915_ttm_move() 649 dst_rsgt, true, &deps); in i915_ttm_move() 650 i915_deps_fini(&deps); in i915_ttm_move() 715 struct i915_deps deps; in i915_gem_obj_copy_ttm() local 734 ret = i915_deps_add_resv(&deps, dst_bo->base.resv, &ctx); in i915_gem_obj_copy_ttm() 738 ret = i915_deps_add_resv(&deps, src_bo->base.resv, &ctx); in i915_gem_obj_copy_ttm() 745 &deps); in i915_gem_obj_copy_ttm() [all …]
|
| /drivers/gpu/drm/i915/gt/ |
| A D | intel_migrate.h | 26 const struct i915_deps *deps, 36 const struct i915_deps *deps, 48 const struct i915_deps *deps, 56 const struct i915_deps *deps,
|
| A D | intel_migrate.c | 679 const struct i915_deps *deps, in intel_context_migrate_copy() argument 759 if (deps) { in intel_context_migrate_copy() 760 err = i915_request_await_deps(rq, deps); in intel_context_migrate_copy() 770 deps = NULL; in intel_context_migrate_copy() 985 const struct i915_deps *deps, in intel_context_migrate_clear() argument 1016 if (deps) { in intel_context_migrate_clear() 1017 err = i915_request_await_deps(rq, deps); in intel_context_migrate_clear() 1027 deps = NULL; in intel_context_migrate_clear() 1080 const struct i915_deps *deps, in intel_migrate_copy() argument 1105 err = intel_context_migrate_copy(ce, deps, in intel_migrate_copy() [all …]
|
| A D | selftest_migrate.c | 138 const struct i915_deps *deps, in intel_context_copy_ccs() argument 169 if (deps) { in intel_context_copy_ccs() 170 err = i915_request_await_deps(rq, deps); in intel_context_copy_ccs() 180 deps = NULL; in intel_context_copy_ccs() 224 const struct i915_deps *deps, in intel_migrate_ccs_copy() argument 246 err = intel_context_copy_ccs(ce, deps, sg, pat_index, in intel_migrate_ccs_copy()
|
| /drivers/power/sequencing/ |
| A D | core.c | 83 struct list_head deps; member 104 INIT_LIST_HEAD(&unit->deps); in pwrseq_unit_new() 170 pwrseq_unit_free_deps(&unit->deps); in pwrseq_unit_release() 310 for (cur = data->deps; cur && *cur; cur++) { in pwrseq_check_unit_deps() 364 if (data->deps) { in pwrseq_unit_setup() 365 ret = pwrseq_unit_setup_deps(data->deps, &unit->deps, in pwrseq_unit_setup() 806 ret = pwrseq_unit_enable_deps(pwrseq, &unit->deps); in pwrseq_unit_enable() 820 pwrseq_unit_disable_deps(pwrseq, &unit->deps); in pwrseq_unit_enable() 859 ret = pwrseq_unit_disable_deps(pwrseq, &unit->deps); in pwrseq_unit_disable() 1030 if (list_empty(&unit->deps)) in pwrseq_debugfs_seq_show_unit() [all …]
|
| A D | pwrseq-qcom-wcn.c | 152 .deps = pwrseq_qcom_wcn_unit_deps, 159 .deps = pwrseq_qcom_wcn6855_unit_deps, 186 .deps = pwrseq_qcom_wcn_unit_deps, 193 .deps = pwrseq_qcom_wcn6855_unit_deps,
|
| /drivers/gpu/drm/i915/gem/selftests/ |
| A D | i915_gem_migrate.c | 184 struct i915_deps *deps, in __igt_lmem_pages_migrate() argument 221 err = intel_migrate_clear(>->migrate, &ww, deps, in __igt_lmem_pages_migrate() 395 struct i915_deps deps; in igt_async_migrate() local 418 i915_deps_init(&deps, GFP_KERNEL); in igt_async_migrate() 419 err = i915_deps_add_dependency(&deps, &rq->fence, &ctx); in igt_async_migrate() 425 err = __igt_lmem_pages_migrate(gt, &ppgtt->vm, &deps, &spin, in igt_async_migrate() 427 i915_deps_fini(&deps); in igt_async_migrate()
|
| /drivers/net/ethernet/ibm/emac/ |
| A D | core.c | 2352 deps[i].phandle = 0; in emac_check_deps() 2356 if (deps[i].node == NULL) in emac_check_deps() 2359 if (deps[i].node == NULL) in emac_check_deps() 2360 deps[i].node = of_find_node_by_phandle(deps[i].phandle); in emac_check_deps() 2361 if (deps[i].node == NULL) in emac_check_deps() 2363 if (deps[i].ofdev == NULL) in emac_check_deps() 2364 deps[i].ofdev = of_find_device_by_node(deps[i].node); in emac_check_deps() 2365 if (deps[i].ofdev == NULL) in emac_check_deps() 2368 deps[i].drvdata = platform_get_drvdata(deps[i].ofdev); in emac_check_deps() 2391 memset(&deps, 0, sizeof(deps)); in emac_wait_deps() [all …]
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_cs.c | 412 struct drm_amdgpu_cs_chunk_dep *deps = chunk->kdata; in amdgpu_cs_p2_dependencies() local 425 ctx = amdgpu_ctx_get(fpriv, deps[i].ctx_id); in amdgpu_cs_p2_dependencies() 429 r = amdgpu_ctx_get_entity(ctx, deps[i].ip_type, in amdgpu_cs_p2_dependencies() 430 deps[i].ip_instance, in amdgpu_cs_p2_dependencies() 431 deps[i].ring, &entity); in amdgpu_cs_p2_dependencies() 437 fence = amdgpu_ctx_get_fence(ctx, entity, deps[i].handle); in amdgpu_cs_p2_dependencies() 484 struct drm_amdgpu_cs_chunk_sem *deps = chunk->kdata; in amdgpu_cs_p2_syncobj_in() local 491 r = amdgpu_syncobj_lookup_and_add(p, deps[i].handle, 0, 0); in amdgpu_cs_p2_syncobj_in() 522 struct drm_amdgpu_cs_chunk_sem *deps = chunk->kdata; in amdgpu_cs_p2_syncobj_out() local 542 drm_syncobj_find(p->filp, deps[i].handle); in amdgpu_cs_p2_syncobj_out()
|
| /drivers/firmware/efi/libstub/ |
| A D | Makefile | 73 libfdt-deps := fdt_rw.c fdt_ro.c fdt_wip.c fdt.c \ 77 $(patsubst %.c,lib-%.o,$(libfdt-deps))
|
| /drivers/md/ |
| A D | dm-ioctl.c | 1649 struct dm_target_deps *deps; in retrieve_deps() local 1653 deps = get_result_buffer(param, param_size, &len); in retrieve_deps() 1664 needed = struct_size(deps, dev, count); in retrieve_deps() 1673 deps->count = count; in retrieve_deps() 1676 deps->dev[count++] = huge_encode_dev(dd->dm_dev->bdev->bd_dev); in retrieve_deps()
|
| /drivers/gpu/drm/xe/ |
| A D | xe_oa.c | 629 static struct dma_fence *xe_oa_submit_bb(struct xe_oa_stream *stream, enum xe_oa_submit_deps deps, in xe_oa_submit_bb() argument 646 if (deps == XE_OA_SUBMIT_ADD_DEPS) { in xe_oa_submit_bb()
|