| /drivers/gpu/drm/i915/gt/ |
| A D | intel_gtt.c | 237 container_of(work, struct i915_address_space, release_work); in __i915_vm_release() 258 queue_work(vm->i915->wq, &vm->release_work); in i915_vm_release() 273 INIT_WORK(&vm->release_work, __i915_vm_release); in i915_address_space_init()
|
| A D | intel_gtt.h | 249 struct work_struct release_work; member
|
| /drivers/gpu/drm/imagination/ |
| A D | pvr_queue.h | 69 struct work_struct release_work; member
|
| A D | pvr_queue.c | 114 struct pvr_queue_fence *fence = container_of(w, struct pvr_queue_fence, release_work); in pvr_queue_fence_release_work() 125 queue_work(pvr_dev->sched_wq, &fence->release_work); in pvr_queue_fence_release() 279 INIT_WORK(&fence->release_work, pvr_queue_fence_release_work); in pvr_queue_fence_init()
|
| /drivers/gpu/drm/i915/gem/ |
| A D | i915_gem_context_types.h | 325 struct work_struct release_work; member
|
| A D | i915_gem_context.c | 1284 release_work); in i915_gem_context_release_work() 1320 queue_work(ctx->i915->wq, &ctx->release_work); in i915_gem_context_release() 1622 INIT_WORK(&ctx->release_work, i915_gem_context_release_work); in i915_gem_create_context()
|
| /drivers/gpu/drm/i915/gem/selftests/ |
| A D | mock_context.c | 27 INIT_WORK(&ctx->release_work, i915_gem_context_release_work); in mock_context()
|
| /drivers/infiniband/ulp/isert/ |
| A D | ib_isert.h | 188 struct work_struct release_work; member
|
| A D | ib_isert.c | 313 INIT_WORK(&isert_conn->release_work, isert_release_work); in isert_init_conn() 564 queue_work(isert_release_wq, &isert_conn->release_work); in isert_handle_unbound_conn() 2484 release_work); in isert_release_work() 2571 queue_work(isert_release_wq, &isert_conn->release_work); in isert_wait_conn()
|
| /drivers/infiniband/ulp/srpt/ |
| A D | ib_srpt.h | 330 struct work_struct release_work; member
|
| A D | ib_srpt.c | 947 schedule_work(&ch->release_work); in srpt_zerolength_write_done() 1983 schedule_work(&ch->release_work); in srpt_close_ch() 2147 ch = container_of(w, struct srpt_rdma_ch, release_work); in srpt_release_channel_work() 2286 INIT_WORK(&ch->release_work, srpt_release_channel_work); in srpt_cm_req_recv()
|
| /drivers/gpu/drm/amd/amdkfd/ |
| A D | kfd_svm.h | 50 struct work_struct release_work; member
|
| A D | kfd_process.c | 1160 release_work); in kfd_process_wq_release() 1203 INIT_WORK(&p->release_work, kfd_process_wq_release); in kfd_process_ref_release() 1204 queue_work(kfd_process_wq, &p->release_work); in kfd_process_ref_release()
|
| A D | kfd_priv.h | 901 struct work_struct release_work; member
|
| A D | kfd_svm.c | 441 svm_bo = container_of(work, struct svm_range_bo, release_work); in svm_range_bo_wq_release() 451 INIT_WORK(&svm_bo->release_work, svm_range_bo_wq_release); in svm_range_bo_release_async() 452 schedule_work(&svm_bo->release_work); in svm_range_bo_release_async()
|
| /drivers/infiniband/ulp/iser/ |
| A D | iscsi_iser.c | 889 INIT_WORK(&iser_conn->release_work, iser_release_work); in iscsi_iser_ep_disconnect() 890 queue_work(release_wq, &iser_conn->release_work); in iscsi_iser_ep_disconnect()
|
| A D | iscsi_iser.h | 411 struct work_struct release_work; member
|
| A D | iser_verbs.c | 347 iser_conn = container_of(work, struct iser_conn, release_work); in iser_release_work()
|
| /drivers/nvme/target/ |
| A D | rdma.c | 106 struct work_struct release_work; member 1371 container_of(w, struct nvmet_rdma_queue, release_work); in nvmet_rdma_release_queue_work() 1454 INIT_WORK(&queue->release_work, nvmet_rdma_release_queue_work); in nvmet_rdma_alloc_queue() 1688 queue_work(nvmet_wq, &queue->release_work); in __nvmet_rdma_queue_disconnect() 1718 queue_work(nvmet_wq, &queue->release_work); in nvmet_rdma_queue_connect_fail()
|
| A D | tcp.c | 187 struct work_struct release_work; member 1360 queue_work(nvmet_wq, &queue->release_work); in nvmet_tcp_release_queue() 1564 container_of(w, struct nvmet_tcp_queue, release_work); in nvmet_tcp_release_queue_work() 1881 INIT_WORK(&queue->release_work, nvmet_tcp_release_queue_work); in nvmet_tcp_alloc_queue()
|
| /drivers/thunderbolt/ |
| A D | tb.c | 1562 container_of(work, typeof(*group), release_work.work); in tb_bandwidth_group_release_work() 1582 INIT_DELAYED_WORK(&group->release_work, in tb_init_bandwidth_groups() 1680 cancel_delayed_work(&group->release_work); in tb_detach_bandwidth_group() 2631 mod_delayed_work(system_wq, &group->release_work, in tb_alloc_dp_bandwidth() 2954 cancel_delayed_work_sync(&tcm->groups[i].release_work); in tb_deinit()
|
| A D | tb.h | 243 struct delayed_work release_work; member
|
| /drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
| A D | ipsec.c | 824 goto release_work; in mlx5e_xfrm_add_state() 876 release_work: in mlx5e_xfrm_add_state()
|
| /drivers/gpu/drm/panthor/ |
| A D | panthor_sched.c | 685 struct work_struct release_work; member 911 release_work); in group_release_work() 936 queue_work(panthor_cleanup_wq, &group->release_work); in group_release() 3468 INIT_WORK(&group->release_work, group_release_work); in panthor_group_create()
|