| /drivers/gpu/drm/nouveau/nvkm/core/ |
| A D | uevent.c | 101 return uevent; in nvkm_uevent_dtor() 115 struct nvkm_uevent *uevent = container_of(ntfy, typeof(*uevent), ntfy); in nvkm_uevent_ntfy() local 118 if (uevent->func) in nvkm_uevent_ntfy() 119 return uevent->func(uevent->parent, uevent->object.object, bits); in nvkm_uevent_ntfy() 128 if (WARN_ON(uevent->func)) in nvkm_uevent_add() 131 nvkm_event_ntfy_add(event, id, bits, uevent->wait, nvkm_uevent_ntfy, &uevent->ntfy); in nvkm_uevent_add() 132 uevent->func = func; in nvkm_uevent_add() 147 if (!(uevent = kzalloc(sizeof(*uevent), GFP_KERNEL))) in nvkm_uevent_new() 152 uevent->parent = parent; in nvkm_uevent_new() 153 uevent->func = NULL; in nvkm_uevent_new() [all …]
|
| A D | oproxy.c | 79 struct nvkm_uevent *uevent) in nvkm_oproxy_uevent() argument 83 if (!oproxy->object->func->uevent) in nvkm_oproxy_uevent() 86 return oproxy->object->func->uevent(oproxy->object, argv, argc, uevent); in nvkm_oproxy_uevent() 166 .uevent = nvkm_oproxy_uevent,
|
| A D | ioctl.c | 44 if ( object->func->uevent && in nvkm_ioctl_sclass_() 45 !object->func->uevent(object, NULL, 0, NULL) && index-- == 0) { in nvkm_ioctl_sclass_() 110 if (!parent->func->sclass && !parent->func->uevent) { in nvkm_ioctl_new()
|
| /drivers/infiniband/core/ |
| A D | uverbs_std_types_wq.c | 16 container_of(uobject, struct ib_uwq_object, uevent.uobject); in uverbs_free_wq() 23 ib_uverbs_release_uevent(&uwq->uevent); in uverbs_free_wq() 32 typeof(*obj), uevent.uobject); in UVERBS_HANDLER() 66 obj->uevent.event_file = ib_uverbs_get_async_event(attrs, in UVERBS_HANDLER() 68 obj->uevent.uobject.user_handle = user_handle; in UVERBS_HANDLER() 69 INIT_LIST_HEAD(&obj->uevent.event_list); in UVERBS_HANDLER() 80 obj->uevent.uobject.object = wq; in UVERBS_HANDLER() 110 if (obj->uevent.event_file) in UVERBS_HANDLER() 111 uverbs_uobject_put(&obj->uevent.event_file->uobj); in UVERBS_HANDLER() 168 &obj->uevent.events_reported, in UVERBS_HANDLER() [all …]
|
| A D | ucma.c | 265 uevent = kzalloc(sizeof(*uevent), GFP_KERNEL); in ucma_create_uevent() 266 if (!uevent) in ucma_create_uevent() 275 uevent->resp.uid = uevent->mc->uid; in ucma_create_uevent() 276 uevent->resp.id = uevent->mc->id; in ucma_create_uevent() 294 return uevent; in ucma_create_uevent() 312 if (!uevent) in ucma_connect_event_handler() 352 if (!uevent) in ucma_event_handler() 412 if (uevent->mc) in ucma_get_event() 418 kfree(uevent); in ucma_get_event() 517 kfree(uevent); in ucma_cleanup_mc_events() [all …]
|
| A D | uverbs_std_types_srq.c | 15 struct ib_uevent_object *uevent = in uverbs_free_srq() local 27 uevent.uobject); in uverbs_free_srq() 32 ib_uverbs_release_uevent(uevent); in uverbs_free_srq() 41 typeof(*obj), uevent.uobject); in UVERBS_HANDLER() 101 obj->uevent.event_file = ib_uverbs_get_async_event(attrs, in UVERBS_HANDLER() 103 INIT_LIST_HEAD(&obj->uevent.event_list); in UVERBS_HANDLER() 105 obj->uevent.uobject.user_handle = user_handle; in UVERBS_HANDLER() 113 obj->uevent.uobject.object = srq; in UVERBS_HANDLER() 139 if (obj->uevent.event_file) in UVERBS_HANDLER() 140 uverbs_uobject_put(&obj->uevent.event_file->uobj); in UVERBS_HANDLER() [all …]
|
| A D | uverbs_std_types_cq.c | 45 container_of(uobject, struct ib_ucq_object, uevent.uobject); in uverbs_free_cq() 66 typeof(*obj), uevent.uobject); in UVERBS_HANDLER() 110 obj->uevent.event_file = ib_uverbs_get_async_event( in UVERBS_HANDLER() 119 INIT_LIST_HEAD(&obj->uevent.event_list); in UVERBS_HANDLER() 199 obj->uevent.uobject.object = cq; in UVERBS_HANDLER() 200 obj->uevent.uobject.user_handle = user_handle; in UVERBS_HANDLER() 212 if (obj->uevent.event_file) in UVERBS_HANDLER() 213 uverbs_uobject_put(&obj->uevent.event_file->uobj); in UVERBS_HANDLER() 266 container_of(uobj, struct ib_ucq_object, uevent.uobject); in UVERBS_HANDLER() 269 .async_events_reported = obj->uevent.events_reported in UVERBS_HANDLER()
|
| A D | uverbs_std_types_qp.c | 17 container_of(uobject, struct ib_uqp_object, uevent.uobject); in uverbs_free_qp() 41 ib_uverbs_release_uevent(&uqp->uevent); in uverbs_free_qp() 87 typeof(*obj), uevent.uobject); in UVERBS_HANDLER() 228 obj->uevent.event_file = ib_uverbs_get_async_event(attrs, in UVERBS_HANDLER() 230 INIT_LIST_HEAD(&obj->uevent.event_list); in UVERBS_HANDLER() 232 obj->uevent.uobject.user_handle = user_handle; in UVERBS_HANDLER() 265 obj->uevent.uobject.object = qp; in UVERBS_HANDLER() 281 if (obj->uevent.event_file) in UVERBS_HANDLER() 282 uverbs_uobject_put(&obj->uevent.event_file->uobj); in UVERBS_HANDLER() 351 container_of(uobj, struct ib_uqp_object, uevent.uobject); in UVERBS_HANDLER() [all …]
|
| A D | uverbs_cmd.c | 1050 INIT_LIST_HEAD(&obj->uevent.event_list); in create_cq() 1076 obj->uevent.uobject.object = cq; in create_cq() 1078 if (obj->uevent.event_file) in create_cq() 1437 INIT_LIST_HEAD(&obj->uevent.event_list); in create_qp() 1470 obj->uevent.uobject.object = qp; in create_qp() 1472 if (obj->uevent.event_file) in create_qp() 1620 obj->uevent.uobject.object = qp; in ib_uverbs_open_qp() 2961 obj->uevent.uobject.object = wq; in ib_uverbs_ex_create_wq() 2970 if (obj->uevent.event_file) in ib_uverbs_ex_create_wq() 3456 obj->uevent.uobject.object = srq; in __uverbs_create_xsrq() [all …]
|
| A D | uverbs.h | 166 struct ib_uevent_object uevent; member 171 struct ib_uevent_object uevent; member 179 struct ib_uevent_object uevent; member 183 struct ib_uevent_object uevent; member
|
| /drivers/net/wwan/iosm/ |
| A D | iosm_ipc_uevent.c | 20 envp[0] = info->uevent; in ipc_uevent_work() 23 pr_err("uevent %s failed to sent", info->uevent); in ipc_uevent_work() 28 void ipc_uevent_send(struct device *dev, char *uevent) in ipc_uevent_send() argument 40 snprintf(info->uevent, MAX_UEVENT_LEN, "IOSM_EVENT=%s", uevent); in ipc_uevent_send()
|
| A D | iosm_ipc_uevent.h | 29 char uevent[MAX_UEVENT_LEN]; member 39 void ipc_uevent_send(struct device *dev, char *uevent);
|
| /drivers/gpu/drm/nouveau/nvkm/engine/sw/ |
| A D | nvsw.c | 30 nvkm_nvsw_uevent(struct nvkm_object *object, void *argv, u32 argc, struct nvkm_uevent *uevent) in nvkm_nvsw_uevent() argument 34 if (!uevent) in nvkm_nvsw_uevent() 39 return nvkm_uevent_add(uevent, &nvkm_nvsw(object)->chan->event, 0, in nvkm_nvsw_uevent() 57 .uevent = nvkm_nvsw_uevent,
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/fault/ |
| A D | user.c | 32 nvkm_ufault_uevent(struct nvkm_object *object, void *argv, u32 argc, struct nvkm_uevent *uevent) in nvkm_ufault_uevent() argument 37 if (!uevent) in nvkm_ufault_uevent() 42 return nvkm_uevent_add(uevent, &buffer->fault->event, buffer->id, in nvkm_ufault_uevent() 86 .uevent = nvkm_ufault_uevent,
|
| /drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
| A D | uconn.c | 95 nvkm_uconn_uevent(struct nvkm_object *object, void *argv, u32 argc, struct nvkm_uevent *uevent) in nvkm_uconn_uevent() argument 104 if (!uevent) { in nvkm_uconn_uevent() 126 return nvkm_uevent_add(uevent, &disp->rm.event, outp->index, bits, in nvkm_uconn_uevent() 135 return nvkm_uevent_add(uevent, &device->i2c->event, outp->dp.aux->id, bits, in nvkm_uconn_uevent() 147 return nvkm_uevent_add(uevent, &device->gpio->event, conn->info.hpd, bits, in nvkm_uconn_uevent() 166 .uevent = nvkm_uconn_uevent,
|
| A D | uhead.c | 31 nvkm_uhead_uevent(struct nvkm_object *object, void *argv, u32 argc, struct nvkm_uevent *uevent) in nvkm_uhead_uevent() argument 36 if (!uevent) in nvkm_uhead_uevent() 41 return nvkm_uevent_add(uevent, &head->disp->vblank, head->id, in nvkm_uhead_uevent() 102 .uevent = nvkm_uhead_uevent,
|
| A D | base.c | 181 nvkm_event_fini(&disp->uevent); in nvkm_disp_dtor() 260 return nvkm_event_init(func->uevent, &disp->engine.subdev, 1, ARRAY_SIZE(disp->chan), in nvkm_disp_new_() 261 &disp->uevent); in nvkm_disp_new_()
|
| /drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
| A D | uchan.c | 42 nvkm_uchan_uevent(struct nvkm_object *object, void *argv, u32 argc, struct nvkm_uevent *uevent) in nvkm_uchan_uevent() argument 48 if (!uevent) in nvkm_uchan_uevent() 55 return nvkm_uevent_add(uevent, &runl->fifo->nonstall.event, runl->id, in nvkm_uchan_uevent() 58 return nvkm_uevent_add(uevent, &runl->chid->event, chan->id, in nvkm_uchan_uevent() 317 .uevent = nvkm_uchan_uevent,
|
| /drivers/acpi/ |
| A D | dock.c | 92 acpi_hp_uevent uevent = NULL; in dock_hotplug_event() local 101 uevent = adev->hp->uevent; in dock_hotplug_event() 110 else if (uevent) in dock_hotplug_event() 111 uevent(adev, event); in dock_hotplug_event()
|
| /drivers/xen/xenbus/ |
| A D | xenbus_probe_backend.c | 125 if (drv && drv->uevent) in xenbus_uevent_backend() 126 return drv->uevent(xdev, env); in xenbus_uevent_backend() 205 .uevent = xenbus_uevent_backend,
|
| /drivers/base/ |
| A D | Kconfig | 8 bool "Support for uevent helper" 10 The uevent helper program is forked by the kernel for 11 every uevent. 12 Before the switch to the netlink-based uevent source, this was 22 string "path to uevent helper"
|
| /drivers/gpu/drm/ci/xfails/ |
| A D | msm-apq8096-fails.txt | 2 kms_lease@lease-uevent,Fail
|
| A D | msm-apq8016-fails.txt | 4 kms_lease@lease-uevent,Fail
|
| A D | msm-sm8350-hdk-fails.txt | 9 kms_lease@lease-uevent,Fail
|
| A D | meson-g12b-fails.txt | 8 kms_lease@lease-uevent,Fail
|