| /linux/kernel/cgroup/ |
| A D | freezer.c | 24 while ((cgrp = cgroup_parent(cgrp))) { in cgroup_propagate_frozen() 30 cgrp->nr_descendants) { in cgroup_propagate_frozen() 64 cgrp->freezer.nr_frozen_tasks == __cgroup_task_count(cgrp); in cgroup_update_frozen() 109 struct cgroup *cgrp; in cgroup_enter_frozen() local 117 cgroup_inc_frozen_cnt(cgrp); in cgroup_enter_frozen() 118 cgroup_update_frozen(cgrp); in cgroup_enter_frozen() 133 struct cgroup *cgrp; in cgroup_leave_frozen() local 138 cgroup_dec_frozen_cnt(cgrp); in cgroup_leave_frozen() 139 cgroup_update_frozen(cgrp); in cgroup_leave_frozen() 213 if (cgrp->nr_descendants == cgrp->freezer.nr_frozen_descendants) in cgroup_do_freeze() [all …]
|
| A D | rstat.c | 118 rstatc->updated_next = cgrp; in cgroup_rstat_updated() 126 cgrp = parent; in cgroup_rstat_updated() 353 __cgroup_rstat_lock(cgrp, -1); in cgroup_rstat_flush() 390 if (!cgrp->rstat_cpu) { in cgroup_rstat_init() 392 if (!cgrp->rstat_cpu) in cgroup_rstat_init() 411 cgroup_rstat_flush(cgrp); in cgroup_rstat_exit() 422 free_percpu(cgrp->rstat_cpu); in cgroup_rstat_exit() 423 cgrp->rstat_cpu = NULL; in cgroup_rstat_exit() 486 delta = cgrp->bstat; in cgroup_base_stat_flush() 613 if (cgroup_parent(cgrp)) { in cgroup_base_stat_cputime_show() [all …]
|
| A D | cgroup.c | 362 return cgrp->dom_cgrp != cgrp; in cgroup_is_threaded() 428 while ((cgrp = cgroup_parent(cgrp))) { in cgroup_is_valid_domain() 520 cgrp = cgroup_parent(cgrp); in cgroup_e_css_by_mask() 554 cgrp = cgroup_parent(cgrp); in cgroup_e_css() 586 cgrp = cgroup_parent(cgrp); in cgroup_get_e_css() 818 cgrp = cgroup_parent(cgrp); in cgroup_update_populated() 1149 link->cgrp = cgrp; in link_css_set() 1323 struct cgroup *cgrp = &root->cgrp; in cgroup_destroy_root() local 2050 cgrp->self.cgroup = cgrp; in init_cgroup_housekeeping() 2052 cgrp->dom_cgrp = cgrp; in init_cgroup_housekeeping() [all …]
|
| A D | cgroup-v1.c | 324 l->owner = cgrp; in cgroup_pidlist_find_create() 498 struct cgroup *cgrp; in __cgroup1_procs_write() local 505 if (!cgrp) in __cgroup1_procs_write() 553 struct cgroup *cgrp; in cgroup_release_agent_write() local 568 if (!cgrp) in cgroup_release_agent_write() 726 if (!cgrp || !cgroup_tryget(cgrp)) { in cgroupstats_build() 755 cgroup_put(cgrp); in cgroupstats_build() 761 if (notify_on_release(cgrp) && !cgroup_is_populated(cgrp) && in cgroup1_check_for_release() 762 !css_has_online_children(&cgrp->self) && !cgroup_is_dead(cgrp)) in cgroup1_check_for_release() 1301 if (!cgrp || !cgroup_tryget(cgrp)) in task_get_cgroup1() [all …]
|
| A D | cgroup-internal.h | 27 #define TRACE_CGROUP_PATH(type, cgrp, ...) \ argument 33 cgroup_path(cgrp, trace_cgroup_path, \ 35 trace_cgroup_##type(cgrp, trace_cgroup_path, \ 98 struct cgroup *cgrp; member 188 return !(cgrp->self.flags & CSS_ONLINE); in cgroup_is_dead() 224 bool cgroup_on_dfl(const struct cgroup *cgrp); 266 int __cgroup_task_count(const struct cgroup *cgrp); 267 int cgroup_task_count(const struct cgroup *cgrp); 272 int cgroup_rstat_init(struct cgroup *cgrp); 273 void cgroup_rstat_exit(struct cgroup *cgrp); [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
| A D | cgrp.c | 181 nvkm_chid_put(runl->cgid, cgrp->id, &cgrp->lock); in nvkm_cgrp_del() 185 kfree(cgrp); in nvkm_cgrp_del() 193 if (!cgrp) in nvkm_cgrp_unref() 203 if (cgrp) in nvkm_cgrp_ref() 206 return cgrp; in nvkm_cgrp_ref() 214 if (!cgrp) in nvkm_cgrp_put() 227 if (!(cgrp = *pcgrp = kmalloc(sizeof(*cgrp), GFP_KERNEL))) in nvkm_cgrp_new() 230 cgrp->func = runl->fifo->func->cgrp.func; in nvkm_cgrp_new() 231 strscpy(cgrp->name, name, sizeof(cgrp->name)); in nvkm_cgrp_new() 234 cgrp->hw = hw; in nvkm_cgrp_new() [all …]
|
| A D | chan.c | 43 struct nvkm_cgrp *cgrp = chan->cgrp; in nvkm_chan_cctx_bind() local 55 if (cgrp->hw) in nvkm_chan_cctx_bind() 65 if (cgrp->hw) in nvkm_chan_cctx_bind() 95 struct nvkm_cgrp *cgrp = chan->cgrp; in nvkm_chan_cctx_get() local 165 struct nvkm_cgrp *cgrp = chan->cgrp; in nvkm_chan_remove_locked() local 197 struct nvkm_cgrp *cgrp = chan->cgrp; in nvkm_chan_insert() local 210 list_add_tail(&cgrp->head, &cgrp->runl->cgrps); in nvkm_chan_insert() 395 if (!cgrp) { in nvkm_chan_new_() 402 cgrp = chan->cgrp; in nvkm_chan_new_() 404 if (cgrp->runl != runl || cgrp->vmm != vmm) { in nvkm_chan_new_() [all …]
|
| A D | gk110.c | 38 struct nvkm_cgrp *cgrp = chan->cgrp; in gk110_chan_preempt() local 40 if (cgrp->hw) { in gk110_chan_preempt() 41 cgrp->func->preempt(cgrp); in gk110_chan_preempt() 61 gk110_cgrp_preempt(struct nvkm_cgrp *cgrp) in gk110_cgrp_preempt() argument 63 nvkm_wr32(cgrp->runl->fifo->engine.subdev.device, 0x002634, 0x01000000 | cgrp->id); in gk110_cgrp_preempt() 72 gk110_runl_insert_cgrp(struct nvkm_cgrp *cgrp, struct nvkm_memory *memory, u64 offset) in gk110_runl_insert_cgrp() argument 74 nvkm_wo32(memory, offset + 0, (cgrp->chan_nr << 26) | (128 << 18) | in gk110_runl_insert_cgrp() 75 (3 << 14) | 0x00002000 | cgrp->id); in gk110_runl_insert_cgrp() 123 .cgrp = {{ 0, 0, KEPLER_CHANNEL_GROUP_A }, &gk110_cgrp },
|
| A D | runl.c | 48 cgrp = chan->cgrp; in nvkm_engn_cgrp_get() 53 WARN_ON(!cgrp); in nvkm_engn_cgrp_get() 54 return cgrp; in nvkm_engn_cgrp_get() 103 if (!cgrp) { in nvkm_runl_rc() 157 struct nvkm_cgrp *cgrp; in nvkm_runl_rc_engn() local 162 if (!cgrp) { in nvkm_runl_rc_engn() 167 nvkm_runl_rc_cgrp(cgrp); in nvkm_runl_rc_engn() 231 struct nvkm_cgrp *cgrp; in nvkm_runl_cgrp_get_cgid() local 236 cgrp = cgid->data[id]; in nvkm_runl_cgrp_get_cgid() 237 if (likely(cgrp)) { in nvkm_runl_cgrp_get_cgid() [all …]
|
| A D | ucgrp.c | 33 struct nvkm_cgrp *cgrp; member 40 struct nvkm_cgrp *cgrp = nvkm_ucgrp(oclass->parent)->cgrp; in nvkm_ucgrp_chan_new() local 42 return nvkm_uchan_new(cgrp->runl->fifo, cgrp, oclass, argv, argc, pobject); in nvkm_ucgrp_chan_new() 48 struct nvkm_cgrp *cgrp = nvkm_ucgrp(object)->cgrp; in nvkm_ucgrp_sclass() local 49 struct nvkm_fifo *fifo = cgrp->runl->fifo; in nvkm_ucgrp_sclass() 70 nvkm_cgrp_unref(&ucgrp->cgrp); in nvkm_ucgrp_dtor() 115 ret = nvkm_cgrp_new(runl, args->v0.name, vmm, true, &ucgrp->cgrp); in nvkm_ucgrp_new() 120 args->v0.cgid = ucgrp->cgrp->id; in nvkm_ucgrp_new()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| A D | cgrp_kfunc_failure.c | 24 status = cgrps_kfunc_map_insert(cgrp); in insert_lookup_cgrp() 38 v = insert_lookup_cgrp(cgrp); in BPF_PROG() 56 acquired = bpf_cgroup_acquire(cgrp); in BPF_PROG() 87 acquired = bpf_cgroup_acquire(cgrp); in BPF_PROG() 143 v = insert_lookup_cgrp(cgrp); in BPF_PROG() 163 v = insert_lookup_cgrp(cgrp); in BPF_PROG() 168 kptr = v->cgrp; in BPF_PROG() 183 v = insert_lookup_cgrp(cgrp); in BPF_PROG() 188 bpf_cgroup_release(v->cgrp); in BPF_PROG() 218 local.cgrp = NULL; in BPF_PROG() [all …]
|
| A D | cgrp_ls_sleepable.c | 34 if (cgrp == NULL) in cgroup_iter() 40 cgroup_id = cgrp->kn->id; in cgroup_iter() 54 cgroup_id = cgrp->kn->id; in __no_rcu_lock() 61 struct cgroup *cgrp; in cgrp1_no_rcu_lock() local 69 if (!cgrp) in cgrp1_no_rcu_lock() 72 __no_rcu_lock(cgrp); in cgrp1_no_rcu_lock() 73 bpf_cgroup_release(cgrp); in cgrp1_no_rcu_lock() 95 struct cgroup *cgrp; in yes_rcu_lock() local 105 if (!cgrp) { in yes_rcu_lock() 113 bpf_cgroup_release(cgrp); in yes_rcu_lock() [all …]
|
| A D | cgrp_kfunc_success.c | 40 acquired = bpf_cgroup_acquire(cgrp); in BPF_PROG() 57 status = cgrps_kfunc_map_insert(cgrp); in BPF_PROG() 74 status = cgrps_kfunc_map_insert(cgrp); in BPF_PROG() 86 kptr = v->cgrp; in BPF_PROG() 96 kptr = bpf_kptr_xchg(&v->cgrp, NULL); in BPF_PROG() 117 status = cgrps_kfunc_map_insert(cgrp); in BPF_PROG() 130 kptr = v->cgrp; in BPF_PROG() 146 self = bpf_cgroup_ancestor(cgrp, cgrp->level); in BPF_PROG() 152 if (self->self.id != cgrp->self.id) { in BPF_PROG() 159 ancestor1 = bpf_cgroup_ancestor(cgrp, cgrp->level - 1); in BPF_PROG() [all …]
|
| A D | cgrp_ls_recursion.c | 30 static void __on_update(struct cgroup *cgrp) in __on_update() argument 47 struct cgroup *cgrp; in BPF_PROG() local 50 cgrp = bpf_task_get_cgroup1(task, target_hid); in BPF_PROG() 51 if (!cgrp) in BPF_PROG() 54 __on_update(cgrp); in BPF_PROG() 55 bpf_cgroup_release(cgrp); in BPF_PROG() 80 struct cgroup *cgrp; in BPF_PROG() local 83 cgrp = bpf_task_get_cgroup1(task, target_hid); in BPF_PROG() 84 if (!cgrp) in BPF_PROG() 87 __on_enter(regs, id, cgrp); in BPF_PROG() [all …]
|
| A D | cgrp_ls_tp_btf.c | 42 ptr = bpf_cgrp_storage_get(&map_a, cgrp, 0, in __on_enter() 48 err = bpf_cgrp_storage_delete(&map_a, cgrp); in __on_enter() 58 ptr = bpf_cgrp_storage_get(&map_a, cgrp, 0, in __on_enter() 70 struct cgroup *cgrp; in BPF_PROG() local 78 if (!cgrp) in BPF_PROG() 81 __on_enter(regs, id, cgrp); in BPF_PROG() 82 bpf_cgroup_release(cgrp); in BPF_PROG() 108 struct cgroup *cgrp; in BPF_PROG() local 116 if (!cgrp) in BPF_PROG() 119 __on_exit(regs, id, cgrp); in BPF_PROG() [all …]
|
| A D | iters_css_task.c | 28 struct cgroup *cgrp; in BPF_PROG() local 33 cgrp = bpf_cgroup_from_id(cg_id); in BPF_PROG() 35 if (!cgrp) in BPF_PROG() 38 css = &cgrp->self; in BPF_PROG() 45 bpf_cgroup_release(cgrp); in BPF_PROG() 52 return cgrp->kn->id; in cgroup_id() 64 if (cgrp == NULL) { in cgroup_id_printer() 75 css = &cgrp->self; in cgroup_id_printer() 93 if (cgrp == NULL) in BPF_PROG() 95 css = &cgrp->self; in BPF_PROG() [all …]
|
| A D | test_task_under_cgroup.c | 23 struct cgroup *cgrp = NULL; in BPF_PROG() local 36 cgrp = bpf_cgroup_from_id(cgid); in BPF_PROG() 37 if (!cgrp) in BPF_PROG() 40 if (bpf_task_under_cgroup(acquired, cgrp)) in BPF_PROG() 44 if (cgrp) in BPF_PROG() 45 bpf_cgroup_release(cgrp); in BPF_PROG() 54 struct cgroup *cgrp = NULL; in BPF_PROG() local 66 cgrp = bpf_cgroup_from_id(1); in BPF_PROG() 67 if (!cgrp) in BPF_PROG() 69 if (!bpf_task_under_cgroup(task, cgrp)) in BPF_PROG() [all …]
|
| A D | iters_task_failure.c | 34 struct cgroup *cgrp = bpf_cgroup_from_id(cg_id); in BPF_PROG() local 37 if (!cgrp) in BPF_PROG() 39 root_css = &cgrp->self; in BPF_PROG() 44 bpf_cgroup_release(cgrp); in BPF_PROG() 69 struct cgroup *cgrp = bpf_cgroup_from_id(cg_id); in BPF_PROG() local 72 if (!cgrp) in BPF_PROG() 74 root_css = &cgrp->self; in BPF_PROG() 83 bpf_cgroup_release(cgrp); in BPF_PROG() 96 if (cgrp == NULL) in BPF_PROG() 98 css = &cgrp->self; in BPF_PROG() [all …]
|
| A D | cgrp_kfunc_common.h | 13 struct cgroup __kptr * cgrp; member 25 struct cgroup *bpf_cgroup_ancestor(struct cgroup *cgrp, int level) __ksym; 30 static inline struct __cgrps_kfunc_map_value *cgrps_kfunc_map_value_lookup(struct cgroup *cgrp) in cgrps_kfunc_map_value_lookup() argument 35 status = bpf_probe_read_kernel(&id, sizeof(id), &cgrp->self.id); in cgrps_kfunc_map_value_lookup() 42 static inline int cgrps_kfunc_map_insert(struct cgroup *cgrp) in cgrps_kfunc_map_insert() argument 49 status = bpf_probe_read_kernel(&id, sizeof(id), &cgrp->self.id); in cgrps_kfunc_map_insert() 53 local.cgrp = NULL; in cgrps_kfunc_map_insert() 64 acquired = bpf_cgroup_acquire(cgrp); in cgrps_kfunc_map_insert() 70 old = bpf_kptr_xchg(&v->cgrp, acquired); in cgrps_kfunc_map_insert()
|
| A D | test_cgroup1_hierarchy.c | 14 struct cgroup *bpf_cgroup_ancestor(struct cgroup *cgrp, int level) __ksym; 15 void bpf_cgroup_release(struct cgroup *cgrp) __ksym; 19 struct cgroup *cgrp, *ancestor; in bpf_link_create_verify() local 32 cgrp = bpf_task_get_cgroup1(task, target_hid); in bpf_link_create_verify() 33 if (!cgrp) in bpf_link_create_verify() 37 if (cgrp->kn->id == target_ancestor_cgid) in bpf_link_create_verify() 40 ancestor = bpf_cgroup_ancestor(cgrp, target_ancestor_level); in bpf_link_create_verify() 49 bpf_cgroup_release(cgrp); in bpf_link_create_verify()
|
| /linux/include/trace/events/ |
| A D | cgroup.h | 58 TP_ARGS(cgrp, path), 70 __entry->level = cgrp->level; 82 TP_ARGS(cgrp, path) 89 TP_ARGS(cgrp, path) 96 TP_ARGS(cgrp, path) 103 TP_ARGS(cgrp, path) 110 TP_ARGS(cgrp, path) 117 TP_ARGS(cgrp, path) 170 TP_ARGS(cgrp, path, val), 197 TP_ARGS(cgrp, path, val) [all …]
|
| /linux/tools/sched_ext/ |
| A D | scx_flatcg.bpf.c | 184 cgrp = bpf_cgroup_ancestor(cgrp, level); in find_ancestor_cgrp_ctx() 185 if (!cgrp) { in find_ancestor_cgrp_ctx() 353 struct cgroup *cgrp; in BPF_STRUCT_OPS() local 512 struct cgroup *cgrp; in BPF_STRUCT_OPS() local 521 struct cgroup *cgrp; in BPF_STRUCT_OPS() local 545 struct cgroup *cgrp; in BPF_STRUCT_OPS() local 582 struct cgroup *cgrp; in BPF_STRUCT_OPS() local 597 if (cgrp->level) { in BPF_STRUCT_OPS() 598 pcgc = find_ancestor_cgrp_ctx(cgrp, cgrp->level - 1); in BPF_STRUCT_OPS() 654 if (!cgrp) { in try_pick_next_cgroup() [all …]
|
| /linux/tools/perf/util/ |
| A D | cgroup.c | 159 if (!cgrp) in add_cgroup() 175 counter->cgrp = cgrp; in add_cgroup() 189 if (cgrp && refcount_dec_and_test(&cgrp->refcnt)) { in cgroup__put() 400 cgrp = counter->cgrp; in parse_cgroups() 402 counter->cgrp = cgrp; in parse_cgroups() 481 evsel->cgrp = cgroup__get(cgrp); in evlist__expand_cgroup() 544 cgrp = malloc(sizeof(*cgrp)); in __cgroup__findnew() 550 free(cgrp); in __cgroup__findnew() 561 return cgrp; in __cgroup__findnew() 572 return cgrp; in cgroup__findnew() [all …]
|
| /linux/include/linux/ |
| A D | cgroup.h | 326 return cgrp->kn->id; in cgroup_id() 351 css_get(&cgrp->self); in cgroup_get() 361 css_put(&cgrp->self); in cgroup_put() 514 if (cgrp->root != ancestor->root || cgrp->level < ancestor->level) in cgroup_is_descendant() 558 return cgrp->nr_populated_csets + cgrp->nr_populated_domain_children + in cgroup_is_populated() 565 return kernfs_ino(cgrp->kn); in cgroup_ino() 714 struct cgroup *cgrp; in cgroup_account_cputime() local 718 cgrp = task_dfl_cgroup(task); in cgroup_account_cputime() 719 if (cgroup_parent(cgrp)) in cgroup_account_cputime() 727 struct cgroup *cgrp; in cgroup_account_cputime_field() local [all …]
|
| /linux/kernel/bpf/ |
| A D | cgroup.c | 88 struct cgroup *cgrp; in __cgroup_bpf_run_lsm_sock() local 98 if (likely(cgrp)) in __cgroup_bpf_run_lsm_sock() 110 struct cgroup *cgrp; in __cgroup_bpf_run_lsm_socket() local 120 if (likely(cgrp)) in __cgroup_bpf_run_lsm_socket() 131 struct cgroup *cgrp; in __cgroup_bpf_run_lsm_current() local 139 if (likely(cgrp)) in __cgroup_bpf_run_lsm_current() 211 cgroup_get(cgrp); in cgroup_bpf_offline() 341 cgroup_put(cgrp); in cgroup_bpf_release() 441 p = cgrp; in compute_effective_progs() 1155 if (IS_ERR(cgrp)) in cgroup_bpf_prog_attach() [all …]
|