| /linux/drivers/video/fbdev/omap2/omapfb/ |
| A D | vrfb.c | 67 static struct vrfb_ctx *ctxs; variable 88 omap2_sms_write_rot_control(ctxs[ctx].control, ctx); in restore_hw_context() 89 omap2_sms_write_rot_size(ctxs[ctx].size, ctx); in restore_hw_context() 90 omap2_sms_write_rot_physical_ba(ctxs[ctx].physical_ba, ctx); in restore_hw_context() 211 ctxs[ctx].physical_ba = paddr; in omap_vrfb_setup() 212 ctxs[ctx].size = size; in omap_vrfb_setup() 213 ctxs[ctx].control = control; in omap_vrfb_setup() 308 paddr = ctxs[ctx].base + SMS_ROT_VIRT_BASE(rot); in omap_vrfb_request_ctx() 348 ctxs = devm_kcalloc(&pdev->dev, in vrfb_probe() 352 if (!ctxs) in vrfb_probe() [all …]
|
| /linux/arch/x86/mm/ |
| A D | tlb.c | 208 this_cpu_write(cpu_tlbstate.ctxs[asid].ctx_id, 0); in clear_asid_other() 231 if (this_cpu_read(cpu_tlbstate.ctxs[asid].ctx_id) != in choose_new_asid() 236 *need_flush = (this_cpu_read(cpu_tlbstate.ctxs[asid].tlb_gen) < in choose_new_asid() 558 VM_WARN_ON(this_cpu_read(cpu_tlbstate.ctxs[prev_asid].ctx_id) != in switch_mm_irqs_off() 591 if (this_cpu_read(cpu_tlbstate.ctxs[prev_asid].tlb_gen) == in switch_mm_irqs_off() 634 this_cpu_write(cpu_tlbstate.ctxs[new_asid].tlb_gen, next_tlb_gen); in switch_mm_irqs_off() 722 this_cpu_write(cpu_tlbstate.ctxs[0].ctx_id, mm->context.ctx_id); in initialize_tlbstate_and_flush() 723 this_cpu_write(cpu_tlbstate.ctxs[0].tlb_gen, tlb_gen); in initialize_tlbstate_and_flush() 727 this_cpu_write(cpu_tlbstate.ctxs[i].ctx_id, 0); in initialize_tlbstate_and_flush() 771 VM_WARN_ON(this_cpu_read(cpu_tlbstate.ctxs[loaded_mm_asid].ctx_id) != in flush_tlb_func() [all …]
|
| /linux/block/ |
| A D | blk-mq-sysfs.c | 18 struct blk_mq_ctxs *ctxs = container_of(kobj, struct blk_mq_ctxs, kobj); in blk_mq_sysfs_release() local 20 free_percpu(ctxs->queue_ctx); in blk_mq_sysfs_release() 21 kfree(ctxs); in blk_mq_sysfs_release() 29 kobject_put(&ctx->ctxs->kobj); in blk_mq_ctx_sysfs_release() 40 kfree(hctx->ctxs); in blk_mq_hw_sysfs_release()
|
| A D | blk-mq.c | 3778 if (!hctx->ctxs) in blk_mq_alloc_hctx() 3801 kfree(hctx->ctxs); in blk_mq_alloc_hctx() 3948 hctx->ctxs[hctx->nr_ctx++] = ctx; in blk_mq_map_swqueue() 4081 struct blk_mq_ctxs *ctxs; in blk_mq_alloc_ctxs() local 4084 ctxs = kzalloc(sizeof(*ctxs), GFP_KERNEL); in blk_mq_alloc_ctxs() 4085 if (!ctxs) in blk_mq_alloc_ctxs() 4089 if (!ctxs->queue_ctx) in blk_mq_alloc_ctxs() 4094 ctx->ctxs = ctxs; in blk_mq_alloc_ctxs() 4097 q->mq_kobj = &ctxs->kobj; in blk_mq_alloc_ctxs() 4098 q->queue_ctx = ctxs->queue_ctx; in blk_mq_alloc_ctxs() [all …]
|
| A D | blk-mq.h | 29 struct blk_mq_ctxs *ctxs; member
|
| A D | blk-mq-sched.c | 204 return hctx->ctxs[idx]; in blk_mq_next_ctx()
|
| /linux/tools/testing/selftests/drivers/net/hw/ |
| A D | rss_ctx.py | 337 ctxs = cfg.ethnl.rss_get({}, dump=True) 338 tuples = [(c['header']['dev-name'], c.get('context', -1)) for c in ctxs] 344 for data in ctxs: 356 ctxs = cfg.ethnl.rss_get({'header': {'dev-name': cfg.ifname}}, dump=True) 357 tuples = [(c['header']['dev-name'], c.get('context', -1)) for c in ctxs] 365 ctxs = cfg.ethnl.rss_get({'start-context': 1}, dump=True) 366 tuples = [(c['header']['dev-name'], c.get('context', -1)) for c in ctxs] 372 ctxs = cfg.ethnl.rss_get({'header': {'dev-name': cfg.ifname}, 'start-context': 1}, dump=True) 373 ctx_tuples = set([(c['header']['dev-name'], c.get('context', -1)) for c in ctxs])
|
| /linux/drivers/iommu/arm/arm-smmu/ |
| A D | qcom_iommu.c | 54 struct qcom_iommu_ctx *ctxs[]; /* indexed by asid */ member 87 return qcom_iommu->ctxs[asid]; in to_ctx() 573 WARN_ON(qcom_iommu->ctxs[asid] == NULL)) { in qcom_iommu_of_xlate() 733 qcom_iommu->ctxs[ctx->asid] = ctx; in qcom_iommu_ctx_probe() 745 qcom_iommu->ctxs[ctx->asid] = NULL; in qcom_iommu_ctx_remove() 795 qcom_iommu = devm_kzalloc(dev, struct_size(qcom_iommu, ctxs, max_asid + 1), in qcom_iommu_device_probe()
|
| /linux/include/linux/ |
| A D | damon.h | 779 int damon_start(struct damon_ctx **ctxs, int nr_ctxs, bool exclusive); 780 int damon_stop(struct damon_ctx **ctxs, int nr_ctxs);
|
| A D | blk-mq.h | 392 struct blk_mq_ctx **ctxs; member 978 ({ ctx = (hctx)->ctxs[(i)]; 1; }); (i)++)
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
| A D | chan.h | 39 unsigned ctxs:5; member
|
| A D | nv04.c | 66 u32 cm = ((1ULL << c->bits) - 1) << c->ctxs; in nv04_chan_stop() 69 nvkm_wo32(fctx, c->ctxp + data, cv | (rv << c->ctxs)); in nv04_chan_stop()
|
| /linux/arch/x86/include/asm/ |
| A D | tlbflush.h | 150 struct tlb_context ctxs[TLB_NR_DYN_ASIDS]; member
|
| /linux/mm/damon/ |
| A D | core.c | 1089 int damon_start(struct damon_ctx **ctxs, int nr_ctxs, bool exclusive) in damon_start() argument 1102 err = __damon_start(ctxs[i]); in damon_start() 1144 int damon_stop(struct damon_ctx **ctxs, int nr_ctxs) in damon_stop() argument 1150 err = __damon_stop(ctxs[i]); in damon_stop()
|
| /linux/drivers/gpu/drm/xe/ |
| A D | xe_gt_sriov_pf_debugfs.c | 202 DEFINE_SRIOV_GT_CONFIG_DEBUGFS_ATTRIBUTE(ctxs, u32, "%llu\n");
|
| /linux/kernel/ |
| A D | workqueue.c | 6869 LIST_HEAD(ctxs); in workqueue_apply_unbound_cpumask() 6886 list_add_tail(&ctx->list, &ctxs); in workqueue_apply_unbound_cpumask() 6889 list_for_each_entry_safe(ctx, n, &ctxs, list) { in workqueue_apply_unbound_cpumask()
|