| /drivers/gpu/drm/i915/gt/ |
| A D | intel_context.h | 79 return ce; in intel_context_to_parent() 85 return intel_context_is_child(ce) || intel_context_is_parent(ce); in intel_context_is_parallel() 131 return ce->ops->cancel_request(ce, rq); in intel_context_cancel_request() 198 ce->ops->sched_disable(ce); in intel_context_unpin() 214 ce->ops->enter(ce); in intel_context_enter() 215 ce->wakeref = intel_gt_pm_get(ce->vm->gt); in intel_context_enter() 232 intel_gt_pm_put_async(ce->vm->gt, ce->wakeref); in intel_context_exit() 233 ce->ops->exit(ce); in intel_context_exit() 239 return ce; in intel_context_get() 244 kref_put(&ce->ref, ce->ops->destroy); in intel_context_put() [all …]
|
| A D | intel_context.c | 46 if (!ce) in intel_context_create() 68 err = ce->ops->alloc(ce); in intel_context_alloc_state() 248 err = ce->ops->pre_pin(ce, ww, &vaddr); in __intel_context_do_pin_ww() 272 err = ce->ops->pin(ce, vaddr); in __intel_context_do_pin_ww() 280 ce->ring->head, ce->ring->tail); in __intel_context_do_pin_ww() 297 ce->ops->post_unpin(ce); in __intel_context_do_pin_ww() 335 ce->ops->unpin(ce); in __intel_context_do_unpin() 336 ce->ops->post_unpin(ce); in __intel_context_do_unpin() 602 ce->ops->update_stats(ce); in intel_context_get_total_runtime_ns() 632 ce->ops->revoke(ce, rq, in intel_context_ban() [all …]
|
| A D | intel_lrc.c | 955 __lrc_init_regs(ce->lrc_reg_state, ce, engine, inhibit); in lrc_init_regs() 1044 return i915_ggtt_offset(ce->state) + context_wa_bb_offset(ce); in lrc_indirect_bb() 1167 ce->ring = ring; in lrc_alloc() 1168 ce->state = vma; in lrc_alloc() 1183 intel_ring_reset(ce->ring, ce->ring->emit); in lrc_reset() 1186 lrc_init_regs(ce, ce->engine, true); in lrc_reset() 1187 ce->lrc.lrca = lrc_update_regs(ce, ce->engine, ce->ring->tail); in lrc_reset() 1218 ce->lrc.lrca = lrc_update_regs(ce, engine, ce->ring->tail); in lrc_pin() 1239 if (!ce->state) in lrc_fini() 1248 struct intel_context *ce = container_of(kref, typeof(*ce), ref); in lrc_destroy() local [all …]
|
| A D | intel_engine_pm.c | 40 if (ce->state) { in dbg_poison_ce() 62 struct intel_context *ce; in __engine_unpark() local 69 ce = engine->kernel_context; in __engine_unpark() 70 if (ce) { in __engine_unpark() 78 dbg_poison_ce(ce); in __engine_unpark() 81 ce->ops->reset(ce); in __engine_unpark() 84 ce->timeline->seqno, in __engine_unpark() 86 ce->ring->emit); in __engine_unpark() 313 struct intel_context *ce; in intel_engine_reset_pinned_contexts() local 321 dbg_poison_ce(ce); in intel_engine_reset_pinned_contexts() [all …]
|
| A D | intel_breadcrumbs.c | 86 struct intel_context *ce) in add_signaling_context() argument 100 if (!list_empty(&ce->signals)) in remove_signaling_context() 113 if (rq->context != ce) in check_signal_order() 175 struct intel_context *ce; in signal_irq_work() local 236 intel_context_put(ce); in signal_irq_work() 368 intel_context_get(ce); in insert_breadcrumb() 370 pos = &ce->signals; in insert_breadcrumb() 427 spin_lock(&ce->signal_lock); in i915_request_enable_breadcrumb() 451 intel_context_put(ce); in i915_request_cancel_breadcrumb() 486 intel_context_put(ce); in intel_context_remove_breadcrumbs() [all …]
|
| A D | intel_lrc.h | 36 int lrc_alloc(struct intel_context *ce, 38 void lrc_reset(struct intel_context *ce); 39 void lrc_fini(struct intel_context *ce); 43 lrc_pre_pin(struct intel_context *ce, 48 lrc_pin(struct intel_context *ce, 51 void lrc_unpin(struct intel_context *ce); 52 void lrc_post_unpin(struct intel_context *ce); 54 void lrc_init_state(struct intel_context *ce, 67 void lrc_update_offsets(struct intel_context *ce, 120 if (intel_context_is_barrier(ce)) in lrc_runtime_start() [all …]
|
| A D | selftest_mocs.c | 26 struct intel_context *ce; in mocs_context_create() local 29 if (IS_ERR(ce)) in mocs_context_create() 30 return ce; in mocs_context_create() 33 ce->ring_size = SZ_16K; in mocs_context_create() 35 return ce; in mocs_context_create() 304 if (IS_ERR(ce)) { in live_mocs_clean() 305 err = PTR_ERR(ce); in live_mocs_clean() 310 intel_context_put(ce); in live_mocs_clean() 415 if (IS_ERR(ce)) { in live_mocs_reset() 416 err = PTR_ERR(ce); in live_mocs_reset() [all …]
|
| A D | selftest_lrc.c | 420 if (IS_ERR(ce)) in __live_lrc_state() 421 return PTR_ERR(ce); in __live_lrc_state() 634 if (IS_ERR(ce)) in __live_lrc_gpr() 635 return PTR_ERR(ce); in __live_lrc_gpr() 899 data.ce[i] = tmp; in live_lrc_timestamp() 917 if (!data.ce[i]) in live_lrc_timestamp() 1627 setup_per_ctx_bb(ce, ce->engine, emit_per_ctx_bb_canary); in wabb_ctx_setup() 1629 setup_indirect_ctx_bb(ce, ce->engine, emit_indirect_ctx_bb_canary); in wabb_ctx_setup() 1806 if (IS_ERR(ce)) in __lrc_garbage() 1807 return PTR_ERR(ce); in __lrc_garbage() [all …]
|
| A D | intel_context_sseu.c | 17 const struct intel_context *ce, in gen8_emit_rpcs_config() argument 27 offset = i915_ggtt_offset(ce->state) + in gen8_emit_rpcs_config() 46 lockdep_assert_held(&ce->pin_mutex); in gen8_modify_rpcs() 54 if (!intel_context_pin_if_active(ce)) in gen8_modify_rpcs() 66 ret = gen8_emit_rpcs_config(rq, ce, sseu); in gen8_modify_rpcs() 70 intel_context_unpin(ce); in gen8_modify_rpcs() 82 ret = intel_context_lock_pinned(ce); in intel_context_reconfigure_sseu() 87 if (!memcmp(&ce->sseu, &sseu, sizeof(sseu))) in intel_context_reconfigure_sseu() 90 ret = gen8_modify_rpcs(ce, sseu); in intel_context_reconfigure_sseu() 92 ce->sseu = sseu; in intel_context_reconfigure_sseu() [all …]
|
| /drivers/crypto/gemini/ |
| A D | sl3516-ce-core.c | 35 ce->tx = dma_alloc_coherent(ce->dev, sz, &ce->dtx, GFP_KERNEL); in sl3516_ce_desc_init() 38 ce->rx = dma_alloc_coherent(ce->dev, sz, &ce->drx, GFP_KERNEL); in sl3516_ce_desc_init() 61 dma_free_coherent(ce->dev, sz, ce->rx, ce->drx); in sl3516_ce_desc_init() 63 dma_free_coherent(ce->dev, sz, ce->tx, ce->dtx); in sl3516_ce_desc_init() 71 dma_free_coherent(ce->dev, sz, ce->tx, ce->dtx); in sl3516_ce_free_descs() 72 dma_free_coherent(ce->dev, sz, ce->rx, ce->drx); in sl3516_ce_free_descs() 103 dd = &ce->tx[ce->ctx]; in get_desc_tx() 114 rdd = &ce->rx[ce->crx]; in get_desc_rx() 254 ce->hwrng_stat_req, ce->hwrng_stat_bytes); in sl3516_ce_debugfs_show() 288 ce_algs[i].ce = ce; in sl3516_ce_register_algs() [all …]
|
| A D | sl3516-ce-rng.c | 15 struct sl3516_ce_dev *ce; in sl3516_ce_rng_read() local 23 ce->hwrng_stat_req++; in sl3516_ce_rng_read() 24 ce->hwrng_stat_bytes += max; in sl3516_ce_rng_read() 27 err = pm_runtime_get_sync(ce->dev); in sl3516_ce_rng_read() 29 pm_runtime_put_noidle(ce->dev); in sl3516_ce_rng_read() 39 pm_runtime_put(ce->dev); in sl3516_ce_rng_read() 48 ce->trng.name = "SL3516 Crypto Engine RNG"; in sl3516_ce_rng_register() 49 ce->trng.read = sl3516_ce_rng_read; in sl3516_ce_rng_register() 50 ce->trng.quality = 700; in sl3516_ce_rng_register() 52 ret = hwrng_register(&ce->trng); in sl3516_ce_rng_register() [all …]
|
| A D | sl3516-ce-cipher.c | 29 struct sl3516_ce_dev *ce = op->ce; in sl3516_ce_need_fallback() local 35 ce->fallback_mod16++; in sl3516_ce_need_fallback() 56 ce->fallback_mod16++; in sl3516_ce_need_fallback() 60 ce->fallback_mod16++; in sl3516_ce_need_fallback() 64 ce->fallback_align16++; in sl3516_ce_need_fallback() 72 ce->fallback_mod16++; in sl3516_ce_need_fallback() 76 ce->fallback_mod16++; in sl3516_ce_need_fallback() 80 ce->fallback_align16++; in sl3516_ce_need_fallback() 131 struct sl3516_ce_dev *ce = op->ce; in sl3516_ce_cipher() local 326 op->ce = algt->ce; in sl3516_ce_cipher_init() [all …]
|
| /drivers/gpu/drm/i915/gt/uc/ |
| A D | intel_guc_submission.c | 522 return ce; in __get_context() 1906 ce = list_next_entry(ce, parallel.child_link); in __guc_reset_context() 2976 ret = deregister_context(ce, ce->guc_id.id); in try_context_registration() 3173 !ctx_id_mapped(ce_to_guc(ce), ce->guc_id.id) || in context_cant_unblock() 3453 ret = deregister_context(ce, ce->guc_id.id); in guc_lrc_desc_unpin() 3619 return lrc_alloc(ce, ce->engine); in guc_context_alloc() 5208 ce->guc_id.id, ce->engine->name); in capture_error_state() 5212 intel_engine_set_hung_context(ce->engine, ce); in capture_error_state() 5225 __guc_reset_context(ce, ce->engine->mask); in guc_context_replay() 5238 ce->guc_id.id, ce->engine->name, in guc_handle_context_reset() [all …]
|
| A D | selftest_guc.c | 57 struct intel_context *ce; in intel_guc_scrub_ctbs() local 68 if (IS_ERR(ce)) { in intel_guc_scrub_ctbs() 69 ret = PTR_ERR(ce); in intel_guc_scrub_ctbs() 87 intel_context_put(ce); in intel_guc_scrub_ctbs() 151 struct intel_context **ce; in intel_guc_steal_guc_ids() local 156 ce = kcalloc(GUC_MAX_CONTEXT_ID, sizeof(*ce), GFP_KERNEL); in intel_guc_steal_guc_ids() 157 if (!ce) { in intel_guc_steal_guc_ids() 283 kfree(ce); in intel_guc_steal_guc_ids() 327 if (IS_ERR(ce)) { in intel_guc_fast_request() 328 ret = PTR_ERR(ce); in intel_guc_fast_request() [all …]
|
| /drivers/crypto/allwinner/sun8i-ce/ |
| A D | sun8i-ce-core.c | 199 writel(desc_addr_val(ce, ce->chanlist[flow].t_phy), ce->base + CE_TDQ); in sun8i_ce_run_task() 705 ce->hwrng_stat_req, ce->hwrng_stat_bytes); in sun8i_ce_debugfs_show() 731 ce->chanlist = devm_kcalloc(ce->dev, MAXFLOW, in sun8i_ce_allocate_chanlist() 739 ce->chanlist[i].engine = crypto_engine_alloc_init(ce->dev, true); in sun8i_ce_allocate_chanlist() 751 ce->chanlist[i].tl = dma_alloc_coherent(ce->dev, in sun8i_ce_allocate_chanlist() 761 ce->chanlist[i].bounce_iv = devm_kmalloc(ce->dev, AES_BLOCK_SIZE, in sun8i_ce_allocate_chanlist() 767 ce->chanlist[i].backup_iv = devm_kmalloc(ce->dev, AES_BLOCK_SIZE, in sun8i_ce_allocate_chanlist() 851 ce->ceclks[i] = devm_clk_get(ce->dev, ce->variant->ce_clks[i].name); in sun8i_ce_get_clks() 889 ce_algs[i].ce = ce; in sun8i_ce_register_algs() 997 ce = devm_kzalloc(&pdev->dev, sizeof(*ce), GFP_KERNEL); in sun8i_ce_probe() [all …]
|
| A D | sun8i-ce-trng.c | 25 struct sun8i_ce_dev *ce; in sun8i_ce_trng_read() local 46 ce->hwrng_stat_req++; in sun8i_ce_trng_read() 47 ce->hwrng_stat_bytes += todo; in sun8i_ce_trng_read() 61 mutex_lock(&ce->rnglock); in sun8i_ce_trng_read() 62 chan = &ce->chanlist[flow]; in sun8i_ce_trng_read() 82 ce->chanlist[flow].timeout = todo; in sun8i_ce_trng_read() 85 mutex_unlock(&ce->rnglock); in sun8i_ce_trng_read() 87 pm_runtime_put(ce->dev); in sun8i_ce_trng_read() 110 ce->trng.read = sun8i_ce_trng_read; in sun8i_ce_hwrng_register() 112 ret = hwrng_register(&ce->trng); in sun8i_ce_hwrng_register() [all …]
|
| A D | sun8i-ce-prng.c | 61 struct sun8i_ce_dev *ce; in sun8i_ce_prng_generate() local 72 ce = algt->ce; in sun8i_ce_prng_generate() 75 dev_err(ce->dev, "not seeded\n"); in sun8i_ce_prng_generate() 98 if (dma_mapping_error(ce->dev, dma_iv)) { in sun8i_ce_prng_generate() 115 mutex_lock(&ce->rnglock); in sun8i_ce_prng_generate() 116 chan = &ce->chanlist[flow]; in sun8i_ce_prng_generate() 126 if (ce->variant->prng_t_dlen_in_bytes) in sun8i_ce_prng_generate() 140 ce->chanlist[flow].timeout = 2000; in sun8i_ce_prng_generate() 142 err = sun8i_ce_run_task(ce, 3, "PRNG"); in sun8i_ce_prng_generate() 143 mutex_unlock(&ce->rnglock); in sun8i_ce_prng_generate() [all …]
|
| A D | sun8i-ce-cipher.c | 139 struct sun8i_ce_dev *ce = op->ce; in sun8i_ce_cipher_prepare() local 168 chan = &ce->chanlist[flow]; in sun8i_ce_cipher_prepare() 178 if (ce->variant->cipher_t_dlen_in_bytes) in sun8i_ce_cipher_prepare() 324 struct sun8i_ce_dev *ce = op->ce; in sun8i_ce_cipher_unprepare() local 334 chan = &ce->chanlist[flow]; in sun8i_ce_cipher_unprepare() 369 struct sun8i_ce_dev *ce = op->ce; in sun8i_ce_cipher_run() local 404 e = sun8i_ce_get_engine_number(op->ce); in sun8i_ce_skdecrypt() 406 engine = op->ce->chanlist[e].engine; in sun8i_ce_skdecrypt() 425 engine = op->ce->chanlist[e].engine; in sun8i_ce_skencrypt() 442 op->ce = algt->ce; in sun8i_ce_cipher_init() [all …]
|
| /drivers/base/power/ |
| A D | clock_ops.c | 175 if (!ce->clk) in pm_clk_acquire() 176 ce->clk = clk_get(dev, ce->con_id); in pm_clk_acquire() 192 ce->clk, ce->con_id); in pm_clk_acquire() 204 ce = kzalloc(sizeof(*ce), GFP_KERNEL); in __pm_clk_add() 205 if (!ce) in __pm_clk_add() 211 kfree(ce); in __pm_clk_add() 216 kfree(ce); in __pm_clk_add() 219 ce->clk = clk; in __pm_clk_add() 323 if (!ce) in __pm_clk_remove() 342 kfree(ce->con_id); in __pm_clk_remove() [all …]
|
| /drivers/clocksource/ |
| A D | timer-sun5i.c | 71 sun5i_clkevt_sync(ce); in sun5i_clkevt_time_stop() 96 sun5i_clkevt_time_stop(ce, 0); in sun5i_clkevt_shutdown() 104 sun5i_clkevt_time_stop(ce, 0); in sun5i_clkevt_set_oneshot() 113 sun5i_clkevt_time_stop(ce, 0); in sun5i_clkevt_set_periodic() 114 sun5i_clkevt_time_setup(ce, 0, ce->ticks_per_jiffy); in sun5i_clkevt_set_periodic() 124 sun5i_clkevt_time_stop(ce, 0); in sun5i_clkevt_next_event() 136 ce->clkevt.event_handler(&ce->clkevt); in sun5i_timer_interrupt() 203 void __iomem *base = ce->base; in sun5i_setup_clockevent() 214 ce->clkevt.rating = 340; in sun5i_setup_clockevent() 215 ce->clkevt.irq = irq; in sun5i_setup_clockevent() [all …]
|
| A D | mps2-timer.c | 83 struct clockevent_mps2 *ce = dev_id; in mps2_timer_interrupt() local 91 writel_relaxed(1, ce->reg + TIMER_INT); in mps2_timer_interrupt() 93 ce->clkevt.event_handler(&ce->clkevt); in mps2_timer_interrupt() 102 struct clockevent_mps2 *ce; in mps2_clockevent_init() local 139 ce = kzalloc(sizeof(*ce), GFP_KERNEL); in mps2_clockevent_init() 140 if (!ce) { in mps2_clockevent_init() 145 ce->reg = base; in mps2_clockevent_init() 147 ce->clkevt.irq = irq; in mps2_clockevent_init() 148 ce->clkevt.name = name; in mps2_clockevent_init() 149 ce->clkevt.rating = 200; in mps2_clockevent_init() [all …]
|
| A D | timer-digicolor.c | 58 struct clock_event_device ce; member 66 return container_of(ce, struct digicolor_timer, ce); in dc_timer() 90 dc_timer_disable(ce); in digicolor_clkevt_shutdown() 96 dc_timer_disable(ce); in digicolor_clkevt_set_oneshot() 105 dc_timer_disable(ce); in digicolor_clkevt_set_periodic() 112 struct clock_event_device *ce) in digicolor_clkevt_next_event() argument 114 dc_timer_disable(ce); in digicolor_clkevt_next_event() 115 dc_timer_set_count(ce, evt); in digicolor_clkevt_next_event() 122 .ce = { 190 &dc_timer_dev.ce); in digicolor_timer_init() [all …]
|
| A D | timer-rockchip.c | 43 struct clock_event_device ce; member 52 return &container_of(ce, struct rk_clkevt, ce)->timer; in rk_timer() 114 if (clockevent_state_oneshot(ce)) in rk_timer_interrupt() 117 ce->event_handler(ce); in rk_timer_interrupt() 207 struct clock_event_device *ce; in rk_clkevt_init() local 220 ce = &rk_clkevt->ce; in rk_clkevt_init() 221 ce->name = TIMER_NAME; in rk_clkevt_init() 227 ce->irq = rk_clkevt->timer.irq; in rk_clkevt_init() 228 ce->cpumask = cpu_possible_mask; in rk_clkevt_init() 229 ce->rating = 250; in rk_clkevt_init() [all …]
|
| /drivers/of/ |
| A D | dynamic.c | 528 kref_read(&ce->np->kobj.kref), ce->np); in __of_changeset_entry_destroy() 536 kfree(ce); in __of_changeset_entry_destroy() 585 rd.dn = ce->np; in __of_changeset_entry_notify() 591 ret = of_property_notify(ce->action, ce->np, ce->prop, ce->old_prop); in __of_changeset_entry_notify() 608 of_changeset_action_debug("apply: ", ce->action, ce->np, ce->prop); in __of_changeset_entry_apply() 618 ret = __of_add_property(ce->np, ce->prop); in __of_changeset_entry_apply() 621 ret = __of_remove_property(ce->np, ce->prop); in __of_changeset_entry_apply() 625 ret = __of_update_property(ce->np, ce->prop, &ce->old_prop); in __of_changeset_entry_apply() 632 of_changeset_action_err("apply failed: ", ce->action, ce->np, ce->prop); in __of_changeset_entry_apply() 911 ce = kzalloc(sizeof(*ce), GFP_KERNEL); in of_changeset_action() [all …]
|
| /drivers/gpu/drm/i915/selftests/ |
| A D | i915_request.c | 654 if (IS_ERR(ce)) { in __cancel_inactive() 707 if (IS_ERR(ce)) { in __cancel_active() 768 if (IS_ERR(ce)) { in __cancel_completed() 837 if (IS_ERR(ce)) { in __cancel_reset() 2383 ce, ce->engine->kernel_context in measure_context_switch() 2871 ps->ce[idx++] = ce; in perf_series_engines() 2887 struct intel_context *ce = ps->ce[idx]; in perf_series_engines() local 2906 struct intel_context *ce = ps->ce[idx]; in perf_series_engines() local 2976 if (IS_ERR(ce)) { in p_sync0() 3051 if (IS_ERR(ce)) { in p_sync1() [all …]
|