Home
last modified time | relevance | path

Searched refs:ce (Results 1 – 25 of 194) sorted by relevance

12345678

/drivers/gpu/drm/i915/gt/
A Dintel_context.h79 return ce; in intel_context_to_parent()
85 return intel_context_is_child(ce) || intel_context_is_parent(ce); in intel_context_is_parallel()
131 return ce->ops->cancel_request(ce, rq); in intel_context_cancel_request()
198 ce->ops->sched_disable(ce); in intel_context_unpin()
214 ce->ops->enter(ce); in intel_context_enter()
215 ce->wakeref = intel_gt_pm_get(ce->vm->gt); in intel_context_enter()
232 intel_gt_pm_put_async(ce->vm->gt, ce->wakeref); in intel_context_exit()
233 ce->ops->exit(ce); in intel_context_exit()
239 return ce; in intel_context_get()
244 kref_put(&ce->ref, ce->ops->destroy); in intel_context_put()
[all …]
A Dintel_context.c46 if (!ce) in intel_context_create()
68 err = ce->ops->alloc(ce); in intel_context_alloc_state()
248 err = ce->ops->pre_pin(ce, ww, &vaddr); in __intel_context_do_pin_ww()
272 err = ce->ops->pin(ce, vaddr); in __intel_context_do_pin_ww()
280 ce->ring->head, ce->ring->tail); in __intel_context_do_pin_ww()
297 ce->ops->post_unpin(ce); in __intel_context_do_pin_ww()
335 ce->ops->unpin(ce); in __intel_context_do_unpin()
336 ce->ops->post_unpin(ce); in __intel_context_do_unpin()
602 ce->ops->update_stats(ce); in intel_context_get_total_runtime_ns()
632 ce->ops->revoke(ce, rq, in intel_context_ban()
[all …]
A Dintel_lrc.c955 __lrc_init_regs(ce->lrc_reg_state, ce, engine, inhibit); in lrc_init_regs()
1044 return i915_ggtt_offset(ce->state) + context_wa_bb_offset(ce); in lrc_indirect_bb()
1167 ce->ring = ring; in lrc_alloc()
1168 ce->state = vma; in lrc_alloc()
1183 intel_ring_reset(ce->ring, ce->ring->emit); in lrc_reset()
1186 lrc_init_regs(ce, ce->engine, true); in lrc_reset()
1187 ce->lrc.lrca = lrc_update_regs(ce, ce->engine, ce->ring->tail); in lrc_reset()
1218 ce->lrc.lrca = lrc_update_regs(ce, engine, ce->ring->tail); in lrc_pin()
1239 if (!ce->state) in lrc_fini()
1248 struct intel_context *ce = container_of(kref, typeof(*ce), ref); in lrc_destroy() local
[all …]
A Dintel_engine_pm.c40 if (ce->state) { in dbg_poison_ce()
62 struct intel_context *ce; in __engine_unpark() local
69 ce = engine->kernel_context; in __engine_unpark()
70 if (ce) { in __engine_unpark()
78 dbg_poison_ce(ce); in __engine_unpark()
81 ce->ops->reset(ce); in __engine_unpark()
84 ce->timeline->seqno, in __engine_unpark()
86 ce->ring->emit); in __engine_unpark()
313 struct intel_context *ce; in intel_engine_reset_pinned_contexts() local
321 dbg_poison_ce(ce); in intel_engine_reset_pinned_contexts()
[all …]
A Dintel_breadcrumbs.c86 struct intel_context *ce) in add_signaling_context() argument
100 if (!list_empty(&ce->signals)) in remove_signaling_context()
113 if (rq->context != ce) in check_signal_order()
175 struct intel_context *ce; in signal_irq_work() local
236 intel_context_put(ce); in signal_irq_work()
368 intel_context_get(ce); in insert_breadcrumb()
370 pos = &ce->signals; in insert_breadcrumb()
427 spin_lock(&ce->signal_lock); in i915_request_enable_breadcrumb()
451 intel_context_put(ce); in i915_request_cancel_breadcrumb()
486 intel_context_put(ce); in intel_context_remove_breadcrumbs()
[all …]
A Dintel_lrc.h36 int lrc_alloc(struct intel_context *ce,
38 void lrc_reset(struct intel_context *ce);
39 void lrc_fini(struct intel_context *ce);
43 lrc_pre_pin(struct intel_context *ce,
48 lrc_pin(struct intel_context *ce,
51 void lrc_unpin(struct intel_context *ce);
52 void lrc_post_unpin(struct intel_context *ce);
54 void lrc_init_state(struct intel_context *ce,
67 void lrc_update_offsets(struct intel_context *ce,
120 if (intel_context_is_barrier(ce)) in lrc_runtime_start()
[all …]
A Dselftest_mocs.c26 struct intel_context *ce; in mocs_context_create() local
29 if (IS_ERR(ce)) in mocs_context_create()
30 return ce; in mocs_context_create()
33 ce->ring_size = SZ_16K; in mocs_context_create()
35 return ce; in mocs_context_create()
304 if (IS_ERR(ce)) { in live_mocs_clean()
305 err = PTR_ERR(ce); in live_mocs_clean()
310 intel_context_put(ce); in live_mocs_clean()
415 if (IS_ERR(ce)) { in live_mocs_reset()
416 err = PTR_ERR(ce); in live_mocs_reset()
[all …]
A Dselftest_lrc.c420 if (IS_ERR(ce)) in __live_lrc_state()
421 return PTR_ERR(ce); in __live_lrc_state()
634 if (IS_ERR(ce)) in __live_lrc_gpr()
635 return PTR_ERR(ce); in __live_lrc_gpr()
899 data.ce[i] = tmp; in live_lrc_timestamp()
917 if (!data.ce[i]) in live_lrc_timestamp()
1627 setup_per_ctx_bb(ce, ce->engine, emit_per_ctx_bb_canary); in wabb_ctx_setup()
1629 setup_indirect_ctx_bb(ce, ce->engine, emit_indirect_ctx_bb_canary); in wabb_ctx_setup()
1806 if (IS_ERR(ce)) in __lrc_garbage()
1807 return PTR_ERR(ce); in __lrc_garbage()
[all …]
A Dintel_context_sseu.c17 const struct intel_context *ce, in gen8_emit_rpcs_config() argument
27 offset = i915_ggtt_offset(ce->state) + in gen8_emit_rpcs_config()
46 lockdep_assert_held(&ce->pin_mutex); in gen8_modify_rpcs()
54 if (!intel_context_pin_if_active(ce)) in gen8_modify_rpcs()
66 ret = gen8_emit_rpcs_config(rq, ce, sseu); in gen8_modify_rpcs()
70 intel_context_unpin(ce); in gen8_modify_rpcs()
82 ret = intel_context_lock_pinned(ce); in intel_context_reconfigure_sseu()
87 if (!memcmp(&ce->sseu, &sseu, sizeof(sseu))) in intel_context_reconfigure_sseu()
90 ret = gen8_modify_rpcs(ce, sseu); in intel_context_reconfigure_sseu()
92 ce->sseu = sseu; in intel_context_reconfigure_sseu()
[all …]
/drivers/crypto/gemini/
A Dsl3516-ce-core.c35 ce->tx = dma_alloc_coherent(ce->dev, sz, &ce->dtx, GFP_KERNEL); in sl3516_ce_desc_init()
38 ce->rx = dma_alloc_coherent(ce->dev, sz, &ce->drx, GFP_KERNEL); in sl3516_ce_desc_init()
61 dma_free_coherent(ce->dev, sz, ce->rx, ce->drx); in sl3516_ce_desc_init()
63 dma_free_coherent(ce->dev, sz, ce->tx, ce->dtx); in sl3516_ce_desc_init()
71 dma_free_coherent(ce->dev, sz, ce->tx, ce->dtx); in sl3516_ce_free_descs()
72 dma_free_coherent(ce->dev, sz, ce->rx, ce->drx); in sl3516_ce_free_descs()
103 dd = &ce->tx[ce->ctx]; in get_desc_tx()
114 rdd = &ce->rx[ce->crx]; in get_desc_rx()
254 ce->hwrng_stat_req, ce->hwrng_stat_bytes); in sl3516_ce_debugfs_show()
288 ce_algs[i].ce = ce; in sl3516_ce_register_algs()
[all …]
A Dsl3516-ce-rng.c15 struct sl3516_ce_dev *ce; in sl3516_ce_rng_read() local
23 ce->hwrng_stat_req++; in sl3516_ce_rng_read()
24 ce->hwrng_stat_bytes += max; in sl3516_ce_rng_read()
27 err = pm_runtime_get_sync(ce->dev); in sl3516_ce_rng_read()
29 pm_runtime_put_noidle(ce->dev); in sl3516_ce_rng_read()
39 pm_runtime_put(ce->dev); in sl3516_ce_rng_read()
48 ce->trng.name = "SL3516 Crypto Engine RNG"; in sl3516_ce_rng_register()
49 ce->trng.read = sl3516_ce_rng_read; in sl3516_ce_rng_register()
50 ce->trng.quality = 700; in sl3516_ce_rng_register()
52 ret = hwrng_register(&ce->trng); in sl3516_ce_rng_register()
[all …]
A Dsl3516-ce-cipher.c29 struct sl3516_ce_dev *ce = op->ce; in sl3516_ce_need_fallback() local
35 ce->fallback_mod16++; in sl3516_ce_need_fallback()
56 ce->fallback_mod16++; in sl3516_ce_need_fallback()
60 ce->fallback_mod16++; in sl3516_ce_need_fallback()
64 ce->fallback_align16++; in sl3516_ce_need_fallback()
72 ce->fallback_mod16++; in sl3516_ce_need_fallback()
76 ce->fallback_mod16++; in sl3516_ce_need_fallback()
80 ce->fallback_align16++; in sl3516_ce_need_fallback()
131 struct sl3516_ce_dev *ce = op->ce; in sl3516_ce_cipher() local
326 op->ce = algt->ce; in sl3516_ce_cipher_init()
[all …]
/drivers/gpu/drm/i915/gt/uc/
A Dintel_guc_submission.c522 return ce; in __get_context()
1906 ce = list_next_entry(ce, parallel.child_link); in __guc_reset_context()
2976 ret = deregister_context(ce, ce->guc_id.id); in try_context_registration()
3173 !ctx_id_mapped(ce_to_guc(ce), ce->guc_id.id) || in context_cant_unblock()
3453 ret = deregister_context(ce, ce->guc_id.id); in guc_lrc_desc_unpin()
3619 return lrc_alloc(ce, ce->engine); in guc_context_alloc()
5208 ce->guc_id.id, ce->engine->name); in capture_error_state()
5212 intel_engine_set_hung_context(ce->engine, ce); in capture_error_state()
5225 __guc_reset_context(ce, ce->engine->mask); in guc_context_replay()
5238 ce->guc_id.id, ce->engine->name, in guc_handle_context_reset()
[all …]
A Dselftest_guc.c57 struct intel_context *ce; in intel_guc_scrub_ctbs() local
68 if (IS_ERR(ce)) { in intel_guc_scrub_ctbs()
69 ret = PTR_ERR(ce); in intel_guc_scrub_ctbs()
87 intel_context_put(ce); in intel_guc_scrub_ctbs()
151 struct intel_context **ce; in intel_guc_steal_guc_ids() local
156 ce = kcalloc(GUC_MAX_CONTEXT_ID, sizeof(*ce), GFP_KERNEL); in intel_guc_steal_guc_ids()
157 if (!ce) { in intel_guc_steal_guc_ids()
283 kfree(ce); in intel_guc_steal_guc_ids()
327 if (IS_ERR(ce)) { in intel_guc_fast_request()
328 ret = PTR_ERR(ce); in intel_guc_fast_request()
[all …]
/drivers/crypto/allwinner/sun8i-ce/
A Dsun8i-ce-core.c199 writel(desc_addr_val(ce, ce->chanlist[flow].t_phy), ce->base + CE_TDQ); in sun8i_ce_run_task()
705 ce->hwrng_stat_req, ce->hwrng_stat_bytes); in sun8i_ce_debugfs_show()
731 ce->chanlist = devm_kcalloc(ce->dev, MAXFLOW, in sun8i_ce_allocate_chanlist()
739 ce->chanlist[i].engine = crypto_engine_alloc_init(ce->dev, true); in sun8i_ce_allocate_chanlist()
751 ce->chanlist[i].tl = dma_alloc_coherent(ce->dev, in sun8i_ce_allocate_chanlist()
761 ce->chanlist[i].bounce_iv = devm_kmalloc(ce->dev, AES_BLOCK_SIZE, in sun8i_ce_allocate_chanlist()
767 ce->chanlist[i].backup_iv = devm_kmalloc(ce->dev, AES_BLOCK_SIZE, in sun8i_ce_allocate_chanlist()
851 ce->ceclks[i] = devm_clk_get(ce->dev, ce->variant->ce_clks[i].name); in sun8i_ce_get_clks()
889 ce_algs[i].ce = ce; in sun8i_ce_register_algs()
997 ce = devm_kzalloc(&pdev->dev, sizeof(*ce), GFP_KERNEL); in sun8i_ce_probe()
[all …]
A Dsun8i-ce-trng.c25 struct sun8i_ce_dev *ce; in sun8i_ce_trng_read() local
46 ce->hwrng_stat_req++; in sun8i_ce_trng_read()
47 ce->hwrng_stat_bytes += todo; in sun8i_ce_trng_read()
61 mutex_lock(&ce->rnglock); in sun8i_ce_trng_read()
62 chan = &ce->chanlist[flow]; in sun8i_ce_trng_read()
82 ce->chanlist[flow].timeout = todo; in sun8i_ce_trng_read()
85 mutex_unlock(&ce->rnglock); in sun8i_ce_trng_read()
87 pm_runtime_put(ce->dev); in sun8i_ce_trng_read()
110 ce->trng.read = sun8i_ce_trng_read; in sun8i_ce_hwrng_register()
112 ret = hwrng_register(&ce->trng); in sun8i_ce_hwrng_register()
[all …]
A Dsun8i-ce-prng.c61 struct sun8i_ce_dev *ce; in sun8i_ce_prng_generate() local
72 ce = algt->ce; in sun8i_ce_prng_generate()
75 dev_err(ce->dev, "not seeded\n"); in sun8i_ce_prng_generate()
98 if (dma_mapping_error(ce->dev, dma_iv)) { in sun8i_ce_prng_generate()
115 mutex_lock(&ce->rnglock); in sun8i_ce_prng_generate()
116 chan = &ce->chanlist[flow]; in sun8i_ce_prng_generate()
126 if (ce->variant->prng_t_dlen_in_bytes) in sun8i_ce_prng_generate()
140 ce->chanlist[flow].timeout = 2000; in sun8i_ce_prng_generate()
142 err = sun8i_ce_run_task(ce, 3, "PRNG"); in sun8i_ce_prng_generate()
143 mutex_unlock(&ce->rnglock); in sun8i_ce_prng_generate()
[all …]
A Dsun8i-ce-cipher.c139 struct sun8i_ce_dev *ce = op->ce; in sun8i_ce_cipher_prepare() local
168 chan = &ce->chanlist[flow]; in sun8i_ce_cipher_prepare()
178 if (ce->variant->cipher_t_dlen_in_bytes) in sun8i_ce_cipher_prepare()
324 struct sun8i_ce_dev *ce = op->ce; in sun8i_ce_cipher_unprepare() local
334 chan = &ce->chanlist[flow]; in sun8i_ce_cipher_unprepare()
369 struct sun8i_ce_dev *ce = op->ce; in sun8i_ce_cipher_run() local
404 e = sun8i_ce_get_engine_number(op->ce); in sun8i_ce_skdecrypt()
406 engine = op->ce->chanlist[e].engine; in sun8i_ce_skdecrypt()
425 engine = op->ce->chanlist[e].engine; in sun8i_ce_skencrypt()
442 op->ce = algt->ce; in sun8i_ce_cipher_init()
[all …]
/drivers/base/power/
A Dclock_ops.c175 if (!ce->clk) in pm_clk_acquire()
176 ce->clk = clk_get(dev, ce->con_id); in pm_clk_acquire()
192 ce->clk, ce->con_id); in pm_clk_acquire()
204 ce = kzalloc(sizeof(*ce), GFP_KERNEL); in __pm_clk_add()
205 if (!ce) in __pm_clk_add()
211 kfree(ce); in __pm_clk_add()
216 kfree(ce); in __pm_clk_add()
219 ce->clk = clk; in __pm_clk_add()
323 if (!ce) in __pm_clk_remove()
342 kfree(ce->con_id); in __pm_clk_remove()
[all …]
/drivers/clocksource/
A Dtimer-sun5i.c71 sun5i_clkevt_sync(ce); in sun5i_clkevt_time_stop()
96 sun5i_clkevt_time_stop(ce, 0); in sun5i_clkevt_shutdown()
104 sun5i_clkevt_time_stop(ce, 0); in sun5i_clkevt_set_oneshot()
113 sun5i_clkevt_time_stop(ce, 0); in sun5i_clkevt_set_periodic()
114 sun5i_clkevt_time_setup(ce, 0, ce->ticks_per_jiffy); in sun5i_clkevt_set_periodic()
124 sun5i_clkevt_time_stop(ce, 0); in sun5i_clkevt_next_event()
136 ce->clkevt.event_handler(&ce->clkevt); in sun5i_timer_interrupt()
203 void __iomem *base = ce->base; in sun5i_setup_clockevent()
214 ce->clkevt.rating = 340; in sun5i_setup_clockevent()
215 ce->clkevt.irq = irq; in sun5i_setup_clockevent()
[all …]
A Dmps2-timer.c83 struct clockevent_mps2 *ce = dev_id; in mps2_timer_interrupt() local
91 writel_relaxed(1, ce->reg + TIMER_INT); in mps2_timer_interrupt()
93 ce->clkevt.event_handler(&ce->clkevt); in mps2_timer_interrupt()
102 struct clockevent_mps2 *ce; in mps2_clockevent_init() local
139 ce = kzalloc(sizeof(*ce), GFP_KERNEL); in mps2_clockevent_init()
140 if (!ce) { in mps2_clockevent_init()
145 ce->reg = base; in mps2_clockevent_init()
147 ce->clkevt.irq = irq; in mps2_clockevent_init()
148 ce->clkevt.name = name; in mps2_clockevent_init()
149 ce->clkevt.rating = 200; in mps2_clockevent_init()
[all …]
A Dtimer-digicolor.c58 struct clock_event_device ce; member
66 return container_of(ce, struct digicolor_timer, ce); in dc_timer()
90 dc_timer_disable(ce); in digicolor_clkevt_shutdown()
96 dc_timer_disable(ce); in digicolor_clkevt_set_oneshot()
105 dc_timer_disable(ce); in digicolor_clkevt_set_periodic()
112 struct clock_event_device *ce) in digicolor_clkevt_next_event() argument
114 dc_timer_disable(ce); in digicolor_clkevt_next_event()
115 dc_timer_set_count(ce, evt); in digicolor_clkevt_next_event()
122 .ce = {
190 &dc_timer_dev.ce); in digicolor_timer_init()
[all …]
A Dtimer-rockchip.c43 struct clock_event_device ce; member
52 return &container_of(ce, struct rk_clkevt, ce)->timer; in rk_timer()
114 if (clockevent_state_oneshot(ce)) in rk_timer_interrupt()
117 ce->event_handler(ce); in rk_timer_interrupt()
207 struct clock_event_device *ce; in rk_clkevt_init() local
220 ce = &rk_clkevt->ce; in rk_clkevt_init()
221 ce->name = TIMER_NAME; in rk_clkevt_init()
227 ce->irq = rk_clkevt->timer.irq; in rk_clkevt_init()
228 ce->cpumask = cpu_possible_mask; in rk_clkevt_init()
229 ce->rating = 250; in rk_clkevt_init()
[all …]
/drivers/of/
A Ddynamic.c528 kref_read(&ce->np->kobj.kref), ce->np); in __of_changeset_entry_destroy()
536 kfree(ce); in __of_changeset_entry_destroy()
585 rd.dn = ce->np; in __of_changeset_entry_notify()
591 ret = of_property_notify(ce->action, ce->np, ce->prop, ce->old_prop); in __of_changeset_entry_notify()
608 of_changeset_action_debug("apply: ", ce->action, ce->np, ce->prop); in __of_changeset_entry_apply()
618 ret = __of_add_property(ce->np, ce->prop); in __of_changeset_entry_apply()
621 ret = __of_remove_property(ce->np, ce->prop); in __of_changeset_entry_apply()
625 ret = __of_update_property(ce->np, ce->prop, &ce->old_prop); in __of_changeset_entry_apply()
632 of_changeset_action_err("apply failed: ", ce->action, ce->np, ce->prop); in __of_changeset_entry_apply()
911 ce = kzalloc(sizeof(*ce), GFP_KERNEL); in of_changeset_action()
[all …]
/drivers/gpu/drm/i915/selftests/
A Di915_request.c654 if (IS_ERR(ce)) { in __cancel_inactive()
707 if (IS_ERR(ce)) { in __cancel_active()
768 if (IS_ERR(ce)) { in __cancel_completed()
837 if (IS_ERR(ce)) { in __cancel_reset()
2383 ce, ce->engine->kernel_context in measure_context_switch()
2871 ps->ce[idx++] = ce; in perf_series_engines()
2887 struct intel_context *ce = ps->ce[idx]; in perf_series_engines() local
2906 struct intel_context *ce = ps->ce[idx]; in perf_series_engines() local
2976 if (IS_ERR(ce)) { in p_sync0()
3051 if (IS_ERR(ce)) { in p_sync1()
[all …]

Completed in 94 milliseconds

12345678