Home
last modified time | relevance | path

Searched refs:submit (Results 1 – 25 of 185) sorted by relevance

12345678

/linux-6.3-rc2/drivers/gpu/drm/msm/
A Dmsm_gem_submit.c41 if (!submit) in submit_create()
46 kfree(submit); in submit_create()
51 submit->dev = dev; in submit_create()
54 submit->cmd = (void *)&submit->bos[nr_bos]; in submit_create()
65 return submit; in submit_create()
76 idr_remove(&submit->queue->fence_idr, submit->fence_id); in __msm_gem_submit_destroy()
89 kfree(submit); in __msm_gem_submit_destroy()
851 ret = submit_bo(submit, submit->cmd[i].idx, in msm_ioctl_gem_submit()
857 ((submit->cmd[i].size + submit->cmd[i].offset) > in msm_ioctl_gem_submit()
870 submit->cmd[i].nr_relocs, submit->cmd[i].relocs); in msm_ioctl_gem_submit()
[all …]
A Dmsm_rd.c87 struct msm_gem_submit *submit; member
321 offset = iova - submit->bos[idx].iova; in snapshot_buf()
323 iova = submit->bos[idx].iova; in snapshot_buf()
387 pid_nr(submit->pid), submit->seqno); in msm_rd_dump_submit()
390 pid_nr(submit->pid), submit->seqno); in msm_rd_dump_submit()
396 for (i = 0; i < submit->nr_bos; i++) in msm_rd_dump_submit()
397 snapshot_buf(rd, submit, i, 0, 0, should_dump(submit, i)); in msm_rd_dump_submit()
403 if (!should_dump(submit, i)) { in msm_rd_dump_submit()
404 snapshot_buf(rd, submit, submit->cmd[i].idx, in msm_rd_dump_submit()
410 uint64_t iova = submit->cmd[i].iova; in msm_rd_dump_submit()
[all …]
A Dmsm_gpu.c282 if (submit) { in msm_gpu_crashstate_capture()
323 return submit; in find_submit()
372 if (submit) { in recover_worker()
374 submit->queue->faults++; in recover_worker()
375 if (submit->aspace) in recover_worker()
440 gpu->funcs->submit(gpu, submit); in recover_worker()
462 if (submit && submit->fault_dumped) in fault_worker()
465 if (submit) { in fault_worker()
719 if (submit && dma_fence_is_signaled(submit->hw_fence)) { in retire_submits()
763 submit->seqno = submit->hw_fence->seqno; in msm_gpu_submit()
[all …]
A Dmsm_ringbuffer.c16 struct msm_gem_submit *submit = to_msm_submit(job); in msm_job_run() local
17 struct msm_fence_context *fctx = submit->ring->fctx; in msm_job_run()
18 struct msm_gpu *gpu = submit->gpu; in msm_job_run()
21 submit->hw_fence = msm_fence_alloc(fctx); in msm_job_run()
23 for (i = 0; i < submit->nr_bos; i++) { in msm_job_run()
24 struct drm_gem_object *obj = &submit->bos[i].obj->base; in msm_job_run()
27 msm_gem_unpin_vma_fenced(submit->bos[i].vma, fctx); in msm_job_run()
36 msm_gpu_submit(gpu, submit); in msm_job_run()
40 return dma_fence_get(submit->hw_fence); in msm_job_run()
45 struct msm_gem_submit *submit = to_msm_submit(job); in msm_job_free() local
[all …]
A Dmsm_gpu_trace.h34 TP_PROTO(struct msm_gem_submit *submit, u64 ticks),
35 TP_ARGS(submit, ticks),
44 __entry->pid = pid_nr(submit->pid);
45 __entry->id = submit->ident;
46 __entry->ringid = submit->ring->id;
47 __entry->seqno = submit->seqno;
59 TP_ARGS(submit, elapsed, clock, start, end),
71 __entry->pid = pid_nr(submit->pid);
72 __entry->id = submit->ident;
73 __entry->ringid = submit->ring->id;
[all …]
A Dmsm_gem.h314 static inline void msm_gem_submit_get(struct msm_gem_submit *submit) in msm_gem_submit_get() argument
316 kref_get(&submit->ref); in msm_gem_submit_get()
319 static inline void msm_gem_submit_put(struct msm_gem_submit *submit) in msm_gem_submit_put() argument
321 kref_put(&submit->ref, __msm_gem_submit_destroy); in msm_gem_submit_put()
324 void msm_submit_retire(struct msm_gem_submit *submit);
330 should_dump(struct msm_gem_submit *submit, int idx) in should_dump() argument
333 return rd_full || (submit->bos[idx].flags & MSM_SUBMIT_BO_DUMP); in should_dump()
/linux-6.3-rc2/drivers/gpu/drm/etnaviv/
A Detnaviv_gem_submit.c35 size_t sz = size_vstruct(nr_bos, sizeof(submit->bos[0]), sizeof(*submit)); in submit_create()
38 if (!submit) in submit_create()
43 if (!submit->pmrs) { in submit_create()
44 kfree(submit); in submit_create()
49 submit->gpu = gpu; in submit_create()
52 return submit; in submit_create()
105 submit->nr_bos = i; in submit_lookup_objects()
400 xa_erase(&submit->gpu->user_fences, submit->out_fence_id); in submit_cleanup()
407 kfree(submit); in submit_cleanup()
522 if (!submit) { in etnaviv_ioctl_gem_submit()
[all …]
A Detnaviv_sched.c26 fence = etnaviv_gpu_submit(submit); in etnaviv_sched_run_job()
37 struct etnaviv_gpu *gpu = submit->gpu; in etnaviv_sched_timedout_job()
48 if (dma_fence_is_signaled(submit->out_fence)) in etnaviv_sched_timedout_job()
69 etnaviv_core_dump(submit); in etnaviv_sched_timedout_job()
70 etnaviv_gpu_recover_hang(submit); in etnaviv_sched_timedout_job()
89 etnaviv_submit_put(submit); in etnaviv_sched_free_job()
100 struct etnaviv_gpu *gpu = submit->gpu; in etnaviv_sched_push_job()
110 drm_sched_job_arm(&submit->sched_job); in etnaviv_sched_push_job()
112 submit->out_fence = dma_fence_get(&submit->sched_job.s_fence->finished); in etnaviv_sched_push_job()
114 submit->out_fence, xa_limit_32b, in etnaviv_sched_push_job()
[all …]
A Detnaviv_dump.c118 void etnaviv_core_dump(struct etnaviv_gem_submit *submit) in etnaviv_core_dump() argument
120 struct etnaviv_gpu *gpu = submit->gpu; in etnaviv_core_dump()
142 mmu_size + gpu->buffer.size + submit->cmdbuf.size; in etnaviv_core_dump()
145 for (i = 0; i < submit->nr_bos; i++) { in etnaviv_core_dump()
146 obj = submit->bos[i].obj; in etnaviv_core_dump()
184 submit->cmdbuf.vaddr, submit->cmdbuf.size, in etnaviv_core_dump()
185 etnaviv_cmdbuf_get_va(&submit->cmdbuf, in etnaviv_core_dump()
201 for (i = 0; i < submit->nr_bos; i++) { in etnaviv_core_dump()
206 obj = submit->bos[i].obj; in etnaviv_core_dump()
207 vram = submit->bos[i].mapping; in etnaviv_core_dump()
A Detnaviv_gpu.c1310 const struct etnaviv_gem_submit *submit = event->submit; in sync_point_perfmon_sample() local
1342 const struct etnaviv_gem_submit *submit = event->submit; in sync_point_perfmon_sample_post() local
1390 if (submit->nr_pmrs) in etnaviv_gpu_submit()
1413 if (submit->prev_mmu_context) in etnaviv_gpu_submit()
1417 if (submit->nr_pmrs) { in etnaviv_gpu_submit()
1419 kref_get(&submit->refcount); in etnaviv_gpu_submit()
1420 gpu->event[event[1]].submit = submit; in etnaviv_gpu_submit()
1425 submit->cmdbuf.user_size = submit->cmdbuf.size - 8; in etnaviv_gpu_submit()
1426 etnaviv_buffer_queue(gpu, submit->exec_state, submit->mmu_context, in etnaviv_gpu_submit()
1429 if (submit->nr_pmrs) { in etnaviv_gpu_submit()
[all …]
/linux-6.3-rc2/crypto/async_tx/
A Dasync_raid6_recov.c39 if (submit->flags & ASYNC_TX_FENCE) in async_sum_product()
58 async_tx_submit(chan, tx, submit); in async_sum_product()
89 struct async_submit_ctl *submit) in async_mult() argument
155 struct async_submit_ctl *submit) in __2data_recov_4() argument
165 void *cb_param = submit->cb_param; in __2data_recov_4()
205 struct async_submit_ctl *submit) in __2data_recov_5() argument
423 async_tx_sync_epilog(submit); in async_raid6_2data_recov()
445 blocks, offs, submit); in async_raid6_2data_recov()
453 blocks, offs, submit); in async_raid6_2data_recov()
456 blocks, offs, submit); in async_raid6_2data_recov()
[all …]
A Dasync_xor.c40 submit->flags = flags_orig; in do_async_xor()
48 submit->cb_fn = NULL; in do_async_xor()
49 submit->cb_param = NULL; in do_async_xor()
54 if (submit->cb_fn) in do_async_xor()
84 submit->depend_tx = tx; in do_async_xor()
110 if (submit->scribble) in do_sync_xor_offs()
111 srcs = submit->scribble; in do_sync_xor_offs()
244 src_cnt, len, submit); in async_xor_offs()
277 src_cnt, len, submit); in async_xor()
337 if (submit->cb_fn) in async_xor_val_offs()
[all …]
A Dasync_pq.c52 submit->flags = flags_orig; in do_async_gen_syndrome()
61 submit->cb_fn = NULL; in do_async_gen_syndrome()
62 submit->cb_param = NULL; in do_async_gen_syndrome()
64 submit->cb_fn = cb_fn_orig; in do_async_gen_syndrome()
91 submit->depend_tx = tx; in do_async_gen_syndrome()
114 if (submit->scribble) in do_sync_gen_syndrome()
115 srcs = submit->scribble; in do_sync_gen_syndrome()
139 async_tx_sync_epilog(submit); in do_sync_gen_syndrome()
417 len, submit); in async_syndrome_val()
430 submit->cb_fn = cb_fn_orig; in async_syndrome_val()
[all …]
A Draid6test.c60 struct async_submit_ctl submit; in raid6_dual_recov() local
71 init_async_submit(&submit, 0, NULL, NULL, NULL, addr_conv); in raid6_dual_recov()
73 disks, bytes, &submit); in raid6_dual_recov()
91 init_async_submit(&submit, ASYNC_TX_XOR_ZERO_DST, NULL, in raid6_dual_recov()
93 tx = async_xor(dest, blocks, 0, count, bytes, &submit); in raid6_dual_recov()
95 init_async_submit(&submit, 0, tx, NULL, NULL, addr_conv); in raid6_dual_recov()
97 disks, bytes, &submit); in raid6_dual_recov()
104 faila, ptrs, offs, &submit); in raid6_dual_recov()
109 faila, failb, ptrs, offs, &submit); in raid6_dual_recov()
115 disks, bytes, &result, spare, 0, &submit); in raid6_dual_recov()
[all …]
A Dasync_tx.c43 __async_tx_find_channel(struct async_submit_ctl *submit, in __async_tx_find_channel() argument
46 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in __async_tx_find_channel()
144 struct async_submit_ctl *submit) in async_tx_submit() argument
146 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in async_tx_submit()
148 tx->callback = submit->cb_fn; in async_tx_submit()
149 tx->callback_param = submit->cb_param; in async_tx_submit()
204 if (submit->flags & ASYNC_TX_ACK) in async_tx_submit()
221 async_trigger_callback(struct async_submit_ctl *submit) in async_trigger_callback() argument
245 async_tx_submit(chan, tx, submit); in async_trigger_callback()
250 async_tx_quiesce(&submit->depend_tx); in async_trigger_callback()
[all …]
A Dasync_memcpy.c34 struct async_submit_ctl *submit) in async_memcpy() argument
36 struct dma_chan *chan = async_tx_find_channel(submit, DMA_MEMCPY, in async_memcpy()
48 if (submit->cb_fn) in async_memcpy()
50 if (submit->flags & ASYNC_TX_FENCE) in async_memcpy()
70 async_tx_submit(chan, tx, submit); in async_memcpy()
76 async_tx_quiesce(&submit->depend_tx); in async_memcpy()
86 async_tx_sync_epilog(submit); in async_memcpy()
/linux-6.3-rc2/include/linux/
A Dasync_tx.h118 async_tx_find_channel(struct async_submit_ctl *submit, in async_tx_find_channel() argument
133 async_tx_sync_epilog(struct async_submit_ctl *submit) in async_tx_sync_epilog() argument
135 if (submit->cb_fn) in async_tx_sync_epilog()
136 submit->cb_fn(submit->cb_param); in async_tx_sync_epilog()
159 struct async_submit_ctl *submit);
173 struct async_submit_ctl *submit);
179 struct async_submit_ctl *submit);
184 struct async_submit_ctl *submit);
190 size_t len, struct async_submit_ctl *submit);
200 struct async_submit_ctl *submit);
[all …]
/linux-6.3-rc2/drivers/dma/ti/
A Dcppi41.c119 u16 submit; member
157 [ 0] = { .submit = 32, .complete = 93},
158 [ 1] = { .submit = 34, .complete = 94},
159 [ 2] = { .submit = 36, .complete = 95},
160 [ 3] = { .submit = 38, .complete = 96},
161 [ 4] = { .submit = 40, .complete = 97},
162 [ 5] = { .submit = 42, .complete = 98},
235 [0] = { .submit = 1, .complete = 26},
236 [1] = { .submit = 3, .complete = 26},
237 [2] = { .submit = 5, .complete = 26},
[all …]
/linux-6.3-rc2/drivers/gpu/drm/lima/
A Dlima_gem.c282 if (!submit->in_sync[i]) in lima_gem_add_deps()
308 struct lima_bo **bos = submit->lbos; in lima_gem_submit()
310 if (submit->out_sync) { in lima_gem_submit()
341 submit->nr_bos, &ctx); in lima_gem_submit()
346 submit->task, submit->ctx->context + submit->pipe, in lima_gem_submit()
347 bos, submit->nr_bos, vm); in lima_gem_submit()
357 submit->task, bos[i], in lima_gem_submit()
373 submit->nr_bos, &ctx); in lima_gem_submit()
375 for (i = 0; i < submit->nr_bos; i++) in lima_gem_submit()
388 lima_sched_task_fini(submit->task); in lima_gem_submit()
[all …]
A Dlima_drv.c110 struct lima_submit submit = {0}; in lima_ioctl_gem_submit() local
124 bos = kvcalloc(args->nr_bos, sizeof(*submit.bos) + sizeof(*submit.lbos), GFP_KERNEL); in lima_ioctl_gem_submit()
156 submit.pipe = args->pipe; in lima_ioctl_gem_submit()
157 submit.bos = bos; in lima_ioctl_gem_submit()
158 submit.lbos = (void *)bos + size; in lima_ioctl_gem_submit()
159 submit.nr_bos = args->nr_bos; in lima_ioctl_gem_submit()
160 submit.task = task; in lima_ioctl_gem_submit()
161 submit.ctx = ctx; in lima_ioctl_gem_submit()
162 submit.flags = args->flags; in lima_ioctl_gem_submit()
163 submit.in_sync[0] = args->in_sync[0]; in lima_ioctl_gem_submit()
[all …]
/linux-6.3-rc2/fs/iomap/
A Ddirect-io.c45 } submit; member
71 dio->submit.poll_bio = bio; in iomap_dio_submit_bio()
168 WRITE_ONCE(dio->submit.waiter, NULL); in iomap_dio_bio_end_io()
281 iov_iter_truncate(dio->submit.iter, length); in iomap_dio_bio_iter()
283 if (!iov_iter_count(dio->submit.iter)) in iomap_dio_bio_iter()
311 iov_iter_revert(dio->submit.iter, copied); in iomap_dio_bio_iter()
395 struct iov_iter *iter = dio->submit.iter; in iomap_dio_inline_iter()
512 dio->submit.iter = iter; in __iomap_dio_rw()
513 dio->submit.waiter = current; in __iomap_dio_rw()
514 dio->submit.poll_bio = NULL; in __iomap_dio_rw()
[all …]
/linux-6.3-rc2/drivers/gpu/drm/msm/adreno/
A Da2xx_gpu.c13 static void a2xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit) in a2xx_submit() argument
15 struct msm_ringbuffer *ring = submit->ring; in a2xx_submit()
18 for (i = 0; i < submit->nr_cmds; i++) { in a2xx_submit()
19 switch (submit->cmd[i].type) { in a2xx_submit()
25 if (gpu->cur_ctx_seqno == submit->queue->ctx->seqno) in a2xx_submit()
30 OUT_RING(ring, lower_32_bits(submit->cmd[i].iova)); in a2xx_submit()
31 OUT_RING(ring, submit->cmd[i].size); in a2xx_submit()
38 OUT_RING(ring, submit->seqno); in a2xx_submit()
47 OUT_RING(ring, submit->seqno); in a2xx_submit()
498 .submit = a2xx_submit,
A Da5xx_gpu.c73 for (i = 0; i < submit->nr_cmds; i++) { in a5xx_submit_in_rb()
74 switch (submit->cmd[i].type) { in a5xx_submit_in_rb()
83 obj = submit->bos[submit->cmd[i].idx].obj; in a5xx_submit_in_rb()
84 dwords = submit->cmd[i].size; in a5xx_submit_in_rb()
120 ring->memptrs->fence = submit->seqno; in a5xx_submit_in_rb()
133 a5xx_submit_in_rb(gpu, submit); in a5xx_submit()
163 switch (submit->cmd[i].type) { in a5xx_submit()
174 OUT_RING(ring, submit->cmd[i].size); in a5xx_submit()
208 OUT_RING(ring, submit->seqno); in a5xx_submit()
219 OUT_RING(ring, submit->seqno); in a5xx_submit()
[all …]
/linux-6.3-rc2/drivers/gpu/drm/i915/selftests/
A Di915_active.c81 struct i915_sw_fence *submit; in __live_active_setup() local
90 submit = heap_fence_create(GFP_KERNEL); in __live_active_setup()
91 if (!submit) { in __live_active_setup()
109 err = i915_sw_fence_await_sw_fence_gfp(&rq->submit, in __live_active_setup()
110 submit, in __live_active_setup()
135 i915_sw_fence_commit(submit); in __live_active_setup()
136 heap_fence_put(submit); in __live_active_setup()
/linux-6.3-rc2/drivers/gpu/drm/tegra/
A Dfirewall.c9 struct tegra_drm_submit_data *submit; member
31 for (i = 0; i < fw->submit->num_used_mappings; i++) { in fw_check_addr_valid()
32 struct tegra_drm_mapping *m = fw->submit->used_mappings[i].mapping; in fw_check_addr_valid()
145 u32 words, struct tegra_drm_submit_data *submit, in tegra_drm_fw_validate() argument
149 .submit = submit, in tegra_drm_fw_validate()

Completed in 46 milliseconds

12345678