| /linux/drivers/staging/media/atomisp/pci/base/circbuf/src/ |
| A D | circbuf.c | 80 cb->desc = desc; in ia_css_circbuf_create() 83 cb->desc->end = 0; in ia_css_circbuf_create() 89 cb->elems = elems; in ia_css_circbuf_create() 98 cb->desc = NULL; in ia_css_circbuf_destroy() 100 cb->elems = NULL; in ia_css_circbuf_destroy() 133 max_offset = ia_css_circbuf_get_offset(cb, cb->desc->start, cb->desc->end); in ia_css_circbuf_extract() 178 pos = ia_css_circbuf_get_pos_at_offset(cb, cb->desc->end, offset); in ia_css_circbuf_peek() 277 elem = cb->elems[cb->desc->start]; in ia_css_circbuf_read() 280 ia_css_circbuf_elem_init(&cb->elems[cb->desc->start]); in ia_css_circbuf_read() 283 cb->desc->start = ia_css_circbuf_get_pos_at_offset(cb, cb->desc->start, 1); in ia_css_circbuf_read() [all …]
|
| /linux/drivers/staging/media/atomisp/pci/base/circbuf/interface/ |
| A D | ia_css_circbuf.h | 153 OP___assert(cb); in ia_css_circbuf_get_pos_at_offset() 186 OP___assert(cb); in ia_css_circbuf_get_offset() 207 OP___assert(cb); in ia_css_circbuf_get_size() 225 OP___assert(cb); in ia_css_circbuf_get_num_elems() 228 num = ia_css_circbuf_get_offset(cb, cb->desc->start, cb->desc->end); in ia_css_circbuf_get_num_elems() 245 OP___assert(cb); in ia_css_circbuf_is_empty() 262 OP___assert(cb); in ia_css_circbuf_is_full() 281 OP___assert(cb); in ia_css_circbuf_write() 287 ia_css_circbuf_elem_cpy(&elem, &cb->elems[cb->desc->end]); in ia_css_circbuf_write() 289 cb->desc->end = ia_css_circbuf_get_pos_at_offset(cb, cb->desc->end, 1); in ia_css_circbuf_write() [all …]
|
| /linux/drivers/accel/habanalabs/common/ |
| A D | command_buffer.c | 33 cb->roundup_size = roundup(cb->size, page_size); in cb_map_mem() 43 rc = hl_mmu_map_contiguous(ctx, cb->virtual_addr, cb->bus_address, cb->roundup_size); in cb_map_mem() 86 hl_asic_dma_free_coherent(hdev, cb->size, cb->kernel_address, cb->bus_address); in cb_fini() 119 cb = kzalloc(sizeof(*cb), GFP_ATOMIC); in hl_cb_alloc() 122 cb = kzalloc(sizeof(*cb), GFP_KERNEL); in hl_cb_alloc() 175 cb_unmap_mem(cb->ctx, cb); in hl_cb_mmap_mem_release() 179 cb_do_release(cb->hdev, cb); in hl_cb_mmap_mem_release() 220 cb->buf->mappable_size = cb->size; in hl_cb_mmap_mem_alloc() 221 cb->buf->private = cb; in hl_cb_mmap_mem_alloc() 254 return cb->hdev->asic_funcs->mmap(cb->hdev, vma, cb->kernel_address, in hl_cb_mmap() [all …]
|
| /linux/drivers/scsi/ |
| A D | myrb.c | 172 cb->prev_cmd_mbox2 = cb->prev_cmd_mbox1; in myrb_qcmd() 193 cb->qcmd(cb, cmd_blk); in myrb_exec_cmd() 705 cb->old_ev_seq, cb->new_ev_seq, in myrb_hba_enquiry() 860 cb->next_cmd_mbox = cb->first_cmd_mbox; in myrb_enable_mmio() 1215 cb->enquiry, cb->enquiry_addr); in myrb_unmap() 1242 cb->disable_intr(cb->io_base); in myrb_cleanup() 1246 free_irq(cb->irq, cb); in myrb_cleanup() 1259 cb->reset(cb->io_base); in myrb_host_reset() 1333 cb->qcmd(cb, cmd_blk); in myrb_pthru_queuecommand() 1600 cb->qcmd(cb, cmd_blk); in myrb_ldev_queuecommand() [all …]
|
| /linux/drivers/mfd/ |
| A D | lm3533-ctrlbank.c | 30 return base + cb->id; in lm3533_ctrlbank_get_reg() 38 dev_dbg(cb->dev, "%s - %d\n", __func__, cb->id); in lm3533_ctrlbank_enable() 40 mask = 1 << cb->id; in lm3533_ctrlbank_enable() 44 dev_err(cb->dev, "failed to enable ctrlbank %d\n", cb->id); in lm3533_ctrlbank_enable() 55 dev_dbg(cb->dev, "%s - %d\n", __func__, cb->id); in lm3533_ctrlbank_disable() 57 mask = 1 << cb->id; in lm3533_ctrlbank_disable() 60 dev_err(cb->dev, "failed to disable ctrlbank %d\n", cb->id); in lm3533_ctrlbank_disable() 83 ret = lm3533_write(cb->lm3533, reg, val); in lm3533_ctrlbank_set_max_current() 97 ret = lm3533_write(cb->lm3533, reg, val); in lm3533_ctrlbank_set_brightness() 111 ret = lm3533_read(cb->lm3533, reg, val); in lm3533_ctrlbank_get_brightness() [all …]
|
| /linux/drivers/irqchip/ |
| A D | irq-crossbar.c | 47 writel(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]); in crossbar_writel() 52 writew(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]); in crossbar_writew() 57 writeb(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]); in crossbar_writeb() 163 cb->write(d->hwirq, cb->safe_map); in crossbar_domain_free() 202 cb = kzalloc(sizeof(*cb), GFP_KERNEL); in crossbar_of_init() 204 if (!cb) in crossbar_of_init() 226 if (!cb->irq_map) in crossbar_of_init() 229 cb->int_max = max; in crossbar_of_init() 313 cb->write(i, cb->safe_map); in crossbar_of_init() 327 kfree(cb); in crossbar_of_init() [all …]
|
| /linux/block/ |
| A D | blk-stat.c | 79 struct blk_stat_callback *cb = from_timer(cb, t, timer); in blk_stat_timer_fn() local 96 cb->timer_fn(cb); in blk_stat_timer_fn() 106 cb = kmalloc(sizeof(*cb), GFP_KERNEL); in blk_stat_alloc_callback() 107 if (!cb) in blk_stat_alloc_callback() 112 if (!cb->stat) { in blk_stat_alloc_callback() 113 kfree(cb); in blk_stat_alloc_callback() 120 kfree(cb); in blk_stat_alloc_callback() 126 cb->data = data; in blk_stat_alloc_callback() 130 return cb; in blk_stat_alloc_callback() 175 kfree(cb); in blk_stat_free_callback_rcu() [all …]
|
| /linux/drivers/net/mdio/ |
| A D | mdio-mux.c | 132 while (cb) { in mdio_mux_uninit_children() 134 mdiobus_free(cb->mii_bus); in mdio_mux_uninit_children() 135 cb = cb->next; in mdio_mux_uninit_children() 198 cb = devm_kzalloc(dev, sizeof(*cb), GFP_KERNEL); in mdio_mux_init() 199 if (!cb) { in mdio_mux_init() 203 cb->bus_number = v; in mdio_mux_init() 204 cb->parent = pb; in mdio_mux_init() 207 if (!cb->mii_bus) { in mdio_mux_init() 211 cb->mii_bus->priv = cb; in mdio_mux_init() 232 devm_kfree(dev, cb); in mdio_mux_init() [all …]
|
| /linux/fs/nfsd/ |
| A D | nfs4callback.c | 1207 if (cb->cb_ops && cb->cb_ops->release) in nfsd41_destroy_cb() 1208 cb->cb_ops->release(cb); in nfsd41_destroy_cb() 1228 cb->cb_status = 0; in nfsd4_cb_prepare() 1344 switch (cb->cb_ops->done(cb, task)) { in nfsd4_cb_done() 1507 if (cb->cb_ops && cb->cb_ops->prepare) in nfsd4_run_cb_work() 1508 cb->cb_ops->prepare(cb); in nfsd4_run_cb_work() 1513 cb->cb_ops ? &nfsd4_cb_ops : &nfsd4_cb_probe_ops, cb); in nfsd4_run_cb_work() 1519 cb->cb_clp = clp; in nfsd4_init_cb() 1521 cb->cb_msg.rpc_argp = cb; in nfsd4_init_cb() 1522 cb->cb_msg.rpc_resp = cb; in nfsd4_init_cb() [all …]
|
| /linux/fs/dlm/ |
| A D | ast.c | 43 dlm_run_callback(cb->ls_id, cb->lkb_id, cb->mode, cb->flags, in dlm_do_callback() 44 cb->sb_flags, cb->sb_status, cb->lkb_lksb, in dlm_do_callback() 45 cb->astfn, cb->bastfn, cb->astparam, in dlm_do_callback() 46 cb->res_name, cb->res_length); in dlm_do_callback() 47 dlm_free_cb(cb); in dlm_do_callback() 54 dlm_do_callback(cb); in dlm_callback_work() 129 if (WARN_ON_ONCE(!*cb)) in dlm_get_cb() 138 (*cb)->flags = flags; in dlm_get_cb() 139 (*cb)->mode = mode; in dlm_get_cb() 250 list_del(&cb->list); in dlm_callback_resume() [all …]
|
| /linux/drivers/misc/mei/ |
| A D | interrupt.c | 35 cl = cb->cl; in mei_irq_compl_handler() 116 if (!cb) { in mei_cl_irq_read_msg() 122 if (!cb) in mei_cl_irq_read_msg() 162 if (cb->vtag && cb->vtag != vtag_hdr->vtag) { in mei_cl_irq_read_msg() 213 cb->buf.size, length, cb->buf_idx); in mei_cl_irq_read_msg() 219 mei_dma_ring_read(dev, cb->buf.data + cb->buf_idx, length); in mei_cl_irq_read_msg() 221 mei_read_slots(dev, cb->buf.data + cb->buf_idx, 0); in mei_cl_irq_read_msg() 223 mei_read_slots(dev, cb->buf.data + cb->buf_idx, length); in mei_cl_irq_read_msg() 239 if (cb) in mei_cl_irq_read_msg() 531 cl = cb->cl; in mei_irq_write_handler() [all …]
|
| A D | client.c | 326 kfree(cb); in mei_io_cb_free() 395 cb = kzalloc(sizeof(*cb), GFP_KERNEL); in mei_io_cb_init() 396 if (!cb) in mei_io_cb_init() 407 return cb; in mei_io_cb_init() 444 if (cl == cb->cl && (!fp || fp == cb->fp)) in mei_io_tx_list_free_cl() 494 if (!cb) in mei_cl_alloc_cb() 507 return cb; in mei_cl_alloc_cb() 533 if (!cb) in mei_cl_enqueue_ctrl_wr_cb() 1703 if (!cb) in mei_cl_read_start() 1773 if (!cb) in mei_msg_hdr_init() [all …]
|
| /linux/drivers/misc/sgi-gru/ |
| A D | gru_instructions.h | 22 extern int gru_check_status_proc(void *cb); 23 extern int gru_wait_proc(void *cb); 24 extern void gru_wait_abort_proc(void *cb); 69 unsigned long cb; member 390 struct gru_instruction *ins = (void *)cb; in gru_vstore() 403 struct gru_instruction *ins = (void *)cb; in gru_ivload() 416 struct gru_instruction *ins = (void *)cb; in gru_ivstore() 662 ret = gru_check_status_proc(cb); in gru_check_status() 673 static inline int gru_wait(void *cb) in gru_wait() argument 675 return gru_wait_proc(cb); in gru_wait() [all …]
|
| A D | grukservices.c | 365 if (cb) in gru_lock_async_resource() 511 ret = gru_wait_proc(cb); in gru_wait_abort_proc() 590 ret = gru_wait(cb); in send_noop_message() 798 void *cb; in gru_send_message_gpa() local 905 void *cb; in gru_read_gpa() local 914 ret = gru_wait(cb); in gru_read_gpa() 929 void *cb; in gru_copy_gpa() local 938 ret = gru_wait(cb); in gru_copy_gpa() 951 void *cb; in quicktest0() local 1048 void *cb0, *cb; in quicktest2() local [all …]
|
| /linux/arch/s390/kernel/ |
| A D | runtime_instr.c | 55 cb->rla = 0xfff; in init_runtime_instr_cb() 56 cb->s = 1; in init_runtime_instr_cb() 57 cb->k = 1; in init_runtime_instr_cb() 58 cb->ps = 1; in init_runtime_instr_cb() 59 cb->pc = 1; in init_runtime_instr_cb() 61 cb->v = 1; in init_runtime_instr_cb() 86 cb = kzalloc(sizeof(*cb), GFP_KERNEL); in SYSCALL_DEFINE2() 87 if (!cb) in SYSCALL_DEFINE2() 91 memset(cb, 0, sizeof(*cb)); in SYSCALL_DEFINE2() 94 init_runtime_instr_cb(cb); in SYSCALL_DEFINE2() [all …]
|
| /linux/drivers/net/ethernet/netronome/nfp/ |
| A D | ccm_mbox.c | 64 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_msg_init() local 75 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_maxlen() local 82 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_done() local 89 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_in_progress() local 97 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_set_busy() local 104 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_is_posted() local 111 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_mark_posted() local 137 cb = (void *)skb->cb; in nfp_ccm_mbox_mark_next_runner() 268 cb = (void *)skb->cb; in nfp_ccm_mbox_copy_out() 333 cb = (void *)skb->cb; in nfp_ccm_mbox_copy_out() [all …]
|
| /linux/include/trace/events/ |
| A D | notifier.h | 12 TP_PROTO(void *cb), 14 TP_ARGS(cb), 17 __field(void *, cb) 21 __entry->cb = cb; 24 TP_printk("%ps", __entry->cb) 35 TP_PROTO(void *cb), 37 TP_ARGS(cb) 48 TP_PROTO(void *cb), 50 TP_ARGS(cb) 61 TP_PROTO(void *cb), [all …]
|
| /linux/drivers/dma-buf/ |
| A D | st-dma-fence.c | 43 wake_up_process(container_of(cb, struct wait_cb, cb)->task); in mock_wakeup() 51 if (dma_fence_add_callback(f, &cb.cb, mock_wakeup)) in mock_wait() 67 if (!dma_fence_remove_callback(f, &cb.cb)) in mock_wait() 170 if (dma_fence_add_callback(f, &cb.cb, simple_callback)) { in test_add_callback() 201 if (!dma_fence_add_callback(f, &cb.cb, simple_callback)) { in test_late_add_callback() 207 if (cb.seen) { in test_late_add_callback() 228 if (dma_fence_add_callback(f, &cb.cb, simple_callback)) { in test_rm_callback() 233 if (!dma_fence_remove_callback(f, &cb.cb)) { in test_rm_callback() 239 if (cb.seen) { in test_rm_callback() 260 if (dma_fence_add_callback(f, &cb.cb, simple_callback)) { in test_late_rm_callback() [all …]
|
| /linux/drivers/dma/ |
| A D | dmaengine.h | 115 struct dmaengine_desc_callback *cb) in dmaengine_desc_get_callback() argument 117 cb->callback = tx->callback; in dmaengine_desc_get_callback() 118 cb->callback_result = tx->callback_result; in dmaengine_desc_get_callback() 119 cb->callback_param = tx->callback_param; in dmaengine_desc_get_callback() 140 if (cb->callback_result) { in dmaengine_desc_callback_invoke() 143 cb->callback_result(cb->callback_param, result); in dmaengine_desc_callback_invoke() 144 } else if (cb->callback) { in dmaengine_desc_callback_invoke() 145 cb->callback(cb->callback_param); in dmaengine_desc_callback_invoke() 163 struct dmaengine_desc_callback cb; in dmaengine_desc_get_callback_invoke() local 165 dmaengine_desc_get_callback(tx, &cb); in dmaengine_desc_get_callback_invoke() [all …]
|
| /linux/tools/testing/selftests/bpf/verifier/ |
| A D | ctx_skb.c | 396 offsetof(struct __sk_buff, cb[0])), 404 offsetof(struct __sk_buff, cb[1])), 412 offsetof(struct __sk_buff, cb[2])), 420 offsetof(struct __sk_buff, cb[3])), 428 offsetof(struct __sk_buff, cb[4])), 436 offsetof(struct __sk_buff, cb[0])), 444 offsetof(struct __sk_buff, cb[1])), 452 offsetof(struct __sk_buff, cb[2])), 460 offsetof(struct __sk_buff, cb[3])), 468 offsetof(struct __sk_buff, cb[4])), [all …]
|
| /linux/drivers/net/wireless/marvell/mwifiex/ |
| A D | util.h | 43 struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; in MWIFIEX_SKB_RXCB() local 45 BUILD_BUG_ON(sizeof(struct mwifiex_cb) > sizeof(skb->cb)); in MWIFIEX_SKB_RXCB() 46 return &cb->rx_info; in MWIFIEX_SKB_RXCB() 51 struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; in MWIFIEX_SKB_TXCB() local 53 return &cb->tx_info; in MWIFIEX_SKB_TXCB() 59 struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; in mwifiex_store_mapping() local 61 memcpy(&cb->dma_mapping, mapping, sizeof(*mapping)); in mwifiex_store_mapping() 67 struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; in mwifiex_get_mapping() local 69 memcpy(mapping, &cb->dma_mapping, sizeof(*mapping)); in mwifiex_get_mapping()
|
| /linux/drivers/gpu/drm/i915/ |
| A D | i915_sw_fence.c | 421 struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base); in dma_i915_sw_fence_wake() local 425 kfree(cb); in dma_i915_sw_fence_wake() 430 struct i915_sw_dma_fence_cb_timer *cb = from_timer(cb, t, timer); in timer_i915_sw_fence_wake() local 438 cb->dma->ops->get_driver_name(cb->dma), in timer_i915_sw_fence_wake() 439 cb->dma->ops->get_timeline_name(cb->dma), in timer_i915_sw_fence_wake() 440 cb->dma->seqno, in timer_i915_sw_fence_wake() 471 kfree_rcu(cb, rcu); in irq_i915_sw_fence_work() 495 if (!cb) { in i915_sw_fence_await_dma_fence() 507 cb->fence = fence; in i915_sw_fence_await_dma_fence() 540 struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base); in __dma_i915_sw_fence_wake() local [all …]
|
| /linux/drivers/s390/net/ |
| A D | smsgiucv.c | 55 struct smsg_callback *cb; in smsg_message_pending() local 79 if (strncmp(buffer + 8, cb->prefix, cb->len) == 0) { in smsg_message_pending() 91 struct smsg_callback *cb; in smsg_register_callback() local 94 if (!cb) in smsg_register_callback() 96 cb->prefix = prefix; in smsg_register_callback() 97 cb->len = strlen(prefix); in smsg_register_callback() 98 cb->callback = callback; in smsg_register_callback() 112 cb = NULL; in smsg_unregister_callback() 116 cb = tmp; in smsg_unregister_callback() 117 list_del(&cb->list); in smsg_unregister_callback() [all …]
|
| /linux/arch/x86/kernel/cpu/ |
| A D | scattered.c | 62 const struct cpuid_bit *cb; in init_scattered_cpuid_features() local 64 for (cb = cpuid_bits; cb->feature; cb++) { in init_scattered_cpuid_features() 67 max_level = cpuid_eax(cb->level & 0xffff0000); in init_scattered_cpuid_features() 68 if (max_level < cb->level || in init_scattered_cpuid_features() 69 max_level > (cb->level | 0xffff)) in init_scattered_cpuid_features() 72 cpuid_count(cb->level, cb->sub_leaf, ®s[CPUID_EAX], in init_scattered_cpuid_features() 76 if (regs[cb->reg] & (1 << cb->bit)) in init_scattered_cpuid_features() 77 set_cpu_cap(c, cb->feature); in init_scattered_cpuid_features()
|
| /linux/fs/btrfs/ |
| A D | compression.c | 126 switch (cb->compress_type) { in compression_decompress_bio() 286 unsigned long end_index = (cb->start + cb->len - 1) >> PAGE_SHIFT; in end_compressed_writeback() 307 cb->start, cb->len); in end_compressed_writeback() 319 btrfs_finish_ordered_extent(cb->bbio.ordered, NULL, cb->start, cb->len, in btrfs_finish_compressed_write_work() 322 if (cb->writeback) in btrfs_finish_compressed_write_work() 327 bio_put(&cb->bbio.bio); in btrfs_finish_compressed_write_work() 379 struct compressed_bio *cb; in btrfs_submit_compressed_write() local 605 cb->orig_bbio = bbio; in btrfs_submit_compressed_read() 610 cb->compressed_folios = kcalloc(cb->nr_folios, sizeof(struct page *), GFP_NOFS); in btrfs_submit_compressed_read() 616 ret2 = btrfs_alloc_folio_array(cb->nr_folios, cb->compressed_folios); in btrfs_submit_compressed_read() [all …]
|