| /linux/drivers/bus/mhi/ep/ |
| A D | ring.c | 26 ring->rd_offset = (ring->rd_offset + 1) % ring->ring_size; in mhi_ep_ring_inc_index() 122 if (ring->rd_offset < ring->wr_offset) in mhi_ep_ring_add_element() 123 num_free_elem = (ring->wr_offset - ring->rd_offset) - 1; in mhi_ep_ring_add_element() 125 num_free_elem = ((ring->ring_size - ring->rd_offset) + ring->wr_offset) - 1; in mhi_ep_ring_add_element() 133 old_offset = ring->rd_offset; in mhi_ep_ring_add_element() 136 dev_dbg(dev, "Adding an element to ring at offset (%zu)\n", ring->rd_offset); in mhi_ep_ring_add_element() 139 rp = cpu_to_le64(ring->rd_offset * sizeof(*el) + ring->rbase); in mhi_ep_ring_add_element() 200 ring->rd_offset = mhi_ep_ring_addr2offset(ring, le64_to_cpu(val)); in mhi_ep_ring_start()
|
| A D | main.c | 203 mhi_chan->rd_offset = ch_ring->rd_offset; in mhi_ep_process_cmd_ring() 333 return !!(mhi_chan->rd_offset == ring->wr_offset); in mhi_ep_queue_is_empty() 343 struct mhi_ring_element *el = &ring->ring_cache[ring->rd_offset]; in mhi_ep_read_completion() 424 el = &ring->ring_cache[mhi_chan->rd_offset]; in mhi_ep_read_channel() 469 mhi_chan->rd_offset = (mhi_chan->rd_offset + 1) % ring->ring_size; in mhi_ep_read_channel() 581 el = &ring->ring_cache[mhi_chan->rd_offset]; in mhi_ep_queue_skb() 618 mhi_chan->rd_offset = (mhi_chan->rd_offset + 1) % ring->ring_size; in mhi_ep_queue_skb() 796 if (ring->rd_offset == ring->wr_offset) in mhi_ep_cmd_ring_worker() 803 while (ring->rd_offset != ring->wr_offset) { in mhi_ep_cmd_ring_worker() 804 el = &ring->ring_cache[ring->rd_offset]; in mhi_ep_cmd_ring_worker() [all …]
|
| A D | internal.h | 131 size_t rd_offset; member 165 size_t rd_offset; member
|
| /linux/drivers/target/ |
| A D | target_core_rd.c | 425 u32 rd_offset; in rd_execute_rw() local 438 rd_offset = do_div(tmp, PAGE_SIZE); in rd_execute_rw() 451 cmd->t_task_lba, rd_size, rd_page, rd_offset); in rd_execute_rw() 460 src_len = PAGE_SIZE - rd_offset; in rd_execute_rw() 479 rd_page, rd_offset, rd_size); in rd_execute_rw() 484 rd_addr = sg_virt(rd_sg) + rd_offset; in rd_execute_rw() 497 rd_offset += len; in rd_execute_rw() 503 rd_offset = 0; in rd_execute_rw()
|
| /linux/drivers/gpu/drm/xe/ |
| A D | xe_gsc.c | 132 u32 rd_offset; in query_compatibility_version() local 163 &rd_offset); in query_compatibility_version() 169 compat->major = version_query_rd(xe, &bo->vmap, rd_offset, proj_major); in query_compatibility_version() 170 compat->minor = version_query_rd(xe, &bo->vmap, rd_offset, compat_major); in query_compatibility_version() 171 compat->patch = version_query_rd(xe, &bo->vmap, rd_offset, compat_minor); in query_compatibility_version()
|
| A D | xe_huc.c | 159 u32 rd_offset; in huc_auth_via_gsccs() local 195 sizeof(struct pxp43_huc_auth_out), &rd_offset); in huc_auth_via_gsccs() 206 out_status = huc_auth_msg_rd(xe, &pkt->vmap, rd_offset, header.status); in huc_auth_via_gsccs()
|
| /linux/drivers/gpu/drm/imagination/ |
| A D | pvr_ccb.c | 219 u32 rd_offset = READ_ONCE(ctrl->read_offset); in pvr_kccb_used_slot_count_locked() local 224 if (wr_offset >= rd_offset) in pvr_kccb_used_slot_count_locked() 225 used_count = wr_offset - rd_offset; in pvr_kccb_used_slot_count_locked() 227 used_count = wr_offset + pvr_dev->kccb.slot_count - rd_offset; in pvr_kccb_used_slot_count_locked()
|
| /linux/drivers/spi/ |
| A D | spi-mtk-snfi.c | 869 u32 rd_offset = 0; in mtk_snand_read_page_cache() local 892 rd_offset = op_addr & mask; in mtk_snand_read_page_cache() 896 if (rd_offset == 0 && op->data.nbytes >= snf->nfi_cfg.page_size) in mtk_snand_read_page_cache() 1021 } else if (rd_offset < snf->buf_len) { in mtk_snand_read_page_cache() 1022 u32 cap_len = snf->buf_len - rd_offset; in mtk_snand_read_page_cache() 1026 memcpy(op->data.buf.in, snf->buf + rd_offset, cap_len); in mtk_snand_read_page_cache()
|
| /linux/drivers/crypto/inside-secure/ |
| A D | safexcel.c | 565 writel(EIP197_xDR_DESC_MODE_64BIT | (priv->config.rd_offset << 14) | in safexcel_hw_setup_rdesc_rings() 571 (rd_fetch_cnt * (priv->config.rd_offset / sizeof(u32))), in safexcel_hw_setup_rdesc_rings() 764 writel((EIP197_DEFAULT_RING_SIZE * priv->config.rd_offset), in safexcel_hw_init() 890 writel((rdesc * priv->config.rd_offset), in safexcel_dequeue() 1053 (tot_descs * priv->config.rd_offset), in safexcel_handle_result_descriptor() 1343 priv->config.rd_offset = (priv->config.rd_size + mask) & ~mask; in safexcel_configure() 1348 priv->config.rd_offset *= sizeof(u32); in safexcel_configure()
|
| A D | safexcel_ring.c | 57 rdr->offset = priv->config.rd_offset; in safexcel_init_ring_descriptors()
|
| A D | safexcel.h | 684 u32 rd_offset; member
|
| /linux/drivers/scsi/lpfc/ |
| A D | lpfc_bsg.h | 279 uint32_t rd_offset; member
|
| A D | lpfc_bsg.c | 5262 u32 rd_offset, rd_index, offset; in lpfc_bsg_get_ras_fwlog() local 5295 rd_offset = ras_req->read_offset; in lpfc_bsg_get_ras_fwlog() 5304 rd_index = (rd_offset / LPFC_RAS_MAX_ENTRY_SIZE); in lpfc_bsg_get_ras_fwlog() 5305 offset = (rd_offset % LPFC_RAS_MAX_ENTRY_SIZE); in lpfc_bsg_get_ras_fwlog()
|
| /linux/fs/nfsd/ |
| A D | nfs4proc.c | 949 read->rd_offset, read->rd_length); in nfsd4_read() 952 if (read->rd_offset > (u64)OFFSET_MAX) in nfsd4_read() 953 read->rd_offset = (u64)OFFSET_MAX; in nfsd4_read() 954 if (read->rd_offset + read->rd_length > (u64)OFFSET_MAX) in nfsd4_read() 955 read->rd_length = (u64)OFFSET_MAX - read->rd_offset; in nfsd4_read() 988 u->read.rd_offset, u->read.rd_length); in nfsd4_read_release()
|
| A D | xdr4.h | 421 u64 rd_offset; /* request */ member
|
| A D | nfs4xdr.c | 1258 if (xdr_stream_decode_u64(argp->xdr, &read->rd_offset) < 0) in nfsd4_decode_read() 4337 file, read->rd_offset, &maxcount, in nfsd4_encode_splice_read() 4389 read->rd_offset, &maxcount, base, in nfsd4_encode_readv() 5309 p = xdr_encode_hyper(p, read->rd_offset); in nfsd4_encode_read_plus_data() 5335 read->rd_eof = read->rd_offset >= i_size_read(file_inode(file)); in nfsd4_encode_read_plus()
|