| /drivers/net/ethernet/broadcom/bnxt/ |
| A D | bnxt_hwrm.h | 33 u32 req_len; member 128 int __hwrm_req_init(struct bnxt *bp, void **req, u16 req_type, u32 req_len); 149 u32 req_len; in bnxt_hwrm_func_cfg_short_req_init() local 151 req_len = min_t(u32, sizeof(**req), bp->hwrm_max_ext_req_len); in bnxt_hwrm_func_cfg_short_req_init() 152 return __hwrm_req_init(bp, (void **)req, HWRM_FUNC_CFG, req_len); in bnxt_hwrm_func_cfg_short_req_init()
|
| A D | bnxt_hwrm.c | 61 int __hwrm_req_init(struct bnxt *bp, void **req, u16 req_type, u32 req_len) in __hwrm_req_init() argument 67 if (req_len > BNXT_HWRM_CTX_OFFSET) in __hwrm_req_init() 78 ctx->req_len = req_len; in __hwrm_req_init() 220 ctx->req_len = len; in hwrm_req_replace() 469 u32 msg_len = ctx->req_len; in __hwrm_send() 794 if (addr < req_addr + max_offset && req_addr + ctx->req_len <= addr) { in hwrm_req_dma_slice()
|
| /drivers/watchdog/ |
| A D | mei_wdt.c | 169 const size_t req_len = sizeof(req); in mei_wdt_ping() local 172 memset(&req, 0, req_len); in mei_wdt_ping() 174 req.hdr.bytecount = req_len - offsetof(struct mei_mc_hdr, subcommand); in mei_wdt_ping() 179 ret = mei_cldev_send(wdt->cldev, (u8 *)&req, req_len); in mei_wdt_ping() 197 const size_t req_len = sizeof(req); in mei_wdt_stop() local 200 memset(&req, 0, req_len); in mei_wdt_stop() 202 req.hdr.bytecount = req_len - offsetof(struct mei_mc_hdr, subcommand); in mei_wdt_stop() 206 ret = mei_cldev_send(wdt->cldev, (u8 *)&req, req_len); in mei_wdt_stop()
|
| /drivers/char/tpm/ |
| A D | tpm_vtpm_proxy.c | 43 size_t req_len; /* length of queued TPM request */ member 80 proxy_dev->req_len != 0 || in vtpm_proxy_fops_read() 92 len = proxy_dev->req_len; in vtpm_proxy_fops_read() 103 proxy_dev->req_len = 0; in vtpm_proxy_fops_read() 147 proxy_dev->req_len = 0; in vtpm_proxy_fops_write() 182 if (proxy_dev->req_len) in vtpm_proxy_fops_poll() 355 proxy_dev->req_len = count; in vtpm_proxy_tpm_op_send()
|
| /drivers/misc/ |
| A D | rpmb-core.c | 51 unsigned int req_len, u8 *rsp, unsigned int rsp_len) in rpmb_route_frames() argument 53 if (!req || !req_len || !rsp || !rsp_len) in rpmb_route_frames() 56 return rdev->descr.route_frames(rdev->dev.parent, req, req_len, in rpmb_route_frames()
|
| /drivers/net/ethernet/broadcom/bnge/ |
| A D | bnge_hwrm.c | 22 u32 req_len) in bnge_hwrm_req_create() argument 28 if (req_len > BNGE_HWRM_CTX_OFFSET) in bnge_hwrm_req_create() 39 ctx->req_len = req_len; in bnge_hwrm_req_create() 257 u32 msg_len = ctx->req_len; in __hwrm_send_ctx() 459 if (addr < req_addr + max_offset && req_addr + ctx->req_len <= addr) { in bnge_hwrm_req_dma_slice()
|
| A D | bnge_hwrm.h | 26 u32 req_len; member 95 u32 req_len);
|
| /drivers/dma/ |
| A D | bcm-sba-raid.c | 731 size_t req_len; in sba_prep_dma_memcpy() local 741 req_len, flags); in sba_prep_dma_memcpy() 753 off += req_len; in sba_prep_dma_memcpy() 754 len -= req_len; in sba_prep_dma_memcpy() 870 size_t req_len; in sba_prep_dma_xor() local 884 req_len, flags); in sba_prep_dma_xor() 896 off += req_len; in sba_prep_dma_xor() 897 len -= req_len; in sba_prep_dma_xor() 1338 size_t req_len; in sba_prep_dma_pq() local 1420 off += req_len; in sba_prep_dma_pq() [all …]
|
| A D | tegra20-apb-dma.c | 150 unsigned int req_len; member 517 return sg_req->req_len - (status & TEGRA_APBDMA_STATUS_COUNT_MASK) - 4; in get_current_xferred_count() 579 dma_desc->bytes_transferred += sgreq->req_len; in handle_once_dma_done() 615 (dma_desc->bytes_transferred + sgreq->req_len) % in handle_cont_sngl_cycle_dma_done() 844 return sg_req->req_len; in tegra_dma_sg_bytes_xferred() 864 wcount = sg_req->req_len - 4; in tegra_dma_sg_bytes_xferred() 878 wcount = sg_req->req_len - 4; in tegra_dma_sg_bytes_xferred() 1134 sg_req->req_len = len; in tegra_dma_prep_slave_sg() 1269 sg_req->req_len = len; in tegra_dma_prep_dma_cyclic()
|
| /drivers/net/wireless/zydas/zd1211rw/ |
| A D | zd_usb.c | 1698 int r, i, req_len, actual_req_len, try_count = 0; in zd_usb_ioread16v() local 1743 if (req_len != actual_req_len) { in zd_usb_ioread16v() 1746 req_len, actual_req_len); in zd_usb_ioread16v() 1865 int i, req_len; in zd_usb_iowrite16v_async() local 1891 req_len = struct_size(req, reg_writes, count); in zd_usb_iowrite16v_async() 1892 req = kmalloc(req_len, GFP_KERNEL); in zd_usb_iowrite16v_async() 1910 req, req_len, iowrite16v_urb_complete, usb, in zd_usb_iowrite16v_async() 1914 req, req_len, iowrite16v_urb_complete, usb); in zd_usb_iowrite16v_async() 1956 int i, req_len, actual_req_len; in zd_usb_rfwrite() local 2020 if (req_len != actual_req_len) { in zd_usb_rfwrite() [all …]
|
| /drivers/infiniband/hw/hfi1/ |
| A D | pin_system.c | 204 size_t req_start, size_t req_len) in get_system_cache_entry() argument 208 u64 end = PFN_ALIGN(req_start + req_len); in get_system_cache_entry() 214 req_start, req_len, start, end); in get_system_cache_entry() 218 SDMA_DBG(req, "req_start %lx req_len %lu", req_start, req_len); in get_system_cache_entry()
|
| /drivers/tee/tstee/ |
| A D | core.c | 169 u32 req_len, ffa_args[FFA_DIRECT_REQ_ARG_NUM] = {}; in tstee_invoke_func() local 191 req_len = lower_32_bits(param[0].u.value.b); in tstee_invoke_func() 198 if (shm->size < req_len) { in tstee_invoke_func() 214 ffa_args[TS_RPC_SERVICE_REQ_LEN] = req_len; in tstee_invoke_func()
|
| /drivers/usb/gadget/udc/bdc/ |
| A D | bdc_ep.c | 387 u32 req_len; in setup_first_bd_ep0() local 390 req_len = req->usb_req.length; in setup_first_bd_ep0() 399 if ((wValue > req_len) && in setup_first_bd_ep0() 400 (req_len % bdc->gadget.ep0->maxpacket == 0)) { in setup_first_bd_ep0() 402 wValue, req_len, in setup_first_bd_ep0() 434 u32 req_len; in setup_bd_list_xfr() local 442 req_len = req->usb_req.length; in setup_bd_list_xfr() 465 if (req_len > BD_MAX_BUFF_SIZE) { in setup_bd_list_xfr() 467 req_len -= BD_MAX_BUFF_SIZE; in setup_bd_list_xfr() 470 dword2 |= req_len; in setup_bd_list_xfr() [all …]
|
| /drivers/scsi/snic/ |
| A D | snic_io.c | 241 rqi->req_len = sizeof(struct snic_host_req); in snic_req_init() 249 rqi->req_len += (sg_cnt * sizeof(struct snic_sg_desc)); in snic_req_init() 258 memset(rqi->req, 0, rqi->req_len); in snic_req_init() 361 rqi->req_len, in snic_req_free()
|
| A D | snic_io.h | 67 u16 req_len; /* buf len passing to fw (req + sgl)*/ member
|
| /drivers/usb/gadget/function/ |
| A D | u_audio.c | 595 int req_len, i, ret; in u_audio_start_capture() local 606 req_len = ep->maxpacket; in u_audio_start_capture() 625 req->length = req_len; in u_audio_start_capture() 653 req_len = ep_fback->maxpacket; in u_audio_start_capture() 662 req_fback->length = req_len; in u_audio_start_capture() 665 req_fback->buf = kzalloc(req_len, GFP_ATOMIC); in u_audio_start_capture() 708 int req_len, i, ret; in u_audio_start_playback() local 741 req_len = p_pktsize; in u_audio_start_playback() 761 req->length = req_len; in u_audio_start_playback()
|
| /drivers/platform/arm64/ |
| A D | lenovo-yoga-c630.c | 38 static int yoga_c630_ec_request(struct yoga_c630_ec *ec, u8 *req, size_t req_len, in yoga_c630_ec_request() argument 46 req_len, req); in yoga_c630_ec_request()
|
| /drivers/vfio/platform/ |
| A D | vfio_platform_common.c | 555 u64 req_len, pgoff, req_start; in vfio_platform_mmap_mmio() local 557 req_len = vma->vm_end - vma->vm_start; in vfio_platform_mmap_mmio() 562 if (region.size < PAGE_SIZE || req_start + req_len > region.size) in vfio_platform_mmap_mmio() 569 req_len, vma->vm_page_prot); in vfio_platform_mmap_mmio()
|
| /drivers/net/ethernet/amd/pds_core/ |
| A D | auxbus.c | 94 size_t req_len, in pds_client_adminq_cmd() argument 113 cp_len = min_t(size_t, req_len, sizeof(cmd.client_request.client_cmd)); in pds_client_adminq_cmd()
|
| /drivers/iio/accel/ |
| A D | mma9551_core.c | 116 u8 req_len, err_code; in mma9551_transfer() local 124 req_len = 1 + MMA9551_MAILBOX_CTRL_REGS + num_inbytes; in mma9551_transfer() 139 out.len = req_len; in mma9551_transfer()
|
| /drivers/net/ethernet/freescale/enetc/ |
| A D | ntmp.c | 239 u8 tbl_ver, u32 entry_id, u32 req_len, in ntmp_delete_entry_by_id() argument 244 .size = max(req_len, resp_len), in ntmp_delete_entry_by_id() 255 ntmp_fill_request_hdr(&cbd, data.dma, NTMP_LEN(req_len, resp_len), in ntmp_delete_entry_by_id()
|
| /drivers/vfio/pci/nvgrace-gpu/ |
| A D | main.c | 141 u64 req_len, pgoff, end; in nvgrace_gpu_mmap() local 159 if (check_sub_overflow(vma->vm_end, vma->vm_start, &req_len) || in nvgrace_gpu_mmap() 161 check_add_overflow(PFN_PHYS(pgoff), req_len, &end)) in nvgrace_gpu_mmap() 199 req_len, vma->vm_page_prot); in nvgrace_gpu_mmap()
|
| /drivers/net/ethernet/microsoft/mana/ |
| A D | hw_channel.c | 834 int mana_hwc_send_request(struct hw_channel_context *hwc, u32 req_len, in mana_hwc_send_request() argument 851 if (req_len > tx_wr->buf_len) { in mana_hwc_send_request() 852 dev_err(hwc->dev, "HWC: req msg size: %d > %d\n", req_len, in mana_hwc_send_request() 864 memcpy(req_msg, req, req_len); in mana_hwc_send_request() 868 tx_wr->msg_size = req_len; in mana_hwc_send_request()
|
| /drivers/scsi/mvsas/ |
| A D | mv_sas.c | 300 u32 req_len, resp_len, tag = tei->tag; in mvs_task_prep_smp() local 315 req_len = sg_dma_len(sg_req); in mvs_task_prep_smp() 326 if ((req_len & 0x3) || (resp_len & 0x3)) { in mvs_task_prep_smp() 374 hdr->lens = cpu_to_le32(((resp_len / 4) << 16) | ((req_len - 4) / 4)); in mvs_task_prep_smp() 432 u32 i, req_len, resp_len; in mvs_task_prep_ata() local 510 req_len = sizeof(struct host_to_dev_fis); in mvs_task_prep_ata() 516 hdr->lens = cpu_to_le32(((resp_len / 4) << 16) | (req_len / 4)); in mvs_task_prep_ata() 560 u32 resp_len, req_len, i, tag = tei->tag; in mvs_task_prep_ssp() local 625 req_len = sizeof(struct ssp_frame_hdr) + 28; in mvs_task_prep_ssp() 628 hdr->lens = cpu_to_le32(((resp_len / 4) << 16) | (req_len / 4)); in mvs_task_prep_ssp()
|
| /drivers/scsi/isci/ |
| A D | sas.h | 203 u8 req_len; /* byte 3 */ member
|