| /drivers/gpu/drm/exynos/ |
| A D | exynos_drm_g2d.c | 629 struct g2d_buf_desc *buf_desc, in g2d_check_buf_desc_is_valid() argument 644 width = (int)buf_desc->right_x - (int)buf_desc->left_x; in g2d_check_buf_desc_is_valid() 651 height = (int)buf_desc->bottom_y - (int)buf_desc->top_y; in g2d_check_buf_desc_is_valid() 662 (unsigned long)buf_desc->stride + in g2d_check_buf_desc_is_valid() 693 struct g2d_buf_desc *buf_desc; in g2d_map_cmdlist_gem() local 776 struct g2d_buf_desc *buf_desc; in g2d_unmap_cmdlist_gem() local 793 memset(buf_desc, 0x00, sizeof(*buf_desc)); in g2d_unmap_cmdlist_gem() 1034 struct g2d_buf_desc *buf_desc; in g2d_check_reg_offset() local 1085 buf_desc->format = value & 0xf; in g2d_check_reg_offset() 1097 buf_desc->left_x = value & 0x1fff; in g2d_check_reg_offset() [all …]
|
| /drivers/net/ethernet/broadcom/ |
| A D | bcm4908_enet.c | 72 struct bcm4908_enet_dma_ring_bd *buf_desc; member 268 struct bcm4908_enet_dma_ring_bd *buf_desc = &enet->rx_ring.buf_desc[idx]; in bcm4908_enet_dma_alloc_rx_buf() local 292 buf_desc->ctl = cpu_to_le32(tmp); in bcm4908_enet_dma_alloc_rx_buf() 527 struct bcm4908_enet_dma_ring_bd *buf_desc; in bcm4908_enet_start_xmit() local 547 buf_desc = &ring->buf_desc[ring->write_idx]; in bcm4908_enet_start_xmit() 571 buf_desc->ctl = cpu_to_le32(tmp); in bcm4908_enet_start_xmit() 589 struct bcm4908_enet_dma_ring_bd *buf_desc; in bcm4908_enet_poll_rx() local 596 buf_desc = &enet->rx_ring.buf_desc[enet->rx_ring.read_idx]; in bcm4908_enet_poll_rx() 597 ctl = le32_to_cpu(buf_desc->ctl); in bcm4908_enet_poll_rx() 655 struct bcm4908_enet_dma_ring_bd *buf_desc; in bcm4908_enet_poll_tx() local [all …]
|
| /drivers/net/wireless/realtek/rtw88/ |
| A D | pci.c | 214 struct rtw_pci_rx_buffer_desc *buf_desc; in rtw_pci_reset_rx_desc() local 228 memset(buf_desc, 0, sizeof(*buf_desc)); in rtw_pci_reset_rx_desc() 230 buf_desc->dma = cpu_to_le32(dma); in rtw_pci_reset_rx_desc() 240 struct rtw_pci_rx_buffer_desc *buf_desc; in rtw_pci_sync_rx_desc_device() local 247 memset(buf_desc, 0, sizeof(*buf_desc)); in rtw_pci_sync_rx_desc_device() 249 buf_desc->dma = cpu_to_le32(dma); in rtw_pci_sync_rx_desc_device() 689 struct rtw_pci_rx_buffer_desc *buf_desc; in rtw_pci_dma_check() local 814 struct rtw_pci_tx_buffer_desc *buf_desc; in rtw_pci_tx_write_data() local 836 memset(buf_desc, 0, tx_buf_desc_sz); in rtw_pci_tx_write_data() 843 buf_desc[0].dma = cpu_to_le32(dma); in rtw_pci_tx_write_data() [all …]
|
| A D | pci.h | 274 u8 *buf_desc; in get_tx_buffer_desc() local 276 buf_desc = ring->r.head + ring->r.wp * size; in get_tx_buffer_desc() 277 return (struct rtw_pci_tx_buffer_desc *)buf_desc; in get_tx_buffer_desc()
|
| /drivers/spi/ |
| A D | spi-pic32-sqi.c | 97 struct buf_desc { struct 129 struct buf_desc *bd; 264 struct buf_desc *bd; in pic32_sqi_one_transfer() 457 struct buf_desc *bd; in ring_desc_ring_alloc() 478 bd = (struct buf_desc *)sqi->bd; in ring_desc_ring_alloc() 502 sizeof(struct buf_desc) * PESQI_BD_COUNT, in ring_desc_ring_free()
|
| /drivers/net/ethernet/huawei/hinic/ |
| A D | hinic_hw_qp.c | 928 sge->hi_addr = be32_to_cpu(rq_wqe->buf_desc.hi_addr); in hinic_rq_get_sge() 929 sge->lo_addr = be32_to_cpu(rq_wqe->buf_desc.lo_addr); in hinic_rq_get_sge() 944 struct hinic_rq_bufdesc *buf_desc = &rq_wqe->buf_desc; in hinic_rq_prepare_wqe() local 953 HINIC_RQ_CTRL_SET(SIZE_8BYTES(sizeof(*buf_desc)), in hinic_rq_prepare_wqe() 959 buf_desc->hi_addr = sge->hi_addr; in hinic_rq_prepare_wqe() 960 buf_desc->lo_addr = sge->lo_addr; in hinic_rq_prepare_wqe()
|
| A D | hinic_hw_wqe.h | 340 struct hinic_scmd_bufdesc buf_desc; member 348 struct hinic_lcmd_bufdesc buf_desc; member 418 struct hinic_rq_bufdesc buf_desc; member
|
| A D | hinic_hw_cmdq.c | 231 hinic_set_sge(&wqe_lcmd->buf_desc.sge, buf_in->dma_addr, buf_in->size); in cmdq_set_lcmd_bufdesc() 239 wqe_scmd->buf_desc.buf_len = in_size; in cmdq_set_direct_wqe_data() 240 memcpy(wqe_scmd->buf_desc.data, buf_in, in_size); in cmdq_set_direct_wqe_data()
|
| /drivers/net/ethernet/huawei/hinic3/ |
| A D | hinic3_tx.c | 69 struct hinic3_sq_bufdesc *buf_desc = wqe_combo->bds_head; in hinic3_tx_map_skb() local 92 buf_desc = wqe_combo->bds_sec2; in hinic3_tx_map_skb() 104 hinic3_set_buf_desc(buf_desc, dma_info[idx].dma, in hinic3_tx_map_skb() 106 buf_desc++; in hinic3_tx_map_skb()
|
| /drivers/net/ethernet/intel/idpf/ |
| A D | idpf_txrx.c | 3527 struct virtchnl2_splitq_rx_buf_desc *buf_desc) in idpf_rx_update_bufq_desc() argument 3541 buf_desc->pkt_addr = cpu_to_le64(addr); in idpf_rx_update_bufq_desc() 3542 buf_desc->qword0.buf_id = cpu_to_le16(buf_id); in idpf_rx_update_bufq_desc() 3555 buf_desc->hdr_addr = cpu_to_le64(addr); in idpf_rx_update_bufq_desc() 3570 struct virtchnl2_splitq_rx_buf_desc *buf_desc; in idpf_rx_clean_refillq() local 3575 buf_desc = &bufq->split_buf[bufq_nta]; in idpf_rx_clean_refillq() 3587 failure = idpf_rx_update_bufq_desc(bufq, buf_id, buf_desc); in idpf_rx_clean_refillq() 3597 buf_desc = &bufq->split_buf[0]; in idpf_rx_clean_refillq() 3600 buf_desc++; in idpf_rx_clean_refillq()
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | psp_gfx_if.h | 173 struct psp_gfx_buf_desc buf_desc[GFX_BUF_MAX_DESC]; /* list of buffer descriptors */ member
|
| /drivers/net/ethernet/sfc/falcon/ |
| A D | farch.c | 170 ef4_qword_t buf_desc; in ef4_init_special_buffer() local 184 EF4_POPULATE_QWORD_3(buf_desc, in ef4_init_special_buffer() 188 ef4_write_buf_tbl(efx, &buf_desc, index); in ef4_init_special_buffer()
|
| /drivers/net/ethernet/sfc/siena/ |
| A D | farch.c | 174 efx_qword_t buf_desc; in efx_init_special_buffer() local 188 EFX_POPULATE_QWORD_3(buf_desc, in efx_init_special_buffer() 192 efx_write_buf_tbl(efx, &buf_desc, index); in efx_init_special_buffer()
|