| /drivers/infiniband/ulp/ipoib/ |
| A D | ipoib_ib.c | 279 u64 *mapping = tx_req->mapping; in ipoib_dma_map_tx() 319 struct ipoib_tx_buf *tx_req) in ipoib_dma_unmap_tx() argument 322 u64 *mapping = tx_req->mapping; in ipoib_dma_unmap_tx() 390 struct ipoib_tx_buf *tx_req; in ipoib_ib_handle_tx_wc() local 401 tx_req = &priv->tx_ring[wr_id]; in ipoib_ib_handle_tx_wc() 403 ipoib_dma_unmap_tx(priv, tx_req); in ipoib_ib_handle_tx_wc() 408 dev_kfree_skb_any(tx_req->skb); in ipoib_ib_handle_tx_wc() 574 ipoib_build_sge(priv, tx_req); in post_send() 595 struct ipoib_tx_buf *tx_req; in ipoib_send() local 652 tx_req->skb = skb; in ipoib_send() [all …]
|
| A D | ipoib_cm.c | 699 struct ipoib_tx_buf *tx_req) in post_send() argument 701 ipoib_build_sge(priv, tx_req); in post_send() 711 struct ipoib_tx_buf *tx_req; in ipoib_cm_send() local 751 tx_req->skb = skb; in ipoib_cm_send() 782 ipoib_dma_unmap_tx(priv, tx_req); in ipoib_cm_send() 799 struct ipoib_tx_buf *tx_req; in ipoib_cm_handle_tx_wc() local 811 tx_req = &tx->tx_ring[wr_id]; in ipoib_cm_handle_tx_wc() 813 ipoib_dma_unmap_tx(priv, tx_req); in ipoib_cm_handle_tx_wc() 819 dev_kfree_skb_any(tx_req->skb); in ipoib_cm_handle_tx_wc() 1202 struct ipoib_tx_buf *tx_req; in ipoib_cm_tx_destroy() local [all …]
|
| A D | ipoib.h | 533 int ipoib_dma_map_tx(struct ib_device *ca, struct ipoib_tx_buf *tx_req); 535 struct ipoib_tx_buf *tx_req); 540 struct ipoib_tx_buf *tx_req) in ipoib_build_sge() argument 543 struct sk_buff *skb = tx_req->skb; in ipoib_build_sge() 546 u64 *mapping = tx_req->mapping; in ipoib_build_sge()
|
| /drivers/net/wireless/realtek/rtw89/ |
| A D | core.c | 1034 if (tx_req->sta) in rtw89_core_tx_update_desc_info() 1036 else if (tx_req->vif) in rtw89_core_tx_update_desc_info() 1042 tx_req->tx_type = tx_type; in rtw89_core_tx_update_desc_info() 1045 tx_req->rtwsta_link); in rtw89_core_tx_update_desc_info() 1059 switch (tx_req->tx_type) { in rtw89_core_tx_update_desc_info() 1131 tx_req.skb = skb; in rtw89_h2c_tx() 1134 tx_req.desc_info.fw_dl = true; in rtw89_h2c_tx() 1168 tx_req.skb = skb; in rtw89_core_tx_write_link() 1169 tx_req.vif = vif; in rtw89_core_tx_write_link() 1170 tx_req.sta = sta; in rtw89_core_tx_write_link() [all …]
|
| A D | usb.c | 319 struct rtw89_core_tx_request *tx_req) in rtw89_usb_tx_write_fwcmd() argument 321 struct rtw89_tx_desc_info *desc_info = &tx_req->desc_info; in rtw89_usb_tx_write_fwcmd() 323 struct sk_buff *skb = tx_req->skb; in rtw89_usb_tx_write_fwcmd() 346 tx_req->skb = skb512; in rtw89_usb_tx_write_fwcmd() 361 struct rtw89_core_tx_request *tx_req) in rtw89_usb_ops_tx_write() argument 363 struct rtw89_tx_desc_info *desc_info = &tx_req->desc_info; in rtw89_usb_ops_tx_write() 365 struct sk_buff *skb = tx_req->skb; in rtw89_usb_ops_tx_write() 370 tx_req->tx_type == RTW89_CORE_TX_TYPE_FWCMD) && in rtw89_usb_ops_tx_write() 372 tx_req->tx_type != RTW89_CORE_TX_TYPE_FWCMD)) { in rtw89_usb_ops_tx_write() 374 desc_info->ch_dma, tx_req->tx_type); in rtw89_usb_ops_tx_write() [all …]
|
| A D | txrx.h | 717 rtw89_core_get_qsel_mgmt(struct rtw89_dev *rtwdev, struct rtw89_core_tx_request *tx_req) in rtw89_core_get_qsel_mgmt() argument 719 struct rtw89_tx_desc_info *desc_info = &tx_req->desc_info; in rtw89_core_get_qsel_mgmt() 720 struct rtw89_vif_link *rtwvif_link = tx_req->rtwvif_link; in rtw89_core_get_qsel_mgmt()
|
| A D | pci.c | 1364 struct rtw89_core_tx_request *tx_req) in rtw89_pci_txwd_submit() argument 1368 struct rtw89_tx_desc_info *desc_info = &tx_req->desc_info; in rtw89_pci_txwd_submit() 1372 struct sk_buff *skb = tx_req->skb; in rtw89_pci_txwd_submit() 1423 struct rtw89_core_tx_request *tx_req) in rtw89_pci_fwcmd_submit() argument 1427 struct rtw89_tx_desc_info *desc_info = &tx_req->desc_info; in rtw89_pci_fwcmd_submit() 1431 struct sk_buff *skb = tx_req->skb; in rtw89_pci_fwcmd_submit() 1462 struct rtw89_core_tx_request *tx_req) in rtw89_pci_txbd_submit() argument 1482 ret = rtw89_pci_txwd_submit(rtwdev, tx_ring, txwd, tx_req); in rtw89_pci_txbd_submit() 1517 tx_req->tx_type == RTW89_CORE_TX_TYPE_FWCMD) && in rtw89_pci_tx_write() 1519 tx_req->tx_type != RTW89_CORE_TX_TYPE_FWCMD)) { in rtw89_pci_tx_write() [all …]
|
| A D | core.h | 3634 int (*tx_write)(struct rtw89_dev *rtwdev, struct rtw89_core_tx_request *tx_req); 6185 struct rtw89_core_tx_request *tx_req) in rtw89_hci_tx_write() argument 6187 return rtwdev->hci.ops->tx_write(rtwdev, tx_req); in rtw89_hci_tx_write()
|
| /drivers/usb/gadget/udc/ |
| A D | fsl_qe_udc.c | 1141 if (ep->tx_req != NULL) { in txcomplete() 1168 done(ep, ep->tx_req, 0); in txcomplete() 1169 ep->tx_req = NULL; in txcomplete() 1176 if (ep->tx_req == NULL) { in txcomplete() 1317 tx_req = ep->tx_req; in ep0_txcomplete() 1318 if (tx_req != NULL) { in ep0_txcomplete() 1331 ep->tx_req = NULL; in ep0_txcomplete() 1460 if (ep->tx_req == NULL) { in ep_req_send() 1633 ep->tx_req = NULL; in qe_ep_disable() 1904 ep->tx_req = NULL; in udc_reset_ep_queue() [all …]
|
| A D | fsl_qe_udc.h | 285 struct qe_req *tx_req; member
|
| /drivers/media/platform/qcom/venus/ |
| A D | hfi_venus.c | 68 u32 tx_req; member 199 qhdr->tx_req = 1; in venus_write_queue() 205 qhdr->tx_req = 0; in venus_write_queue() 280 *tx_req = 0; in venus_read_queue() 334 *tx_req = qhdr->tx_req ? 1 : 0; in venus_read_queue() 691 u32 tx_req; in venus_iface_msgq_read_nolock() local 699 ret = venus_read_queue(hdev, queue, pkt, &tx_req); in venus_iface_msgq_read_nolock() 703 if (tx_req) in venus_iface_msgq_read_nolock() 724 u32 tx_req; in venus_iface_dbgq_read_nolock() local 737 if (tx_req) in venus_iface_dbgq_read_nolock() [all …]
|
| /drivers/net/wwan/t7xx/ |
| A D | t7xx_hif_cldma.c | 847 struct cldma_gpd *gpd = tx_req->gpd; in t7xx_cldma_gpd_handle_tx_request() 851 tx_req->mapped_buff = dma_map_single(md_ctrl->dev, skb->data, skb->len, DMA_TO_DEVICE); in t7xx_cldma_gpd_handle_tx_request() 853 if (dma_mapping_error(md_ctrl->dev, tx_req->mapped_buff)) { in t7xx_cldma_gpd_handle_tx_request() 858 t7xx_cldma_gpd_set_data_ptr(gpd, tx_req->mapped_buff); in t7xx_cldma_gpd_handle_tx_request() 870 tx_req->skb = skb; in t7xx_cldma_gpd_handle_tx_request() 923 struct cldma_request *tx_req; in t7xx_cldma_send_skb() local 948 tx_req = queue->tx_next; in t7xx_cldma_send_skb() 949 if (queue->budget > 0 && !tx_req->skb) { in t7xx_cldma_send_skb() 953 t7xx_cldma_gpd_handle_tx_request(queue, tx_req, skb); in t7xx_cldma_send_skb() 954 queue->tx_next = list_next_entry_circular(tx_req, gpd_ring, entry); in t7xx_cldma_send_skb() [all …]
|
| /drivers/media/platform/qcom/iris/ |
| A D | iris_hfi_queue.h | 136 u32 tx_req; member
|
| A D | iris_hfi_queue.c | 27 queue->tx_req = 0; in iris_hfi_queue_write() 212 iface_q->qhdr->tx_req = 0x0; in iris_hfi_queue_set_header()
|
| /drivers/perf/hisilicon/ |
| A D | hisi_uncore_pa_pmu.c | 320 HISI_PMU_EVENT_ATTR(tx_req, 0x5c), 331 HISI_PMU_EVENT_ATTR(tx_req, 0x0),
|
| A D | hisi_uncore_sllc_pmu.c | 385 HISI_PMU_EVENT_ATTR(tx_req, 0x34),
|
| /drivers/rapidio/ |
| A D | rio_cm.c | 138 struct tx_req { struct 668 struct tx_req *req, *_req; in rio_txcq_handler() 703 struct tx_req *treq; in riocm_queue_req()
|
| /drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
| A D | hclge_main.c | 9963 struct hclge_tx_vlan_type_cfg_cmd *tx_req; in hclge_set_vlan_protocol_type() local 9988 tx_req = (struct hclge_tx_vlan_type_cfg_cmd *)desc.data; in hclge_set_vlan_protocol_type() 9989 tx_req->ot_vlan_type = cpu_to_le16(hdev->vlan_type_cfg.tx_ot_vlan_type); in hclge_set_vlan_protocol_type() 9990 tx_req->in_vlan_type = cpu_to_le16(hdev->vlan_type_cfg.tx_in_vlan_type); in hclge_set_vlan_protocol_type()
|