| /drivers/target/iscsi/cxgbit/ |
| A D | cxgbit_main.c | 231 gl->frags[i].page, in cxgbit_copy_frags() 232 gl->frags[i].offset, in cxgbit_copy_frags() 233 gl->frags[i].size); in cxgbit_copy_frags() 238 get_page(gl->frags[gl->nfrags - 1].page); in cxgbit_copy_frags() 260 if (unlikely(gl->nfrags > 1)) in cxgbit_lro_add_packet_gl() 308 pdu_cb->frags += gl->nfrags; in cxgbit_lro_add_packet_gl() 413 if ((gl && (((skb_shinfo(skb)->nr_frags + gl->nfrags) > in cxgbit_lro_receive() 420 if (gl) in cxgbit_lro_receive() 462 rpl = gl ? (struct cpl_tx_data *)gl->va : in cxgbit_uld_lro_rx_handler() 474 if (!gl) { in cxgbit_uld_lro_rx_handler() [all …]
|
| /drivers/net/ethernet/chelsio/cxgb4vf/ |
| A D | sge.c | 1486 get_page(gl->frags[gl->nfrags - 1].page); in copy_frags() 1521 skb_copy_to_linear_data(skb, gl->va, gl->tot_len); in t4vf_pktgl_to_skb() 1550 frag = gl->nfrags - 1; in t4vf_pktgl_free() 1575 t4vf_pktgl_free(gl); in do_gro() 1630 do_gro(rxq, gl, pkt); in t4vf_ethrx_handler() 1639 t4vf_pktgl_free(gl); in t4vf_ethrx_handler() 1777 struct pkt_gl gl; in process_responses() local 1799 gl.tot_len = len; in process_responses() 1817 gl.nfrags = frag+1; in process_responses() 1827 gl.va = (page_address(gl.frags[0].page) + in process_responses() [all …]
|
| A D | cxgb4vf_main.c | 488 const struct pkt_gl *gl) in fwevtq_handler() argument
|
| /drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
| A D | chtls_main.c | 339 static struct sk_buff *copy_gl_to_skb_pkt(const struct pkt_gl *gl, in copy_gl_to_skb_pkt() argument 349 skb = alloc_skb(size_add(gl->tot_len, in copy_gl_to_skb_pkt() 354 __skb_put(skb, gl->tot_len + sizeof(struct cpl_pass_accept_req) in copy_gl_to_skb_pkt() 359 , gl->va + pktshift, in copy_gl_to_skb_pkt() 360 gl->tot_len - pktshift); in copy_gl_to_skb_pkt() 366 const struct pkt_gl *gl, const __be64 *rsp) in chtls_recv_packet() argument 372 skb = copy_gl_to_skb_pkt(gl, rsp, cdev->lldi->sge_pktshift); in chtls_recv_packet() 440 const struct pkt_gl *gl) in chtls_uld_rx_handler() argument 449 if (chtls_recv_packet(cdev, gl, rsp) < 0) in chtls_uld_rx_handler() 454 if (!gl) in chtls_uld_rx_handler() [all …]
|
| /drivers/infiniband/hw/cxgb4/ |
| A D | device.c | 1116 skb = alloc_skb(size_add(gl->tot_len, in copy_gl_to_skb_pkt() 1138 gl->va + pktshift, in copy_gl_to_skb_pkt() 1139 gl->tot_len - pktshift); in copy_gl_to_skb_pkt() 1168 const struct pkt_gl *gl) in c4iw_uld_rx_handler() argument 1175 if (gl == NULL) { in c4iw_uld_rx_handler() 1184 } else if (gl == CXGB4_MSG_AN) { in c4iw_uld_rx_handler() 1191 if (recv_rx_pkt(dev, gl, rsp)) in c4iw_uld_rx_handler() 1195 pci_name(ctx->lldi.pdev), gl->va, in c4iw_uld_rx_handler() 1197 be64_to_cpu(*(__force __be64 *)gl->va), in c4iw_uld_rx_handler() 1198 gl->tot_len); in c4iw_uld_rx_handler() [all …]
|
| /drivers/gpu/drm/ci/ |
| A D | container.yml | 10 debian/x86_64_test-gl: 27 debian/baremetal_arm32_test-gl:
|
| /drivers/net/ethernet/chelsio/cxgb4/ |
| A D | sge.c | 3317 gl->frags[i].offset, in copy_frags() 3318 gl->frags[i].size); in copy_frags() 3321 get_page(gl->frags[gl->nfrags - 1].page); in copy_frags() 3347 __skb_put(skb, gl->tot_len); in cxgb4_pktgl_to_skb() 3348 skb_copy_to_linear_data(skb, gl->va, gl->tot_len); in cxgb4_pktgl_to_skb() 3356 copy_frags(skb, gl, pull_len); in cxgb4_pktgl_to_skb() 3357 skb->len = gl->tot_len; in cxgb4_pktgl_to_skb() 3377 for (p = gl->frags, n = gl->nfrags - 1; n--; p++) in t4_pktgl_free() 3392 t4_pktgl_free(gl); in handle_trace_pkt() 3441 t4_pktgl_free(gl); in do_gro() [all …]
|
| A D | cxgb4_uld.h | 483 const struct pkt_gl *gl); 487 const struct pkt_gl *gl, 521 struct sk_buff *cxgb4_pktgl_to_skb(const struct pkt_gl *gl,
|
| A D | cxgb4_uld.c | 75 const struct pkt_gl *gl) in uldrx_handler() argument 88 rsp, gl, &q->lro_mgr, in uldrx_handler() 92 rsp, gl); in uldrx_handler() 99 if (!gl) in uldrx_handler() 101 else if (gl == CXGB4_MSG_AN) in uldrx_handler()
|
| A D | cxgb4.h | 763 const struct pkt_gl *gl); 1616 const struct pkt_gl *gl);
|
| A D | cxgb4_main.c | 556 const struct pkt_gl *gl) in fwevtq_handler() argument
|
| /drivers/scsi/qedi/ |
| A D | qedi_main.c | 1519 struct global_queue **gl = qedi->global_queues; in qedi_free_global_queues() local 1522 if (!gl[i]) in qedi_free_global_queues() 1525 if (gl[i]->cq) in qedi_free_global_queues() 1526 dma_free_coherent(&qedi->pdev->dev, gl[i]->cq_mem_size, in qedi_free_global_queues() 1527 gl[i]->cq, gl[i]->cq_dma); in qedi_free_global_queues() 1528 if (gl[i]->cq_pbl) in qedi_free_global_queues() 1529 dma_free_coherent(&qedi->pdev->dev, gl[i]->cq_pbl_size, in qedi_free_global_queues() 1530 gl[i]->cq_pbl, gl[i]->cq_pbl_dma); in qedi_free_global_queues() 1532 kfree(gl[i]); in qedi_free_global_queues()
|
| /drivers/scsi/qedf/ |
| A D | qedf_main.c | 2947 struct global_queue **gl = qedf->global_queues; in qedf_free_global_queues() local 2950 if (!gl[i]) in qedf_free_global_queues() 2953 if (gl[i]->cq) in qedf_free_global_queues() 2955 gl[i]->cq_mem_size, gl[i]->cq, gl[i]->cq_dma); in qedf_free_global_queues() 2956 if (gl[i]->cq_pbl) in qedf_free_global_queues() 2957 dma_free_coherent(&qedf->pdev->dev, gl[i]->cq_pbl_size, in qedf_free_global_queues() 2958 gl[i]->cq_pbl, gl[i]->cq_pbl_dma); in qedf_free_global_queues() 2960 kfree(gl[i]); in qedf_free_global_queues()
|
| /drivers/net/ethernet/fungible/funeth/ |
| A D | funeth_main.c | 186 struct fun_dataop_gl gl; in fun_config_rss() member 208 fun_dataop_gl_init(&cmd.gl, 0, 0, len, fp->rss_dma_addr); in fun_config_rss()
|