| /drivers/net/ethernet/intel/ice/ |
| A D | ice_flow.c | 1302 if (segs[i].hdrs != p->segs[i].hdrs || in ice_flow_find_prof_conds() 1304 segs[i].match != p->segs[i].match)) in ice_flow_find_prof_conds() 1415 memcpy(¶ms->prof->segs[i], &segs[i], sizeof(*segs)); in ice_flow_add_prof_sync() 1656 if (!segs) in ice_flow_add_prof() 1998 seg = &segs[seg_cnt - 1]; in ice_flow_set_rss_seg_info() 2348 segs = kcalloc(segs_cnt, sizeof(*segs), GFP_KERNEL); in ice_add_rss_cfg_sync() 2349 if (!segs) in ice_add_rss_cfg_sync() 2426 kfree(segs); in ice_add_rss_cfg_sync() 2495 segs = kcalloc(segs_cnt, sizeof(*segs), GFP_KERNEL); in ice_rem_rss_cfg_sync() 2496 if (!segs) in ice_rem_rss_cfg_sync() [all …]
|
| A D | ice_flow.h | 401 struct ice_flow_seg_info segs[ICE_FLOW_SEG_MAX]; member 418 struct ice_flow_seg_info *segs, u8 segs_cnt,
|
| /drivers/infiniband/sw/rdmavt/ |
| A D | mr.c | 384 mr->mr.map[m]->segs[n].vaddr = vaddr; in rvt_reg_user_mr() 385 mr->mr.map[m]->segs[n].length = PAGE_SIZE; in rvt_reg_user_mr() 574 mr->mr.map[m]->segs[n].vaddr = (void *)addr; in rvt_set_page() 575 mr->mr.map[m]->segs[n].length = ps; in rvt_set_page() 792 while (off >= mr->map[m]->segs[n].length) { in rvt_lkey_ok() 793 off -= mr->map[m]->segs[n].length; in rvt_lkey_ok() 802 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_lkey_ok() 899 while (off >= mr->map[m]->segs[n].length) { in rvt_rkey_ok() 900 off -= mr->map[m]->segs[n].length; in rvt_rkey_ok() 909 sge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_rkey_ok() [all …]
|
| /drivers/net/ethernet/broadcom/bnxt/ |
| A D | bnxt_coredump.c | 96 info->segs = le16_to_cpu(*((__le16 *)(resp + in bnxt_hwrm_dbg_dma_data() 98 if (!info->segs) { in bnxt_hwrm_dbg_dma_data() 103 info->dest_buf_size = info->segs * in bnxt_hwrm_dbg_dma_data() 167 coredump->total_segs = info.segs; in bnxt_hwrm_dbg_coredump_list() 346 u32 *segs) in bnxt_get_ctx_coredump() argument 356 *segs = 0; in bnxt_get_ctx_coredump() 395 *segs += 1; in bnxt_get_ctx_coredump() 433 u32 drv_len, segs = 0; in __bnxt_get_coredump() local 435 drv_len = bnxt_get_ctx_coredump(bp, buf, offset, &segs); in __bnxt_get_coredump() 439 coredump.total_segs += segs; in __bnxt_get_coredump()
|
| A D | bnxt_coredump.h | 128 u16 segs; member
|
| /drivers/net/wireguard/ |
| A D | device.c | 179 struct sk_buff *segs = skb_gso_segment(skb, 0); in wg_xmit() local 181 if (IS_ERR(segs)) { in wg_xmit() 182 ret = PTR_ERR(segs); in wg_xmit() 186 skb = segs; in wg_xmit()
|
| /drivers/net/ethernet/marvell/octeontx2/nic/ |
| A D | otx2_txrx.c | 259 for (seg = 0; seg < sg->segs; seg++, seg_addr++) in otx2_free_rcv_seg() 325 if (cqe->sg.segs) in otx2_check_rcv_errors() 368 for (seg = 0; seg < sg->segs; seg++, seg_addr++) { in otx2_rcv_pkt_handler() 677 sg->segs = 0; in otx2_sqe_add_sg() 693 sg->segs++; in otx2_sqe_add_sg() 911 sg->segs = 0; in otx2_sqe_tso_add_sg() 924 sg->segs++; in otx2_sqe_tso_add_sg() 1304 if (cqe->sg.segs > 1) { in otx2_cleanup_rx_cqes() 1426 sg->segs = 1; in otx2_xdp_sqe_add_sg()
|
| A D | otx2_struct.h | 153 u64 segs : 2; member 228 u64 segs : 2; member
|
| A D | cn10k_ipsec.h | 203 u64 segs : 2; member
|
| A D | cn10k_ipsec.c | 883 sg->segs = 0; in otx2_sqe_add_sg_ipsec() 902 sg->segs++; in otx2_sqe_add_sg_ipsec()
|
| /drivers/usb/host/ |
| A D | xhci-mem.c | 2295 xhci_alloc_interrupter(struct xhci_hcd *xhci, unsigned int segs, gfp_t flags) in xhci_alloc_interrupter() argument 2302 if (!segs) in xhci_alloc_interrupter() 2303 segs = ERST_DEFAULT_SEGS; in xhci_alloc_interrupter() 2306 segs = min(segs, max_segs); in xhci_alloc_interrupter() 2312 ir->event_ring = xhci_ring_alloc(xhci, segs, TYPE_EVENT, 0, flags); in xhci_alloc_interrupter() 2359 xhci_create_secondary_interrupter(struct usb_hcd *hcd, unsigned int segs, in xhci_create_secondary_interrupter() argument 2371 ir = xhci_alloc_interrupter(xhci, segs, GFP_KERNEL); in xhci_create_secondary_interrupter()
|
| /drivers/net/ |
| A D | tap.c | 281 struct sk_buff *segs = __skb_gso_segment(skb, features, false); in tap_handle_frame() local 284 if (IS_ERR(segs)) { in tap_handle_frame() 289 if (!segs) { in tap_handle_frame() 298 skb_list_walk_safe(segs, skb, next) { in tap_handle_frame()
|
| /drivers/media/dvb-frontends/ |
| A D | mb86a20s.c | 1428 const struct linear_segments *segs; in mb86a20s_get_blk_error_layer_CNR() local 1480 segs = cnr_qpsk_table; in mb86a20s_get_blk_error_layer_CNR() 1484 segs = cnr_16qam_table; in mb86a20s_get_blk_error_layer_CNR() 1489 segs = cnr_64qam_table; in mb86a20s_get_blk_error_layer_CNR() 1493 cnr = interpolate_value(mer, segs, segs_len); in mb86a20s_get_blk_error_layer_CNR()
|
| /drivers/staging/octeon/ |
| A D | ethernet-tx.c | 252 pko_command.s.segs = 1; in cvm_oct_xmit() 281 pko_command.s.segs = skb_shinfo(skb)->nr_frags + 1; in cvm_oct_xmit()
|
| /drivers/scsi/ |
| A D | vmw_pvscsi.c | 363 int segs = scsi_dma_map(cmd); in pvscsi_map_buffers() local 365 if (segs == -ENOMEM) { in pvscsi_map_buffers() 369 } else if (segs > 1) { in pvscsi_map_buffers() 370 pvscsi_create_sg(ctx, sg, segs); in pvscsi_map_buffers()
|
| /drivers/net/ethernet/qlogic/qed/ |
| A D | qed_cxt.c | 196 struct qed_tid_seg *segs = p_cfg->tid_seg; in qed_cxt_tm_iids() local 202 iids->pf_tids[j] += segs[j].count; in qed_cxt_tm_iids() 208 iids->per_vf_tids += segs[NUM_TASK_PF_SEGMENTS].count; in qed_cxt_tm_iids() 226 struct qed_tid_seg *segs; in qed_cxt_qm_iids() local 234 segs = p_mngr->conn_cfg[type].tid_seg; in qed_cxt_qm_iids() 239 iids->tids += segs[j].count; in qed_cxt_qm_iids() 245 vf_tids += segs[NUM_TASK_PF_SEGMENTS].count; in qed_cxt_qm_iids()
|
| /drivers/net/ethernet/intel/e1000/ |
| A D | e1000.h | 131 unsigned short segs; member
|
| A D | e1000_main.c | 2834 unsigned int f, bytecount, segs; in e1000_tx_map() local 2942 segs = skb_shinfo(skb)->gso_segs ?: 1; in e1000_tx_map() 2944 bytecount = ((segs - 1) * skb_headlen(skb)) + skb->len; in e1000_tx_map() 2947 tx_ring->buffer_info[i].segs = segs; in e1000_tx_map() 3856 total_tx_packets += buffer_info->segs; in e1000_clean_tx_irq()
|
| /drivers/net/ethernet/netronome/nfp/nfdk/ |
| A D | dp.c | 43 u32 segs, hdrlen, l3_offset, l4_offset, l4_hdrlen; in nfp_nfdk_tx_tso() local 60 segs = skb_shinfo(skb)->gso_segs; in nfp_nfdk_tx_tso() 68 txd.lso_totsegs = segs; in nfp_nfdk_tx_tso() 70 txbuf->pkt_cnt = segs; in nfp_nfdk_tx_tso()
|
| /drivers/net/wireless/realtek/rtw88/ |
| A D | main.c | 415 const struct rtw_fwcd_segs *segs = chip->fwcd_segs; in rtw_fwcd_prep() local 419 if (segs) { in rtw_fwcd_prep() 420 prep_size += segs->num * sizeof(struct rtw_fwcd_hdr); in rtw_fwcd_prep() 422 for (i = 0; i < segs->num; i++) in rtw_fwcd_prep() 423 prep_size += segs->segs[i]; in rtw_fwcd_prep()
|
| /drivers/net/ethernet/sun/ |
| A D | sunvnet_common.c | 1227 struct sk_buff *segs, *curr, *next; in vnet_handle_offloads() local 1278 segs = skb_gso_segment(skb, dev->features & ~NETIF_F_TSO); in vnet_handle_offloads() 1279 if (IS_ERR(segs)) in vnet_handle_offloads() 1286 skb_list_walk_safe(segs, curr, next) { in vnet_handle_offloads()
|
| /drivers/block/ |
| A D | xen-blkfront.c | 1211 int i, j, segs; in blkif_free_ring() local 1252 segs = rinfo->shadow[i].req.operation == BLKIF_OP_INDIRECT ? in blkif_free_ring() 1255 for (j = 0; j < segs; j++) { in blkif_free_ring() 1270 for (j = 0; j < INDIRECT_GREFS(segs); j++) { in blkif_free_ring()
|
| /drivers/net/ethernet/intel/e1000e/ |
| A D | e1000.h | 140 unsigned int segs; member
|
| /drivers/net/ethernet/intel/igbvf/ |
| A D | netdev.c | 808 unsigned int segs, bytecount; in igbvf_clean_tx_irq() local 811 segs = skb_shinfo(skb)->gso_segs ?: 1; in igbvf_clean_tx_irq() 813 bytecount = ((segs - 1) * skb_headlen(skb)) + in igbvf_clean_tx_irq() 815 total_packets += segs; in igbvf_clean_tx_irq()
|
| /drivers/nvme/target/ |
| A D | pci-epf.c | 704 struct nvmet_pci_epf_segment *segs; in nvmet_pci_epf_alloc_iod_data_segs() local 707 segs = krealloc(iod->data_segs, in nvmet_pci_epf_alloc_iod_data_segs() 710 if (!segs) in nvmet_pci_epf_alloc_iod_data_segs() 714 iod->data_segs = segs; in nvmet_pci_epf_alloc_iod_data_segs()
|