| /drivers/net/slip/ |
| A D | slhc.c | 268 if (th->doff < sizeof(struct tcphdr) / 4) in slhc_compress() 270 hlen = nlen + th->doff * 4; in slhc_compress() 357 || th->doff != cs->cs_tcp.doff in slhc_compress() 359 || (th->doff > 5 && memcmp(th+1,cs->cs_tcpopt,((th->doff)-5)*4) != 0)){ in slhc_compress() 480 if (th->doff > 5) in slhc_compress() 547 hdrlen = ip->ihl * 4 + thp->doff * 4; in slhc_uncompress() 631 if (thp->doff > 5) { in slhc_uncompress() 633 cp += ((thp->doff) - 5) * 4; in slhc_uncompress() 685 if (isize < ihl * 4 + th->doff * 4) in slhc_remember() 694 if (th->doff > 5) in slhc_remember() [all …]
|
| /drivers/dma/ |
| A D | fsl-edma-common.c | 449 edma_cp_tcd_to_reg(fsl_chan, tcd, doff); in fsl_edma_set_tcd_regs() 532 fsl_edma_set_tcd_to_le(fsl_chan, tcd, doff, doff); in fsl_edma_fill_tcd() 597 u16 soff, doff, iter; in fsl_edma_prep_dma_cyclic() local 639 doff = fsl_chan->is_multi_fifo ? 4 : 0; in fsl_edma_prep_dma_cyclic() 641 doff = fsl_chan->cfg.dst_addr_width; in fsl_edma_prep_dma_cyclic() 646 doff = fsl_chan->cfg.src_addr_width; in fsl_edma_prep_dma_cyclic() 653 soff = doff = 0; in fsl_edma_prep_dma_cyclic() 675 u16 soff, doff, iter; in fsl_edma_prep_slave_sg() local 708 doff = 0; in fsl_edma_prep_slave_sg() 713 doff = fsl_chan->cfg.src_addr_width; in fsl_edma_prep_slave_sg() [all …]
|
| A D | fsl-edma-trace.h | 72 __field(u16, doff) 86 __entry->doff = fsl_edma_get_tcd_to_cpu(chan, tcd, doff), 110 __entry->doff,
|
| A D | mpc512x_dma.c | 153 u32 doff:16; /* Signed destination address offset */ member 636 tcd->doff = 32; in mpc_dma_prep_memcpy() 642 tcd->doff = 16; in mpc_dma_prep_memcpy() 647 tcd->doff = 4; in mpc_dma_prep_memcpy() 652 tcd->doff = 2; in mpc_dma_prep_memcpy() 657 tcd->doff = 1; in mpc_dma_prep_memcpy() 749 tcd->doff = mchan->dwidth; in mpc_dma_prep_slave_sg() 758 tcd->doff = 0; in mpc_dma_prep_slave_sg()
|
| A D | fsl-edma-common.h | 98 __le16 doff; member 113 __le16 doff; member
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/clk/ |
| A D | gf100.c | 98 read_div(struct gf100_clk *clk, int doff, u32 dsrc, u32 dctl) in read_div() argument 101 u32 ssrc = nvkm_rd32(device, dsrc + (doff * 4)); in read_div() 112 sclk = read_vco(clk, dsrc + (doff * 4)); in read_div() 115 if (doff <= 2) { in read_div() 116 sctl = nvkm_rd32(device, dctl + (doff * 4)); in read_div()
|
| A D | gk104.c | 106 read_div(struct gk104_clk *clk, int doff, u32 dsrc, u32 dctl) in read_div() argument 109 u32 ssrc = nvkm_rd32(device, dsrc + (doff * 4)); in read_div() 110 u32 sctl = nvkm_rd32(device, dctl + (doff * 4)); in read_div() 121 u32 sclk = read_vco(clk, dsrc + (doff * 4)); in read_div() 126 return read_vco(clk, dsrc + (doff * 4)); in read_div()
|
| /drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
| A D | nv50.c | 278 const u32 doff = nv50_ior_base(dac); in nv50_dac_clock() local 280 nvkm_mask(device, 0x614280 + doff, 0x07070707, 0x00000000); in nv50_dac_clock() 287 const u32 doff = nv50_ior_base(dac); in nv50_dac_sense() local 291 nvkm_wr32(device, 0x61a00c + doff, 0x00100000 | loadval); in nv50_dac_sense() 294 loadval = nvkm_mask(device, 0x61a00c + doff, 0xffffffff, 0x00000000); in nv50_dac_sense() 304 nv50_dac_power_wait(struct nvkm_device *device, const u32 doff) in nv50_dac_power_wait() argument 307 if (!(nvkm_rd32(device, 0x61a004 + doff) & 0x80000000)) in nv50_dac_power_wait() 316 const u32 doff = nv50_ior_base(dac); in nv50_dac_power() local 324 nv50_dac_power_wait(device, doff); in nv50_dac_power() 325 nvkm_mask(device, 0x61a004 + doff, field, state); in nv50_dac_power() [all …]
|
| A D | gf119.c | 355 const u32 doff = nv50_ior_base(dac); in gf119_dac_clock() local 356 nvkm_mask(device, 0x612280 + doff, 0x07070707, 0x00000000); in gf119_dac_clock()
|
| /drivers/net/ethernet/sfc/ |
| A D | tx_tso.c | 162 (tcp_hdr(skb)->doff << 2u)) > in efx_tso_check_protocol() 179 header_len = st->tcp_off + (tcp_hdr(skb)->doff << 2u); in tso_start()
|
| A D | tx_common.c | 361 (inner_tcp_hdr(skb)->doff << 2u); in efx_tx_tso_header_length() 364 (tcp_hdr(skb)->doff << 2u); in efx_tx_tso_header_length()
|
| /drivers/net/ethernet/sfc/siena/ |
| A D | tx_common.c | 321 (inner_tcp_hdr(skb)->doff << 2u); in efx_tx_tso_header_length() 324 (tcp_hdr(skb)->doff << 2u); in efx_tx_tso_header_length()
|
| /drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
| A D | sdio.c | 1522 u8 doff; in brcmf_sdio_rxglom() local 1699 skb_pull(pfirst, doff); in brcmf_sdio_rxglom() 1833 bus->rxctl = buf + doff; in brcmf_sdio_read_control() 1835 bus->rxlen = len - doff; in brcmf_sdio_read_control() 2398 u8 doff; in brcmf_sdio_tx_ctrlframe() local 2411 doff = ((unsigned long)frame % bus->head_align); in brcmf_sdio_tx_ctrlframe() 2412 if (doff) { in brcmf_sdio_tx_ctrlframe() 2413 frame -= doff; in brcmf_sdio_tx_ctrlframe() 2414 len += doff; in brcmf_sdio_tx_ctrlframe() 2415 memset(frame + bus->tx_hdrlen, 0, doff); in brcmf_sdio_tx_ctrlframe() [all …]
|
| /drivers/md/ |
| A D | md-bitmap.c | 447 sector_t doff; in __write_sb_page() local 461 doff = rdev->data_offset; in __write_sb_page() 466 if (sboff + ps > doff && in __write_sb_page() 467 sboff < (doff + mddev->dev_sectors + PAGE_SIZE / SECTOR_SIZE)) in __write_sb_page() 476 if (doff + mddev->dev_sectors > sboff) in __write_sb_page() 481 size = bitmap_io_size(size, opt_size, sboff + ps, doff); in __write_sb_page()
|
| /drivers/crypto/ |
| A D | hifn_795x.c | 1310 unsigned int soff, doff; in hifn_setup_dma() local 1331 doff = 0; in hifn_setup_dma() 1336 doff = dst->offset; in hifn_setup_dma() 1341 hifn_setup_dst_desc(dev, dpage, doff, len, n - len == 0); in hifn_setup_dma()
|
| /drivers/infiniband/hw/irdma/ |
| A D | cm.c | 376 tcph->doff = (u16)((sizeof(*tcph) + opts_len + 3) >> 2); in irdma_form_ah_cm_frame() 377 sqbuf->tcphlen = tcph->doff << 2; in irdma_form_ah_cm_frame() 564 tcph->doff = (u16)((sizeof(*tcph) + opts_len + 3) >> 2); in irdma_form_uda_cm_frame() 565 sqbuf->tcphlen = tcph->doff << 2; in irdma_form_uda_cm_frame() 2684 optionsize = (tcph->doff << 2) - sizeof(struct tcphdr); in irdma_handle_syn_pkt() 2750 optionsize = (tcph->doff << 2) - sizeof(struct tcphdr); in irdma_handle_synack_pkt() 2824 optionsize = (tcph->doff << 2) - sizeof(struct tcphdr); in irdma_handle_ack_pkt()
|
| A D | utils.c | 1480 buf->tcphlen = tcph->doff << 2; in irdma_gen1_puda_get_tcpip_info() 1519 buf->tcphlen = tcph->doff << 2; in irdma_puda_get_tcpip_info()
|
| /drivers/net/ethernet/qlogic/qlcnic/ |
| A D | qlcnic_io.c | 1324 length = (th->doff << 2) + lro_length; in qlcnic_process_lro() 1329 length = (iph->ihl << 2) + (th->doff << 2) + lro_length; in qlcnic_process_lro() 1846 length = (th->doff << 2) + lro_length; in qlcnic_83xx_process_lro() 1851 length = (iph->ihl << 2) + (th->doff << 2) + lro_length; in qlcnic_83xx_process_lro()
|
| /drivers/net/ethernet/huawei/hinic3/ |
| A D | hinic3_tx.c | 176 *offset = (l4->tcp->doff << TCP_HDR_DATA_OFF_UNIT_SHIFT) + in get_inner_l4_info()
|
| /drivers/net/ethernet/intel/iavf/ |
| A D | iavf_txrx.c | 1815 *hdr_len = (u8)((l4.tcp->doff * 4) + l4_offset); in iavf_tso() 1975 offset |= l4.tcp->doff << IAVF_TX_DESC_LENGTH_L4_FC_LEN_SHIFT; in iavf_tx_enable_csum()
|
| /drivers/net/ethernet/intel/idpf/ |
| A D | idpf_singleq_txrx.c | 154 l4_len = l4.tcp->doff; in idpf_tx_singleq_csum()
|
| /drivers/net/ethernet/intel/ice/ |
| A D | ice_txrx.c | 2048 l4_len = l4.tcp->doff; in ice_tx_csum() 2208 off->header_len = (u8)((l4.tcp->doff * 4) + l4_start); in ice_tso()
|
| /drivers/net/ethernet/huawei/hinic/ |
| A D | hinic_tx.c | 241 *l4_len = l4->tcp->doff * 4; in get_inner_l4_info()
|
| /drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | en_rx.c | 1194 check = csum_partial(tcp, tcp->doff * 4, in mlx5e_lro_update_hdr() 1210 check = csum_partial(tcp, tcp->doff * 4, in mlx5e_lro_update_hdr() 1217 return (unsigned int)((unsigned char *)tcp + tcp->doff * 4 - skb->data); in mlx5e_lro_update_hdr()
|
| /drivers/net/ethernet/intel/fm10k/ |
| A D | fm10k_main.c | 766 hdrlen = (th - skb->data) + (((struct tcphdr *)th)->doff << 2); in fm10k_tso()
|