Home
last modified time | relevance | path

Searched refs:rx_buf_sz (Results 1 – 25 of 31) sorted by relevance

12

/drivers/net/ethernet/sis/
A Dsis190.c270 u32 rx_buf_sz; member
472 u32 rx_buf_sz) in sis190_map_to_asic() argument
475 sis190_give_to_asic(desc, rx_buf_sz); in sis190_map_to_asic()
490 u32 rx_buf_sz = tp->rx_buf_sz; in sis190_alloc_rx_skb() local
494 skb = netdev_alloc_skb(tp->dev, rx_buf_sz); in sis190_alloc_rx_skb()
597 sis190_give_to_asic(desc, tp->rx_buf_sz); in sis190_rx_interrupt()
616 tp->rx_buf_sz, in sis190_rx_interrupt()
621 tp->rx_buf_sz, in sis190_rx_interrupt()
1057 if (tp->rx_buf_sz & 0x07) { in sis190_set_rxbufsize()
1058 tp->rx_buf_sz += 8; in sis190_set_rxbufsize()
[all …]
/drivers/net/ethernet/packetengines/
A Dhamachi.c496 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member
1127 hmp->rx_buf_sz, DMA_FROM_DEVICE); in hamachi_tx_timeout()
1143 hmp->rx_buf_sz, in hamachi_tx_timeout()
1180 hmp->rx_buf_sz = (dev->mtu <= 1492 ? PKT_BUF_SZ : in hamachi_init_ring()
1197 hmp->rx_buf_sz, in hamachi_init_ring()
1431 hmp->rx_buf_sz, DMA_FROM_DEVICE); in hamachi_rx()
1500 hmp->rx_buf_sz, in hamachi_rx()
1513 hmp->rx_buf_sz, in hamachi_rx()
1518 hmp->rx_buf_sz, in hamachi_rx()
1604 hmp->rx_buf_sz, in hamachi_rx()
[all …]
A Dyellowfin.c324 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member
735 yp->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in yellowfin_init_ring()
739 cpu_to_le32(CMD_RX_BUF | INTR_ALWAYS | yp->rx_buf_sz); in yellowfin_init_ring()
745 struct sk_buff *skb = netdev_alloc_skb(dev, yp->rx_buf_sz + 2); in yellowfin_init_ring()
752 yp->rx_buf_sz, in yellowfin_init_ring()
1072 yp->rx_buf_sz, DMA_FROM_DEVICE); in yellowfin_rx()
1139 yp->rx_buf_sz, in yellowfin_rx()
1151 yp->rx_buf_sz, in yellowfin_rx()
1173 yp->rx_buf_sz, in yellowfin_rx()
1180 cpu_to_le32(CMD_RX_BUF | INTR_ALWAYS | yp->rx_buf_sz); in yellowfin_rx()
[all …]
/drivers/net/ethernet/dlink/
A Ddl2k.c512 skb = netdev_alloc_skb_ip_align(dev, np->rx_buf_sz); in alloc_list()
525 np->rx_buf_sz, DMA_FROM_DEVICE)); in alloc_list()
681 np->rx_buf_sz); in rio_timer()
692 np->rx_buf_sz, DMA_FROM_DEVICE)); in rio_timer()
695 cpu_to_le64((u64)np->rx_buf_sz << 48); in rio_timer()
971 np->rx_buf_sz, in receive_packet()
978 np->rx_buf_sz, in receive_packet()
986 np->rx_buf_sz, in receive_packet()
1009 skb = netdev_alloc_skb_ip_align(dev, np->rx_buf_sz); in receive_packet()
1021 np->rx_buf_sz, DMA_FROM_DEVICE)); in receive_packet()
[all …]
A Ddl2k.h379 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member
/drivers/net/ethernet/smsc/
A Depic100.c271 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member
907 ep->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in epic_init_ring()
912 ep->rx_ring[i].buflength = ep->rx_buf_sz; in epic_init_ring()
922 struct sk_buff *skb = netdev_alloc_skb(dev, ep->rx_buf_sz + 2); in epic_init_ring()
929 ep->rx_buf_sz, in epic_init_ring()
1193 ep->rx_buf_sz, in epic_rx()
1199 ep->rx_buf_sz, in epic_rx()
1204 ep->rx_buf_sz, in epic_rx()
1223 skb = ep->rx_skbuff[entry] = netdev_alloc_skb(dev, ep->rx_buf_sz + 2); in epic_rx()
1229 ep->rx_buf_sz, in epic_rx()
[all …]
/drivers/net/ethernet/
A Dfealnx.c400 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member
1057 skb = netdev_alloc_skb(dev, np->rx_buf_sz); in allocate_rx_buffers()
1067 np->rx_buf_sz, in allocate_rx_buffers()
1232 np->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in init_ring()
1240 np->rx_ring[i].control = np->rx_buf_sz << RBSShift; in init_ring()
1253 struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz); in init_ring()
1264 np->rx_buf_sz, in init_ring()
1702 np->rx_buf_sz, in netdev_rx()
1716 np->rx_buf_sz, in netdev_rx()
1721 np->rx_buf_sz, in netdev_rx()
[all …]
/drivers/net/ethernet/realtek/
A D8139cp.c347 unsigned rx_buf_sz; member
413 cp->rx_buf_sz = mtu + ETH_HLEN + 8; in cp_set_rxbufsize()
415 cp->rx_buf_sz = PKT_BUF_SZ; in cp_set_rxbufsize()
477 const unsigned buflen = cp->rx_buf_sz; in cp_rx_poll()
546 cp->rx_buf_sz); in cp_rx_poll()
548 desc->opts1 = cpu_to_le32(DescOwn | cp->rx_buf_sz); in cp_rx_poll()
1067 skb = netdev_alloc_skb_ip_align(dev, cp->rx_buf_sz); in cp_refill_rx()
1072 cp->rx_buf_sz, DMA_FROM_DEVICE); in cp_refill_rx()
1083 cpu_to_le32(DescOwn | RingEnd | cp->rx_buf_sz); in cp_refill_rx()
1086 cpu_to_le32(DescOwn | cp->rx_buf_sz); in cp_refill_rx()
[all …]
/drivers/net/ethernet/dec/tulip/
A Dde2104x.c304 unsigned rx_buf_sz; member
436 buflen = copying_skb ? (len + RX_OFFSET) : de->rx_buf_sz; in de_rx()
479 cpu_to_le32(RingEnd | de->rx_buf_sz); in de_rx()
481 de->rx_ring[rx_tail].opts2 = cpu_to_le32(de->rx_buf_sz); in de_rx()
1283 skb = netdev_alloc_skb(de->dev, de->rx_buf_sz); in de_refill_rx()
1289 de->rx_buf_sz, in de_refill_rx()
1296 cpu_to_le32(RingEnd | de->rx_buf_sz); in de_refill_rx()
1298 de->rx_ring[i].opts2 = cpu_to_le32(de->rx_buf_sz); in de_refill_rx()
1345 de->rx_skb[i].mapping, de->rx_buf_sz, in de_clean_rings()
1390 de->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in de_open()
A Dwinbond-840.c303 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member
791 np->rx_ring[i].length = np->rx_buf_sz; in init_rxtx_rings()
800 struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz); in init_rxtx_rings()
805 np->rx_buf_sz, in init_rxtx_rings()
962 np->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in alloc_ringdesc()
1256 skb = netdev_alloc_skb(dev, np->rx_buf_sz); in netdev_rx()
1262 np->rx_buf_sz, in netdev_rx()
/drivers/net/ethernet/realtek/rtase/
A Drtase_main.c245 static void rtase_mark_to_asic(union rtase_rx_desc *desc, u32 rx_buf_sz) in rtase_mark_to_asic() argument
253 cpu_to_le32(RTASE_DESC_OWN | eor | rx_buf_sz)); in rtase_mark_to_asic()
355 u32 rx_buf_sz) in rtase_map_to_asic() argument
359 rtase_mark_to_asic(desc, rx_buf_sz); in rtase_map_to_asic()
387 rtase_map_to_asic(desc, mapping, tp->rx_buf_sz); in rtase_alloc_rx_data_buf()
527 rtase_mark_to_asic(desc, tp->rx_buf_sz); in rx_handler()
542 rtase_mark_to_asic(desc, tp->rx_buf_sz); in rx_handler()
548 tp->rx_buf_sz, DMA_FROM_DEVICE); in rx_handler()
553 rtase_mark_to_asic(desc, tp->rx_buf_sz); in rx_handler()
983 rtase_w16(tp, RTASE_RMS, tp->rx_buf_sz); in rtase_hw_config()
[all …]
A Drtase.h327 u32 rx_buf_sz; member
/drivers/iio/adc/
A Dstm32-adc.c281 unsigned int rx_buf_sz; member
1734 unsigned int rx_buf_sz = STM32_DMA_BUFFER_SIZE; in stm32_adc_set_watermark() local
1743 adc->rx_buf_sz = min(rx_buf_sz, watermark * 2 * adc->num_conv); in stm32_adc_set_watermark()
1838 unsigned int i = adc->rx_buf_sz - state.residue; in stm32_adc_dma_residue()
1845 size = adc->rx_buf_sz + i - adc->bufi; in stm32_adc_dma_residue()
1877 if (adc->bufi >= adc->rx_buf_sz) in stm32_adc_dma_buffer_done()
1893 adc->rx_buf_sz, adc->rx_buf_sz / 2); in stm32_adc_dma_start()
1898 adc->rx_buf_sz, adc->rx_buf_sz / 2, in stm32_adc_dma_start()
A Dat91-sama5d2_adc.c556 int rx_buf_sz; member
1187 i = st->dma_st.rx_buf_sz - state.residue; in at91_adc_dma_size_done()
1193 size = st->dma_st.rx_buf_sz + i - st->dma_st.buf_idx; in at91_adc_dma_size_done()
1222 st->dma_st.rx_buf_sz = 0; in at91_adc_dma_start()
1232 st->dma_st.rx_buf_sz += chan->scan_type.storagebits / 8; in at91_adc_dma_start()
1234 st->dma_st.rx_buf_sz *= st->dma_st.watermark; in at91_adc_dma_start()
1239 st->dma_st.rx_buf_sz, in at91_adc_dma_start()
1240 st->dma_st.rx_buf_sz / 2, in at91_adc_dma_start()
1500 sample_size = div_s64(st->dma_st.rx_buf_sz, st->dma_st.watermark); in at91_adc_trigger_handler_dma()
1527 if (st->dma_st.buf_idx >= st->dma_st.rx_buf_sz) in at91_adc_trigger_handler_dma()
/drivers/net/ethernet/adaptec/
A Dstarfire.c545 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member
905 writel((np->rx_buf_sz << RxBufferLenShift) | in netdev_open()
1127 np->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in init_ring()
1131 struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz); in init_ring()
1137 np->rx_buf_sz, in init_ring()
1472 np->rx_buf_sz, DMA_FROM_DEVICE); in __netdev_rx()
1585 skb = netdev_alloc_skb(dev, np->rx_buf_sz); in refill_rx_ring()
1591 np->rx_buf_sz, DMA_FROM_DEVICE); in refill_rx_ring()
1967 np->rx_buf_sz, DMA_FROM_DEVICE); in netdev_close()
/drivers/net/ethernet/sun/
A Dsungem.h939 #define RX_BUF_ALLOC_SIZE(gp) ((gp)->rx_buf_sz + 28 + RX_OFFSET + 64)
994 int rx_buf_sz; member
A Dsungem.c816 skb_put(new_skb, (gp->rx_buf_sz + RX_OFFSET)); in gem_rx()
1616 gp->rx_buf_sz = max(dev->mtu + ETH_HLEN + VLAN_HLEN, in gem_init_rings()
1631 skb_put(skb, (gp->rx_buf_sz + RX_OFFSET)); in gem_init_rings()
1813 writel(0x20000000 | (gp->rx_buf_sz + 4), gp->regs + MAC_MAXFSZ); in gem_init_mac()
1894 int max_frame = (gp->rx_buf_sz + 4 + 64) & ~63; in gem_init_pause_thresholds()
/drivers/net/ethernet/via/
A Dvia-rhine.c470 unsigned int rx_buf_sz; /* Based on MTU+slack. */ member
1205 const int size = rp->rx_buf_sz; in rhine_skb_dma_init()
1249 rp->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in alloc_rbufs()
1255 rp->rx_ring[i].desc_length = cpu_to_le32(rp->rx_buf_sz); in alloc_rbufs()
1294 rp->rx_buf_sz, DMA_FROM_DEVICE); in free_rbufs()
2098 rp->rx_buf_sz, in rhine_rx()
2107 rp->rx_buf_sz, in rhine_rx()
2119 rp->rx_buf_sz, in rhine_rx()
/drivers/net/wireless/ath/ath11k/
A Ddebugfs.c956 int i, ret, rx_buf_sz = 0; in ath11k_write_pktlog_filter() local
999 rx_buf_sz, &tlv_filter); in ath11k_write_pktlog_filter()
1020 rx_buf_sz = DP_RX_BUFFER_SIZE; in ath11k_write_pktlog_filter()
1030 rx_buf_sz = DP_RX_BUFFER_SIZE_LITE; in ath11k_write_pktlog_filter()
1032 rx_buf_sz = DP_RX_BUFFER_SIZE; in ath11k_write_pktlog_filter()
1059 rx_buf_sz, &tlv_filter); in ath11k_write_pktlog_filter()
/drivers/net/ethernet/natsemi/
A Dnatsemi.c561 unsigned int rx_buf_sz; member
1745 if (np->rx_buf_sz > NATSEMI_LONGPKT) in init_registers()
1934 unsigned int buflen = np->rx_buf_sz+NATSEMI_PADDING; in refill_rx()
1949 np->rx_ring[entry].cmd_status = cpu_to_le32(np->rx_buf_sz); in refill_rx()
1962 np->rx_buf_sz = ETH_DATA_LEN + NATSEMI_HEADERS; in set_bufsize()
1964 np->rx_buf_sz = dev->mtu + NATSEMI_HEADERS; in set_bufsize()
2025 unsigned int buflen = np->rx_buf_sz; in drain_rx()
2283 unsigned int buflen = np->rx_buf_sz; in netdev_rx()
2343 } else if (pkt_len > np->rx_buf_sz) { in netdev_rx()
/drivers/atm/
A Diphase.c768 iadev->rx_buf_sz = IA_RX_BUF_SZ;
779 iadev->rx_buf_sz = IA_RX_BUF_SZ;
791 iadev->rx_buf_sz = IA_RX_BUF_SZ;
796 iadev->rx_buf_sz, iadev->rx_pkt_ram);)
1173 if (len > iadev->rx_buf_sz) {
1174 printk("Over %d bytes sdu received, dropped!!!\n", iadev->rx_buf_sz);
1334 if ((length > iadev->rx_buf_sz) || (length >
1481 writew(iadev->rx_buf_sz, iadev->reass_reg+BUF_SIZE);
1495 rx_pkt_start += iadev->rx_buf_sz;
3145 iadev->num_rx_desc, iadev->rx_buf_sz,
/drivers/net/ethernet/renesas/
A Dsh_eth.c1285 ALIGN(mdp->rx_buf_sz, 32), in sh_eth_ring_free()
1327 int skbuff_size = mdp->rx_buf_sz + SH_ETH_RX_ALIGN + 32 - 1; in sh_eth_ring_format()
1348 buf_len = ALIGN(mdp->rx_buf_sz, 32); in sh_eth_ring_format()
1407 mdp->rx_buf_sz = (ndev->mtu <= 1492 ? PKT_BUF_SZ : in sh_eth_ring_init()
1410 mdp->rx_buf_sz += NET_IP_ALIGN; in sh_eth_ring_init()
1594 int skbuff_size = mdp->rx_buf_sz + SH_ETH_RX_ALIGN + 32 - 1; in sh_eth_rx()
1653 ALIGN(mdp->rx_buf_sz, 32), in sh_eth_rx()
1674 buf_len = ALIGN(mdp->rx_buf_sz, 32); in sh_eth_rx()
A Dsh_eth.h547 u32 rx_buf_sz; /* Based on MTU+slack. */ member
/drivers/net/ethernet/nvidia/
A Dforcedeth.c805 unsigned int rx_buf_sz; member
1833 np->put_rx.orig->flaglen = cpu_to_le32(np->rx_buf_sz | NV_RX_AVAIL); in nv_alloc_rx()
1875 np->put_rx.ex->flaglen = cpu_to_le32(np->rx_buf_sz | NV_RX2_AVAIL); in nv_alloc_rx_optimized()
3072 np->rx_buf_sz = ETH_DATA_LEN + NV_RX_HEADERS; in set_bufsize()
3074 np->rx_buf_sz = dev->mtu + NV_RX_HEADERS; in set_bufsize()
3119 writel(np->rx_buf_sz, base + NvRegOffloadConfig); in nv_change_mtu()
4201 writel(np->rx_buf_sz, base + NvRegOffloadConfig); in nv_do_nic_poll()
4760 writel(np->rx_buf_sz, base + NvRegOffloadConfig); in nv_set_ringparam()
5154 writel(np->rx_buf_sz, base + NvRegOffloadConfig); in nv_loopback_test()
5313 writel(np->rx_buf_sz, base + NvRegOffloadConfig); in nv_self_test()
[all …]
/drivers/net/ethernet/huawei/hinic/
A Dhinic_hw_dev.h264 u16 rx_buf_sz; member

Completed in 140 milliseconds

12