| /drivers/net/ethernet/microchip/sparx5/ |
| A D | sparx5_sdlb.c | 116 u32 itr, next; in sparx5_sdlb_group_get_last() local 123 return itr; in sparx5_sdlb_group_get_last() 125 itr = next; in sparx5_sdlb_group_get_last() 142 u32 itr; in sparx5_sdlb_group_get_adjacent() local 147 itr = *first; in sparx5_sdlb_group_get_adjacent() 158 *prev = itr; in sparx5_sdlb_group_get_adjacent() 159 itr = *next; in sparx5_sdlb_group_get_adjacent() 165 u32 itr, next; in sparx5_sdlb_group_get_count() local 175 itr = next; in sparx5_sdlb_group_get_count() 210 u32 itr, next; in sparx5_sdlb_group_get_by_index() local [all …]
|
| A D | sparx5_qos.c | 119 u32 itr, next; in sparx5_lg_get_last() local 125 if (itr == next) in sparx5_lg_get_last() 126 return itr; in sparx5_lg_get_last() 128 itr = next; in sparx5_lg_get_last() 172 u32 itr, next; in sparx5_lg_get_group_by_index() local 191 itr = next; in sparx5_lg_get_group_by_index() 218 u32 itr; in sparx5_lg_get_adjacent() local 223 itr = *first; in sparx5_lg_get_adjacent() 228 if (itr == idx) in sparx5_lg_get_adjacent() 234 *prev = itr; in sparx5_lg_get_adjacent() [all …]
|
| /drivers/crypto/intel/qat/qat_common/ |
| A D | adf_dev_mgr.c | 36 struct list_head *itr; in adf_find_vf() local 38 list_for_each(itr, &vfs_table) { in adf_find_vf() 50 struct list_head *itr; in adf_get_vf_real_id() local 52 list_for_each(itr, &vfs_table) { in adf_get_vf_real_id() 100 struct list_head *itr; in adf_devmgr_update_class_index() local 103 list_for_each(itr, &accel_table) { in adf_devmgr_update_class_index() 142 struct list_head *itr; in adf_devmgr_add_dev() local 290 struct list_head *itr; in adf_devmgr_pci_to_accel_dev() local 309 struct list_head *itr; in adf_devmgr_get_dev_by_id() local 345 struct list_head *itr; in adf_get_num_dettached_vfs() local [all …]
|
| A D | qat_compression.c | 54 struct list_head *itr; in qat_compression_get_instance_node() local 56 list_for_each(itr, adf_devmgr_get_head()) { in qat_compression_get_instance_node() 61 tmp_dev = list_entry(itr, struct adf_accel_dev, list); in qat_compression_get_instance_node() 77 list_for_each(itr, adf_devmgr_get_head()) { in qat_compression_get_instance_node() 80 tmp_dev = list_entry(itr, struct adf_accel_dev, list); in qat_compression_get_instance_node() 93 list_for_each(itr, &accel_dev->compression_list) { in qat_compression_get_instance_node() 97 tmp_inst = list_entry(itr, struct qat_compression_instance, list); in qat_compression_get_instance_node()
|
| /drivers/platform/x86/intel/int1092/ |
| A D | intel_sar.c | 45 int itr = 0; in update_sar_data() local 47 for (itr = 0; itr < config->total_dev_mode; itr++) { in update_sar_data() 49 config->device_mode_info[itr].device_mode) { in update_sar_data() 51 &config->device_mode_info[itr]; in update_sar_data() 80 int value, itr, reg; in parse_package() local 99 for (itr = 0; itr < data->total_dev_mode; itr++) { in parse_package() 102 num = &item->package.elements[itr + 1]; in parse_package() 113 data->device_mode_info[itr] = temp; in parse_package()
|
| /drivers/net/ethernet/intel/e1000e/ |
| A D | param.c | 335 adapter->itr = InterruptThrottleRate[bd]; in e1000e_check_options() 342 if ((adapter->itr > 4) && in e1000e_check_options() 343 e1000_validate_option(&adapter->itr, &opt, adapter)) in e1000e_check_options() 344 adapter->itr = opt.def; in e1000e_check_options() 349 adapter->itr = opt.def; in e1000e_check_options() 354 if (adapter->itr > 4) in e1000e_check_options() 357 adapter->itr); in e1000e_check_options() 360 adapter->itr_setting = adapter->itr; in e1000e_check_options() 361 switch (adapter->itr) { in e1000e_check_options() 369 adapter->itr = 20000; in e1000e_check_options() [all …]
|
| /drivers/net/ethernet/intel/e1000/ |
| A D | e1000_param.c | 436 adapter->itr = InterruptThrottleRate[bd]; in e1000_check_options() 437 switch (adapter->itr) { in e1000_check_options() 444 adapter->itr_setting = adapter->itr; in e1000_check_options() 445 adapter->itr = 20000; in e1000_check_options() 450 adapter->itr_setting = adapter->itr; in e1000_check_options() 451 adapter->itr = 20000; in e1000_check_options() 456 adapter->itr_setting = adapter->itr; in e1000_check_options() 459 e1000_validate_option(&adapter->itr, &opt, in e1000_check_options() 466 adapter->itr_setting = adapter->itr & ~3; in e1000_check_options() 471 adapter->itr = 20000; in e1000_check_options()
|
| /drivers/net/ethernet/microchip/vcap/ |
| A D | vcap_api_private.h | 58 void vcap_iter_init(struct vcap_stream_iter *itr, int sw_width, 60 void vcap_iter_next(struct vcap_stream_iter *itr); 61 void vcap_iter_set(struct vcap_stream_iter *itr, int sw_width, 63 void vcap_iter_update(struct vcap_stream_iter *itr);
|
| A D | vcap_api.c | 52 memset(itr, 0, sizeof(*itr)); in vcap_iter_set() 56 itr->tg = tg; in vcap_iter_set() 64 while (itr->tg->width && itr->offset >= itr->tg->offset) { in vcap_iter_skip_tg() 65 itr->offset += itr->tg->width; in vcap_iter_skip_tg() 75 sw_idx = itr->offset / itr->sw_width; in vcap_iter_update() 76 sw_bitpos = itr->offset % itr->sw_width; in vcap_iter_update() 78 itr->reg_idx = (sw_idx * itr->regs_per_sw) + (sw_bitpos / 32); in vcap_iter_update() 92 itr->offset++; in vcap_iter_next() 114 itr->offset >= itr->tg->offset && in vcap_encode_bit() 115 itr->offset < itr->tg->offset + itr->tg->width) { in vcap_encode_bit() [all …]
|
| /drivers/net/ethernet/intel/iavf/ |
| A D | iavf_txrx.c | 504 itr = rc->target_itr; in iavf_update_itr() 519 itr = IAVF_ITR_ADAPTIVE_LATENCY; in iavf_update_itr() 550 itr &= IAVF_ITR_ADAPTIVE_LATENCY; in iavf_update_itr() 558 itr &= IAVF_ITR_MASK; in iavf_update_itr() 572 itr /= 2; in iavf_update_itr() 573 itr &= IAVF_ITR_MASK; in iavf_update_itr() 586 itr = IAVF_ITR_ADAPTIVE_BULK; in iavf_update_itr() 644 itr += DIV_ROUND_UP(avg_wire_size, in iavf_update_itr() 649 itr &= IAVF_ITR_ADAPTIVE_LATENCY; in iavf_update_itr() 655 rc->target_itr = itr; in iavf_update_itr() [all …]
|
| /drivers/bus/mhi/host/ |
| A D | pm.c | 808 list_del(&itr->node); in mhi_pm_st_worker() 812 switch (itr->state) { in mhi_pm_st_worker() 857 kfree(itr); in mhi_pm_st_worker() 863 struct mhi_chan *itr, *tmp; in mhi_pm_suspend() local 939 mutex_lock(&itr->mutex); in mhi_pm_suspend() 940 if (itr->mhi_dev) in mhi_pm_suspend() 942 mutex_unlock(&itr->mutex); in mhi_pm_suspend() 951 struct mhi_chan *itr, *tmp; in __mhi_pm_resume() local 975 mutex_lock(&itr->mutex); in __mhi_pm_resume() 976 if (itr->mhi_dev) in __mhi_pm_resume() [all …]
|
| /drivers/i2c/busses/ |
| A D | i2c-qcom-geni.c | 166 const struct geni_i2c_clk_fld *itr; in geni_i2c_clk_map_idx() local 169 itr = geni_i2c_clk_map_32mhz; in geni_i2c_clk_map_idx() 171 itr = geni_i2c_clk_map_19p2mhz; in geni_i2c_clk_map_idx() 173 while (itr->clk_freq_out != 0) { in geni_i2c_clk_map_idx() 175 gi2c->clk_fld = itr; in geni_i2c_clk_map_idx() 178 itr++; in geni_i2c_clk_map_idx() 194 val |= itr->t_low_cnt << LOW_COUNTER_SHFT; in qcom_geni_i2c_conf() 195 val |= itr->t_cycle_cnt; in qcom_geni_i2c_conf() 614 peripheral.high_count = itr->t_high_cnt; in geni_i2c_gpi_xfer() 615 peripheral.low_count = itr->t_low_cnt; in geni_i2c_gpi_xfer() [all …]
|
| /drivers/net/ethernet/qualcomm/emac/ |
| A D | emac-sgmii-qdf2432.c | 107 const struct emac_reg_write *itr, size_t size) in emac_reg_write_all() argument 111 for (i = 0; i < size; ++itr, ++i) in emac_reg_write_all() 112 writel(itr->val, base + itr->offset); in emac_reg_write_all()
|
| A D | emac-sgmii-qdf2400.c | 117 const struct emac_reg_write *itr, size_t size) in emac_reg_write_all() argument 121 for (i = 0; i < size; ++itr, ++i) in emac_reg_write_all() 122 writel(itr->val, base + itr->offset); in emac_reg_write_all()
|
| A D | emac-sgmii-fsm9900.c | 140 const struct emac_reg_write *itr, size_t size) in emac_reg_write_all() argument 144 for (i = 0; i < size; ++itr, ++i) in emac_reg_write_all() 145 writel(itr->val, base + itr->offset); in emac_reg_write_all()
|
| /drivers/crypto/virtio/ |
| A D | virtio_crypto_mgr.c | 35 struct list_head *itr; in virtcrypto_devmgr_add_dev() local 45 list_for_each(itr, &virtio_crypto_table) { in virtcrypto_devmgr_add_dev() 47 list_entry(itr, struct virtio_crypto, list); in virtcrypto_devmgr_add_dev()
|
| /drivers/gpu/drm/i915/gvt/ |
| A D | kvmgt.c | 213 struct gvt_dma *itr; in __gvt_cache_find_dma_addr() local 218 if (dma_addr < itr->dma_addr) in __gvt_cache_find_dma_addr() 223 return itr; in __gvt_cache_find_dma_addr() 231 struct gvt_dma *itr; in __gvt_cache_find_gfn() local 236 if (gfn < itr->gfn) in __gvt_cache_find_gfn() 238 else if (gfn > itr->gfn) in __gvt_cache_find_gfn() 241 return itr; in __gvt_cache_find_gfn() 249 struct gvt_dma *new, *itr; in __gvt_cache_add() local 268 if (gfn < itr->gfn) in __gvt_cache_add() 283 if (dma_addr < itr->dma_addr) in __gvt_cache_add() [all …]
|
| /drivers/bus/mhi/ep/ |
| A D | main.c | 818 struct mhi_ep_ring_item *itr, *tmp; in mhi_ep_ch_ring_worker() local 830 list_for_each_entry_safe(itr, tmp, &head, node) { in mhi_ep_ch_ring_worker() 831 list_del(&itr->node); in mhi_ep_ch_ring_worker() 832 ring = itr->ring; in mhi_ep_ch_ring_worker() 843 kfree(itr); in mhi_ep_ch_ring_worker() 882 struct mhi_ep_state_transition *itr, *tmp; in mhi_ep_state_worker() local 891 list_for_each_entry_safe(itr, tmp, &head, node) { in mhi_ep_state_worker() 892 list_del(&itr->node); in mhi_ep_state_worker() 894 mhi_state_str(itr->state)); in mhi_ep_state_worker() 896 switch (itr->state) { in mhi_ep_state_worker() [all …]
|
| /drivers/net/ethernet/intel/fm10k/ |
| A D | fm10k_main.c | 1343 if (!ITR_IS_ADAPTIVE(ring_container->itr)) in fm10k_update_itr() 1392 ring_container->itr = avg_wire_size | FM10K_ITR_ADAPTIVE; in fm10k_update_itr() 1402 u32 itr = FM10K_ITR_ENABLE; in fm10k_qv_enable() local 1411 itr |= (q_vector->tx.itr & FM10K_ITR_MAX); in fm10k_qv_enable() 1414 itr |= (q_vector->rx.itr & FM10K_ITR_MAX) << FM10K_ITR_INTERVAL1_SHIFT; in fm10k_qv_enable() 1417 writel(itr, q_vector->itr); in fm10k_qv_enable() 1618 q_vector->tx.itr = interface->tx_itr; in fm10k_alloc_q_vector() 1647 q_vector->rx.itr = interface->rx_itr; in fm10k_alloc_q_vector()
|
| A D | fm10k.h | 150 u16 itr; /* interrupt throttle rate value */ member 161 #define ITR_IS_ADAPTIVE(itr) (!!(itr & FM10K_ITR_ADAPTIVE)) argument 187 u32 __iomem *itr; /* pointer to ITR register for this vector */ member
|
| /drivers/pci/endpoint/functions/ |
| A D | pci-epf-mhi.c | 455 struct pci_epf_mhi_dma_transfer *itr, *tmp; in pci_epf_mhi_dma_worker() local 464 list_for_each_entry_safe(itr, tmp, &head, node) { in pci_epf_mhi_dma_worker() 465 list_del(&itr->node); in pci_epf_mhi_dma_worker() 466 dma_unmap_single(dma_dev, itr->paddr, itr->size, itr->dir); in pci_epf_mhi_dma_worker() 467 buf_info = &itr->buf_info; in pci_epf_mhi_dma_worker() 469 kfree(itr); in pci_epf_mhi_dma_worker()
|
| /drivers/net/ethernet/intel/i40e/ |
| A D | i40e_txrx.c | 1194 itr = rc->target_itr; in i40e_update_itr() 1209 itr = I40E_ITR_ADAPTIVE_LATENCY; in i40e_update_itr() 1240 itr &= I40E_ITR_ADAPTIVE_LATENCY; in i40e_update_itr() 1248 itr &= I40E_ITR_MASK; in i40e_update_itr() 1262 itr /= 2; in i40e_update_itr() 1263 itr &= I40E_ITR_MASK; in i40e_update_itr() 1265 itr = I40E_ITR_ADAPTIVE_MIN_USECS; in i40e_update_itr() 1276 itr = I40E_ITR_ADAPTIVE_BULK; in i40e_update_itr() 1324 if (itr & I40E_ITR_ADAPTIVE_LATENCY) in i40e_update_itr() 1338 itr &= I40E_ITR_ADAPTIVE_LATENCY; in i40e_update_itr() [all …]
|
| /drivers/net/ethernet/intel/ixgbe/ |
| A D | ixgbe_lib.c | 889 q_vector->tx.itr = IXGBE_ITR_ADAPTIVE_MAX_USECS | in ixgbe_alloc_q_vector() 891 q_vector->rx.itr = IXGBE_ITR_ADAPTIVE_MAX_USECS | in ixgbe_alloc_q_vector() 898 q_vector->itr = IXGBE_12K_ITR; in ixgbe_alloc_q_vector() 900 q_vector->itr = adapter->tx_itr_setting; in ixgbe_alloc_q_vector() 904 q_vector->itr = IXGBE_20K_ITR; in ixgbe_alloc_q_vector() 906 q_vector->itr = adapter->rx_itr_setting; in ixgbe_alloc_q_vector()
|
| /drivers/net/ethernet/intel/idpf/ |
| A D | idpf_txrx.c | 3758 u16 itr = 0; in idpf_vport_intr_buildreg_itr() local 3769 itr = IDPF_ITR_20K; in idpf_vport_intr_buildreg_itr() 3772 itr &= IDPF_ITR_MASK; in idpf_vport_intr_buildreg_itr() 3950 writel(ITR_REG_ALIGN(itr) >> IDPF_ITR_GRAN_S, in idpf_vport_intr_write_itr() 3962 u16 itr; in idpf_vport_intr_ena_irq_all() local 3972 itr : qv->tx_itr_value, in idpf_vport_intr_ena_irq_all() 3980 itr : qv->rx_itr_value, in idpf_vport_intr_ena_irq_all() 4010 u16 itr; in idpf_tx_dim_work() local 4020 itr = vport->tx_itr_profile[dim->profile_ix]; in idpf_tx_dim_work() 4036 u16 itr; in idpf_rx_dim_work() local [all …]
|
| /drivers/net/ethernet/intel/ixgbevf/ |
| A D | ixgbevf_main.c | 1374 q_vector->itr = IXGBE_12K_ITR; in ixgbevf_configure_msix() 1376 q_vector->itr = adapter->tx_itr_setting; in ixgbevf_configure_msix() 1380 q_vector->itr = IXGBE_20K_ITR; in ixgbevf_configure_msix() 1424 u8 itr_setting = ring_container->itr; in ixgbevf_update_itr() 1435 timepassed_us = q_vector->itr >> 2; in ixgbevf_update_itr() 1463 ring_container->itr = itr_setting; in ixgbevf_update_itr() 1468 u32 new_itr = q_vector->itr; in ixgbevf_set_itr() 1474 current_itr = max(q_vector->rx.itr, q_vector->tx.itr); in ixgbevf_set_itr() 1491 if (new_itr != q_vector->itr) { in ixgbevf_set_itr() 1494 ((9 * new_itr) + q_vector->itr); in ixgbevf_set_itr() [all …]
|